var/home/core/zuul-output/0000755000175000017500000000000015111272314014522 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015111317725015474 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000007011617215111317712017701 0ustar rootrootNov 25 09:24:13 crc systemd[1]: Starting Kubernetes Kubelet... Nov 25 09:24:13 crc restorecon[4738]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 09:24:13 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 09:24:14 crc restorecon[4738]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 09:24:14 crc restorecon[4738]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 25 09:24:15 crc kubenswrapper[4776]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 09:24:15 crc kubenswrapper[4776]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 25 09:24:15 crc kubenswrapper[4776]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 09:24:15 crc kubenswrapper[4776]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 09:24:15 crc kubenswrapper[4776]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 25 09:24:15 crc kubenswrapper[4776]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.406462 4776 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410192 4776 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410256 4776 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410267 4776 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410276 4776 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410283 4776 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410291 4776 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410297 4776 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410303 4776 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410309 4776 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410314 4776 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410320 4776 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410327 4776 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410344 4776 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410351 4776 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410358 4776 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410365 4776 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410372 4776 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410388 4776 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410404 4776 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410412 4776 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410419 4776 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410426 4776 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410433 4776 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410441 4776 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410460 4776 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410467 4776 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410473 4776 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410483 4776 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410490 4776 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410498 4776 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410505 4776 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410512 4776 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410518 4776 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410528 4776 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410538 4776 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410546 4776 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410564 4776 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410573 4776 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410580 4776 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410588 4776 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410595 4776 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410603 4776 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410610 4776 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410626 4776 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410635 4776 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410642 4776 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410650 4776 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410659 4776 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410673 4776 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410681 4776 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410688 4776 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410695 4776 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410701 4776 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410708 4776 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410715 4776 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410722 4776 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410729 4776 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410738 4776 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410745 4776 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410752 4776 feature_gate.go:330] unrecognized feature gate: Example Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410766 4776 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410773 4776 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410780 4776 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410787 4776 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410796 4776 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410812 4776 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410819 4776 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410826 4776 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410833 4776 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410841 4776 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.410848 4776 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.412900 4776 flags.go:64] FLAG: --address="0.0.0.0" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413023 4776 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413133 4776 flags.go:64] FLAG: --anonymous-auth="true" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413158 4776 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413179 4776 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413192 4776 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413311 4776 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413399 4776 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413417 4776 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413428 4776 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413468 4776 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413478 4776 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413489 4776 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413498 4776 flags.go:64] FLAG: --cgroup-root="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413507 4776 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413530 4776 flags.go:64] FLAG: --client-ca-file="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413538 4776 flags.go:64] FLAG: --cloud-config="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413547 4776 flags.go:64] FLAG: --cloud-provider="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413555 4776 flags.go:64] FLAG: --cluster-dns="[]" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413571 4776 flags.go:64] FLAG: --cluster-domain="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413579 4776 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413589 4776 flags.go:64] FLAG: --config-dir="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413597 4776 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413613 4776 flags.go:64] FLAG: --container-log-max-files="5" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413627 4776 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413636 4776 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413646 4776 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413656 4776 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413664 4776 flags.go:64] FLAG: --contention-profiling="false" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413672 4776 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413681 4776 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413690 4776 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413704 4776 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413716 4776 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413725 4776 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413732 4776 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413741 4776 flags.go:64] FLAG: --enable-load-reader="false" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413749 4776 flags.go:64] FLAG: --enable-server="true" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413757 4776 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413769 4776 flags.go:64] FLAG: --event-burst="100" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413784 4776 flags.go:64] FLAG: --event-qps="50" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413793 4776 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413804 4776 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413812 4776 flags.go:64] FLAG: --eviction-hard="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413824 4776 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413832 4776 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413840 4776 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413849 4776 flags.go:64] FLAG: --eviction-soft="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413863 4776 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413871 4776 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413879 4776 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413888 4776 flags.go:64] FLAG: --experimental-mounter-path="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413897 4776 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413907 4776 flags.go:64] FLAG: --fail-swap-on="true" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413915 4776 flags.go:64] FLAG: --feature-gates="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413928 4776 flags.go:64] FLAG: --file-check-frequency="20s" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413937 4776 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413952 4776 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413961 4776 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413973 4776 flags.go:64] FLAG: --healthz-port="10248" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413982 4776 flags.go:64] FLAG: --help="false" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413990 4776 flags.go:64] FLAG: --hostname-override="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.413998 4776 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414006 4776 flags.go:64] FLAG: --http-check-frequency="20s" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414015 4776 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414022 4776 flags.go:64] FLAG: --image-credential-provider-config="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414035 4776 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414045 4776 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414053 4776 flags.go:64] FLAG: --image-service-endpoint="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414096 4776 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414105 4776 flags.go:64] FLAG: --kube-api-burst="100" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414114 4776 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414124 4776 flags.go:64] FLAG: --kube-api-qps="50" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414132 4776 flags.go:64] FLAG: --kube-reserved="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414145 4776 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414153 4776 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414161 4776 flags.go:64] FLAG: --kubelet-cgroups="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414167 4776 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414176 4776 flags.go:64] FLAG: --lock-file="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414183 4776 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414575 4776 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414609 4776 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414650 4776 flags.go:64] FLAG: --log-json-split-stream="false" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414657 4776 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414664 4776 flags.go:64] FLAG: --log-text-split-stream="false" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414671 4776 flags.go:64] FLAG: --logging-format="text" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414678 4776 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414685 4776 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414692 4776 flags.go:64] FLAG: --manifest-url="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414698 4776 flags.go:64] FLAG: --manifest-url-header="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414711 4776 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414717 4776 flags.go:64] FLAG: --max-open-files="1000000" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414726 4776 flags.go:64] FLAG: --max-pods="110" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414733 4776 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414739 4776 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414745 4776 flags.go:64] FLAG: --memory-manager-policy="None" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414752 4776 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414759 4776 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414765 4776 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414772 4776 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414797 4776 flags.go:64] FLAG: --node-status-max-images="50" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414804 4776 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414810 4776 flags.go:64] FLAG: --oom-score-adj="-999" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414817 4776 flags.go:64] FLAG: --pod-cidr="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414824 4776 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414834 4776 flags.go:64] FLAG: --pod-manifest-path="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414841 4776 flags.go:64] FLAG: --pod-max-pids="-1" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414847 4776 flags.go:64] FLAG: --pods-per-core="0" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414854 4776 flags.go:64] FLAG: --port="10250" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414861 4776 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414870 4776 flags.go:64] FLAG: --provider-id="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414877 4776 flags.go:64] FLAG: --qos-reserved="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414886 4776 flags.go:64] FLAG: --read-only-port="10255" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414892 4776 flags.go:64] FLAG: --register-node="true" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414899 4776 flags.go:64] FLAG: --register-schedulable="true" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414906 4776 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414920 4776 flags.go:64] FLAG: --registry-burst="10" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414927 4776 flags.go:64] FLAG: --registry-qps="5" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414933 4776 flags.go:64] FLAG: --reserved-cpus="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414939 4776 flags.go:64] FLAG: --reserved-memory="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414949 4776 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414955 4776 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414962 4776 flags.go:64] FLAG: --rotate-certificates="false" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414969 4776 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414975 4776 flags.go:64] FLAG: --runonce="false" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414981 4776 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414988 4776 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.414995 4776 flags.go:64] FLAG: --seccomp-default="false" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.415001 4776 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.415007 4776 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.415015 4776 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.415021 4776 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.415028 4776 flags.go:64] FLAG: --storage-driver-password="root" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.415035 4776 flags.go:64] FLAG: --storage-driver-secure="false" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.415041 4776 flags.go:64] FLAG: --storage-driver-table="stats" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.415048 4776 flags.go:64] FLAG: --storage-driver-user="root" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.415054 4776 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.415085 4776 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.415092 4776 flags.go:64] FLAG: --system-cgroups="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.415099 4776 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.415112 4776 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.415118 4776 flags.go:64] FLAG: --tls-cert-file="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.415125 4776 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.415134 4776 flags.go:64] FLAG: --tls-min-version="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.415140 4776 flags.go:64] FLAG: --tls-private-key-file="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.415149 4776 flags.go:64] FLAG: --topology-manager-policy="none" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.415155 4776 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.415161 4776 flags.go:64] FLAG: --topology-manager-scope="container" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.415168 4776 flags.go:64] FLAG: --v="2" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.415187 4776 flags.go:64] FLAG: --version="false" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.415197 4776 flags.go:64] FLAG: --vmodule="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.415208 4776 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.415215 4776 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415434 4776 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415444 4776 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415452 4776 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415461 4776 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415468 4776 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415473 4776 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415479 4776 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415487 4776 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415495 4776 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415501 4776 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415509 4776 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415515 4776 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415521 4776 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415527 4776 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415533 4776 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415538 4776 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415544 4776 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415549 4776 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415555 4776 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415560 4776 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415566 4776 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415573 4776 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415578 4776 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415584 4776 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415590 4776 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415595 4776 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415601 4776 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415606 4776 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415612 4776 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415618 4776 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415623 4776 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415629 4776 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415634 4776 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415639 4776 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415645 4776 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415650 4776 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415655 4776 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415661 4776 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415666 4776 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415671 4776 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415676 4776 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415684 4776 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415691 4776 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415696 4776 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415702 4776 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415707 4776 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415712 4776 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415718 4776 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415725 4776 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415732 4776 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415738 4776 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415744 4776 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415749 4776 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415754 4776 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415762 4776 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415769 4776 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415775 4776 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415780 4776 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415787 4776 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415793 4776 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415799 4776 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415805 4776 feature_gate.go:330] unrecognized feature gate: Example Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415811 4776 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415816 4776 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415822 4776 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415828 4776 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415833 4776 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415838 4776 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415844 4776 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415849 4776 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.415854 4776 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.415875 4776 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.428924 4776 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.428978 4776 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429158 4776 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429185 4776 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429198 4776 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429210 4776 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429220 4776 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429230 4776 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429239 4776 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429250 4776 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429260 4776 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429271 4776 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429281 4776 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429291 4776 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429300 4776 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429310 4776 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429319 4776 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429328 4776 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429336 4776 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429348 4776 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429358 4776 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429369 4776 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429378 4776 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429388 4776 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429397 4776 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429408 4776 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429465 4776 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429476 4776 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429485 4776 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429494 4776 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429503 4776 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429512 4776 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429521 4776 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429530 4776 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429538 4776 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429547 4776 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429558 4776 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429567 4776 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429576 4776 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429585 4776 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429594 4776 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429603 4776 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429612 4776 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429623 4776 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429634 4776 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429643 4776 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429652 4776 feature_gate.go:330] unrecognized feature gate: Example Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429662 4776 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429671 4776 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429682 4776 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429692 4776 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429702 4776 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429711 4776 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429720 4776 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429729 4776 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429738 4776 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429748 4776 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429757 4776 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429765 4776 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429774 4776 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429784 4776 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429793 4776 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429805 4776 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429815 4776 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429824 4776 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429833 4776 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429842 4776 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429851 4776 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429859 4776 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429868 4776 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429876 4776 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429885 4776 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.429897 4776 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.429911 4776 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430191 4776 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430207 4776 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430217 4776 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430226 4776 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430234 4776 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430243 4776 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430251 4776 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430261 4776 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430269 4776 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430280 4776 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430293 4776 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430302 4776 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430314 4776 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430324 4776 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430334 4776 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430343 4776 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430352 4776 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430364 4776 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430375 4776 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430385 4776 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430395 4776 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430404 4776 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430413 4776 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430424 4776 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430433 4776 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430442 4776 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430452 4776 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430462 4776 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430472 4776 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430481 4776 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430490 4776 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430498 4776 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430507 4776 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430515 4776 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430526 4776 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430534 4776 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430543 4776 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430551 4776 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430559 4776 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430568 4776 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430577 4776 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430585 4776 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430593 4776 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430601 4776 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430610 4776 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430619 4776 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430627 4776 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430635 4776 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430644 4776 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430655 4776 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430666 4776 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430675 4776 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430684 4776 feature_gate.go:330] unrecognized feature gate: Example Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430692 4776 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430701 4776 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430710 4776 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430718 4776 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430728 4776 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430740 4776 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430749 4776 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430759 4776 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430768 4776 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430777 4776 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430788 4776 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430798 4776 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430807 4776 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430816 4776 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430824 4776 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430832 4776 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430841 4776 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.430862 4776 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.430876 4776 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.431228 4776 server.go:940] "Client rotation is on, will bootstrap in background" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.438292 4776 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.438844 4776 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.442626 4776 server.go:997] "Starting client certificate rotation" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.442735 4776 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.443055 4776 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-16 17:12:33.392924789 +0000 UTC Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.443230 4776 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1255h48m17.949700508s for next certificate rotation Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.474023 4776 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.478447 4776 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.501268 4776 log.go:25] "Validated CRI v1 runtime API" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.533849 4776 log.go:25] "Validated CRI v1 image API" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.536389 4776 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.542884 4776 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-25-09-19-14-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.542935 4776 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.573650 4776 manager.go:217] Machine: {Timestamp:2025-11-25 09:24:15.571010926 +0000 UTC m=+0.612070509 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:1c412267-e4cd-44b2-89dc-7f2cc5766618 BootID:a4d2f912-d0fd-42f2-920e-273120324c92 Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:e0:bc:ab Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:e0:bc:ab Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:e1:2b:5e Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:3d:16:e9 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:cb:85:cf Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:a3:6b:c4 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:37:ba:32 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:ca:ef:df:99:c1:82 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:0a:10:f4:cd:cd:11 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.573985 4776 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.574208 4776 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.574647 4776 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.574943 4776 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.574991 4776 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.576032 4776 topology_manager.go:138] "Creating topology manager with none policy" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.576134 4776 container_manager_linux.go:303] "Creating device plugin manager" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.576653 4776 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.576682 4776 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.578016 4776 state_mem.go:36] "Initialized new in-memory state store" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.578160 4776 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.583414 4776 kubelet.go:418] "Attempting to sync node with API server" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.583448 4776 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.583480 4776 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.583499 4776 kubelet.go:324] "Adding apiserver pod source" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.583517 4776 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.588916 4776 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.590247 4776 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.193:6443: connect: connection refused Nov 25 09:24:15 crc kubenswrapper[4776]: E1125 09:24:15.590347 4776 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.193:6443: connect: connection refused" logger="UnhandledError" Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.590496 4776 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.193:6443: connect: connection refused Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.590610 4776 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 25 09:24:15 crc kubenswrapper[4776]: E1125 09:24:15.590626 4776 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.193:6443: connect: connection refused" logger="UnhandledError" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.592374 4776 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.595181 4776 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.595214 4776 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.595224 4776 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.595235 4776 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.595250 4776 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.595261 4776 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.595271 4776 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.595286 4776 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.595297 4776 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.595309 4776 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.595356 4776 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.595365 4776 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.596364 4776 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.597105 4776 server.go:1280] "Started kubelet" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.597230 4776 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.597641 4776 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.598272 4776 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.193:6443: connect: connection refused Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.598694 4776 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 25 09:24:15 crc systemd[1]: Started Kubernetes Kubelet. Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.601198 4776 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.601255 4776 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.601383 4776 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 23:29:22.700120989 +0000 UTC Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.601419 4776 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 926h5m7.098703841s for next certificate rotation Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.601508 4776 server.go:460] "Adding debug handlers to kubelet server" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.601676 4776 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.601702 4776 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 25 09:24:15 crc kubenswrapper[4776]: E1125 09:24:15.601744 4776 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.601844 4776 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.604497 4776 factory.go:55] Registering systemd factory Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.604534 4776 factory.go:221] Registration of the systemd container factory successfully Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.605818 4776 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.193:6443: connect: connection refused Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.607526 4776 factory.go:153] Registering CRI-O factory Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.607581 4776 factory.go:221] Registration of the crio container factory successfully Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.607709 4776 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.607753 4776 factory.go:103] Registering Raw factory Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.607779 4776 manager.go:1196] Started watching for new ooms in manager Nov 25 09:24:15 crc kubenswrapper[4776]: E1125 09:24:15.606258 4776 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.193:6443: connect: connection refused" logger="UnhandledError" Nov 25 09:24:15 crc kubenswrapper[4776]: E1125 09:24:15.604008 4776 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.193:6443: connect: connection refused" interval="200ms" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.611970 4776 manager.go:319] Starting recovery of all containers Nov 25 09:24:15 crc kubenswrapper[4776]: E1125 09:24:15.612578 4776 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.193:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187b35a38c64b0e5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 09:24:15.597031653 +0000 UTC m=+0.638091246,LastTimestamp:2025-11-25 09:24:15.597031653 +0000 UTC m=+0.638091246,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623423 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623487 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623504 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623519 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623531 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623543 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623556 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623567 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623580 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623589 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623599 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623609 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623618 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623657 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623667 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623676 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623686 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623696 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623705 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623715 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623723 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623735 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623746 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623758 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623768 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623776 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623790 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623800 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623813 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623825 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623838 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623877 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623889 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623899 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623908 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623919 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623929 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623939 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623952 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623965 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623978 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.623992 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624005 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624018 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624030 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624045 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624058 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624087 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624100 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624110 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624120 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624130 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624145 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624156 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624167 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624177 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624188 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624198 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624207 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624219 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624230 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624243 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624257 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624270 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624284 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624297 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624311 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624322 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624333 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624344 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624355 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624366 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624375 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624385 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624397 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624409 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624420 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624431 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624441 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624454 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624467 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624480 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624490 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624500 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624510 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624522 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624535 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624547 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624560 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624572 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624584 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624596 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624607 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624620 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624634 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624648 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624664 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624676 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624689 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624707 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624718 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624732 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624743 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624755 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624774 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624789 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624803 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624817 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624831 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624846 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624858 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624871 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624920 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624939 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.624988 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.625025 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.625037 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.625050 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.625079 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.625095 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.625107 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.625119 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.625147 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.625163 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.625175 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.625187 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.625202 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.625216 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.625231 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.625242 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.625254 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.625303 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.625317 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.625330 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.625343 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.625358 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.625375 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627413 4776 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627442 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627472 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627482 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627494 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627504 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627517 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627529 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627540 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627551 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627560 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627570 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627580 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627590 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627599 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627608 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627617 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627626 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627637 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627651 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627661 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627674 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627684 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627696 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627708 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627719 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627731 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627743 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627766 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627776 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627786 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627796 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627807 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627817 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627829 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627839 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627851 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627862 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627872 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627886 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627896 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627906 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627916 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627926 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627937 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627948 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627958 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627968 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627979 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627988 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.627999 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.628008 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.628020 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.628030 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.628039 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.628051 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.628079 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.628090 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.628102 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.628116 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.628131 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.628144 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.628157 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.628173 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.628187 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.628197 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.628217 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.628227 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.628237 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.628248 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.628261 4776 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.628272 4776 reconstruct.go:97] "Volume reconstruction finished" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.628280 4776 reconciler.go:26] "Reconciler: start to sync state" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.653702 4776 manager.go:324] Recovery completed Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.658914 4776 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.660972 4776 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.661023 4776 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.661052 4776 kubelet.go:2335] "Starting kubelet main sync loop" Nov 25 09:24:15 crc kubenswrapper[4776]: E1125 09:24:15.661138 4776 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 25 09:24:15 crc kubenswrapper[4776]: W1125 09:24:15.662939 4776 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.193:6443: connect: connection refused Nov 25 09:24:15 crc kubenswrapper[4776]: E1125 09:24:15.663087 4776 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.193:6443: connect: connection refused" logger="UnhandledError" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.671417 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.673412 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.673464 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.673476 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.674438 4776 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.674468 4776 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.674494 4776 state_mem.go:36] "Initialized new in-memory state store" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.692537 4776 policy_none.go:49] "None policy: Start" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.693802 4776 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.693854 4776 state_mem.go:35] "Initializing new in-memory state store" Nov 25 09:24:15 crc kubenswrapper[4776]: E1125 09:24:15.701910 4776 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.752405 4776 manager.go:334] "Starting Device Plugin manager" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.752468 4776 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.752485 4776 server.go:79] "Starting device plugin registration server" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.753241 4776 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.753264 4776 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.753635 4776 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.753729 4776 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.753741 4776 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 25 09:24:15 crc kubenswrapper[4776]: E1125 09:24:15.760100 4776 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.193:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187b35a38c64b0e5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 09:24:15.597031653 +0000 UTC m=+0.638091246,LastTimestamp:2025-11-25 09:24:15.597031653 +0000 UTC m=+0.638091246,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.761313 4776 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.761397 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.762376 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.762405 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.762415 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.762540 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.762985 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.763007 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.763774 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.763790 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.763798 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.763872 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.764189 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.764205 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.764536 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.764549 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.764557 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.765195 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.765219 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.765229 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.765353 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.765746 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.765774 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.766165 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.766180 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.766189 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.767113 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.767269 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.767385 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.767599 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.767302 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.768081 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.768097 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.767838 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.768297 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.769109 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.769138 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.769151 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.769367 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.769461 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.769587 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.769838 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.769942 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:15 crc kubenswrapper[4776]: E1125 09:24:15.770488 4776 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.770660 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.770706 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.770716 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:15 crc kubenswrapper[4776]: E1125 09:24:15.812122 4776 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.193:6443: connect: connection refused" interval="400ms" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.832234 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.832285 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.833082 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.833207 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.833294 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.833619 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.833692 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.833715 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.833736 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.833813 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.833861 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.834193 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.834304 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.834475 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.834640 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.853897 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.855470 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.855508 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.855519 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.855550 4776 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 09:24:15 crc kubenswrapper[4776]: E1125 09:24:15.856117 4776 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.193:6443: connect: connection refused" node="crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.936046 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.936172 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.936211 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.936246 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.936279 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.936297 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.936388 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.936313 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.936457 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.936493 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.936507 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.936525 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.936585 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.936616 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.936649 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.936643 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.936693 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.936724 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.936753 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.936762 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.936580 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.936453 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.936959 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.936960 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.937118 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.937134 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.936900 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.937199 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.937195 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 09:24:15 crc kubenswrapper[4776]: I1125 09:24:15.937208 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:24:16 crc kubenswrapper[4776]: I1125 09:24:16.057270 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:16 crc kubenswrapper[4776]: I1125 09:24:16.059221 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:16 crc kubenswrapper[4776]: I1125 09:24:16.059288 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:16 crc kubenswrapper[4776]: I1125 09:24:16.059307 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:16 crc kubenswrapper[4776]: I1125 09:24:16.059347 4776 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 09:24:16 crc kubenswrapper[4776]: E1125 09:24:16.060000 4776 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.193:6443: connect: connection refused" node="crc" Nov 25 09:24:16 crc kubenswrapper[4776]: I1125 09:24:16.092743 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 09:24:16 crc kubenswrapper[4776]: I1125 09:24:16.099230 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 09:24:16 crc kubenswrapper[4776]: I1125 09:24:16.116292 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 25 09:24:16 crc kubenswrapper[4776]: I1125 09:24:16.134947 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:24:16 crc kubenswrapper[4776]: I1125 09:24:16.143325 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:24:16 crc kubenswrapper[4776]: W1125 09:24:16.147522 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-64cfb60aa3f2ac6d13812efb5b99b3d34b3570fde7ca5d10d8c6e127c73e9842 WatchSource:0}: Error finding container 64cfb60aa3f2ac6d13812efb5b99b3d34b3570fde7ca5d10d8c6e127c73e9842: Status 404 returned error can't find the container with id 64cfb60aa3f2ac6d13812efb5b99b3d34b3570fde7ca5d10d8c6e127c73e9842 Nov 25 09:24:16 crc kubenswrapper[4776]: W1125 09:24:16.149818 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-acd85b90fd252892fdf86214ca590c10b74a5344b121a7acb4534e3522c17771 WatchSource:0}: Error finding container acd85b90fd252892fdf86214ca590c10b74a5344b121a7acb4534e3522c17771: Status 404 returned error can't find the container with id acd85b90fd252892fdf86214ca590c10b74a5344b121a7acb4534e3522c17771 Nov 25 09:24:16 crc kubenswrapper[4776]: W1125 09:24:16.155639 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-3fa6f70e7fd7283e13c0c8a789473c5a83850ef7c030569381ca0695f0e89dd9 WatchSource:0}: Error finding container 3fa6f70e7fd7283e13c0c8a789473c5a83850ef7c030569381ca0695f0e89dd9: Status 404 returned error can't find the container with id 3fa6f70e7fd7283e13c0c8a789473c5a83850ef7c030569381ca0695f0e89dd9 Nov 25 09:24:16 crc kubenswrapper[4776]: W1125 09:24:16.161551 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-878a26083b9419f630c75d5751ccbb6ad9d61fe0bf563246f4c32493a91a2350 WatchSource:0}: Error finding container 878a26083b9419f630c75d5751ccbb6ad9d61fe0bf563246f4c32493a91a2350: Status 404 returned error can't find the container with id 878a26083b9419f630c75d5751ccbb6ad9d61fe0bf563246f4c32493a91a2350 Nov 25 09:24:16 crc kubenswrapper[4776]: W1125 09:24:16.162795 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-b69031e75248f7801e9553ea8a4ddf658ba350fcfee9c7e60d47c294cd8bf351 WatchSource:0}: Error finding container b69031e75248f7801e9553ea8a4ddf658ba350fcfee9c7e60d47c294cd8bf351: Status 404 returned error can't find the container with id b69031e75248f7801e9553ea8a4ddf658ba350fcfee9c7e60d47c294cd8bf351 Nov 25 09:24:16 crc kubenswrapper[4776]: E1125 09:24:16.213797 4776 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.193:6443: connect: connection refused" interval="800ms" Nov 25 09:24:16 crc kubenswrapper[4776]: I1125 09:24:16.460358 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:16 crc kubenswrapper[4776]: I1125 09:24:16.462264 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:16 crc kubenswrapper[4776]: I1125 09:24:16.462334 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:16 crc kubenswrapper[4776]: I1125 09:24:16.462356 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:16 crc kubenswrapper[4776]: I1125 09:24:16.462395 4776 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 09:24:16 crc kubenswrapper[4776]: E1125 09:24:16.462811 4776 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.193:6443: connect: connection refused" node="crc" Nov 25 09:24:16 crc kubenswrapper[4776]: W1125 09:24:16.565650 4776 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.193:6443: connect: connection refused Nov 25 09:24:16 crc kubenswrapper[4776]: E1125 09:24:16.565786 4776 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.193:6443: connect: connection refused" logger="UnhandledError" Nov 25 09:24:16 crc kubenswrapper[4776]: I1125 09:24:16.600081 4776 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.193:6443: connect: connection refused Nov 25 09:24:16 crc kubenswrapper[4776]: I1125 09:24:16.666964 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"acd85b90fd252892fdf86214ca590c10b74a5344b121a7acb4534e3522c17771"} Nov 25 09:24:16 crc kubenswrapper[4776]: I1125 09:24:16.668474 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"64cfb60aa3f2ac6d13812efb5b99b3d34b3570fde7ca5d10d8c6e127c73e9842"} Nov 25 09:24:16 crc kubenswrapper[4776]: I1125 09:24:16.670320 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b69031e75248f7801e9553ea8a4ddf658ba350fcfee9c7e60d47c294cd8bf351"} Nov 25 09:24:16 crc kubenswrapper[4776]: I1125 09:24:16.672266 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"878a26083b9419f630c75d5751ccbb6ad9d61fe0bf563246f4c32493a91a2350"} Nov 25 09:24:16 crc kubenswrapper[4776]: I1125 09:24:16.673790 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3fa6f70e7fd7283e13c0c8a789473c5a83850ef7c030569381ca0695f0e89dd9"} Nov 25 09:24:16 crc kubenswrapper[4776]: W1125 09:24:16.880562 4776 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.193:6443: connect: connection refused Nov 25 09:24:16 crc kubenswrapper[4776]: E1125 09:24:16.881336 4776 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.193:6443: connect: connection refused" logger="UnhandledError" Nov 25 09:24:17 crc kubenswrapper[4776]: W1125 09:24:17.013326 4776 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.193:6443: connect: connection refused Nov 25 09:24:17 crc kubenswrapper[4776]: E1125 09:24:17.013648 4776 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.193:6443: connect: connection refused" logger="UnhandledError" Nov 25 09:24:17 crc kubenswrapper[4776]: E1125 09:24:17.014803 4776 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.193:6443: connect: connection refused" interval="1.6s" Nov 25 09:24:17 crc kubenswrapper[4776]: W1125 09:24:17.151370 4776 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.193:6443: connect: connection refused Nov 25 09:24:17 crc kubenswrapper[4776]: E1125 09:24:17.151506 4776 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.193:6443: connect: connection refused" logger="UnhandledError" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.263556 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.265720 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.265769 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.265786 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.265820 4776 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 09:24:17 crc kubenswrapper[4776]: E1125 09:24:17.266397 4776 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.193:6443: connect: connection refused" node="crc" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.599578 4776 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.193:6443: connect: connection refused Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.678702 4776 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22" exitCode=0 Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.678809 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22"} Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.678942 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.680111 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.680169 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.680186 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.683389 4776 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f" exitCode=0 Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.683467 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.683498 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f"} Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.684464 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.684483 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.684495 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.685989 4776 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="9f9c968afbab8870d4105cc2b407cb7b2392919716c0a9b5de3ed47f3d12e5a7" exitCode=0 Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.686042 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.686067 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"9f9c968afbab8870d4105cc2b407cb7b2392919716c0a9b5de3ed47f3d12e5a7"} Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.686507 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.687928 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.687951 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.687965 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.688248 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.688633 4776 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="583bd3f6f361dbdc9c02a6eeeee4a445e2d2d08f605adf2d3c8912cb7d416abb" exitCode=0 Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.688680 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"583bd3f6f361dbdc9c02a6eeeee4a445e2d2d08f605adf2d3c8912cb7d416abb"} Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.688697 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.688744 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.689545 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.690853 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.690881 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.690896 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.694067 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc"} Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.694147 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04"} Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.694170 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4"} Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.694265 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.695163 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.695187 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:17 crc kubenswrapper[4776]: I1125 09:24:17.695197 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:18 crc kubenswrapper[4776]: W1125 09:24:18.495286 4776 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.193:6443: connect: connection refused Nov 25 09:24:18 crc kubenswrapper[4776]: E1125 09:24:18.495393 4776 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.193:6443: connect: connection refused" logger="UnhandledError" Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.600419 4776 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.193:6443: connect: connection refused Nov 25 09:24:18 crc kubenswrapper[4776]: E1125 09:24:18.616078 4776 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.193:6443: connect: connection refused" interval="3.2s" Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.698977 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33"} Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.699139 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.700948 4776 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e" exitCode=0 Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.700999 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e"} Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.701291 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.701848 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.701886 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.701898 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.703074 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.703125 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.703140 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.707994 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf"} Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.708023 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea"} Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.708052 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670"} Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.708077 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68"} Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.710306 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"21b232d5cd319131183cefb0df8c537e099bfa1743d7fab3748858d3873a4802"} Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.710397 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.711833 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.711862 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.711900 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.715340 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"bf7b5ddb321a7c0a66d83ecb015a42c43b13ae4af284f4806d8fcf5a5f3ba005"} Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.715394 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"284857ebafc167c12fe70604afa6702c071316774d5976a5e95042968dafe0ba"} Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.715416 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"655c0dff64b7dbe451d7c61bf1e1e63051f02f7587bbb65e71b5ee7d2bec0a2a"} Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.715538 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.719231 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.719283 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.719308 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.867298 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.880516 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.880547 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.880555 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:18 crc kubenswrapper[4776]: I1125 09:24:18.880601 4776 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 09:24:18 crc kubenswrapper[4776]: E1125 09:24:18.880914 4776 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.193:6443: connect: connection refused" node="crc" Nov 25 09:24:18 crc kubenswrapper[4776]: W1125 09:24:18.910188 4776 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.193:6443: connect: connection refused Nov 25 09:24:18 crc kubenswrapper[4776]: E1125 09:24:18.910277 4776 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.193:6443: connect: connection refused" logger="UnhandledError" Nov 25 09:24:18 crc kubenswrapper[4776]: W1125 09:24:18.918911 4776 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.193:6443: connect: connection refused Nov 25 09:24:18 crc kubenswrapper[4776]: E1125 09:24:18.918966 4776 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.193:6443: connect: connection refused" logger="UnhandledError" Nov 25 09:24:19 crc kubenswrapper[4776]: I1125 09:24:19.515703 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:24:19 crc kubenswrapper[4776]: I1125 09:24:19.534681 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:24:19 crc kubenswrapper[4776]: I1125 09:24:19.720174 4776 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454" exitCode=0 Nov 25 09:24:19 crc kubenswrapper[4776]: I1125 09:24:19.720248 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454"} Nov 25 09:24:19 crc kubenswrapper[4776]: I1125 09:24:19.720371 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:19 crc kubenswrapper[4776]: I1125 09:24:19.721263 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:19 crc kubenswrapper[4776]: I1125 09:24:19.721287 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:19 crc kubenswrapper[4776]: I1125 09:24:19.721297 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:19 crc kubenswrapper[4776]: I1125 09:24:19.726538 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:19 crc kubenswrapper[4776]: I1125 09:24:19.727371 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:19 crc kubenswrapper[4776]: I1125 09:24:19.727985 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"be450778465626f9bbb72bd6ee1448c151dbed8c4d3983b96c99c0bf8fb9da5e"} Nov 25 09:24:19 crc kubenswrapper[4776]: I1125 09:24:19.728041 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 09:24:19 crc kubenswrapper[4776]: I1125 09:24:19.728328 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:19 crc kubenswrapper[4776]: I1125 09:24:19.728370 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:24:19 crc kubenswrapper[4776]: I1125 09:24:19.728317 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:19 crc kubenswrapper[4776]: I1125 09:24:19.729337 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:19 crc kubenswrapper[4776]: I1125 09:24:19.729376 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:19 crc kubenswrapper[4776]: I1125 09:24:19.729377 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:19 crc kubenswrapper[4776]: I1125 09:24:19.729412 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:19 crc kubenswrapper[4776]: I1125 09:24:19.729477 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:19 crc kubenswrapper[4776]: I1125 09:24:19.729489 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:19 crc kubenswrapper[4776]: I1125 09:24:19.729499 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:19 crc kubenswrapper[4776]: I1125 09:24:19.729396 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:19 crc kubenswrapper[4776]: I1125 09:24:19.729547 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:19 crc kubenswrapper[4776]: I1125 09:24:19.729623 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:19 crc kubenswrapper[4776]: I1125 09:24:19.729645 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:19 crc kubenswrapper[4776]: I1125 09:24:19.729653 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:20 crc kubenswrapper[4776]: I1125 09:24:20.239992 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:24:20 crc kubenswrapper[4776]: I1125 09:24:20.736822 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:20 crc kubenswrapper[4776]: I1125 09:24:20.736896 4776 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 09:24:20 crc kubenswrapper[4776]: I1125 09:24:20.736910 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6"} Nov 25 09:24:20 crc kubenswrapper[4776]: I1125 09:24:20.737005 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:20 crc kubenswrapper[4776]: I1125 09:24:20.736950 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:20 crc kubenswrapper[4776]: I1125 09:24:20.737044 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26"} Nov 25 09:24:20 crc kubenswrapper[4776]: I1125 09:24:20.737161 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a"} Nov 25 09:24:20 crc kubenswrapper[4776]: I1125 09:24:20.739787 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:20 crc kubenswrapper[4776]: I1125 09:24:20.739827 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:20 crc kubenswrapper[4776]: I1125 09:24:20.739839 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:20 crc kubenswrapper[4776]: I1125 09:24:20.740259 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:20 crc kubenswrapper[4776]: I1125 09:24:20.740304 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:20 crc kubenswrapper[4776]: I1125 09:24:20.740494 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:20 crc kubenswrapper[4776]: I1125 09:24:20.740561 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:20 crc kubenswrapper[4776]: I1125 09:24:20.740590 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:20 crc kubenswrapper[4776]: I1125 09:24:20.740512 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:20 crc kubenswrapper[4776]: I1125 09:24:20.915941 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:24:21 crc kubenswrapper[4776]: I1125 09:24:21.746250 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:21 crc kubenswrapper[4776]: I1125 09:24:21.746305 4776 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 09:24:21 crc kubenswrapper[4776]: I1125 09:24:21.746363 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:21 crc kubenswrapper[4776]: I1125 09:24:21.746240 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f"} Nov 25 09:24:21 crc kubenswrapper[4776]: I1125 09:24:21.746440 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209"} Nov 25 09:24:21 crc kubenswrapper[4776]: I1125 09:24:21.746604 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:21 crc kubenswrapper[4776]: I1125 09:24:21.747834 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:21 crc kubenswrapper[4776]: I1125 09:24:21.747893 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:21 crc kubenswrapper[4776]: I1125 09:24:21.747920 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:21 crc kubenswrapper[4776]: I1125 09:24:21.748157 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:21 crc kubenswrapper[4776]: I1125 09:24:21.748158 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:21 crc kubenswrapper[4776]: I1125 09:24:21.748201 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:21 crc kubenswrapper[4776]: I1125 09:24:21.748211 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:21 crc kubenswrapper[4776]: I1125 09:24:21.748220 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:21 crc kubenswrapper[4776]: I1125 09:24:21.748225 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:22 crc kubenswrapper[4776]: I1125 09:24:22.081569 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:22 crc kubenswrapper[4776]: I1125 09:24:22.083491 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:22 crc kubenswrapper[4776]: I1125 09:24:22.083576 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:22 crc kubenswrapper[4776]: I1125 09:24:22.083607 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:22 crc kubenswrapper[4776]: I1125 09:24:22.083653 4776 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 09:24:22 crc kubenswrapper[4776]: I1125 09:24:22.533459 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:24:22 crc kubenswrapper[4776]: I1125 09:24:22.749892 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:22 crc kubenswrapper[4776]: I1125 09:24:22.749893 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:22 crc kubenswrapper[4776]: I1125 09:24:22.751146 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:22 crc kubenswrapper[4776]: I1125 09:24:22.751162 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:22 crc kubenswrapper[4776]: I1125 09:24:22.751216 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:22 crc kubenswrapper[4776]: I1125 09:24:22.751235 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:22 crc kubenswrapper[4776]: I1125 09:24:22.751174 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:22 crc kubenswrapper[4776]: I1125 09:24:22.751308 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:23 crc kubenswrapper[4776]: I1125 09:24:23.726319 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:24:23 crc kubenswrapper[4776]: I1125 09:24:23.726594 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:23 crc kubenswrapper[4776]: I1125 09:24:23.728196 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:23 crc kubenswrapper[4776]: I1125 09:24:23.728302 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:23 crc kubenswrapper[4776]: I1125 09:24:23.728329 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:23 crc kubenswrapper[4776]: I1125 09:24:23.843542 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:24:23 crc kubenswrapper[4776]: I1125 09:24:23.843795 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:23 crc kubenswrapper[4776]: I1125 09:24:23.845379 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:23 crc kubenswrapper[4776]: I1125 09:24:23.845453 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:23 crc kubenswrapper[4776]: I1125 09:24:23.845473 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:25 crc kubenswrapper[4776]: I1125 09:24:25.533474 4776 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 25 09:24:25 crc kubenswrapper[4776]: I1125 09:24:25.533576 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 25 09:24:25 crc kubenswrapper[4776]: E1125 09:24:25.771777 4776 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 25 09:24:26 crc kubenswrapper[4776]: I1125 09:24:26.231792 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 25 09:24:26 crc kubenswrapper[4776]: I1125 09:24:26.232180 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:26 crc kubenswrapper[4776]: I1125 09:24:26.234031 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:26 crc kubenswrapper[4776]: I1125 09:24:26.234115 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:26 crc kubenswrapper[4776]: I1125 09:24:26.234134 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:26 crc kubenswrapper[4776]: I1125 09:24:26.481398 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 25 09:24:26 crc kubenswrapper[4776]: I1125 09:24:26.761670 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:26 crc kubenswrapper[4776]: I1125 09:24:26.763209 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:26 crc kubenswrapper[4776]: I1125 09:24:26.763295 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:26 crc kubenswrapper[4776]: I1125 09:24:26.763318 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:27 crc kubenswrapper[4776]: I1125 09:24:27.197929 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:24:27 crc kubenswrapper[4776]: I1125 09:24:27.198161 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:27 crc kubenswrapper[4776]: I1125 09:24:27.199730 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:27 crc kubenswrapper[4776]: I1125 09:24:27.199780 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:27 crc kubenswrapper[4776]: I1125 09:24:27.199797 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:29 crc kubenswrapper[4776]: I1125 09:24:29.599367 4776 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 25 09:24:29 crc kubenswrapper[4776]: W1125 09:24:29.675953 4776 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 25 09:24:29 crc kubenswrapper[4776]: I1125 09:24:29.676134 4776 trace.go:236] Trace[1533248396]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 09:24:19.674) (total time: 10001ms): Nov 25 09:24:29 crc kubenswrapper[4776]: Trace[1533248396]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (09:24:29.675) Nov 25 09:24:29 crc kubenswrapper[4776]: Trace[1533248396]: [10.001285962s] [10.001285962s] END Nov 25 09:24:29 crc kubenswrapper[4776]: E1125 09:24:29.676170 4776 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 25 09:24:29 crc kubenswrapper[4776]: I1125 09:24:29.779917 4776 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:53940->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 25 09:24:29 crc kubenswrapper[4776]: I1125 09:24:29.780025 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:53940->192.168.126.11:17697: read: connection reset by peer" Nov 25 09:24:30 crc kubenswrapper[4776]: I1125 09:24:30.108365 4776 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 25 09:24:30 crc kubenswrapper[4776]: I1125 09:24:30.108690 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 25 09:24:30 crc kubenswrapper[4776]: I1125 09:24:30.126762 4776 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Nov 25 09:24:30 crc kubenswrapper[4776]: I1125 09:24:30.126826 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 25 09:24:30 crc kubenswrapper[4776]: I1125 09:24:30.778199 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 25 09:24:30 crc kubenswrapper[4776]: I1125 09:24:30.781728 4776 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="be450778465626f9bbb72bd6ee1448c151dbed8c4d3983b96c99c0bf8fb9da5e" exitCode=255 Nov 25 09:24:30 crc kubenswrapper[4776]: I1125 09:24:30.781812 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"be450778465626f9bbb72bd6ee1448c151dbed8c4d3983b96c99c0bf8fb9da5e"} Nov 25 09:24:30 crc kubenswrapper[4776]: I1125 09:24:30.782385 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:30 crc kubenswrapper[4776]: I1125 09:24:30.791728 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:30 crc kubenswrapper[4776]: I1125 09:24:30.791803 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:30 crc kubenswrapper[4776]: I1125 09:24:30.791823 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:30 crc kubenswrapper[4776]: I1125 09:24:30.792826 4776 scope.go:117] "RemoveContainer" containerID="be450778465626f9bbb72bd6ee1448c151dbed8c4d3983b96c99c0bf8fb9da5e" Nov 25 09:24:31 crc kubenswrapper[4776]: I1125 09:24:31.787951 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 25 09:24:31 crc kubenswrapper[4776]: I1125 09:24:31.790810 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d"} Nov 25 09:24:31 crc kubenswrapper[4776]: I1125 09:24:31.790998 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:31 crc kubenswrapper[4776]: I1125 09:24:31.792288 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:31 crc kubenswrapper[4776]: I1125 09:24:31.792343 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:31 crc kubenswrapper[4776]: I1125 09:24:31.792358 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:32 crc kubenswrapper[4776]: I1125 09:24:32.797160 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 25 09:24:32 crc kubenswrapper[4776]: I1125 09:24:32.797853 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 25 09:24:32 crc kubenswrapper[4776]: I1125 09:24:32.801328 4776 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d" exitCode=255 Nov 25 09:24:32 crc kubenswrapper[4776]: I1125 09:24:32.801412 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d"} Nov 25 09:24:32 crc kubenswrapper[4776]: I1125 09:24:32.801525 4776 scope.go:117] "RemoveContainer" containerID="be450778465626f9bbb72bd6ee1448c151dbed8c4d3983b96c99c0bf8fb9da5e" Nov 25 09:24:32 crc kubenswrapper[4776]: I1125 09:24:32.801724 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:32 crc kubenswrapper[4776]: I1125 09:24:32.803787 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:32 crc kubenswrapper[4776]: I1125 09:24:32.803849 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:32 crc kubenswrapper[4776]: I1125 09:24:32.803875 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:32 crc kubenswrapper[4776]: I1125 09:24:32.804856 4776 scope.go:117] "RemoveContainer" containerID="399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d" Nov 25 09:24:32 crc kubenswrapper[4776]: E1125 09:24:32.805268 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 25 09:24:33 crc kubenswrapper[4776]: I1125 09:24:33.727352 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:24:33 crc kubenswrapper[4776]: I1125 09:24:33.807511 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 25 09:24:33 crc kubenswrapper[4776]: I1125 09:24:33.812463 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:33 crc kubenswrapper[4776]: I1125 09:24:33.814157 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:33 crc kubenswrapper[4776]: I1125 09:24:33.814211 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:33 crc kubenswrapper[4776]: I1125 09:24:33.814235 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:33 crc kubenswrapper[4776]: I1125 09:24:33.815038 4776 scope.go:117] "RemoveContainer" containerID="399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d" Nov 25 09:24:33 crc kubenswrapper[4776]: E1125 09:24:33.815405 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 25 09:24:33 crc kubenswrapper[4776]: I1125 09:24:33.851749 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:24:34 crc kubenswrapper[4776]: I1125 09:24:34.815005 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:34 crc kubenswrapper[4776]: I1125 09:24:34.816309 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:34 crc kubenswrapper[4776]: I1125 09:24:34.816374 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:34 crc kubenswrapper[4776]: I1125 09:24:34.816397 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:34 crc kubenswrapper[4776]: I1125 09:24:34.817587 4776 scope.go:117] "RemoveContainer" containerID="399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d" Nov 25 09:24:34 crc kubenswrapper[4776]: E1125 09:24:34.817966 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 25 09:24:34 crc kubenswrapper[4776]: I1125 09:24:34.822342 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:24:35 crc kubenswrapper[4776]: E1125 09:24:35.104816 4776 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.108596 4776 trace.go:236] Trace[237881518]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 09:24:22.511) (total time: 12596ms): Nov 25 09:24:35 crc kubenswrapper[4776]: Trace[237881518]: ---"Objects listed" error: 12596ms (09:24:35.108) Nov 25 09:24:35 crc kubenswrapper[4776]: Trace[237881518]: [12.596659427s] [12.596659427s] END Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.108628 4776 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.109819 4776 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.109992 4776 trace.go:236] Trace[1131210294]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 09:24:24.895) (total time: 10214ms): Nov 25 09:24:35 crc kubenswrapper[4776]: Trace[1131210294]: ---"Objects listed" error: 10214ms (09:24:35.109) Nov 25 09:24:35 crc kubenswrapper[4776]: Trace[1131210294]: [10.214762972s] [10.214762972s] END Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.110212 4776 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 25 09:24:35 crc kubenswrapper[4776]: E1125 09:24:35.110983 4776 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.112343 4776 trace.go:236] Trace[1029563829]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 09:24:24.920) (total time: 10191ms): Nov 25 09:24:35 crc kubenswrapper[4776]: Trace[1029563829]: ---"Objects listed" error: 10191ms (09:24:35.112) Nov 25 09:24:35 crc kubenswrapper[4776]: Trace[1029563829]: [10.191521438s] [10.191521438s] END Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.112369 4776 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.136522 4776 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.533867 4776 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.533941 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.597133 4776 apiserver.go:52] "Watching apiserver" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.603308 4776 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.603725 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.604227 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.604753 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.604821 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:24:35 crc kubenswrapper[4776]: E1125 09:24:35.604977 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:24:35 crc kubenswrapper[4776]: E1125 09:24:35.605107 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.605213 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.605229 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.605253 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:24:35 crc kubenswrapper[4776]: E1125 09:24:35.605696 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.610871 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.610890 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.611006 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.611025 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.611157 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.611158 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.611217 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.611308 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.610891 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.660044 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.674422 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.686887 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.698652 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.702475 4776 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.709467 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.714244 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.714317 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.714365 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.714401 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.714438 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.714480 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.714527 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.714574 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.714611 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.714645 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.714643 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.714681 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.714685 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.714717 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.714754 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.714796 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.714831 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.714865 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.714902 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.714912 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.714906 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.714905 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.714966 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715038 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715110 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715153 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715194 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715230 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715264 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715298 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715334 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715369 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715402 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715435 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715113 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715470 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715146 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715431 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715439 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715518 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715591 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715652 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715720 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715758 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715795 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715837 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715871 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715907 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715944 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715979 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716012 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716049 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716146 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716183 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716273 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716309 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716341 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716381 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716417 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716457 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716493 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716529 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716565 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716600 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715870 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.715897 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716739 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716179 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716248 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716797 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716842 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716925 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716968 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717004 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717040 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717106 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717142 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717175 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717217 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717251 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717286 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717323 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717357 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717393 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717430 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717472 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717508 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717546 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717655 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717708 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717743 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717780 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717817 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717852 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717889 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717936 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717981 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718017 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718126 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718164 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718205 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718241 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718278 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718314 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718349 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718384 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718424 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718458 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718495 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718531 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718566 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718620 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718701 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718769 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718833 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718879 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718926 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718978 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.719028 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.719121 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.719179 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.719218 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.719253 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.719298 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.719350 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.719400 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.719434 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.719481 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.719528 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.719564 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.719602 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.719642 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.719678 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.719715 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.719751 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.719832 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.719890 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.719946 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.719994 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.720043 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.720137 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.720194 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.720254 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.720307 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.720371 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.720601 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.720692 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.720758 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.720815 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.720869 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.720931 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.720985 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.721037 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.721134 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.721190 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.721241 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.721292 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.721343 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.721406 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.721462 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.721533 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.721593 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.721653 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.721713 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.721785 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.721843 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.721961 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.722033 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.722101 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.722141 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.722197 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.722603 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.722662 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.722702 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.723388 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.723440 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.723478 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.723518 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.723552 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.723584 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.723615 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.723645 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.723693 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.723730 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.723758 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.723787 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.723821 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.723849 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.723882 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.723911 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716267 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716254 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716260 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716394 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716457 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716486 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716604 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716611 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716671 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716689 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716871 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.716969 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717234 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717409 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.724239 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.724281 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.724317 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.724349 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.724381 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.724409 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.724444 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.724478 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.724506 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.724531 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.724603 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.724649 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.724727 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.724775 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.724811 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.724852 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.724885 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.724919 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.725813 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.725959 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726021 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726100 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726140 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726186 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726222 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726259 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726297 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726331 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726372 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726409 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726440 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726477 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726515 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726683 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726715 4776 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726734 4776 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726753 4776 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726773 4776 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726790 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726806 4776 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726822 4776 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726842 4776 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726859 4776 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726881 4776 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726899 4776 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726919 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726954 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726980 4776 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726997 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726693 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.734741 4776 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.737510 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.738023 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.738551 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717436 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.741870 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717656 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717724 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717711 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717746 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.717889 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718121 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718369 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718481 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718540 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718776 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718817 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718950 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.718976 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.719049 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.719552 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.719783 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.720015 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.720009 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.742252 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.720129 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.742276 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.720289 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.720178 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.720198 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.720747 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.721238 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.721417 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.721509 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.721576 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.721599 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.721700 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.721741 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.722824 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.722924 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.723249 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.723599 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.723624 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.723868 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.724006 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.724269 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.724599 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.724623 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.724916 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.725280 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.725873 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.725873 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.725935 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726231 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726226 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726367 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726566 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726749 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726865 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.742783 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726921 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.726924 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.727202 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.727681 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.727703 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.727729 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.727743 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.727950 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.728148 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.728397 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.728671 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.728857 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.729327 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.729436 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.729467 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.729624 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.729624 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.730393 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.730523 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.730517 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.730701 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.731184 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.743183 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.731300 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.732287 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.732372 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.732447 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.732390 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.732515 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.732544 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.733895 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.734178 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.734707 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.735031 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.735111 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.734898 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.735532 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: E1125 09:24:35.735610 4776 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.743423 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.743455 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.736051 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.736281 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.736352 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.736452 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.735651 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.737869 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.738050 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.738136 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.738569 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.738954 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.739392 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: E1125 09:24:35.739725 4776 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.739899 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.740211 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.740436 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.740457 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.740525 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.741057 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.741112 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.741511 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.741549 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.741821 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.743520 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: E1125 09:24:35.743529 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:24:36.243484495 +0000 UTC m=+21.284544088 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.744345 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.744583 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.744629 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.745111 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.745650 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.745671 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.745680 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.746105 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.746353 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.746470 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.746505 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.746864 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.746972 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.747412 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.747536 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.747872 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.747927 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.745268 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.737653 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.748462 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.748567 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.748721 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.749595 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.751838 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.751845 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.752992 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: E1125 09:24:35.755728 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:24:36.255699757 +0000 UTC m=+21.296759340 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:24:35 crc kubenswrapper[4776]: E1125 09:24:35.756341 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:24:36.256295713 +0000 UTC m=+21.297355306 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.758119 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.758190 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.758508 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.758741 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.759195 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: E1125 09:24:35.762057 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:24:35 crc kubenswrapper[4776]: E1125 09:24:35.762094 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:24:35 crc kubenswrapper[4776]: E1125 09:24:35.762106 4776 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:24:35 crc kubenswrapper[4776]: E1125 09:24:35.762173 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 09:24:36.262154892 +0000 UTC m=+21.303214445 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.762613 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.763531 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.764282 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.764373 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.766221 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.766656 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.767287 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.771669 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 09:24:35 crc kubenswrapper[4776]: E1125 09:24:35.771802 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:24:35 crc kubenswrapper[4776]: E1125 09:24:35.771819 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:24:35 crc kubenswrapper[4776]: E1125 09:24:35.771833 4776 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:24:35 crc kubenswrapper[4776]: E1125 09:24:35.771886 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 09:24:36.27186519 +0000 UTC m=+21.312924733 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.771916 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.775378 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.776391 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.777388 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.779441 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.780576 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.781002 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.781062 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.781730 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.782334 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.782776 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.784978 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.785966 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.786294 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.786371 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.793799 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.802895 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.805745 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.807230 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.808296 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.815853 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.826806 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.828017 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.828560 4776 scope.go:117] "RemoveContainer" containerID="399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d" Nov 25 09:24:35 crc kubenswrapper[4776]: E1125 09:24:35.828769 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829235 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829270 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829330 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829347 4776 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829389 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829407 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829427 4776 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829440 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829449 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829458 4776 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829467 4776 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829475 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829484 4776 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829492 4776 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829502 4776 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829511 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829522 4776 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829531 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829541 4776 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829550 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829558 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829567 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829578 4776 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829587 4776 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829598 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829607 4776 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829617 4776 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829626 4776 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829635 4776 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829644 4776 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829652 4776 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829661 4776 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829670 4776 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829679 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829688 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829697 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829706 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829715 4776 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829723 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829731 4776 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829740 4776 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829748 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829757 4776 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829766 4776 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829776 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829786 4776 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829795 4776 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829803 4776 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829812 4776 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829822 4776 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829830 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829838 4776 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829847 4776 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829855 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829864 4776 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829872 4776 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829880 4776 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829888 4776 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829897 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829905 4776 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829913 4776 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829922 4776 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829931 4776 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829938 4776 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829946 4776 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829954 4776 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829989 4776 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.829997 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830006 4776 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830014 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830022 4776 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830030 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830038 4776 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830046 4776 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830053 4776 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830061 4776 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830255 4776 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830266 4776 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830276 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830284 4776 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830294 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830303 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830312 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830320 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830330 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830338 4776 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830347 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830356 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830365 4776 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830375 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830384 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830392 4776 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830400 4776 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830409 4776 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830417 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830425 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830432 4776 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830440 4776 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830448 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830457 4776 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830467 4776 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830478 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830490 4776 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830500 4776 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830510 4776 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830521 4776 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830531 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830539 4776 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830548 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830559 4776 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830570 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830582 4776 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830593 4776 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830602 4776 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830610 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830617 4776 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830625 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830633 4776 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830641 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830650 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830658 4776 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830666 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830674 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830683 4776 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830691 4776 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830699 4776 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830707 4776 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830715 4776 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830723 4776 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830731 4776 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830738 4776 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830747 4776 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830755 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830763 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830771 4776 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830780 4776 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830788 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830796 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830804 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830813 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830821 4776 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830829 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830838 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830846 4776 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830854 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830862 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830869 4776 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830877 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830885 4776 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830892 4776 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830901 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830910 4776 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830918 4776 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830926 4776 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830934 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830942 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830950 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830958 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830967 4776 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830976 4776 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830985 4776 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.830993 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.831001 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.831009 4776 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.831017 4776 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.831025 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.831033 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.831045 4776 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.831053 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.831061 4776 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.831093 4776 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.831101 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.831109 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.831118 4776 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.831125 4776 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.831133 4776 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.831141 4776 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.831149 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.831157 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.831165 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.831172 4776 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.831180 4776 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.918009 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.929700 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 09:24:35 crc kubenswrapper[4776]: I1125 09:24:35.936382 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 09:24:35 crc kubenswrapper[4776]: W1125 09:24:35.948606 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-32a5d7201fc96a359db76d11abf1946695e43dd4ef073d9f5c33197ef49e8175 WatchSource:0}: Error finding container 32a5d7201fc96a359db76d11abf1946695e43dd4ef073d9f5c33197ef49e8175: Status 404 returned error can't find the container with id 32a5d7201fc96a359db76d11abf1946695e43dd4ef073d9f5c33197ef49e8175 Nov 25 09:24:35 crc kubenswrapper[4776]: W1125 09:24:35.954132 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-d2aeb05c772ed98de024d2c15896c834267fd19aa60d045527e9416fc9e10fc5 WatchSource:0}: Error finding container d2aeb05c772ed98de024d2c15896c834267fd19aa60d045527e9416fc9e10fc5: Status 404 returned error can't find the container with id d2aeb05c772ed98de024d2c15896c834267fd19aa60d045527e9416fc9e10fc5 Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.262255 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.273552 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.274377 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.281349 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.284404 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.298295 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.311092 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.329173 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.335498 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:24:36 crc kubenswrapper[4776]: E1125 09:24:36.335666 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:24:37.335624627 +0000 UTC m=+22.376684220 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.335747 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.335837 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.335888 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.335926 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:24:36 crc kubenswrapper[4776]: E1125 09:24:36.336117 4776 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:24:36 crc kubenswrapper[4776]: E1125 09:24:36.336273 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:24:37.336259683 +0000 UTC m=+22.377319236 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:24:36 crc kubenswrapper[4776]: E1125 09:24:36.336124 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:24:36 crc kubenswrapper[4776]: E1125 09:24:36.336478 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:24:36 crc kubenswrapper[4776]: E1125 09:24:36.336172 4776 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:24:36 crc kubenswrapper[4776]: E1125 09:24:36.336624 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:24:37.336608542 +0000 UTC m=+22.377668095 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:24:36 crc kubenswrapper[4776]: E1125 09:24:36.336704 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:24:36 crc kubenswrapper[4776]: E1125 09:24:36.336764 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:24:36 crc kubenswrapper[4776]: E1125 09:24:36.336793 4776 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:24:36 crc kubenswrapper[4776]: E1125 09:24:36.336934 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 09:24:37.33691465 +0000 UTC m=+22.377974243 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:24:36 crc kubenswrapper[4776]: E1125 09:24:36.336580 4776 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:24:36 crc kubenswrapper[4776]: E1125 09:24:36.337011 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 09:24:37.336994552 +0000 UTC m=+22.378054145 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.346331 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.356185 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.365475 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.377108 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.386550 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.394293 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.402157 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.423583 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.433996 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.444384 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.820539 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"d2aeb05c772ed98de024d2c15896c834267fd19aa60d045527e9416fc9e10fc5"} Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.822163 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2"} Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.822191 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87"} Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.822201 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"32a5d7201fc96a359db76d11abf1946695e43dd4ef073d9f5c33197ef49e8175"} Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.823949 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c"} Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.824097 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"7d7dec9c18eae6a8db791dc21e8481b0047e518e1aa19e0b51913c9625e56c87"} Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.824308 4776 scope.go:117] "RemoveContainer" containerID="399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d" Nov 25 09:24:36 crc kubenswrapper[4776]: E1125 09:24:36.824417 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.835261 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:36 crc kubenswrapper[4776]: E1125 09:24:36.837761 4776 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.848458 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.869696 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.884740 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.898951 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.910936 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.923783 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.938967 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.953642 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.973276 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:36 crc kubenswrapper[4776]: I1125 09:24:36.991804 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.004805 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.019868 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.036184 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.049928 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.061738 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.345452 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.345549 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.345596 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.345640 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:24:37 crc kubenswrapper[4776]: E1125 09:24:37.345673 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:24:39.345640469 +0000 UTC m=+24.386700032 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:24:37 crc kubenswrapper[4776]: E1125 09:24:37.345742 4776 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.345753 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:24:37 crc kubenswrapper[4776]: E1125 09:24:37.345783 4776 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:24:37 crc kubenswrapper[4776]: E1125 09:24:37.345818 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:24:39.345796073 +0000 UTC m=+24.386855666 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:24:37 crc kubenswrapper[4776]: E1125 09:24:37.345877 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:24:39.345855194 +0000 UTC m=+24.386914767 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:24:37 crc kubenswrapper[4776]: E1125 09:24:37.346016 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:24:37 crc kubenswrapper[4776]: E1125 09:24:37.346115 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:24:37 crc kubenswrapper[4776]: E1125 09:24:37.346136 4776 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:24:37 crc kubenswrapper[4776]: E1125 09:24:37.346241 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 09:24:39.346199753 +0000 UTC m=+24.387259496 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:24:37 crc kubenswrapper[4776]: E1125 09:24:37.346025 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:24:37 crc kubenswrapper[4776]: E1125 09:24:37.346304 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:24:37 crc kubenswrapper[4776]: E1125 09:24:37.346340 4776 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:24:37 crc kubenswrapper[4776]: E1125 09:24:37.346458 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 09:24:39.346419199 +0000 UTC m=+24.387478822 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.661684 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:24:37 crc kubenswrapper[4776]: E1125 09:24:37.661808 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.661879 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.661904 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:24:37 crc kubenswrapper[4776]: E1125 09:24:37.662037 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:24:37 crc kubenswrapper[4776]: E1125 09:24:37.662192 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.665528 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.666581 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.667976 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.668718 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.670161 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.670688 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.671426 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.672473 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.673121 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.674145 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.674690 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.675992 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.676638 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.677231 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.678227 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.679099 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.680811 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.681260 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.682268 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.682854 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.683394 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.684038 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.684514 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.685199 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.685611 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.686224 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.686847 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.688496 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.690047 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.691191 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.692413 4776 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.692672 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.695775 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.696935 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.699044 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.702806 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.704344 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.705548 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.706445 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.708003 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.708540 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.709331 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.710118 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.710785 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.712851 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.714010 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.715089 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.716783 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.717855 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.718799 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.719774 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.720852 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.722011 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 25 09:24:37 crc kubenswrapper[4776]: I1125 09:24:37.724151 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 25 09:24:38 crc kubenswrapper[4776]: I1125 09:24:38.700511 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:24:38 crc kubenswrapper[4776]: I1125 09:24:38.701456 4776 scope.go:117] "RemoveContainer" containerID="399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d" Nov 25 09:24:38 crc kubenswrapper[4776]: E1125 09:24:38.701688 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Nov 25 09:24:38 crc kubenswrapper[4776]: I1125 09:24:38.831372 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b"} Nov 25 09:24:38 crc kubenswrapper[4776]: I1125 09:24:38.855564 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:38Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:38 crc kubenswrapper[4776]: I1125 09:24:38.869897 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:38Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:38 crc kubenswrapper[4776]: I1125 09:24:38.885448 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:38Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:38 crc kubenswrapper[4776]: I1125 09:24:38.901244 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:38Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:38 crc kubenswrapper[4776]: I1125 09:24:38.918925 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:38Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:38 crc kubenswrapper[4776]: I1125 09:24:38.935267 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:38Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:38 crc kubenswrapper[4776]: I1125 09:24:38.947439 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:38Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:38 crc kubenswrapper[4776]: I1125 09:24:38.958496 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:38Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:39 crc kubenswrapper[4776]: I1125 09:24:39.363236 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:24:39 crc kubenswrapper[4776]: I1125 09:24:39.363351 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:24:39 crc kubenswrapper[4776]: I1125 09:24:39.363393 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:24:39 crc kubenswrapper[4776]: I1125 09:24:39.363430 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:24:39 crc kubenswrapper[4776]: I1125 09:24:39.363464 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:24:39 crc kubenswrapper[4776]: E1125 09:24:39.363612 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:24:39 crc kubenswrapper[4776]: E1125 09:24:39.363616 4776 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:24:39 crc kubenswrapper[4776]: E1125 09:24:39.363638 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:24:39 crc kubenswrapper[4776]: E1125 09:24:39.363668 4776 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:24:39 crc kubenswrapper[4776]: E1125 09:24:39.363705 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:24:39 crc kubenswrapper[4776]: E1125 09:24:39.363758 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:24:39 crc kubenswrapper[4776]: E1125 09:24:39.363775 4776 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:24:39 crc kubenswrapper[4776]: E1125 09:24:39.363733 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:24:43.363703334 +0000 UTC m=+28.404762927 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:24:39 crc kubenswrapper[4776]: E1125 09:24:39.363867 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 09:24:43.363841597 +0000 UTC m=+28.404901160 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:24:39 crc kubenswrapper[4776]: E1125 09:24:39.363890 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 09:24:43.363878538 +0000 UTC m=+28.404938191 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:24:39 crc kubenswrapper[4776]: E1125 09:24:39.364275 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:24:43.364263948 +0000 UTC m=+28.405323501 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:24:39 crc kubenswrapper[4776]: E1125 09:24:39.364300 4776 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:24:39 crc kubenswrapper[4776]: E1125 09:24:39.364395 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:24:43.364370181 +0000 UTC m=+28.405429774 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:24:39 crc kubenswrapper[4776]: I1125 09:24:39.662101 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:24:39 crc kubenswrapper[4776]: I1125 09:24:39.662218 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:24:39 crc kubenswrapper[4776]: E1125 09:24:39.662252 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:24:39 crc kubenswrapper[4776]: I1125 09:24:39.662121 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:24:39 crc kubenswrapper[4776]: E1125 09:24:39.662399 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:24:39 crc kubenswrapper[4776]: E1125 09:24:39.662561 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.512019 4776 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.514626 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.514657 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.514670 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.514734 4776 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.523083 4776 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.523350 4776 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.524358 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.524379 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.524388 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.524401 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.524413 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:41Z","lastTransitionTime":"2025-11-25T09:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:41 crc kubenswrapper[4776]: E1125 09:24:41.547201 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.550997 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.551075 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.551086 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.551109 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.551122 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:41Z","lastTransitionTime":"2025-11-25T09:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:41 crc kubenswrapper[4776]: E1125 09:24:41.564027 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.567633 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.567685 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.567699 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.567720 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.567733 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:41Z","lastTransitionTime":"2025-11-25T09:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:41 crc kubenswrapper[4776]: E1125 09:24:41.585620 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.591841 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.591908 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.591923 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.591980 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.591998 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:41Z","lastTransitionTime":"2025-11-25T09:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:41 crc kubenswrapper[4776]: E1125 09:24:41.614459 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.619619 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.619680 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.619699 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.619723 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.619742 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:41Z","lastTransitionTime":"2025-11-25T09:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:41 crc kubenswrapper[4776]: E1125 09:24:41.633891 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:41 crc kubenswrapper[4776]: E1125 09:24:41.634002 4776 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.635926 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.635952 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.635962 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.635977 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.636010 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:41Z","lastTransitionTime":"2025-11-25T09:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.659308 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-hgvxs"] Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.659630 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-hgvxs" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.661526 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.661586 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:24:41 crc kubenswrapper[4776]: E1125 09:24:41.661650 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:24:41 crc kubenswrapper[4776]: E1125 09:24:41.661783 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.661593 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:24:41 crc kubenswrapper[4776]: E1125 09:24:41.661902 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.663704 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.663943 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.664167 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.684364 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.702455 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.716852 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.730008 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.739269 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.739323 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.739333 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.739352 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.739368 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:41Z","lastTransitionTime":"2025-11-25T09:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.745668 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.756230 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.783040 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.784316 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjv57\" (UniqueName: \"kubernetes.io/projected/e9adab68-3143-470e-9988-a0190c8e1bc2-kube-api-access-pjv57\") pod \"node-resolver-hgvxs\" (UID: \"e9adab68-3143-470e-9988-a0190c8e1bc2\") " pod="openshift-dns/node-resolver-hgvxs" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.784361 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e9adab68-3143-470e-9988-a0190c8e1bc2-hosts-file\") pod \"node-resolver-hgvxs\" (UID: \"e9adab68-3143-470e-9988-a0190c8e1bc2\") " pod="openshift-dns/node-resolver-hgvxs" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.802402 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.814253 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.841060 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.841111 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.841121 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.841139 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.841154 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:41Z","lastTransitionTime":"2025-11-25T09:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.885804 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e9adab68-3143-470e-9988-a0190c8e1bc2-hosts-file\") pod \"node-resolver-hgvxs\" (UID: \"e9adab68-3143-470e-9988-a0190c8e1bc2\") " pod="openshift-dns/node-resolver-hgvxs" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.885889 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjv57\" (UniqueName: \"kubernetes.io/projected/e9adab68-3143-470e-9988-a0190c8e1bc2-kube-api-access-pjv57\") pod \"node-resolver-hgvxs\" (UID: \"e9adab68-3143-470e-9988-a0190c8e1bc2\") " pod="openshift-dns/node-resolver-hgvxs" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.886207 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e9adab68-3143-470e-9988-a0190c8e1bc2-hosts-file\") pod \"node-resolver-hgvxs\" (UID: \"e9adab68-3143-470e-9988-a0190c8e1bc2\") " pod="openshift-dns/node-resolver-hgvxs" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.906813 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjv57\" (UniqueName: \"kubernetes.io/projected/e9adab68-3143-470e-9988-a0190c8e1bc2-kube-api-access-pjv57\") pod \"node-resolver-hgvxs\" (UID: \"e9adab68-3143-470e-9988-a0190c8e1bc2\") " pod="openshift-dns/node-resolver-hgvxs" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.942766 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.942808 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.942818 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.942832 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.942842 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:41Z","lastTransitionTime":"2025-11-25T09:24:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:41 crc kubenswrapper[4776]: I1125 09:24:41.973183 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-hgvxs" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.051100 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.051179 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.051305 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.051323 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.051333 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:42Z","lastTransitionTime":"2025-11-25T09:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.070669 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-5q5c6"] Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.071095 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.073195 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-4hl78"] Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.074009 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-84dqb"] Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.074199 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-4hl78" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.074346 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.076851 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.077519 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.077907 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.078342 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.078519 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.078564 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.078666 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.078971 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.079200 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.079376 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.079484 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.079584 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.103305 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.147114 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.155365 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.155399 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.155408 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.155422 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.155431 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:42Z","lastTransitionTime":"2025-11-25T09:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.164127 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.185263 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.188455 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e5a1bccd-873a-4af1-8217-c60998b613f2-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-4hl78\" (UID: \"e5a1bccd-873a-4af1-8217-c60998b613f2\") " pod="openshift-multus/multus-additional-cni-plugins-4hl78" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.188651 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-multus-socket-dir-parent\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.188734 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e5a1bccd-873a-4af1-8217-c60998b613f2-os-release\") pod \"multus-additional-cni-plugins-4hl78\" (UID: \"e5a1bccd-873a-4af1-8217-c60998b613f2\") " pod="openshift-multus/multus-additional-cni-plugins-4hl78" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.188821 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/fa47ebcc-a95e-4693-876d-7284c28c3ade-rootfs\") pod \"machine-config-daemon-84dqb\" (UID: \"fa47ebcc-a95e-4693-876d-7284c28c3ade\") " pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.188902 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-cni-binary-copy\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.188985 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8ftp\" (UniqueName: \"kubernetes.io/projected/fa47ebcc-a95e-4693-876d-7284c28c3ade-kube-api-access-t8ftp\") pod \"machine-config-daemon-84dqb\" (UID: \"fa47ebcc-a95e-4693-876d-7284c28c3ade\") " pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.189077 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-multus-conf-dir\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.189169 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-os-release\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.189260 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-host-run-netns\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.189350 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-hostroot\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.189438 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-host-run-k8s-cni-cncf-io\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.189541 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-host-run-multus-certs\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.189625 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9d5qr\" (UniqueName: \"kubernetes.io/projected/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-kube-api-access-9d5qr\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.189719 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-host-var-lib-cni-bin\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.189812 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-host-var-lib-kubelet\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.189899 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-host-var-lib-cni-multus\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.189993 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e5a1bccd-873a-4af1-8217-c60998b613f2-cni-binary-copy\") pod \"multus-additional-cni-plugins-4hl78\" (UID: \"e5a1bccd-873a-4af1-8217-c60998b613f2\") " pod="openshift-multus/multus-additional-cni-plugins-4hl78" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.190086 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kk9n\" (UniqueName: \"kubernetes.io/projected/e5a1bccd-873a-4af1-8217-c60998b613f2-kube-api-access-9kk9n\") pod \"multus-additional-cni-plugins-4hl78\" (UID: \"e5a1bccd-873a-4af1-8217-c60998b613f2\") " pod="openshift-multus/multus-additional-cni-plugins-4hl78" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.190181 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fa47ebcc-a95e-4693-876d-7284c28c3ade-mcd-auth-proxy-config\") pod \"machine-config-daemon-84dqb\" (UID: \"fa47ebcc-a95e-4693-876d-7284c28c3ade\") " pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.190269 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-multus-cni-dir\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.190350 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-etc-kubernetes\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.190444 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e5a1bccd-873a-4af1-8217-c60998b613f2-cnibin\") pod \"multus-additional-cni-plugins-4hl78\" (UID: \"e5a1bccd-873a-4af1-8217-c60998b613f2\") " pod="openshift-multus/multus-additional-cni-plugins-4hl78" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.190531 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-system-cni-dir\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.190616 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-cnibin\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.190700 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-multus-daemon-config\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.190788 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e5a1bccd-873a-4af1-8217-c60998b613f2-system-cni-dir\") pod \"multus-additional-cni-plugins-4hl78\" (UID: \"e5a1bccd-873a-4af1-8217-c60998b613f2\") " pod="openshift-multus/multus-additional-cni-plugins-4hl78" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.190869 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e5a1bccd-873a-4af1-8217-c60998b613f2-tuning-conf-dir\") pod \"multus-additional-cni-plugins-4hl78\" (UID: \"e5a1bccd-873a-4af1-8217-c60998b613f2\") " pod="openshift-multus/multus-additional-cni-plugins-4hl78" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.190950 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fa47ebcc-a95e-4693-876d-7284c28c3ade-proxy-tls\") pod \"machine-config-daemon-84dqb\" (UID: \"fa47ebcc-a95e-4693-876d-7284c28c3ade\") " pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.207287 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.221743 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.239713 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.257474 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.257524 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.257535 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.257555 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.257567 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:42Z","lastTransitionTime":"2025-11-25T09:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.259315 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.269378 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.279456 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.292348 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fa47ebcc-a95e-4693-876d-7284c28c3ade-proxy-tls\") pod \"machine-config-daemon-84dqb\" (UID: \"fa47ebcc-a95e-4693-876d-7284c28c3ade\") " pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.292393 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e5a1bccd-873a-4af1-8217-c60998b613f2-system-cni-dir\") pod \"multus-additional-cni-plugins-4hl78\" (UID: \"e5a1bccd-873a-4af1-8217-c60998b613f2\") " pod="openshift-multus/multus-additional-cni-plugins-4hl78" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.292411 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e5a1bccd-873a-4af1-8217-c60998b613f2-tuning-conf-dir\") pod \"multus-additional-cni-plugins-4hl78\" (UID: \"e5a1bccd-873a-4af1-8217-c60998b613f2\") " pod="openshift-multus/multus-additional-cni-plugins-4hl78" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.292431 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e5a1bccd-873a-4af1-8217-c60998b613f2-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-4hl78\" (UID: \"e5a1bccd-873a-4af1-8217-c60998b613f2\") " pod="openshift-multus/multus-additional-cni-plugins-4hl78" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.292495 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e5a1bccd-873a-4af1-8217-c60998b613f2-system-cni-dir\") pod \"multus-additional-cni-plugins-4hl78\" (UID: \"e5a1bccd-873a-4af1-8217-c60998b613f2\") " pod="openshift-multus/multus-additional-cni-plugins-4hl78" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.292577 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-multus-socket-dir-parent\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.292601 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e5a1bccd-873a-4af1-8217-c60998b613f2-os-release\") pod \"multus-additional-cni-plugins-4hl78\" (UID: \"e5a1bccd-873a-4af1-8217-c60998b613f2\") " pod="openshift-multus/multus-additional-cni-plugins-4hl78" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.292621 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/fa47ebcc-a95e-4693-876d-7284c28c3ade-rootfs\") pod \"machine-config-daemon-84dqb\" (UID: \"fa47ebcc-a95e-4693-876d-7284c28c3ade\") " pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.292645 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-cni-binary-copy\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.292665 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8ftp\" (UniqueName: \"kubernetes.io/projected/fa47ebcc-a95e-4693-876d-7284c28c3ade-kube-api-access-t8ftp\") pod \"machine-config-daemon-84dqb\" (UID: \"fa47ebcc-a95e-4693-876d-7284c28c3ade\") " pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.292689 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-multus-conf-dir\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.292711 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-os-release\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.292729 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-host-run-netns\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.292747 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-hostroot\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.292766 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9d5qr\" (UniqueName: \"kubernetes.io/projected/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-kube-api-access-9d5qr\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.292795 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-host-run-k8s-cni-cncf-io\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.292815 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-host-run-multus-certs\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.292837 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-host-var-lib-cni-bin\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.292867 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-host-var-lib-kubelet\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.292917 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kk9n\" (UniqueName: \"kubernetes.io/projected/e5a1bccd-873a-4af1-8217-c60998b613f2-kube-api-access-9kk9n\") pod \"multus-additional-cni-plugins-4hl78\" (UID: \"e5a1bccd-873a-4af1-8217-c60998b613f2\") " pod="openshift-multus/multus-additional-cni-plugins-4hl78" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.292940 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fa47ebcc-a95e-4693-876d-7284c28c3ade-mcd-auth-proxy-config\") pod \"machine-config-daemon-84dqb\" (UID: \"fa47ebcc-a95e-4693-876d-7284c28c3ade\") " pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.292960 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-multus-cni-dir\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.292979 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-host-var-lib-cni-multus\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.293009 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e5a1bccd-873a-4af1-8217-c60998b613f2-cni-binary-copy\") pod \"multus-additional-cni-plugins-4hl78\" (UID: \"e5a1bccd-873a-4af1-8217-c60998b613f2\") " pod="openshift-multus/multus-additional-cni-plugins-4hl78" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.293030 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-etc-kubernetes\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.293061 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e5a1bccd-873a-4af1-8217-c60998b613f2-cnibin\") pod \"multus-additional-cni-plugins-4hl78\" (UID: \"e5a1bccd-873a-4af1-8217-c60998b613f2\") " pod="openshift-multus/multus-additional-cni-plugins-4hl78" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.293103 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-system-cni-dir\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.293121 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-cnibin\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.293139 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-multus-daemon-config\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.293460 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e5a1bccd-873a-4af1-8217-c60998b613f2-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-4hl78\" (UID: \"e5a1bccd-873a-4af1-8217-c60998b613f2\") " pod="openshift-multus/multus-additional-cni-plugins-4hl78" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.293471 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-host-run-multus-certs\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.293528 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-multus-conf-dir\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.293557 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-os-release\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.293580 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-multus-socket-dir-parent\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.293612 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-hostroot\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.293661 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e5a1bccd-873a-4af1-8217-c60998b613f2-os-release\") pod \"multus-additional-cni-plugins-4hl78\" (UID: \"e5a1bccd-873a-4af1-8217-c60998b613f2\") " pod="openshift-multus/multus-additional-cni-plugins-4hl78" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.293695 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/fa47ebcc-a95e-4693-876d-7284c28c3ade-rootfs\") pod \"machine-config-daemon-84dqb\" (UID: \"fa47ebcc-a95e-4693-876d-7284c28c3ade\") " pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.293758 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-multus-daemon-config\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.293832 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-host-var-lib-cni-bin\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.293866 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-host-run-k8s-cni-cncf-io\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.293884 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-host-var-lib-kubelet\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.293586 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-host-run-netns\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.294180 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-etc-kubernetes\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.294212 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-cni-binary-copy\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.294237 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e5a1bccd-873a-4af1-8217-c60998b613f2-cnibin\") pod \"multus-additional-cni-plugins-4hl78\" (UID: \"e5a1bccd-873a-4af1-8217-c60998b613f2\") " pod="openshift-multus/multus-additional-cni-plugins-4hl78" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.294349 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-multus-cni-dir\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.294467 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-system-cni-dir\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.294507 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e5a1bccd-873a-4af1-8217-c60998b613f2-cni-binary-copy\") pod \"multus-additional-cni-plugins-4hl78\" (UID: \"e5a1bccd-873a-4af1-8217-c60998b613f2\") " pod="openshift-multus/multus-additional-cni-plugins-4hl78" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.294520 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-cnibin\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.294551 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-host-var-lib-cni-multus\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.294602 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e5a1bccd-873a-4af1-8217-c60998b613f2-tuning-conf-dir\") pod \"multus-additional-cni-plugins-4hl78\" (UID: \"e5a1bccd-873a-4af1-8217-c60998b613f2\") " pod="openshift-multus/multus-additional-cni-plugins-4hl78" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.294808 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fa47ebcc-a95e-4693-876d-7284c28c3ade-mcd-auth-proxy-config\") pod \"machine-config-daemon-84dqb\" (UID: \"fa47ebcc-a95e-4693-876d-7284c28c3ade\") " pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.296896 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fa47ebcc-a95e-4693-876d-7284c28c3ade-proxy-tls\") pod \"machine-config-daemon-84dqb\" (UID: \"fa47ebcc-a95e-4693-876d-7284c28c3ade\") " pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.299588 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.309019 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9d5qr\" (UniqueName: \"kubernetes.io/projected/d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c-kube-api-access-9d5qr\") pod \"multus-5q5c6\" (UID: \"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\") " pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.310746 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.314681 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8ftp\" (UniqueName: \"kubernetes.io/projected/fa47ebcc-a95e-4693-876d-7284c28c3ade-kube-api-access-t8ftp\") pod \"machine-config-daemon-84dqb\" (UID: \"fa47ebcc-a95e-4693-876d-7284c28c3ade\") " pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.315013 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kk9n\" (UniqueName: \"kubernetes.io/projected/e5a1bccd-873a-4af1-8217-c60998b613f2-kube-api-access-9kk9n\") pod \"multus-additional-cni-plugins-4hl78\" (UID: \"e5a1bccd-873a-4af1-8217-c60998b613f2\") " pod="openshift-multus/multus-additional-cni-plugins-4hl78" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.323916 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.335740 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.346779 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.358223 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.359328 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.359365 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.359378 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.359396 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.359408 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:42Z","lastTransitionTime":"2025-11-25T09:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.369675 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.382597 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.389208 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-5q5c6" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.395232 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-4hl78" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.397878 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: W1125 09:24:42.400893 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd4c37fb2_c7d2_4155_8dd9_13ed9c7d3f8c.slice/crio-dc57ac7a9118fc2776bf9c9d9b25cba6f51c5c5066897877cfe19302c181a777 WatchSource:0}: Error finding container dc57ac7a9118fc2776bf9c9d9b25cba6f51c5c5066897877cfe19302c181a777: Status 404 returned error can't find the container with id dc57ac7a9118fc2776bf9c9d9b25cba6f51c5c5066897877cfe19302c181a777 Nov 25 09:24:42 crc kubenswrapper[4776]: W1125 09:24:42.407211 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode5a1bccd_873a_4af1_8217_c60998b613f2.slice/crio-a89367616c843e8b7da95279e72f12190eb8daa42676f8c05a3a9476f7434573 WatchSource:0}: Error finding container a89367616c843e8b7da95279e72f12190eb8daa42676f8c05a3a9476f7434573: Status 404 returned error can't find the container with id a89367616c843e8b7da95279e72f12190eb8daa42676f8c05a3a9476f7434573 Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.411773 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.411771 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.423328 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: W1125 09:24:42.429478 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa47ebcc_a95e_4693_876d_7284c28c3ade.slice/crio-6c3223a1ea75564da7d74e13d9562dc58417da35b96460381d9088ff6db73ed1 WatchSource:0}: Error finding container 6c3223a1ea75564da7d74e13d9562dc58417da35b96460381d9088ff6db73ed1: Status 404 returned error can't find the container with id 6c3223a1ea75564da7d74e13d9562dc58417da35b96460381d9088ff6db73ed1 Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.433152 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.465001 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.465041 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.465057 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.465097 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.465112 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:42Z","lastTransitionTime":"2025-11-25T09:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.473842 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-r7x57"] Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.474729 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.479857 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.480213 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.480517 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.480757 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.484207 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.484553 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.484838 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.497904 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.511332 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.524855 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.537839 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.540266 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.547439 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.548619 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.553152 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.567028 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.567073 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.567082 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.567095 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.567105 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:42Z","lastTransitionTime":"2025-11-25T09:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.570571 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.582908 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.597338 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.597379 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/34502968-eb8d-47fa-83c2-7f6f6735ce2d-ovnkube-config\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.597397 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-etc-openvswitch\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.597416 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-run-ovn\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.597588 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-cni-bin\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.597670 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/34502968-eb8d-47fa-83c2-7f6f6735ce2d-ovnkube-script-lib\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.597717 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-kubelet\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.597754 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-run-systemd\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.597782 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/34502968-eb8d-47fa-83c2-7f6f6735ce2d-env-overrides\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.597919 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-run-openvswitch\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.597967 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-slash\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.597994 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-cni-netd\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.598041 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-var-lib-openvswitch\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.598085 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmdxk\" (UniqueName: \"kubernetes.io/projected/34502968-eb8d-47fa-83c2-7f6f6735ce2d-kube-api-access-lmdxk\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.598111 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-run-ovn-kubernetes\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.598159 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-systemd-units\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.598181 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-run-netns\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.598229 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-node-log\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.598350 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-log-socket\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.598436 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/34502968-eb8d-47fa-83c2-7f6f6735ce2d-ovn-node-metrics-cert\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.599692 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.611288 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.623376 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.642493 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.655618 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.668847 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.669378 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.669410 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.669422 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.669439 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.669450 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:42Z","lastTransitionTime":"2025-11-25T09:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.680313 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.696627 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.698920 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-run-ovn-kubernetes\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.698946 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-systemd-units\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.698961 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-run-netns\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.698982 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-node-log\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.698996 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-log-socket\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699010 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/34502968-eb8d-47fa-83c2-7f6f6735ce2d-ovn-node-metrics-cert\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699026 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-etc-openvswitch\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699041 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-run-ovn\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699102 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699102 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-systemd-units\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699139 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-run-ovn-kubernetes\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699155 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-etc-openvswitch\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699184 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-run-netns\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699118 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-log-socket\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699158 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-node-log\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699194 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-run-ovn\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699142 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/34502968-eb8d-47fa-83c2-7f6f6735ce2d-ovnkube-config\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699212 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699425 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-cni-bin\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699452 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/34502968-eb8d-47fa-83c2-7f6f6735ce2d-ovnkube-script-lib\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699482 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-kubelet\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699501 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-run-systemd\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699510 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-cni-bin\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699603 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-kubelet\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699518 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/34502968-eb8d-47fa-83c2-7f6f6735ce2d-env-overrides\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699661 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-run-openvswitch\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699678 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-slash\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699694 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-cni-netd\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699709 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-var-lib-openvswitch\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699729 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmdxk\" (UniqueName: \"kubernetes.io/projected/34502968-eb8d-47fa-83c2-7f6f6735ce2d-kube-api-access-lmdxk\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699928 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-run-openvswitch\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699953 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-slash\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699974 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-cni-netd\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699982 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/34502968-eb8d-47fa-83c2-7f6f6735ce2d-env-overrides\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.699996 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-var-lib-openvswitch\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.700009 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/34502968-eb8d-47fa-83c2-7f6f6735ce2d-ovnkube-config\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.700020 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-run-systemd\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.700140 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/34502968-eb8d-47fa-83c2-7f6f6735ce2d-ovnkube-script-lib\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.702402 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/34502968-eb8d-47fa-83c2-7f6f6735ce2d-ovn-node-metrics-cert\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.710256 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.716376 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmdxk\" (UniqueName: \"kubernetes.io/projected/34502968-eb8d-47fa-83c2-7f6f6735ce2d-kube-api-access-lmdxk\") pod \"ovnkube-node-r7x57\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.722765 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.733754 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.745773 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.757557 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.768571 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.772052 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.772099 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.772109 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.772125 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.772133 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:42Z","lastTransitionTime":"2025-11-25T09:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.780534 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.793575 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.801190 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.810769 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.833394 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.843666 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" event={"ID":"34502968-eb8d-47fa-83c2-7f6f6735ce2d","Type":"ContainerStarted","Data":"963473c38a206a0d639545c34a261dca74521650e71b6eedd3a3275052faeb01"} Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.845620 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7"} Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.845646 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb"} Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.845656 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"6c3223a1ea75564da7d74e13d9562dc58417da35b96460381d9088ff6db73ed1"} Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.847009 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" event={"ID":"e5a1bccd-873a-4af1-8217-c60998b613f2","Type":"ContainerStarted","Data":"08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127"} Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.847031 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" event={"ID":"e5a1bccd-873a-4af1-8217-c60998b613f2","Type":"ContainerStarted","Data":"a89367616c843e8b7da95279e72f12190eb8daa42676f8c05a3a9476f7434573"} Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.847922 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-hgvxs" event={"ID":"e9adab68-3143-470e-9988-a0190c8e1bc2","Type":"ContainerStarted","Data":"23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342"} Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.847942 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-hgvxs" event={"ID":"e9adab68-3143-470e-9988-a0190c8e1bc2","Type":"ContainerStarted","Data":"c6f08ba965f7dcf46625a9b265ca040e417e4e3e1f5402f011db8f49f222c412"} Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.850051 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5q5c6" event={"ID":"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c","Type":"ContainerStarted","Data":"c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9"} Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.850086 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5q5c6" event={"ID":"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c","Type":"ContainerStarted","Data":"dc57ac7a9118fc2776bf9c9d9b25cba6f51c5c5066897877cfe19302c181a777"} Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.855762 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.870445 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.874313 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.874354 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.874365 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.874384 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.874397 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:42Z","lastTransitionTime":"2025-11-25T09:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.881944 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.895278 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.913880 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.929528 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.944701 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.958756 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.973752 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.976655 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.976694 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.976703 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.976718 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.976727 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:42Z","lastTransitionTime":"2025-11-25T09:24:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:42 crc kubenswrapper[4776]: I1125 09:24:42.986547 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.013016 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.034882 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.062483 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.075452 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.078647 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.078689 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.078701 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.078719 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.078731 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:43Z","lastTransitionTime":"2025-11-25T09:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.094390 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.109623 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.181915 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.181963 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.181972 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.181988 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.181998 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:43Z","lastTransitionTime":"2025-11-25T09:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.284485 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.284526 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.284539 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.284563 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.284576 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:43Z","lastTransitionTime":"2025-11-25T09:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.387476 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.387525 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.387533 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.387548 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.387558 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:43Z","lastTransitionTime":"2025-11-25T09:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.407389 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.407508 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.407541 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:24:43 crc kubenswrapper[4776]: E1125 09:24:43.407657 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:24:43 crc kubenswrapper[4776]: E1125 09:24:43.407660 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:24:51.407605701 +0000 UTC m=+36.448665274 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:24:43 crc kubenswrapper[4776]: E1125 09:24:43.407677 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:24:43 crc kubenswrapper[4776]: E1125 09:24:43.407740 4776 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:24:43 crc kubenswrapper[4776]: E1125 09:24:43.407762 4776 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.408026 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:24:43 crc kubenswrapper[4776]: E1125 09:24:43.408054 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:24:51.408027672 +0000 UTC m=+36.449087395 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:24:43 crc kubenswrapper[4776]: E1125 09:24:43.408194 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 09:24:51.408153715 +0000 UTC m=+36.449213308 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:24:43 crc kubenswrapper[4776]: E1125 09:24:43.408113 4776 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.408226 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:24:43 crc kubenswrapper[4776]: E1125 09:24:43.408276 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:24:51.408255798 +0000 UTC m=+36.449315361 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:24:43 crc kubenswrapper[4776]: E1125 09:24:43.408436 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:24:43 crc kubenswrapper[4776]: E1125 09:24:43.408476 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:24:43 crc kubenswrapper[4776]: E1125 09:24:43.408498 4776 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:24:43 crc kubenswrapper[4776]: E1125 09:24:43.408575 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 09:24:51.408558575 +0000 UTC m=+36.449618168 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.490167 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.490226 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.490238 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.490258 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.490270 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:43Z","lastTransitionTime":"2025-11-25T09:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.594371 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.594876 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.594920 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.594950 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.594979 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:43Z","lastTransitionTime":"2025-11-25T09:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.661389 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.661510 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:24:43 crc kubenswrapper[4776]: E1125 09:24:43.661571 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.661605 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:24:43 crc kubenswrapper[4776]: E1125 09:24:43.661748 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:24:43 crc kubenswrapper[4776]: E1125 09:24:43.661869 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.697875 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.697932 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.697950 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.697973 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.697989 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:43Z","lastTransitionTime":"2025-11-25T09:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.801490 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.801536 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.801548 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.801566 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.801577 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:43Z","lastTransitionTime":"2025-11-25T09:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.854379 4776 generic.go:334] "Generic (PLEG): container finished" podID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerID="a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b" exitCode=0 Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.854456 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" event={"ID":"34502968-eb8d-47fa-83c2-7f6f6735ce2d","Type":"ContainerDied","Data":"a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b"} Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.855981 4776 generic.go:334] "Generic (PLEG): container finished" podID="e5a1bccd-873a-4af1-8217-c60998b613f2" containerID="08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127" exitCode=0 Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.856024 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" event={"ID":"e5a1bccd-873a-4af1-8217-c60998b613f2","Type":"ContainerDied","Data":"08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127"} Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.885672 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.901552 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.905340 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.905394 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.905410 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.905432 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.905448 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:43Z","lastTransitionTime":"2025-11-25T09:24:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.917141 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.931051 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.942753 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.957788 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.975483 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:43 crc kubenswrapper[4776]: I1125 09:24:43.998353 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.008466 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.008839 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.008849 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.008867 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.008880 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:44Z","lastTransitionTime":"2025-11-25T09:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.012461 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.025672 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.039133 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.060651 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.125710 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.125747 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.125872 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.125890 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.125903 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:44Z","lastTransitionTime":"2025-11-25T09:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.130588 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.146899 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.161419 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.187178 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.204182 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.216277 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.228742 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.229244 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.229368 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.229492 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.229631 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:44Z","lastTransitionTime":"2025-11-25T09:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.234117 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.249111 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.262937 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.284929 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.305206 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.324403 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.332552 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.332591 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.332603 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.332619 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.332629 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:44Z","lastTransitionTime":"2025-11-25T09:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.335787 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.347495 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.368881 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.388430 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.435392 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.435423 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.435431 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.435444 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.435455 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:44Z","lastTransitionTime":"2025-11-25T09:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.538302 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.538336 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.538344 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.538360 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.538370 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:44Z","lastTransitionTime":"2025-11-25T09:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.640303 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.640335 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.640344 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.640358 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.640370 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:44Z","lastTransitionTime":"2025-11-25T09:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.666215 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-6gp8r"] Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.666530 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-6gp8r" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.668394 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.668668 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.668999 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.669038 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.683159 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.698783 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.710169 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.723336 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.742235 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.742278 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.742286 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.742301 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.742310 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:44Z","lastTransitionTime":"2025-11-25T09:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.745274 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.755627 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.770392 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.788935 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.806527 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.817820 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.821522 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/0d194bcf-a8c4-4dd9-8239-d4172c5480e3-serviceca\") pod \"node-ca-6gp8r\" (UID: \"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\") " pod="openshift-image-registry/node-ca-6gp8r" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.821603 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cx8n\" (UniqueName: \"kubernetes.io/projected/0d194bcf-a8c4-4dd9-8239-d4172c5480e3-kube-api-access-4cx8n\") pod \"node-ca-6gp8r\" (UID: \"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\") " pod="openshift-image-registry/node-ca-6gp8r" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.821808 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d194bcf-a8c4-4dd9-8239-d4172c5480e3-host\") pod \"node-ca-6gp8r\" (UID: \"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\") " pod="openshift-image-registry/node-ca-6gp8r" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.830533 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.844710 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.844677 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.844768 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.844789 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.844816 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.844833 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:44Z","lastTransitionTime":"2025-11-25T09:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.870569 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" event={"ID":"34502968-eb8d-47fa-83c2-7f6f6735ce2d","Type":"ContainerStarted","Data":"8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49"} Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.870628 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" event={"ID":"34502968-eb8d-47fa-83c2-7f6f6735ce2d","Type":"ContainerStarted","Data":"55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516"} Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.870656 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" event={"ID":"34502968-eb8d-47fa-83c2-7f6f6735ce2d","Type":"ContainerStarted","Data":"c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295"} Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.870680 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" event={"ID":"34502968-eb8d-47fa-83c2-7f6f6735ce2d","Type":"ContainerStarted","Data":"c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188"} Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.870702 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" event={"ID":"34502968-eb8d-47fa-83c2-7f6f6735ce2d","Type":"ContainerStarted","Data":"954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760"} Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.870729 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" event={"ID":"34502968-eb8d-47fa-83c2-7f6f6735ce2d","Type":"ContainerStarted","Data":"af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1"} Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.870730 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.874318 4776 generic.go:334] "Generic (PLEG): container finished" podID="e5a1bccd-873a-4af1-8217-c60998b613f2" containerID="9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860" exitCode=0 Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.874387 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" event={"ID":"e5a1bccd-873a-4af1-8217-c60998b613f2","Type":"ContainerDied","Data":"9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860"} Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.894787 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.913628 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.922947 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cx8n\" (UniqueName: \"kubernetes.io/projected/0d194bcf-a8c4-4dd9-8239-d4172c5480e3-kube-api-access-4cx8n\") pod \"node-ca-6gp8r\" (UID: \"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\") " pod="openshift-image-registry/node-ca-6gp8r" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.923182 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d194bcf-a8c4-4dd9-8239-d4172c5480e3-host\") pod \"node-ca-6gp8r\" (UID: \"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\") " pod="openshift-image-registry/node-ca-6gp8r" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.923217 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/0d194bcf-a8c4-4dd9-8239-d4172c5480e3-serviceca\") pod \"node-ca-6gp8r\" (UID: \"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\") " pod="openshift-image-registry/node-ca-6gp8r" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.923287 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d194bcf-a8c4-4dd9-8239-d4172c5480e3-host\") pod \"node-ca-6gp8r\" (UID: \"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\") " pod="openshift-image-registry/node-ca-6gp8r" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.924815 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/0d194bcf-a8c4-4dd9-8239-d4172c5480e3-serviceca\") pod \"node-ca-6gp8r\" (UID: \"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\") " pod="openshift-image-registry/node-ca-6gp8r" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.947661 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.947991 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.948024 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.948046 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.948304 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:44Z","lastTransitionTime":"2025-11-25T09:24:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.974117 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cx8n\" (UniqueName: \"kubernetes.io/projected/0d194bcf-a8c4-4dd9-8239-d4172c5480e3-kube-api-access-4cx8n\") pod \"node-ca-6gp8r\" (UID: \"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\") " pod="openshift-image-registry/node-ca-6gp8r" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.974354 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:44 crc kubenswrapper[4776]: I1125 09:24:44.978128 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-6gp8r" Nov 25 09:24:45 crc kubenswrapper[4776]: W1125 09:24:45.002034 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0d194bcf_a8c4_4dd9_8239_d4172c5480e3.slice/crio-fd75e30c237a593735bbd03f754af23f8dc61692822c77a5d8dd116d817520dd WatchSource:0}: Error finding container fd75e30c237a593735bbd03f754af23f8dc61692822c77a5d8dd116d817520dd: Status 404 returned error can't find the container with id fd75e30c237a593735bbd03f754af23f8dc61692822c77a5d8dd116d817520dd Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.013560 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.049569 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.051571 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.051614 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.051626 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.051644 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.051933 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:45Z","lastTransitionTime":"2025-11-25T09:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.089809 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.139179 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.154716 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.154779 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.154798 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.154827 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.154845 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:45Z","lastTransitionTime":"2025-11-25T09:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.169503 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.215171 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.249052 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.257952 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.258010 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.258028 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.258054 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.258102 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:45Z","lastTransitionTime":"2025-11-25T09:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.295809 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.333763 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.360160 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.360203 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.360213 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.360230 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.360242 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:45Z","lastTransitionTime":"2025-11-25T09:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.374709 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.411466 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.453630 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.463817 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.463884 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.463907 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.463937 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.463960 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:45Z","lastTransitionTime":"2025-11-25T09:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.494261 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.535491 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.566244 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.566289 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.566317 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.566334 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.566345 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:45Z","lastTransitionTime":"2025-11-25T09:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.661601 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.661676 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:24:45 crc kubenswrapper[4776]: E1125 09:24:45.661752 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.661757 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:24:45 crc kubenswrapper[4776]: E1125 09:24:45.661850 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:24:45 crc kubenswrapper[4776]: E1125 09:24:45.661956 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.668440 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.668475 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.668484 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.668497 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.668506 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:45Z","lastTransitionTime":"2025-11-25T09:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.677226 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.699921 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.713939 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.728431 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.744712 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.770410 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.770467 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.770486 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.770513 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.770535 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:45Z","lastTransitionTime":"2025-11-25T09:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.790381 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.812382 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.850737 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.873000 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.873044 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.873053 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.873082 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.873095 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:45Z","lastTransitionTime":"2025-11-25T09:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.878095 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-6gp8r" event={"ID":"0d194bcf-a8c4-4dd9-8239-d4172c5480e3","Type":"ContainerStarted","Data":"a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279"} Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.878155 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-6gp8r" event={"ID":"0d194bcf-a8c4-4dd9-8239-d4172c5480e3","Type":"ContainerStarted","Data":"fd75e30c237a593735bbd03f754af23f8dc61692822c77a5d8dd116d817520dd"} Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.880288 4776 generic.go:334] "Generic (PLEG): container finished" podID="e5a1bccd-873a-4af1-8217-c60998b613f2" containerID="415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7" exitCode=0 Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.880319 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" event={"ID":"e5a1bccd-873a-4af1-8217-c60998b613f2","Type":"ContainerDied","Data":"415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7"} Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.895210 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.939976 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.975106 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.975146 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.975157 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.975173 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.975184 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:45Z","lastTransitionTime":"2025-11-25T09:24:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:45 crc kubenswrapper[4776]: I1125 09:24:45.975397 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.014359 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.050979 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.076934 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.076991 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.077019 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.077116 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.077150 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:46Z","lastTransitionTime":"2025-11-25T09:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.106592 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.131740 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.185238 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.185292 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.185309 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.185332 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.185349 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:46Z","lastTransitionTime":"2025-11-25T09:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.211363 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.230043 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.255247 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.292479 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.292518 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.292528 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.292564 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.292575 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:46Z","lastTransitionTime":"2025-11-25T09:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.295707 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.330759 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.377142 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.395260 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.395297 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.395306 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.395320 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.395330 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:46Z","lastTransitionTime":"2025-11-25T09:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.417794 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.462569 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.494731 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.497881 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.497918 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.497926 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.497941 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.497950 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:46Z","lastTransitionTime":"2025-11-25T09:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.531455 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.575014 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.599533 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.599588 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.599610 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.599629 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.599641 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:46Z","lastTransitionTime":"2025-11-25T09:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.618659 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.654134 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.690578 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.702612 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.702689 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.702706 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.702991 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.703050 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:46Z","lastTransitionTime":"2025-11-25T09:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.730483 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.806711 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.806761 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.806774 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.806793 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.806806 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:46Z","lastTransitionTime":"2025-11-25T09:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.888317 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" event={"ID":"34502968-eb8d-47fa-83c2-7f6f6735ce2d","Type":"ContainerStarted","Data":"957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5"} Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.891403 4776 generic.go:334] "Generic (PLEG): container finished" podID="e5a1bccd-873a-4af1-8217-c60998b613f2" containerID="7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76" exitCode=0 Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.891440 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" event={"ID":"e5a1bccd-873a-4af1-8217-c60998b613f2","Type":"ContainerDied","Data":"7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76"} Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.909153 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.909190 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.909205 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.909222 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.909234 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:46Z","lastTransitionTime":"2025-11-25T09:24:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.910642 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.932395 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.953184 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.968566 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.983879 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:46 crc kubenswrapper[4776]: I1125 09:24:46.998626 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.012658 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.012709 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.012721 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.012738 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.012750 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:47Z","lastTransitionTime":"2025-11-25T09:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.013388 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.056735 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.090180 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.115394 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.115426 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.115434 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.115449 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.115457 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:47Z","lastTransitionTime":"2025-11-25T09:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.135823 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.173634 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.214016 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.218730 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.218789 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.218809 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.218841 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.218865 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:47Z","lastTransitionTime":"2025-11-25T09:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.252541 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.291961 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.335494 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.335547 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.335559 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.335579 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.335595 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:47Z","lastTransitionTime":"2025-11-25T09:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.339476 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.438759 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.438825 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.438845 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.438868 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.438886 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:47Z","lastTransitionTime":"2025-11-25T09:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.541993 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.542056 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.542110 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.542140 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.542163 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:47Z","lastTransitionTime":"2025-11-25T09:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.644912 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.644978 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.645002 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.645030 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.645052 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:47Z","lastTransitionTime":"2025-11-25T09:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.663187 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:24:47 crc kubenswrapper[4776]: E1125 09:24:47.663405 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.663460 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.663556 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:24:47 crc kubenswrapper[4776]: E1125 09:24:47.663688 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:24:47 crc kubenswrapper[4776]: E1125 09:24:47.663777 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.747371 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.747403 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.747413 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.747428 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.747439 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:47Z","lastTransitionTime":"2025-11-25T09:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.849436 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.849476 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.849489 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.849504 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.849513 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:47Z","lastTransitionTime":"2025-11-25T09:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.899050 4776 generic.go:334] "Generic (PLEG): container finished" podID="e5a1bccd-873a-4af1-8217-c60998b613f2" containerID="62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb" exitCode=0 Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.899135 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" event={"ID":"e5a1bccd-873a-4af1-8217-c60998b613f2","Type":"ContainerDied","Data":"62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb"} Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.923897 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.944088 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.952446 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.952510 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.952536 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.952569 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.952596 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:47Z","lastTransitionTime":"2025-11-25T09:24:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.965940 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:47 crc kubenswrapper[4776]: I1125 09:24:47.986108 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.013252 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.027584 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.045807 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.057163 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.057193 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.057205 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.057221 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.057231 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:48Z","lastTransitionTime":"2025-11-25T09:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.062565 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.082523 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.100379 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.113339 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.128730 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.143303 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.159741 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.159794 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.159811 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.159837 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.159854 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:48Z","lastTransitionTime":"2025-11-25T09:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.173945 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.186767 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.263521 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.263580 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.263597 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.263620 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.263638 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:48Z","lastTransitionTime":"2025-11-25T09:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.365932 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.365986 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.366000 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.366022 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.366037 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:48Z","lastTransitionTime":"2025-11-25T09:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.469302 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.469383 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.469402 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.469430 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.469449 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:48Z","lastTransitionTime":"2025-11-25T09:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.572812 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.572879 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.572897 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.572921 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.572939 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:48Z","lastTransitionTime":"2025-11-25T09:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.675718 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.675767 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.675782 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.675804 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.675820 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:48Z","lastTransitionTime":"2025-11-25T09:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.792747 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.792828 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.792850 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.792891 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.792911 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:48Z","lastTransitionTime":"2025-11-25T09:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.896487 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.896596 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.896617 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.896684 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.896703 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:48Z","lastTransitionTime":"2025-11-25T09:24:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.907745 4776 generic.go:334] "Generic (PLEG): container finished" podID="e5a1bccd-873a-4af1-8217-c60998b613f2" containerID="3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25" exitCode=0 Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.907803 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" event={"ID":"e5a1bccd-873a-4af1-8217-c60998b613f2","Type":"ContainerDied","Data":"3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25"} Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.927550 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.947752 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.966478 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:48 crc kubenswrapper[4776]: I1125 09:24:48.988403 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.000801 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.001593 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.001645 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.001660 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.001682 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.001698 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:49Z","lastTransitionTime":"2025-11-25T09:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.034380 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.047881 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.061789 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.076281 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.089422 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.104283 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.104329 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.104337 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.104355 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.104363 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:49Z","lastTransitionTime":"2025-11-25T09:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.107312 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.126328 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.144559 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.163519 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.174422 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.206461 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.206498 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.206507 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.206521 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.206530 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:49Z","lastTransitionTime":"2025-11-25T09:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.309685 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.309752 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.309769 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.309792 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.309807 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:49Z","lastTransitionTime":"2025-11-25T09:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.412378 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.412418 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.412431 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.412447 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.412459 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:49Z","lastTransitionTime":"2025-11-25T09:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.515110 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.515178 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.515222 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.515259 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.515283 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:49Z","lastTransitionTime":"2025-11-25T09:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.618681 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.618729 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.618743 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.618762 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.618778 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:49Z","lastTransitionTime":"2025-11-25T09:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.662359 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.662361 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:24:49 crc kubenswrapper[4776]: E1125 09:24:49.662511 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.662389 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:24:49 crc kubenswrapper[4776]: E1125 09:24:49.662631 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:24:49 crc kubenswrapper[4776]: E1125 09:24:49.662739 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.721420 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.721480 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.721499 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.721523 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.721541 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:49Z","lastTransitionTime":"2025-11-25T09:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.825122 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.825179 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.825197 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.825221 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.825244 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:49Z","lastTransitionTime":"2025-11-25T09:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.918330 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" event={"ID":"34502968-eb8d-47fa-83c2-7f6f6735ce2d","Type":"ContainerStarted","Data":"9ea3897d86a1e19fb417f741e9a4d594f143bb3679f0ab458e0ca7b116710880"} Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.918380 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.918450 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.929094 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" event={"ID":"e5a1bccd-873a-4af1-8217-c60998b613f2","Type":"ContainerStarted","Data":"f2ad63967f82781f8cf6d6c9512713cc8927666707203370dc38b031c201475a"} Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.929742 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.929790 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.929815 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.929839 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.929853 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:49Z","lastTransitionTime":"2025-11-25T09:24:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.948258 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.958512 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.960365 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.968710 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:49 crc kubenswrapper[4776]: I1125 09:24:49.985778 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.002327 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.021049 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.032593 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.032636 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.032648 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.032663 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.032674 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:50Z","lastTransitionTime":"2025-11-25T09:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.046707 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.064649 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.086810 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.107541 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.128774 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.135625 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.135663 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.135675 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.135695 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.135708 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:50Z","lastTransitionTime":"2025-11-25T09:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.141430 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.166765 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea3897d86a1e19fb417f741e9a4d594f143bb3679f0ab458e0ca7b116710880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.180110 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.195376 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.207508 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.217933 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.227555 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.238173 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.238226 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.238242 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.238266 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.238287 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:50Z","lastTransitionTime":"2025-11-25T09:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.241541 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ad63967f82781f8cf6d6c9512713cc8927666707203370dc38b031c201475a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.258850 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.272739 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.285551 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.303772 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.331301 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.340816 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.340866 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.340884 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.340942 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.340960 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:50Z","lastTransitionTime":"2025-11-25T09:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.348491 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.365236 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.384790 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.404836 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.424428 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.442963 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.443004 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.443015 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.443031 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.443043 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:50Z","lastTransitionTime":"2025-11-25T09:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.448244 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea3897d86a1e19fb417f741e9a4d594f143bb3679f0ab458e0ca7b116710880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.469008 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.545433 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.545516 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.545535 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.545559 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.545578 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:50Z","lastTransitionTime":"2025-11-25T09:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.648633 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.648701 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.648719 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.648799 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.648817 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:50Z","lastTransitionTime":"2025-11-25T09:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.752736 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.752800 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.752819 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.752846 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.752867 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:50Z","lastTransitionTime":"2025-11-25T09:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.856060 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.856148 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.856165 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.856189 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.856207 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:50Z","lastTransitionTime":"2025-11-25T09:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.932718 4776 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.958800 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.958862 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.958880 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.958905 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:50 crc kubenswrapper[4776]: I1125 09:24:50.958923 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:50Z","lastTransitionTime":"2025-11-25T09:24:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.062801 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.062876 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.062904 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.062932 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.062954 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:51Z","lastTransitionTime":"2025-11-25T09:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.166240 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.166314 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.166338 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.166367 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.166386 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:51Z","lastTransitionTime":"2025-11-25T09:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.269451 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.269503 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.269514 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.269539 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.269552 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:51Z","lastTransitionTime":"2025-11-25T09:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.372151 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.372210 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.372225 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.372246 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.372261 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:51Z","lastTransitionTime":"2025-11-25T09:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.474660 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.474719 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.474731 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.474755 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.474771 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:51Z","lastTransitionTime":"2025-11-25T09:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.497729 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.497885 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:24:51 crc kubenswrapper[4776]: E1125 09:24:51.497968 4776 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:24:51 crc kubenswrapper[4776]: E1125 09:24:51.497974 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:25:07.497937973 +0000 UTC m=+52.538997666 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:24:51 crc kubenswrapper[4776]: E1125 09:24:51.498035 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:25:07.498016045 +0000 UTC m=+52.539075598 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.498102 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.498197 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.498239 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:24:51 crc kubenswrapper[4776]: E1125 09:24:51.498289 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:24:51 crc kubenswrapper[4776]: E1125 09:24:51.498312 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:24:51 crc kubenswrapper[4776]: E1125 09:24:51.498326 4776 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:24:51 crc kubenswrapper[4776]: E1125 09:24:51.498419 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 09:25:07.498407195 +0000 UTC m=+52.539466978 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:24:51 crc kubenswrapper[4776]: E1125 09:24:51.498483 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:24:51 crc kubenswrapper[4776]: E1125 09:24:51.498509 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:24:51 crc kubenswrapper[4776]: E1125 09:24:51.498526 4776 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:24:51 crc kubenswrapper[4776]: E1125 09:24:51.498544 4776 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:24:51 crc kubenswrapper[4776]: E1125 09:24:51.498584 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 09:25:07.498572709 +0000 UTC m=+52.539632272 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:24:51 crc kubenswrapper[4776]: E1125 09:24:51.498718 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:25:07.498683172 +0000 UTC m=+52.539742765 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.577525 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.577577 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.577588 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.577601 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.577610 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:51Z","lastTransitionTime":"2025-11-25T09:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.662226 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.662264 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.662226 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:24:51 crc kubenswrapper[4776]: E1125 09:24:51.662407 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:24:51 crc kubenswrapper[4776]: E1125 09:24:51.662505 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:24:51 crc kubenswrapper[4776]: E1125 09:24:51.662681 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.680269 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.680327 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.680339 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.680356 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.680708 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:51Z","lastTransitionTime":"2025-11-25T09:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.729810 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.729855 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.729867 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.729883 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.729894 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:51Z","lastTransitionTime":"2025-11-25T09:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:51 crc kubenswrapper[4776]: E1125 09:24:51.747118 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:51Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.751463 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.751503 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.751516 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.751534 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.751548 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:51Z","lastTransitionTime":"2025-11-25T09:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:51 crc kubenswrapper[4776]: E1125 09:24:51.770586 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:51Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.779419 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.779472 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.779489 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.779511 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.779530 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:51Z","lastTransitionTime":"2025-11-25T09:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:51 crc kubenswrapper[4776]: E1125 09:24:51.792509 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:51Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.796384 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.796417 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.796426 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.796439 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.796448 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:51Z","lastTransitionTime":"2025-11-25T09:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:51 crc kubenswrapper[4776]: E1125 09:24:51.808620 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:51Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.812370 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.812427 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.812482 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.812507 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.812567 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:51Z","lastTransitionTime":"2025-11-25T09:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:51 crc kubenswrapper[4776]: E1125 09:24:51.824793 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:51Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:51 crc kubenswrapper[4776]: E1125 09:24:51.824906 4776 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.826682 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.826704 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.826712 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.826733 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.826744 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:51Z","lastTransitionTime":"2025-11-25T09:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.928922 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.928955 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.928983 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.928997 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.929006 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:51Z","lastTransitionTime":"2025-11-25T09:24:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:51 crc kubenswrapper[4776]: I1125 09:24:51.936238 4776 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.031797 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.031852 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.031862 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.031879 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.031897 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:52Z","lastTransitionTime":"2025-11-25T09:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.133641 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.133683 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.133691 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.133706 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.133717 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:52Z","lastTransitionTime":"2025-11-25T09:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.236880 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.236916 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.236926 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.236943 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.236955 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:52Z","lastTransitionTime":"2025-11-25T09:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.339923 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.339995 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.340020 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.340050 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.340130 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:52Z","lastTransitionTime":"2025-11-25T09:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.442971 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.443099 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.443177 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.443214 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.443286 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:52Z","lastTransitionTime":"2025-11-25T09:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.547139 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.547267 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.547293 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.547412 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.547471 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:52Z","lastTransitionTime":"2025-11-25T09:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.650130 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.650193 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.650219 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.650251 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.650277 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:52Z","lastTransitionTime":"2025-11-25T09:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.662404 4776 scope.go:117] "RemoveContainer" containerID="399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.753714 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.753750 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.753759 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.753774 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.753784 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:52Z","lastTransitionTime":"2025-11-25T09:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.856554 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.856604 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.856621 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.856644 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.856664 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:52Z","lastTransitionTime":"2025-11-25T09:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.939985 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-r7x57_34502968-eb8d-47fa-83c2-7f6f6735ce2d/ovnkube-controller/0.log" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.943459 4776 generic.go:334] "Generic (PLEG): container finished" podID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerID="9ea3897d86a1e19fb417f741e9a4d594f143bb3679f0ab458e0ca7b116710880" exitCode=1 Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.943555 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" event={"ID":"34502968-eb8d-47fa-83c2-7f6f6735ce2d","Type":"ContainerDied","Data":"9ea3897d86a1e19fb417f741e9a4d594f143bb3679f0ab458e0ca7b116710880"} Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.944434 4776 scope.go:117] "RemoveContainer" containerID="9ea3897d86a1e19fb417f741e9a4d594f143bb3679f0ab458e0ca7b116710880" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.948537 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.951476 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"db22de3ee20f72a81d958f5fa9b10982a4ec446e092817b1880f669656e7a1ef"} Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.951837 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.958356 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.958391 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.958403 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.958417 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.958428 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:52Z","lastTransitionTime":"2025-11-25T09:24:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.961294 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.976023 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:52 crc kubenswrapper[4776]: I1125 09:24:52.986977 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.004479 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea3897d86a1e19fb417f741e9a4d594f143bb3679f0ab458e0ca7b116710880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea3897d86a1e19fb417f741e9a4d594f143bb3679f0ab458e0ca7b116710880\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:24:52Z\\\",\\\"message\\\":\\\"8\\\\nI1125 09:24:52.129147 6086 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 09:24:52.129228 6086 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 09:24:52.129240 6086 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 09:24:52.129264 6086 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 09:24:52.129270 6086 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 09:24:52.129284 6086 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 09:24:52.129705 6086 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 09:24:52.129830 6086 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:24:52.129910 6086 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:24:52.130011 6086 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:24:52.130205 6086 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.018450 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.032261 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.046183 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.061439 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.061488 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.061499 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.061516 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.061528 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:53Z","lastTransitionTime":"2025-11-25T09:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.061987 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.083083 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.094619 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.107299 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.127982 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.143477 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ad63967f82781f8cf6d6c9512713cc8927666707203370dc38b031c201475a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.156155 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.163977 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.164011 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.164027 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.164086 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.164127 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:53Z","lastTransitionTime":"2025-11-25T09:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.175095 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.186997 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.199815 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.214442 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ad63967f82781f8cf6d6c9512713cc8927666707203370dc38b031c201475a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.228492 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.239583 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.256967 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.267112 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.267154 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.267179 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.267200 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.267214 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:53Z","lastTransitionTime":"2025-11-25T09:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.271678 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.310120 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea3897d86a1e19fb417f741e9a4d594f143bb3679f0ab458e0ca7b116710880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea3897d86a1e19fb417f741e9a4d594f143bb3679f0ab458e0ca7b116710880\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:24:52Z\\\",\\\"message\\\":\\\"8\\\\nI1125 09:24:52.129147 6086 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 09:24:52.129228 6086 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 09:24:52.129240 6086 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 09:24:52.129264 6086 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 09:24:52.129270 6086 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 09:24:52.129284 6086 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 09:24:52.129705 6086 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 09:24:52.129830 6086 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:24:52.129910 6086 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:24:52.130011 6086 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:24:52.130205 6086 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.348412 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.363898 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db22de3ee20f72a81d958f5fa9b10982a4ec446e092817b1880f669656e7a1ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.369175 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.369216 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.369225 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.369242 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.369252 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:53Z","lastTransitionTime":"2025-11-25T09:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.375185 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.394619 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.408196 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.421844 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.434718 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.471751 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.471784 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.471792 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.471805 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.471817 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:53Z","lastTransitionTime":"2025-11-25T09:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.574399 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.574479 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.574503 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.574538 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.574571 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:53Z","lastTransitionTime":"2025-11-25T09:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.661845 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.661852 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:24:53 crc kubenswrapper[4776]: E1125 09:24:53.662009 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:24:53 crc kubenswrapper[4776]: E1125 09:24:53.662143 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.662461 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:24:53 crc kubenswrapper[4776]: E1125 09:24:53.662560 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.676985 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.677029 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.677041 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.677060 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.677098 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:53Z","lastTransitionTime":"2025-11-25T09:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.780515 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.780570 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.780582 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.780601 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.780616 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:53Z","lastTransitionTime":"2025-11-25T09:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.883338 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.883424 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.883449 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.883484 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.883507 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:53Z","lastTransitionTime":"2025-11-25T09:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.956745 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-r7x57_34502968-eb8d-47fa-83c2-7f6f6735ce2d/ovnkube-controller/0.log" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.959075 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" event={"ID":"34502968-eb8d-47fa-83c2-7f6f6735ce2d","Type":"ContainerStarted","Data":"d75a86dadfc39269053c91b6ff76d2e14e7368203d6cc51931adfa2e3b986f6c"} Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.959198 4776 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.974114 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.985827 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.985882 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.985890 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.985904 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.985913 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:53Z","lastTransitionTime":"2025-11-25T09:24:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:53 crc kubenswrapper[4776]: I1125 09:24:53.987796 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.000673 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.014188 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.028356 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ad63967f82781f8cf6d6c9512713cc8927666707203370dc38b031c201475a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.048471 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.060257 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.072253 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db22de3ee20f72a81d958f5fa9b10982a4ec446e092817b1880f669656e7a1ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.083988 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.088057 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.088131 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.088140 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.088157 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.088167 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:54Z","lastTransitionTime":"2025-11-25T09:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.098144 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.112581 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.134134 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75a86dadfc39269053c91b6ff76d2e14e7368203d6cc51931adfa2e3b986f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea3897d86a1e19fb417f741e9a4d594f143bb3679f0ab458e0ca7b116710880\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:24:52Z\\\",\\\"message\\\":\\\"8\\\\nI1125 09:24:52.129147 6086 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 09:24:52.129228 6086 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 09:24:52.129240 6086 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 09:24:52.129264 6086 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 09:24:52.129270 6086 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 09:24:52.129284 6086 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 09:24:52.129705 6086 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 09:24:52.129830 6086 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:24:52.129910 6086 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:24:52.130011 6086 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:24:52.130205 6086 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.143262 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.156944 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.168724 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.194263 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.194324 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.194335 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.194363 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.194380 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:54Z","lastTransitionTime":"2025-11-25T09:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.298108 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.298405 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.298498 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.298589 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.298706 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:54Z","lastTransitionTime":"2025-11-25T09:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.401848 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.401894 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.401907 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.401924 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.401936 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:54Z","lastTransitionTime":"2025-11-25T09:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.446369 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj"] Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.446781 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.449578 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.450247 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.460610 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.472812 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a506265-8f64-456c-995f-1bc468cb87dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rzwmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.492406 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.504480 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.504519 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.504530 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.504545 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.504558 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:54Z","lastTransitionTime":"2025-11-25T09:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.509237 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.523542 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.536944 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.556215 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75a86dadfc39269053c91b6ff76d2e14e7368203d6cc51931adfa2e3b986f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea3897d86a1e19fb417f741e9a4d594f143bb3679f0ab458e0ca7b116710880\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:24:52Z\\\",\\\"message\\\":\\\"8\\\\nI1125 09:24:52.129147 6086 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 09:24:52.129228 6086 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 09:24:52.129240 6086 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 09:24:52.129264 6086 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 09:24:52.129270 6086 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 09:24:52.129284 6086 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 09:24:52.129705 6086 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 09:24:52.129830 6086 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:24:52.129910 6086 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:24:52.130011 6086 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:24:52.130205 6086 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.568419 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.581825 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db22de3ee20f72a81d958f5fa9b10982a4ec446e092817b1880f669656e7a1ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.597191 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.606975 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.607189 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.607313 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.607447 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.607585 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:54Z","lastTransitionTime":"2025-11-25T09:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.609664 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.622858 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.629890 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9a506265-8f64-456c-995f-1bc468cb87dd-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-rzwmj\" (UID: \"9a506265-8f64-456c-995f-1bc468cb87dd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.629988 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9a506265-8f64-456c-995f-1bc468cb87dd-env-overrides\") pod \"ovnkube-control-plane-749d76644c-rzwmj\" (UID: \"9a506265-8f64-456c-995f-1bc468cb87dd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.630019 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcf26\" (UniqueName: \"kubernetes.io/projected/9a506265-8f64-456c-995f-1bc468cb87dd-kube-api-access-rcf26\") pod \"ovnkube-control-plane-749d76644c-rzwmj\" (UID: \"9a506265-8f64-456c-995f-1bc468cb87dd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.630086 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9a506265-8f64-456c-995f-1bc468cb87dd-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-rzwmj\" (UID: \"9a506265-8f64-456c-995f-1bc468cb87dd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.636409 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.647059 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.661194 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.679791 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ad63967f82781f8cf6d6c9512713cc8927666707203370dc38b031c201475a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.710576 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.710621 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.710637 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.710658 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.710673 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:54Z","lastTransitionTime":"2025-11-25T09:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.731196 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9a506265-8f64-456c-995f-1bc468cb87dd-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-rzwmj\" (UID: \"9a506265-8f64-456c-995f-1bc468cb87dd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.731263 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9a506265-8f64-456c-995f-1bc468cb87dd-env-overrides\") pod \"ovnkube-control-plane-749d76644c-rzwmj\" (UID: \"9a506265-8f64-456c-995f-1bc468cb87dd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.731280 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcf26\" (UniqueName: \"kubernetes.io/projected/9a506265-8f64-456c-995f-1bc468cb87dd-kube-api-access-rcf26\") pod \"ovnkube-control-plane-749d76644c-rzwmj\" (UID: \"9a506265-8f64-456c-995f-1bc468cb87dd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.731306 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9a506265-8f64-456c-995f-1bc468cb87dd-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-rzwmj\" (UID: \"9a506265-8f64-456c-995f-1bc468cb87dd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.732734 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9a506265-8f64-456c-995f-1bc468cb87dd-env-overrides\") pod \"ovnkube-control-plane-749d76644c-rzwmj\" (UID: \"9a506265-8f64-456c-995f-1bc468cb87dd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.732743 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9a506265-8f64-456c-995f-1bc468cb87dd-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-rzwmj\" (UID: \"9a506265-8f64-456c-995f-1bc468cb87dd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.737082 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9a506265-8f64-456c-995f-1bc468cb87dd-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-rzwmj\" (UID: \"9a506265-8f64-456c-995f-1bc468cb87dd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.750838 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcf26\" (UniqueName: \"kubernetes.io/projected/9a506265-8f64-456c-995f-1bc468cb87dd-kube-api-access-rcf26\") pod \"ovnkube-control-plane-749d76644c-rzwmj\" (UID: \"9a506265-8f64-456c-995f-1bc468cb87dd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.766319 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" Nov 25 09:24:54 crc kubenswrapper[4776]: W1125 09:24:54.798173 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a506265_8f64_456c_995f_1bc468cb87dd.slice/crio-21618de10983fbaf376160ecf47d5e1b4877b0971bd77d341ebeaec7b44e623d WatchSource:0}: Error finding container 21618de10983fbaf376160ecf47d5e1b4877b0971bd77d341ebeaec7b44e623d: Status 404 returned error can't find the container with id 21618de10983fbaf376160ecf47d5e1b4877b0971bd77d341ebeaec7b44e623d Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.813232 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.813318 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.813347 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.813385 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.813411 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:54Z","lastTransitionTime":"2025-11-25T09:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.916657 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.916710 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.916726 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.916747 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.916764 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:54Z","lastTransitionTime":"2025-11-25T09:24:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.963977 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-r7x57_34502968-eb8d-47fa-83c2-7f6f6735ce2d/ovnkube-controller/1.log" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.965350 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-r7x57_34502968-eb8d-47fa-83c2-7f6f6735ce2d/ovnkube-controller/0.log" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.971144 4776 generic.go:334] "Generic (PLEG): container finished" podID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerID="d75a86dadfc39269053c91b6ff76d2e14e7368203d6cc51931adfa2e3b986f6c" exitCode=1 Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.971220 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" event={"ID":"34502968-eb8d-47fa-83c2-7f6f6735ce2d","Type":"ContainerDied","Data":"d75a86dadfc39269053c91b6ff76d2e14e7368203d6cc51931adfa2e3b986f6c"} Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.971259 4776 scope.go:117] "RemoveContainer" containerID="9ea3897d86a1e19fb417f741e9a4d594f143bb3679f0ab458e0ca7b116710880" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.972136 4776 scope.go:117] "RemoveContainer" containerID="d75a86dadfc39269053c91b6ff76d2e14e7368203d6cc51931adfa2e3b986f6c" Nov 25 09:24:54 crc kubenswrapper[4776]: E1125 09:24:54.972360 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-r7x57_openshift-ovn-kubernetes(34502968-eb8d-47fa-83c2-7f6f6735ce2d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.972781 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" event={"ID":"9a506265-8f64-456c-995f-1bc468cb87dd","Type":"ContainerStarted","Data":"21618de10983fbaf376160ecf47d5e1b4877b0971bd77d341ebeaec7b44e623d"} Nov 25 09:24:54 crc kubenswrapper[4776]: I1125 09:24:54.987340 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.002256 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.018389 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ad63967f82781f8cf6d6c9512713cc8927666707203370dc38b031c201475a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.022421 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.022472 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.022487 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.022504 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.022522 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:55Z","lastTransitionTime":"2025-11-25T09:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.039739 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.057768 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.071309 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.084515 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.113151 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.133677 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.133714 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.133723 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.133739 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.133749 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:55Z","lastTransitionTime":"2025-11-25T09:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.134666 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.148658 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a506265-8f64-456c-995f-1bc468cb87dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rzwmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.168599 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db22de3ee20f72a81d958f5fa9b10982a4ec446e092817b1880f669656e7a1ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.184745 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.204657 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.221803 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.236267 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.236309 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.236338 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.236354 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.236363 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:55Z","lastTransitionTime":"2025-11-25T09:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.244943 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75a86dadfc39269053c91b6ff76d2e14e7368203d6cc51931adfa2e3b986f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea3897d86a1e19fb417f741e9a4d594f143bb3679f0ab458e0ca7b116710880\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:24:52Z\\\",\\\"message\\\":\\\"8\\\\nI1125 09:24:52.129147 6086 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 09:24:52.129228 6086 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 09:24:52.129240 6086 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 09:24:52.129264 6086 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 09:24:52.129270 6086 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 09:24:52.129284 6086 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 09:24:52.129705 6086 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 09:24:52.129830 6086 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:24:52.129910 6086 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:24:52.130011 6086 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:24:52.130205 6086 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75a86dadfc39269053c91b6ff76d2e14e7368203d6cc51931adfa2e3b986f6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"message\\\":\\\"\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8441, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8442, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8444, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1125 09:24:54.000812 6252 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.266599 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.338526 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.338802 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.338897 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.339018 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.339164 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:55Z","lastTransitionTime":"2025-11-25T09:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.441931 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.442386 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.442632 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.443226 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.443443 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:55Z","lastTransitionTime":"2025-11-25T09:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.546304 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.546656 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.546808 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.546959 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.547131 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:55Z","lastTransitionTime":"2025-11-25T09:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.651298 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.651360 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.651375 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.651398 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.651413 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:55Z","lastTransitionTime":"2025-11-25T09:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.661878 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.661911 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.661871 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:24:55 crc kubenswrapper[4776]: E1125 09:24:55.662040 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:24:55 crc kubenswrapper[4776]: E1125 09:24:55.662237 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:24:55 crc kubenswrapper[4776]: E1125 09:24:55.662403 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.677544 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.695088 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a506265-8f64-456c-995f-1bc468cb87dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rzwmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.726196 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.739643 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.753898 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.753935 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.753945 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.753969 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.753980 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:55Z","lastTransitionTime":"2025-11-25T09:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.755454 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.777275 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.803982 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75a86dadfc39269053c91b6ff76d2e14e7368203d6cc51931adfa2e3b986f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea3897d86a1e19fb417f741e9a4d594f143bb3679f0ab458e0ca7b116710880\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:24:52Z\\\",\\\"message\\\":\\\"8\\\\nI1125 09:24:52.129147 6086 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 09:24:52.129228 6086 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 09:24:52.129240 6086 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 09:24:52.129264 6086 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 09:24:52.129270 6086 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 09:24:52.129284 6086 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 09:24:52.129705 6086 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 09:24:52.129830 6086 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:24:52.129910 6086 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:24:52.130011 6086 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:24:52.130205 6086 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75a86dadfc39269053c91b6ff76d2e14e7368203d6cc51931adfa2e3b986f6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"message\\\":\\\"\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8441, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8442, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8444, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1125 09:24:54.000812 6252 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.814622 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.825891 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db22de3ee20f72a81d958f5fa9b10982a4ec446e092817b1880f669656e7a1ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.836093 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.846389 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.855700 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.855803 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.855827 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.855862 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.855885 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:55Z","lastTransitionTime":"2025-11-25T09:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.863278 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.877559 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.887998 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.912934 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.924836 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-nvc68"] Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.925612 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:24:55 crc kubenswrapper[4776]: E1125 09:24:55.925708 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.927967 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ad63967f82781f8cf6d6c9512713cc8927666707203370dc38b031c201475a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.949292 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.959200 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.959236 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.959249 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.959277 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.959291 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:55Z","lastTransitionTime":"2025-11-25T09:24:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.963654 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.977668 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a506265-8f64-456c-995f-1bc468cb87dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rzwmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.979395 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" event={"ID":"9a506265-8f64-456c-995f-1bc468cb87dd","Type":"ContainerStarted","Data":"eb2778606c71db6545b88fd0648a8ab08a66863579467e49a5ece3b36d0d21f7"} Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.979442 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" event={"ID":"9a506265-8f64-456c-995f-1bc468cb87dd","Type":"ContainerStarted","Data":"29ead16f9d75513fe9ab6959e2ba0dca4f4b88df8bc65e3b0e6a21f28f2d0d30"} Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.983392 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-r7x57_34502968-eb8d-47fa-83c2-7f6f6735ce2d/ovnkube-controller/1.log" Nov 25 09:24:55 crc kubenswrapper[4776]: I1125 09:24:55.991899 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.010854 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75a86dadfc39269053c91b6ff76d2e14e7368203d6cc51931adfa2e3b986f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea3897d86a1e19fb417f741e9a4d594f143bb3679f0ab458e0ca7b116710880\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:24:52Z\\\",\\\"message\\\":\\\"8\\\\nI1125 09:24:52.129147 6086 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 09:24:52.129228 6086 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 09:24:52.129240 6086 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 09:24:52.129264 6086 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 09:24:52.129270 6086 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 09:24:52.129284 6086 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 09:24:52.129705 6086 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 09:24:52.129830 6086 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:24:52.129910 6086 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:24:52.130011 6086 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:24:52.130205 6086 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75a86dadfc39269053c91b6ff76d2e14e7368203d6cc51931adfa2e3b986f6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"message\\\":\\\"\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8441, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8442, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8444, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1125 09:24:54.000812 6252 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.020796 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.036032 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db22de3ee20f72a81d958f5fa9b10982a4ec446e092817b1880f669656e7a1ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.045051 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5n95\" (UniqueName: \"kubernetes.io/projected/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-kube-api-access-s5n95\") pod \"network-metrics-daemon-nvc68\" (UID: \"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\") " pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.045220 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-metrics-certs\") pod \"network-metrics-daemon-nvc68\" (UID: \"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\") " pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.049274 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.062147 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.062209 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.062227 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.062253 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.062272 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:56Z","lastTransitionTime":"2025-11-25T09:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.063861 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.075289 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.086625 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.096051 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nvc68" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nvc68\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.109267 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.119406 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.129850 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.142960 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ad63967f82781f8cf6d6c9512713cc8927666707203370dc38b031c201475a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.145643 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-metrics-certs\") pod \"network-metrics-daemon-nvc68\" (UID: \"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\") " pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.145761 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5n95\" (UniqueName: \"kubernetes.io/projected/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-kube-api-access-s5n95\") pod \"network-metrics-daemon-nvc68\" (UID: \"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\") " pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:24:56 crc kubenswrapper[4776]: E1125 09:24:56.145816 4776 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:24:56 crc kubenswrapper[4776]: E1125 09:24:56.145875 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-metrics-certs podName:4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087 nodeName:}" failed. No retries permitted until 2025-11-25 09:24:56.645856996 +0000 UTC m=+41.686916549 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-metrics-certs") pod "network-metrics-daemon-nvc68" (UID: "4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.155744 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.160953 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5n95\" (UniqueName: \"kubernetes.io/projected/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-kube-api-access-s5n95\") pod \"network-metrics-daemon-nvc68\" (UID: \"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\") " pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.165020 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.165060 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.165098 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.165115 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.165127 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:56Z","lastTransitionTime":"2025-11-25T09:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.171052 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.182856 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.191857 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.203668 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.217448 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ad63967f82781f8cf6d6c9512713cc8927666707203370dc38b031c201475a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.239143 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.250734 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.262398 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a506265-8f64-456c-995f-1bc468cb87dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ead16f9d75513fe9ab6959e2ba0dca4f4b88df8bc65e3b0e6a21f28f2d0d30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb2778606c71db6545b88fd0648a8ab08a66863579467e49a5ece3b36d0d21f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rzwmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.266976 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.267041 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.267053 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.267084 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.267100 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:56Z","lastTransitionTime":"2025-11-25T09:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.277722 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db22de3ee20f72a81d958f5fa9b10982a4ec446e092817b1880f669656e7a1ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.290281 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.301886 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.312486 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.331370 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75a86dadfc39269053c91b6ff76d2e14e7368203d6cc51931adfa2e3b986f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea3897d86a1e19fb417f741e9a4d594f143bb3679f0ab458e0ca7b116710880\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:24:52Z\\\",\\\"message\\\":\\\"8\\\\nI1125 09:24:52.129147 6086 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 09:24:52.129228 6086 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 09:24:52.129240 6086 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 09:24:52.129264 6086 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 09:24:52.129270 6086 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 09:24:52.129284 6086 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 09:24:52.129705 6086 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 09:24:52.129830 6086 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:24:52.129910 6086 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:24:52.130011 6086 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:24:52.130205 6086 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75a86dadfc39269053c91b6ff76d2e14e7368203d6cc51931adfa2e3b986f6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"message\\\":\\\"\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8441, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8442, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8444, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1125 09:24:54.000812 6252 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.339792 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.350809 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.362172 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.369904 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.369958 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.369967 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.369983 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.369994 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:56Z","lastTransitionTime":"2025-11-25T09:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.371638 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nvc68" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nvc68\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:24:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.472418 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.472450 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.472460 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.472490 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.472499 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:56Z","lastTransitionTime":"2025-11-25T09:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.575409 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.575479 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.575498 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.575524 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.575539 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:56Z","lastTransitionTime":"2025-11-25T09:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.651186 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-metrics-certs\") pod \"network-metrics-daemon-nvc68\" (UID: \"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\") " pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:24:56 crc kubenswrapper[4776]: E1125 09:24:56.651477 4776 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:24:56 crc kubenswrapper[4776]: E1125 09:24:56.651620 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-metrics-certs podName:4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087 nodeName:}" failed. No retries permitted until 2025-11-25 09:24:57.651589871 +0000 UTC m=+42.692649464 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-metrics-certs") pod "network-metrics-daemon-nvc68" (UID: "4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.677969 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.678041 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.678056 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.678098 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.678111 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:56Z","lastTransitionTime":"2025-11-25T09:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.781278 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.781496 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.781610 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.781747 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.781849 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:56Z","lastTransitionTime":"2025-11-25T09:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.884591 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.884636 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.884649 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.884667 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.884680 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:56Z","lastTransitionTime":"2025-11-25T09:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.988381 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.988436 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.988457 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.988480 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:56 crc kubenswrapper[4776]: I1125 09:24:56.988497 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:56Z","lastTransitionTime":"2025-11-25T09:24:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.107844 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.107921 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.107943 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.107973 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.108037 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:57Z","lastTransitionTime":"2025-11-25T09:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.210971 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.211016 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.211031 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.211050 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.211088 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:57Z","lastTransitionTime":"2025-11-25T09:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.314135 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.314171 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.314184 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.314201 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.314248 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:57Z","lastTransitionTime":"2025-11-25T09:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.417479 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.417520 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.417532 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.417547 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.417558 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:57Z","lastTransitionTime":"2025-11-25T09:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.520571 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.520641 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.520658 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.520684 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.520701 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:57Z","lastTransitionTime":"2025-11-25T09:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.622567 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.622607 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.622618 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.622636 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.622649 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:57Z","lastTransitionTime":"2025-11-25T09:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.661683 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.661775 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:24:57 crc kubenswrapper[4776]: E1125 09:24:57.661866 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.661778 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:24:57 crc kubenswrapper[4776]: E1125 09:24:57.661951 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.662010 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:24:57 crc kubenswrapper[4776]: E1125 09:24:57.662216 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:24:57 crc kubenswrapper[4776]: E1125 09:24:57.662278 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.707724 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-metrics-certs\") pod \"network-metrics-daemon-nvc68\" (UID: \"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\") " pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:24:57 crc kubenswrapper[4776]: E1125 09:24:57.707941 4776 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:24:57 crc kubenswrapper[4776]: E1125 09:24:57.708044 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-metrics-certs podName:4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087 nodeName:}" failed. No retries permitted until 2025-11-25 09:24:59.708018489 +0000 UTC m=+44.749078072 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-metrics-certs") pod "network-metrics-daemon-nvc68" (UID: "4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.728008 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.728041 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.728053 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.728093 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.728106 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:57Z","lastTransitionTime":"2025-11-25T09:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.831124 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.831162 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.831170 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.831183 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.831194 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:57Z","lastTransitionTime":"2025-11-25T09:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.933760 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.934006 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.934159 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.934255 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:57 crc kubenswrapper[4776]: I1125 09:24:57.934339 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:57Z","lastTransitionTime":"2025-11-25T09:24:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.036142 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.036174 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.036184 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.036198 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.036210 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:58Z","lastTransitionTime":"2025-11-25T09:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.139335 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.139387 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.139395 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.139408 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.139416 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:58Z","lastTransitionTime":"2025-11-25T09:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.242061 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.242155 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.242173 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.242199 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.242216 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:58Z","lastTransitionTime":"2025-11-25T09:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.345029 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.345142 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.345163 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.345191 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.345211 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:58Z","lastTransitionTime":"2025-11-25T09:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.447850 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.447894 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.447906 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.447923 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.447934 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:58Z","lastTransitionTime":"2025-11-25T09:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.550848 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.550922 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.550939 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.550959 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.550972 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:58Z","lastTransitionTime":"2025-11-25T09:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.654176 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.654245 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.654262 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.654286 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.654303 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:58Z","lastTransitionTime":"2025-11-25T09:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.757465 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.757528 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.757546 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.757569 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.757586 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:58Z","lastTransitionTime":"2025-11-25T09:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.860263 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.860321 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.860338 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.860357 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.860368 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:58Z","lastTransitionTime":"2025-11-25T09:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.962804 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.962855 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.962873 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.962894 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:58 crc kubenswrapper[4776]: I1125 09:24:58.962910 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:58Z","lastTransitionTime":"2025-11-25T09:24:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.065814 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.065953 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.065982 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.066012 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.066033 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:59Z","lastTransitionTime":"2025-11-25T09:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.169303 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.169376 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.169396 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.169418 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.169434 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:59Z","lastTransitionTime":"2025-11-25T09:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.272439 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.272510 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.272527 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.272552 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.272586 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:59Z","lastTransitionTime":"2025-11-25T09:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.375676 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.375742 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.375761 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.375787 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.375808 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:59Z","lastTransitionTime":"2025-11-25T09:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.479281 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.479348 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.479388 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.479421 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.479443 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:59Z","lastTransitionTime":"2025-11-25T09:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.582933 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.583013 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.583034 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.583058 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.583139 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:59Z","lastTransitionTime":"2025-11-25T09:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.665605 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:24:59 crc kubenswrapper[4776]: E1125 09:24:59.665798 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.666170 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.666263 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:24:59 crc kubenswrapper[4776]: E1125 09:24:59.666384 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.666183 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:24:59 crc kubenswrapper[4776]: E1125 09:24:59.667057 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:24:59 crc kubenswrapper[4776]: E1125 09:24:59.667391 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.686282 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.686354 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.686377 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.686404 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.686422 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:59Z","lastTransitionTime":"2025-11-25T09:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.728399 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-metrics-certs\") pod \"network-metrics-daemon-nvc68\" (UID: \"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\") " pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:24:59 crc kubenswrapper[4776]: E1125 09:24:59.728575 4776 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:24:59 crc kubenswrapper[4776]: E1125 09:24:59.728649 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-metrics-certs podName:4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087 nodeName:}" failed. No retries permitted until 2025-11-25 09:25:03.728622898 +0000 UTC m=+48.769682491 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-metrics-certs") pod "network-metrics-daemon-nvc68" (UID: "4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.789710 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.789750 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.789768 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.789793 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.789818 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:59Z","lastTransitionTime":"2025-11-25T09:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.893700 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.894332 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.894353 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.894380 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.894404 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:59Z","lastTransitionTime":"2025-11-25T09:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.998194 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.998290 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.998315 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.998347 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:24:59 crc kubenswrapper[4776]: I1125 09:24:59.998374 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:24:59Z","lastTransitionTime":"2025-11-25T09:24:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.101326 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.101396 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.101408 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.101425 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.101433 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:00Z","lastTransitionTime":"2025-11-25T09:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.204916 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.204974 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.204986 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.205006 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.205020 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:00Z","lastTransitionTime":"2025-11-25T09:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.308905 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.308973 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.308984 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.309002 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.309014 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:00Z","lastTransitionTime":"2025-11-25T09:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.413046 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.413144 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.413162 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.413185 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.413202 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:00Z","lastTransitionTime":"2025-11-25T09:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.427206 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.428183 4776 scope.go:117] "RemoveContainer" containerID="d75a86dadfc39269053c91b6ff76d2e14e7368203d6cc51931adfa2e3b986f6c" Nov 25 09:25:00 crc kubenswrapper[4776]: E1125 09:25:00.428391 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-r7x57_openshift-ovn-kubernetes(34502968-eb8d-47fa-83c2-7f6f6735ce2d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.452784 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:00Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.473857 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:00Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.488950 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nvc68" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nvc68\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:00Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.504539 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:00Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.516316 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.516386 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.516396 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.516413 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.516425 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:00Z","lastTransitionTime":"2025-11-25T09:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.522516 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:00Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.538346 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:00Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.553741 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:00Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.574945 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ad63967f82781f8cf6d6c9512713cc8927666707203370dc38b031c201475a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:00Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.596495 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:00Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.611618 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:00Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.619721 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.619786 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.619801 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.619825 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.619843 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:00Z","lastTransitionTime":"2025-11-25T09:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.625668 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a506265-8f64-456c-995f-1bc468cb87dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ead16f9d75513fe9ab6959e2ba0dca4f4b88df8bc65e3b0e6a21f28f2d0d30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb2778606c71db6545b88fd0648a8ab08a66863579467e49a5ece3b36d0d21f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rzwmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:00Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.641052 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db22de3ee20f72a81d958f5fa9b10982a4ec446e092817b1880f669656e7a1ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:00Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.655168 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:00Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.672194 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:00Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.688641 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:00Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.717723 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75a86dadfc39269053c91b6ff76d2e14e7368203d6cc51931adfa2e3b986f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75a86dadfc39269053c91b6ff76d2e14e7368203d6cc51931adfa2e3b986f6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"message\\\":\\\"\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8441, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8442, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8444, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1125 09:24:54.000812 6252 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-r7x57_openshift-ovn-kubernetes(34502968-eb8d-47fa-83c2-7f6f6735ce2d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:00Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.726697 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.726776 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.726872 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.726900 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.726918 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:00Z","lastTransitionTime":"2025-11-25T09:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.743660 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:00Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.830735 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.830804 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.830821 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.830853 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.830872 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:00Z","lastTransitionTime":"2025-11-25T09:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.934200 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.934253 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.934264 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.934282 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:00 crc kubenswrapper[4776]: I1125 09:25:00.934305 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:00Z","lastTransitionTime":"2025-11-25T09:25:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.037778 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.037866 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.037892 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.037925 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.037949 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:01Z","lastTransitionTime":"2025-11-25T09:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.140982 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.141050 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.141098 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.141122 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.141138 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:01Z","lastTransitionTime":"2025-11-25T09:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.244486 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.244530 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.244544 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.244564 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.244578 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:01Z","lastTransitionTime":"2025-11-25T09:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.347791 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.348396 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.348536 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.348635 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.348719 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:01Z","lastTransitionTime":"2025-11-25T09:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.452088 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.452173 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.452199 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.452231 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.452251 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:01Z","lastTransitionTime":"2025-11-25T09:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.554173 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.554697 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.554835 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.554984 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.555161 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:01Z","lastTransitionTime":"2025-11-25T09:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.658526 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.659111 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.659227 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.659384 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.659532 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:01Z","lastTransitionTime":"2025-11-25T09:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.661884 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.661948 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.661890 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.662059 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:01 crc kubenswrapper[4776]: E1125 09:25:01.662208 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:01 crc kubenswrapper[4776]: E1125 09:25:01.662445 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:01 crc kubenswrapper[4776]: E1125 09:25:01.662594 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:01 crc kubenswrapper[4776]: E1125 09:25:01.662652 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.772808 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.772909 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.772931 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.772960 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.772980 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:01Z","lastTransitionTime":"2025-11-25T09:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.836613 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.837098 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.837228 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.837362 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.837461 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:01Z","lastTransitionTime":"2025-11-25T09:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:01 crc kubenswrapper[4776]: E1125 09:25:01.860500 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:01Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.867643 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.867718 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.867742 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.867774 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.867798 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:01Z","lastTransitionTime":"2025-11-25T09:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:01 crc kubenswrapper[4776]: E1125 09:25:01.890544 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:01Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.896444 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.896504 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.896527 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.896556 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.896576 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:01Z","lastTransitionTime":"2025-11-25T09:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:01 crc kubenswrapper[4776]: E1125 09:25:01.915862 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:01Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.921553 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.921611 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.921622 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.921640 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.921656 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:01Z","lastTransitionTime":"2025-11-25T09:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:01 crc kubenswrapper[4776]: E1125 09:25:01.939866 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:01Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.944166 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.944214 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.944226 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.944243 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.944254 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:01Z","lastTransitionTime":"2025-11-25T09:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:01 crc kubenswrapper[4776]: E1125 09:25:01.963897 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:01Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:01 crc kubenswrapper[4776]: E1125 09:25:01.964058 4776 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.966213 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.966271 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.966287 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.966342 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:01 crc kubenswrapper[4776]: I1125 09:25:01.966359 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:01Z","lastTransitionTime":"2025-11-25T09:25:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.069951 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.070125 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.070151 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.070183 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.070204 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:02Z","lastTransitionTime":"2025-11-25T09:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.220813 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.220893 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.220913 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.220942 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.220962 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:02Z","lastTransitionTime":"2025-11-25T09:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.324870 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.324965 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.324994 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.325033 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.325060 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:02Z","lastTransitionTime":"2025-11-25T09:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.429498 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.429570 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.429589 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.429618 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.429637 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:02Z","lastTransitionTime":"2025-11-25T09:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.533849 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.533936 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.533962 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.533998 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.534021 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:02Z","lastTransitionTime":"2025-11-25T09:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.638208 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.638276 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.638301 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.638335 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.638363 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:02Z","lastTransitionTime":"2025-11-25T09:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.741746 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.741834 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.741860 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.741893 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.741919 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:02Z","lastTransitionTime":"2025-11-25T09:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.845758 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.845845 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.845867 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.845896 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.845952 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:02Z","lastTransitionTime":"2025-11-25T09:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.949883 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.949929 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.949938 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.949957 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:02 crc kubenswrapper[4776]: I1125 09:25:02.949971 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:02Z","lastTransitionTime":"2025-11-25T09:25:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.054336 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.054432 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.054463 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.054490 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.054512 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:03Z","lastTransitionTime":"2025-11-25T09:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.157517 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.157584 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.157607 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.157640 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.157663 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:03Z","lastTransitionTime":"2025-11-25T09:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.260659 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.260718 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.260731 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.260749 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.260761 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:03Z","lastTransitionTime":"2025-11-25T09:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.364698 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.364763 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.364781 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.364807 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.364825 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:03Z","lastTransitionTime":"2025-11-25T09:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.468714 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.468777 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.468791 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.468813 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.468826 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:03Z","lastTransitionTime":"2025-11-25T09:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.571947 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.572010 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.572027 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.572049 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.572104 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:03Z","lastTransitionTime":"2025-11-25T09:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.662423 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.662426 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:03 crc kubenswrapper[4776]: E1125 09:25:03.662637 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.662745 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:03 crc kubenswrapper[4776]: E1125 09:25:03.662795 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.662857 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:03 crc kubenswrapper[4776]: E1125 09:25:03.662990 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:03 crc kubenswrapper[4776]: E1125 09:25:03.663295 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.675386 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.675433 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.675453 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.675476 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.675492 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:03Z","lastTransitionTime":"2025-11-25T09:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.733983 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.759935 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db22de3ee20f72a81d958f5fa9b10982a4ec446e092817b1880f669656e7a1ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:03Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.779497 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.779567 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.779589 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.779622 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.779636 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-metrics-certs\") pod \"network-metrics-daemon-nvc68\" (UID: \"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\") " pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.779650 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:03Z","lastTransitionTime":"2025-11-25T09:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:03 crc kubenswrapper[4776]: E1125 09:25:03.779910 4776 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:25:03 crc kubenswrapper[4776]: E1125 09:25:03.780052 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-metrics-certs podName:4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087 nodeName:}" failed. No retries permitted until 2025-11-25 09:25:11.780012268 +0000 UTC m=+56.821071861 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-metrics-certs") pod "network-metrics-daemon-nvc68" (UID: "4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.784214 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:03Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.807199 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:03Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.830444 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:03Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.856553 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75a86dadfc39269053c91b6ff76d2e14e7368203d6cc51931adfa2e3b986f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75a86dadfc39269053c91b6ff76d2e14e7368203d6cc51931adfa2e3b986f6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"message\\\":\\\"\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8441, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8442, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8444, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1125 09:24:54.000812 6252 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-r7x57_openshift-ovn-kubernetes(34502968-eb8d-47fa-83c2-7f6f6735ce2d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:03Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.870840 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:03Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.882808 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.882867 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.882882 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.882906 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.882924 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:03Z","lastTransitionTime":"2025-11-25T09:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.887816 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:03Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.900596 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:03Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.910005 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nvc68" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nvc68\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:03Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.920725 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:03Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.933759 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:03Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.946223 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:03Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.960098 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:03Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.979316 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ad63967f82781f8cf6d6c9512713cc8927666707203370dc38b031c201475a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:03Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.990773 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.990863 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.990885 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.990973 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:03 crc kubenswrapper[4776]: I1125 09:25:03.991022 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:03Z","lastTransitionTime":"2025-11-25T09:25:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.012289 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.025546 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.040948 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a506265-8f64-456c-995f-1bc468cb87dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ead16f9d75513fe9ab6959e2ba0dca4f4b88df8bc65e3b0e6a21f28f2d0d30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb2778606c71db6545b88fd0648a8ab08a66863579467e49a5ece3b36d0d21f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rzwmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.094368 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.094428 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.094441 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.094464 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.094481 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:04Z","lastTransitionTime":"2025-11-25T09:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.197605 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.197675 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.197697 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.197721 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.197738 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:04Z","lastTransitionTime":"2025-11-25T09:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.301334 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.301389 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.301401 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.301420 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.301432 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:04Z","lastTransitionTime":"2025-11-25T09:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.405755 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.405854 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.405872 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.405927 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.405945 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:04Z","lastTransitionTime":"2025-11-25T09:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.510907 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.510985 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.511010 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.511041 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.511062 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:04Z","lastTransitionTime":"2025-11-25T09:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.614856 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.614907 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.614917 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.614934 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.614945 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:04Z","lastTransitionTime":"2025-11-25T09:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.717586 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.717639 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.717651 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.717674 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.717686 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:04Z","lastTransitionTime":"2025-11-25T09:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.821113 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.821160 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.821173 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.821192 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.821205 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:04Z","lastTransitionTime":"2025-11-25T09:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.924671 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.924732 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.924751 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.924775 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:04 crc kubenswrapper[4776]: I1125 09:25:04.924795 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:04Z","lastTransitionTime":"2025-11-25T09:25:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.028155 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.028266 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.028289 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.028313 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.028402 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:05Z","lastTransitionTime":"2025-11-25T09:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.131808 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.131879 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.131897 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.131924 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.131942 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:05Z","lastTransitionTime":"2025-11-25T09:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.235099 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.235129 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.235137 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.235152 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.235161 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:05Z","lastTransitionTime":"2025-11-25T09:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.337279 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.337309 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.337318 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.337330 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.337339 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:05Z","lastTransitionTime":"2025-11-25T09:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.439171 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.439237 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.439257 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.439286 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.439346 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:05Z","lastTransitionTime":"2025-11-25T09:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.543394 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.543455 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.543478 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.543502 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.543518 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:05Z","lastTransitionTime":"2025-11-25T09:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.646894 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.646967 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.646985 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.647014 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.647030 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:05Z","lastTransitionTime":"2025-11-25T09:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.661736 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.661838 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.661936 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:05 crc kubenswrapper[4776]: E1125 09:25:05.662127 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.662191 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:05 crc kubenswrapper[4776]: E1125 09:25:05.662351 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:05 crc kubenswrapper[4776]: E1125 09:25:05.662456 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:05 crc kubenswrapper[4776]: E1125 09:25:05.662582 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.698832 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:05Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.716568 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:05Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.730917 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a506265-8f64-456c-995f-1bc468cb87dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ead16f9d75513fe9ab6959e2ba0dca4f4b88df8bc65e3b0e6a21f28f2d0d30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb2778606c71db6545b88fd0648a8ab08a66863579467e49a5ece3b36d0d21f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rzwmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:05Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.747885 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db22de3ee20f72a81d958f5fa9b10982a4ec446e092817b1880f669656e7a1ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:05Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.749940 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.749978 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.749991 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.750011 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.750023 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:05Z","lastTransitionTime":"2025-11-25T09:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.767088 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:05Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.785021 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:05Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.803326 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:05Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.825835 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75a86dadfc39269053c91b6ff76d2e14e7368203d6cc51931adfa2e3b986f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75a86dadfc39269053c91b6ff76d2e14e7368203d6cc51931adfa2e3b986f6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"message\\\":\\\"\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8441, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8442, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8444, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1125 09:24:54.000812 6252 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-r7x57_openshift-ovn-kubernetes(34502968-eb8d-47fa-83c2-7f6f6735ce2d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:05Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.838671 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:05Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.853477 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.853562 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.853579 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.853602 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.853619 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:05Z","lastTransitionTime":"2025-11-25T09:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.853835 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:05Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.869673 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:05Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.883910 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nvc68" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nvc68\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:05Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.908558 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ad63967f82781f8cf6d6c9512713cc8927666707203370dc38b031c201475a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:05Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.926667 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:05Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.940966 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:05Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.950739 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:05Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.956663 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.956704 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.956715 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.956736 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.956751 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:05Z","lastTransitionTime":"2025-11-25T09:25:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:05 crc kubenswrapper[4776]: I1125 09:25:05.966128 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:05Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.060027 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.060095 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.060112 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.060134 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.060150 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:06Z","lastTransitionTime":"2025-11-25T09:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.163057 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.163165 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.163185 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.163208 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.163225 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:06Z","lastTransitionTime":"2025-11-25T09:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.266832 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.266955 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.266983 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.267011 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.267029 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:06Z","lastTransitionTime":"2025-11-25T09:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.369941 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.370028 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.370042 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.370057 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.370094 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:06Z","lastTransitionTime":"2025-11-25T09:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.457695 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.473252 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.474375 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.474439 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.474457 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.474483 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.474500 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:06Z","lastTransitionTime":"2025-11-25T09:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.478603 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:06Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.505643 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:06Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.535481 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d75a86dadfc39269053c91b6ff76d2e14e7368203d6cc51931adfa2e3b986f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75a86dadfc39269053c91b6ff76d2e14e7368203d6cc51931adfa2e3b986f6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"message\\\":\\\"\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8441, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8442, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8444, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1125 09:24:54.000812 6252 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-r7x57_openshift-ovn-kubernetes(34502968-eb8d-47fa-83c2-7f6f6735ce2d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:06Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.546812 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:06Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.561534 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db22de3ee20f72a81d958f5fa9b10982a4ec446e092817b1880f669656e7a1ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:06Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.577414 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.577467 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.577478 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.577495 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.577526 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:06Z","lastTransitionTime":"2025-11-25T09:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.579301 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:06Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.592281 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nvc68" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nvc68\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:06Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.610935 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:06Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.626756 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:06Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.638124 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:06Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.654568 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:06Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.664001 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:06Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.675148 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:06Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.680342 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.680388 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.680399 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.680417 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.680427 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:06Z","lastTransitionTime":"2025-11-25T09:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.692658 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ad63967f82781f8cf6d6c9512713cc8927666707203370dc38b031c201475a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:06Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.707812 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a506265-8f64-456c-995f-1bc468cb87dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ead16f9d75513fe9ab6959e2ba0dca4f4b88df8bc65e3b0e6a21f28f2d0d30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb2778606c71db6545b88fd0648a8ab08a66863579467e49a5ece3b36d0d21f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rzwmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:06Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.726537 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:06Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.736051 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:06Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.783441 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.783491 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.783501 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.783518 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.783529 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:06Z","lastTransitionTime":"2025-11-25T09:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.885510 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.885568 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.885580 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.885597 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.885610 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:06Z","lastTransitionTime":"2025-11-25T09:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.988421 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.988459 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.988467 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.988483 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:06 crc kubenswrapper[4776]: I1125 09:25:06.988494 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:06Z","lastTransitionTime":"2025-11-25T09:25:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.091615 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.091680 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.091695 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.091717 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.091731 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:07Z","lastTransitionTime":"2025-11-25T09:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.194395 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.194429 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.194437 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.194451 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.194461 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:07Z","lastTransitionTime":"2025-11-25T09:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.297807 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.297872 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.297885 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.297908 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.297925 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:07Z","lastTransitionTime":"2025-11-25T09:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.401521 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.401574 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.401588 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.401610 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.401625 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:07Z","lastTransitionTime":"2025-11-25T09:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.506015 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.506110 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.506135 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.506165 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.506192 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:07Z","lastTransitionTime":"2025-11-25T09:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.527734 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.527942 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.528011 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.528058 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.528153 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:07 crc kubenswrapper[4776]: E1125 09:25:07.528328 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:25:07 crc kubenswrapper[4776]: E1125 09:25:07.528365 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:25:07 crc kubenswrapper[4776]: E1125 09:25:07.528386 4776 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:25:07 crc kubenswrapper[4776]: E1125 09:25:07.528382 4776 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:25:07 crc kubenswrapper[4776]: E1125 09:25:07.528456 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 09:25:39.528433502 +0000 UTC m=+84.569493095 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:25:07 crc kubenswrapper[4776]: E1125 09:25:07.528455 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:25:07 crc kubenswrapper[4776]: E1125 09:25:07.528507 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:25:07 crc kubenswrapper[4776]: E1125 09:25:07.528508 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:25:39.528472473 +0000 UTC m=+84.569532116 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:25:07 crc kubenswrapper[4776]: E1125 09:25:07.528534 4776 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:25:07 crc kubenswrapper[4776]: E1125 09:25:07.528582 4776 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:25:07 crc kubenswrapper[4776]: E1125 09:25:07.528605 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 09:25:39.528585226 +0000 UTC m=+84.569644809 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:25:07 crc kubenswrapper[4776]: E1125 09:25:07.528761 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:25:39.528724449 +0000 UTC m=+84.569784042 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:25:07 crc kubenswrapper[4776]: E1125 09:25:07.528830 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:25:39.52877812 +0000 UTC m=+84.569837703 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.609340 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.609419 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.609439 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.609465 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.609483 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:07Z","lastTransitionTime":"2025-11-25T09:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.662260 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.662356 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.662395 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.662269 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:07 crc kubenswrapper[4776]: E1125 09:25:07.662534 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:07 crc kubenswrapper[4776]: E1125 09:25:07.662664 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:07 crc kubenswrapper[4776]: E1125 09:25:07.662804 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:07 crc kubenswrapper[4776]: E1125 09:25:07.662907 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.712640 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.712739 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.712843 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.712878 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.712901 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:07Z","lastTransitionTime":"2025-11-25T09:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.816592 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.816659 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.816676 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.816711 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.816730 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:07Z","lastTransitionTime":"2025-11-25T09:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.920009 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.920119 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.920138 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.920166 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:07 crc kubenswrapper[4776]: I1125 09:25:07.920185 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:07Z","lastTransitionTime":"2025-11-25T09:25:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.023421 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.023489 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.023500 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.023521 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.023532 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:08Z","lastTransitionTime":"2025-11-25T09:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.127624 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.127680 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.127700 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.127728 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.127748 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:08Z","lastTransitionTime":"2025-11-25T09:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.231253 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.231321 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.231336 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.231362 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.231381 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:08Z","lastTransitionTime":"2025-11-25T09:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.335204 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.335271 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.335290 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.335317 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.335334 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:08Z","lastTransitionTime":"2025-11-25T09:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.439616 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.439678 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.439696 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.439726 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.439745 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:08Z","lastTransitionTime":"2025-11-25T09:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.543613 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.543983 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.544211 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.544364 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.544503 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:08Z","lastTransitionTime":"2025-11-25T09:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.647641 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.648112 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.648285 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.648426 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.648554 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:08Z","lastTransitionTime":"2025-11-25T09:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.758469 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.758547 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.758566 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.758594 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.758614 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:08Z","lastTransitionTime":"2025-11-25T09:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.861965 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.862022 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.862037 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.862060 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.862090 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:08Z","lastTransitionTime":"2025-11-25T09:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.965374 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.965457 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.965479 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.965508 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:08 crc kubenswrapper[4776]: I1125 09:25:08.965531 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:08Z","lastTransitionTime":"2025-11-25T09:25:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.068045 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.068140 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.068163 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.068192 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.068216 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:09Z","lastTransitionTime":"2025-11-25T09:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.170609 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.170675 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.170696 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.170726 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.170750 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:09Z","lastTransitionTime":"2025-11-25T09:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.273110 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.273158 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.273180 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.273200 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.273213 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:09Z","lastTransitionTime":"2025-11-25T09:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.376476 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.376544 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.376567 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.376600 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.376623 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:09Z","lastTransitionTime":"2025-11-25T09:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.480273 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.480367 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.480386 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.480448 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.480470 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:09Z","lastTransitionTime":"2025-11-25T09:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.584960 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.585042 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.585087 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.585109 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.585124 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:09Z","lastTransitionTime":"2025-11-25T09:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.661974 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:09 crc kubenswrapper[4776]: E1125 09:25:09.662170 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.662686 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:09 crc kubenswrapper[4776]: E1125 09:25:09.662772 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.662862 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:09 crc kubenswrapper[4776]: E1125 09:25:09.662958 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.663039 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:09 crc kubenswrapper[4776]: E1125 09:25:09.663140 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.689296 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.689355 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.689375 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.689418 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.689439 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:09Z","lastTransitionTime":"2025-11-25T09:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.792459 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.792509 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.792560 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.792582 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.792594 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:09Z","lastTransitionTime":"2025-11-25T09:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.895210 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.895255 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.895269 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.895286 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.895300 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:09Z","lastTransitionTime":"2025-11-25T09:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.999361 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.999428 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.999447 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.999469 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:09 crc kubenswrapper[4776]: I1125 09:25:09.999487 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:09Z","lastTransitionTime":"2025-11-25T09:25:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.102764 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.102822 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.102840 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.102868 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.102885 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:10Z","lastTransitionTime":"2025-11-25T09:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.205255 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.205338 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.205363 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.205399 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.205423 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:10Z","lastTransitionTime":"2025-11-25T09:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.308365 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.308424 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.308441 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.308465 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.308482 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:10Z","lastTransitionTime":"2025-11-25T09:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.411179 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.411214 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.411224 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.411239 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.411250 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:10Z","lastTransitionTime":"2025-11-25T09:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.514409 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.514649 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.514718 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.514749 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.514848 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:10Z","lastTransitionTime":"2025-11-25T09:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.618110 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.618163 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.618175 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.618189 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.618199 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:10Z","lastTransitionTime":"2025-11-25T09:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.721227 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.721294 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.721313 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.721384 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.721454 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:10Z","lastTransitionTime":"2025-11-25T09:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.824897 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.824954 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.824971 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.824999 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.825018 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:10Z","lastTransitionTime":"2025-11-25T09:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.928557 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.928651 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.928671 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.928695 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:10 crc kubenswrapper[4776]: I1125 09:25:10.928755 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:10Z","lastTransitionTime":"2025-11-25T09:25:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.036435 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.036500 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.036514 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.036540 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.036555 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:11Z","lastTransitionTime":"2025-11-25T09:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.141240 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.141314 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.141334 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.141363 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.141386 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:11Z","lastTransitionTime":"2025-11-25T09:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.244995 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.245053 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.245113 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.245147 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.245212 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:11Z","lastTransitionTime":"2025-11-25T09:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.348978 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.349414 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.349632 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.349792 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.349965 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:11Z","lastTransitionTime":"2025-11-25T09:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.453516 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.453572 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.453590 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.453617 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.453637 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:11Z","lastTransitionTime":"2025-11-25T09:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.557937 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.557998 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.558017 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.558043 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.558060 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:11Z","lastTransitionTime":"2025-11-25T09:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.660317 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.660382 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.660394 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.660409 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.660420 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:11Z","lastTransitionTime":"2025-11-25T09:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.661649 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.661692 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:11 crc kubenswrapper[4776]: E1125 09:25:11.661756 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.661649 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:11 crc kubenswrapper[4776]: E1125 09:25:11.661868 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.661897 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:11 crc kubenswrapper[4776]: E1125 09:25:11.661927 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:11 crc kubenswrapper[4776]: E1125 09:25:11.661988 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.762756 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.762851 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.762870 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.762893 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.763098 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:11Z","lastTransitionTime":"2025-11-25T09:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.785291 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-metrics-certs\") pod \"network-metrics-daemon-nvc68\" (UID: \"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\") " pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:11 crc kubenswrapper[4776]: E1125 09:25:11.785599 4776 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:25:11 crc kubenswrapper[4776]: E1125 09:25:11.785740 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-metrics-certs podName:4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087 nodeName:}" failed. No retries permitted until 2025-11-25 09:25:27.785709349 +0000 UTC m=+72.826769022 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-metrics-certs") pod "network-metrics-daemon-nvc68" (UID: "4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.866161 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.866213 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.866230 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.866252 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.866270 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:11Z","lastTransitionTime":"2025-11-25T09:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.969626 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.969692 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.969710 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.969737 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.969753 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:11Z","lastTransitionTime":"2025-11-25T09:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.971666 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.971756 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.971773 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.971792 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:11 crc kubenswrapper[4776]: I1125 09:25:11.971809 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:11Z","lastTransitionTime":"2025-11-25T09:25:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:11 crc kubenswrapper[4776]: E1125 09:25:11.994461 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:11Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.001384 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.001447 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.001465 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.001495 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.001521 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:12Z","lastTransitionTime":"2025-11-25T09:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:12 crc kubenswrapper[4776]: E1125 09:25:12.021715 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:12Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.027360 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.027411 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.027428 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.027452 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.027470 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:12Z","lastTransitionTime":"2025-11-25T09:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:12 crc kubenswrapper[4776]: E1125 09:25:12.046814 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:12Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.051431 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.051467 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.051479 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.051495 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.051507 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:12Z","lastTransitionTime":"2025-11-25T09:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:12 crc kubenswrapper[4776]: E1125 09:25:12.068534 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:12Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.073538 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.073584 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.073598 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.073615 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.073628 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:12Z","lastTransitionTime":"2025-11-25T09:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:12 crc kubenswrapper[4776]: E1125 09:25:12.093747 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:12Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:12 crc kubenswrapper[4776]: E1125 09:25:12.093889 4776 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.095912 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.095950 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.095961 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.095979 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.095992 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:12Z","lastTransitionTime":"2025-11-25T09:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.200300 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.200369 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.200384 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.200400 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.200413 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:12Z","lastTransitionTime":"2025-11-25T09:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.304057 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.304336 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.304480 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.304603 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.304716 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:12Z","lastTransitionTime":"2025-11-25T09:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.408907 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.408976 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.409015 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.409046 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.409100 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:12Z","lastTransitionTime":"2025-11-25T09:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.512547 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.512616 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.512633 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.512657 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.512675 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:12Z","lastTransitionTime":"2025-11-25T09:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.615671 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.615740 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.615763 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.615793 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.615817 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:12Z","lastTransitionTime":"2025-11-25T09:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.718827 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.718874 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.718886 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.718902 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.718915 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:12Z","lastTransitionTime":"2025-11-25T09:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.825736 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.825796 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.825813 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.825837 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.825856 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:12Z","lastTransitionTime":"2025-11-25T09:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.928310 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.928374 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.928387 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.928406 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:12 crc kubenswrapper[4776]: I1125 09:25:12.928420 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:12Z","lastTransitionTime":"2025-11-25T09:25:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.031508 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.031563 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.031576 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.031598 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.031612 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:13Z","lastTransitionTime":"2025-11-25T09:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.134305 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.134377 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.134394 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.134419 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.134436 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:13Z","lastTransitionTime":"2025-11-25T09:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.238190 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.238262 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.238287 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.238315 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.238336 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:13Z","lastTransitionTime":"2025-11-25T09:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.341483 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.341578 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.341598 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.341621 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.341638 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:13Z","lastTransitionTime":"2025-11-25T09:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.445038 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.445179 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.445207 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.445247 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.445285 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:13Z","lastTransitionTime":"2025-11-25T09:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.548305 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.548411 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.548435 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.548471 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.548536 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:13Z","lastTransitionTime":"2025-11-25T09:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.651505 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.651571 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.651589 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.651614 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.651633 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:13Z","lastTransitionTime":"2025-11-25T09:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.662309 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.662334 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.662394 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.662443 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:13 crc kubenswrapper[4776]: E1125 09:25:13.662643 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:13 crc kubenswrapper[4776]: E1125 09:25:13.662907 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:13 crc kubenswrapper[4776]: E1125 09:25:13.663440 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:13 crc kubenswrapper[4776]: E1125 09:25:13.663756 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.664051 4776 scope.go:117] "RemoveContainer" containerID="d75a86dadfc39269053c91b6ff76d2e14e7368203d6cc51931adfa2e3b986f6c" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.755055 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.755171 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.755196 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.755231 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.755253 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:13Z","lastTransitionTime":"2025-11-25T09:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.858077 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.858108 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.858119 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.858135 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.858147 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:13Z","lastTransitionTime":"2025-11-25T09:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.960864 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.961143 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.961155 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.961170 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:13 crc kubenswrapper[4776]: I1125 09:25:13.961183 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:13Z","lastTransitionTime":"2025-11-25T09:25:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.060121 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-r7x57_34502968-eb8d-47fa-83c2-7f6f6735ce2d/ovnkube-controller/1.log" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.062606 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.062631 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.062642 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.062660 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.062671 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:14Z","lastTransitionTime":"2025-11-25T09:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.063747 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" event={"ID":"34502968-eb8d-47fa-83c2-7f6f6735ce2d","Type":"ContainerStarted","Data":"5282edc3c7afa8b6281b182576c679d19b20f87a8e32ab32e64d775a17bb705a"} Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.064162 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.083817 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.101028 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.117136 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.143997 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5282edc3c7afa8b6281b182576c679d19b20f87a8e32ab32e64d775a17bb705a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75a86dadfc39269053c91b6ff76d2e14e7368203d6cc51931adfa2e3b986f6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"message\\\":\\\"\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8441, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8442, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8444, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1125 09:24:54.000812 6252 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.160726 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.167372 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.167426 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.167441 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.167461 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.167483 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:14Z","lastTransitionTime":"2025-11-25T09:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.179516 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db22de3ee20f72a81d958f5fa9b10982a4ec446e092817b1880f669656e7a1ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.192682 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.203902 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nvc68" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nvc68\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.215227 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.225287 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.234344 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.242618 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.254571 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.269977 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.270032 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.270050 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.270095 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.270109 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:14Z","lastTransitionTime":"2025-11-25T09:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.270824 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ad63967f82781f8cf6d6c9512713cc8927666707203370dc38b031c201475a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.283962 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26759602-0b52-410a-939d-c2df88f2b9fb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655c0dff64b7dbe451d7c61bf1e1e63051f02f7587bbb65e71b5ee7d2bec0a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284857ebafc167c12fe70604afa6702c071316774d5976a5e95042968dafe0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7b5ddb321a7c0a66d83ecb015a42c43b13ae4af284f4806d8fcf5a5f3ba005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583bd3f6f361dbdc9c02a6eeeee4a445e2d2d08f605adf2d3c8912cb7d416abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://583bd3f6f361dbdc9c02a6eeeee4a445e2d2d08f605adf2d3c8912cb7d416abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.296888 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.311766 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a506265-8f64-456c-995f-1bc468cb87dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ead16f9d75513fe9ab6959e2ba0dca4f4b88df8bc65e3b0e6a21f28f2d0d30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb2778606c71db6545b88fd0648a8ab08a66863579467e49a5ece3b36d0d21f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rzwmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.330340 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.372430 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.372468 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.372477 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.372493 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.372502 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:14Z","lastTransitionTime":"2025-11-25T09:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.475024 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.475096 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.475109 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.475127 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.475140 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:14Z","lastTransitionTime":"2025-11-25T09:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.577495 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.577533 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.577546 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.577564 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.577576 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:14Z","lastTransitionTime":"2025-11-25T09:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.680031 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.680125 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.680144 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.681474 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.681756 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:14Z","lastTransitionTime":"2025-11-25T09:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.784956 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.785014 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.785032 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.785053 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.785090 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:14Z","lastTransitionTime":"2025-11-25T09:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.888060 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.888110 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.888118 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.888130 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.888140 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:14Z","lastTransitionTime":"2025-11-25T09:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.991036 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.991307 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.991373 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.991451 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:14 crc kubenswrapper[4776]: I1125 09:25:14.991518 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:14Z","lastTransitionTime":"2025-11-25T09:25:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.069394 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-r7x57_34502968-eb8d-47fa-83c2-7f6f6735ce2d/ovnkube-controller/2.log" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.070540 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-r7x57_34502968-eb8d-47fa-83c2-7f6f6735ce2d/ovnkube-controller/1.log" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.074367 4776 generic.go:334] "Generic (PLEG): container finished" podID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerID="5282edc3c7afa8b6281b182576c679d19b20f87a8e32ab32e64d775a17bb705a" exitCode=1 Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.074440 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" event={"ID":"34502968-eb8d-47fa-83c2-7f6f6735ce2d","Type":"ContainerDied","Data":"5282edc3c7afa8b6281b182576c679d19b20f87a8e32ab32e64d775a17bb705a"} Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.074481 4776 scope.go:117] "RemoveContainer" containerID="d75a86dadfc39269053c91b6ff76d2e14e7368203d6cc51931adfa2e3b986f6c" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.076432 4776 scope.go:117] "RemoveContainer" containerID="5282edc3c7afa8b6281b182576c679d19b20f87a8e32ab32e64d775a17bb705a" Nov 25 09:25:15 crc kubenswrapper[4776]: E1125 09:25:15.076924 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-r7x57_openshift-ovn-kubernetes(34502968-eb8d-47fa-83c2-7f6f6735ce2d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.094751 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.094808 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.094825 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.094849 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.094869 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:15Z","lastTransitionTime":"2025-11-25T09:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.105843 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.125116 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.143182 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a506265-8f64-456c-995f-1bc468cb87dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ead16f9d75513fe9ab6959e2ba0dca4f4b88df8bc65e3b0e6a21f28f2d0d30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb2778606c71db6545b88fd0648a8ab08a66863579467e49a5ece3b36d0d21f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rzwmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.160362 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db22de3ee20f72a81d958f5fa9b10982a4ec446e092817b1880f669656e7a1ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.186591 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.197315 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.197385 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.197413 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.197439 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.197457 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:15Z","lastTransitionTime":"2025-11-25T09:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.205654 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.216504 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.234250 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5282edc3c7afa8b6281b182576c679d19b20f87a8e32ab32e64d775a17bb705a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75a86dadfc39269053c91b6ff76d2e14e7368203d6cc51931adfa2e3b986f6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"message\\\":\\\"\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8441, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8442, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8444, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1125 09:24:54.000812 6252 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5282edc3c7afa8b6281b182576c679d19b20f87a8e32ab32e64d775a17bb705a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:25:14Z\\\",\\\"message\\\":\\\"pods:v4/a13607449821398607916) with []\\\\nI1125 09:25:14.558154 6499 factory.go:1336] Added *v1.Node event handler 7\\\\nI1125 09:25:14.558204 6499 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1125 09:25:14.558314 6499 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 09:25:14.558339 6499 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 09:25:14.558375 6499 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 09:25:14.558427 6499 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 09:25:14.558441 6499 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 09:25:14.558440 6499 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 09:25:14.558467 6499 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 09:25:14.558469 6499 factory.go:656] Stopping watch factory\\\\nI1125 09:25:14.558484 6499 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 09:25:14.558510 6499 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 09:25:14.558633 6499 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 09:25:14.558670 6499 ovnkube.go:599] Stopped ovnkube\\\\nI1125 09:25:14.558702 6499 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 09:25:14.558785 6499 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.243992 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.256364 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.271411 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.283501 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nvc68" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nvc68\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.304573 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.304649 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.304668 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.304703 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.304720 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:15Z","lastTransitionTime":"2025-11-25T09:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.307407 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ad63967f82781f8cf6d6c9512713cc8927666707203370dc38b031c201475a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.318902 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26759602-0b52-410a-939d-c2df88f2b9fb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655c0dff64b7dbe451d7c61bf1e1e63051f02f7587bbb65e71b5ee7d2bec0a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284857ebafc167c12fe70604afa6702c071316774d5976a5e95042968dafe0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7b5ddb321a7c0a66d83ecb015a42c43b13ae4af284f4806d8fcf5a5f3ba005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583bd3f6f361dbdc9c02a6eeeee4a445e2d2d08f605adf2d3c8912cb7d416abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://583bd3f6f361dbdc9c02a6eeeee4a445e2d2d08f605adf2d3c8912cb7d416abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.331308 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.346667 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.357403 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.372130 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.407477 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.407517 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.407527 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.407565 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.407575 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:15Z","lastTransitionTime":"2025-11-25T09:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.510423 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.510500 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.510521 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.510549 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.510571 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:15Z","lastTransitionTime":"2025-11-25T09:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.613302 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.613425 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.613451 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.613521 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.613545 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:15Z","lastTransitionTime":"2025-11-25T09:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.661754 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.661799 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.661909 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:15 crc kubenswrapper[4776]: E1125 09:25:15.661898 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:15 crc kubenswrapper[4776]: E1125 09:25:15.662062 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.662122 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:15 crc kubenswrapper[4776]: E1125 09:25:15.662186 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:15 crc kubenswrapper[4776]: E1125 09:25:15.662260 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.681596 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.699702 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ad63967f82781f8cf6d6c9512713cc8927666707203370dc38b031c201475a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.714647 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26759602-0b52-410a-939d-c2df88f2b9fb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655c0dff64b7dbe451d7c61bf1e1e63051f02f7587bbb65e71b5ee7d2bec0a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284857ebafc167c12fe70604afa6702c071316774d5976a5e95042968dafe0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7b5ddb321a7c0a66d83ecb015a42c43b13ae4af284f4806d8fcf5a5f3ba005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583bd3f6f361dbdc9c02a6eeeee4a445e2d2d08f605adf2d3c8912cb7d416abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://583bd3f6f361dbdc9c02a6eeeee4a445e2d2d08f605adf2d3c8912cb7d416abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.716231 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.716287 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.716307 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.716332 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.716348 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:15Z","lastTransitionTime":"2025-11-25T09:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.730357 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.744509 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.757663 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.777063 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.790765 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.803531 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a506265-8f64-456c-995f-1bc468cb87dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ead16f9d75513fe9ab6959e2ba0dca4f4b88df8bc65e3b0e6a21f28f2d0d30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb2778606c71db6545b88fd0648a8ab08a66863579467e49a5ece3b36d0d21f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rzwmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.814249 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.818452 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.818501 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.818516 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.818538 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.818555 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:15Z","lastTransitionTime":"2025-11-25T09:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.827527 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db22de3ee20f72a81d958f5fa9b10982a4ec446e092817b1880f669656e7a1ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.840420 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.852009 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.863257 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.890579 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5282edc3c7afa8b6281b182576c679d19b20f87a8e32ab32e64d775a17bb705a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d75a86dadfc39269053c91b6ff76d2e14e7368203d6cc51931adfa2e3b986f6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"message\\\":\\\"\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-controllers_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-controllers\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8441, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8442, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.167\\\\\\\", Port:8444, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1125 09:24:54.000812 6252 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5282edc3c7afa8b6281b182576c679d19b20f87a8e32ab32e64d775a17bb705a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:25:14Z\\\",\\\"message\\\":\\\"pods:v4/a13607449821398607916) with []\\\\nI1125 09:25:14.558154 6499 factory.go:1336] Added *v1.Node event handler 7\\\\nI1125 09:25:14.558204 6499 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1125 09:25:14.558314 6499 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 09:25:14.558339 6499 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 09:25:14.558375 6499 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 09:25:14.558427 6499 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 09:25:14.558441 6499 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 09:25:14.558440 6499 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 09:25:14.558467 6499 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 09:25:14.558469 6499 factory.go:656] Stopping watch factory\\\\nI1125 09:25:14.558484 6499 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 09:25:14.558510 6499 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 09:25:14.558633 6499 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 09:25:14.558670 6499 ovnkube.go:599] Stopped ovnkube\\\\nI1125 09:25:14.558702 6499 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 09:25:14.558785 6499 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:25:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.914047 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.920647 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.920862 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.921029 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.921227 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.921446 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:15Z","lastTransitionTime":"2025-11-25T09:25:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.933938 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:15 crc kubenswrapper[4776]: I1125 09:25:15.943323 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nvc68" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nvc68\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:15Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.023926 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.023961 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.023969 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.023982 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.023990 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:16Z","lastTransitionTime":"2025-11-25T09:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.080009 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-r7x57_34502968-eb8d-47fa-83c2-7f6f6735ce2d/ovnkube-controller/2.log" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.086098 4776 scope.go:117] "RemoveContainer" containerID="5282edc3c7afa8b6281b182576c679d19b20f87a8e32ab32e64d775a17bb705a" Nov 25 09:25:16 crc kubenswrapper[4776]: E1125 09:25:16.086360 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-r7x57_openshift-ovn-kubernetes(34502968-eb8d-47fa-83c2-7f6f6735ce2d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.098833 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:16Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.114623 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ad63967f82781f8cf6d6c9512713cc8927666707203370dc38b031c201475a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:16Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.128034 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.128353 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.128422 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.128465 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.128489 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:16Z","lastTransitionTime":"2025-11-25T09:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.130192 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26759602-0b52-410a-939d-c2df88f2b9fb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655c0dff64b7dbe451d7c61bf1e1e63051f02f7587bbb65e71b5ee7d2bec0a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284857ebafc167c12fe70604afa6702c071316774d5976a5e95042968dafe0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7b5ddb321a7c0a66d83ecb015a42c43b13ae4af284f4806d8fcf5a5f3ba005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583bd3f6f361dbdc9c02a6eeeee4a445e2d2d08f605adf2d3c8912cb7d416abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://583bd3f6f361dbdc9c02a6eeeee4a445e2d2d08f605adf2d3c8912cb7d416abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:16Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.147981 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:16Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.165297 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:16Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.179151 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:16Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.202327 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:16Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.217235 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:16Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.228734 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a506265-8f64-456c-995f-1bc468cb87dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ead16f9d75513fe9ab6959e2ba0dca4f4b88df8bc65e3b0e6a21f28f2d0d30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb2778606c71db6545b88fd0648a8ab08a66863579467e49a5ece3b36d0d21f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rzwmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:16Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.230838 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.231462 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.231484 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.231506 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.231520 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:16Z","lastTransitionTime":"2025-11-25T09:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.240416 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:16Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.254240 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db22de3ee20f72a81d958f5fa9b10982a4ec446e092817b1880f669656e7a1ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:16Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.267845 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:16Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.281006 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:16Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.294385 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:16Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.310586 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5282edc3c7afa8b6281b182576c679d19b20f87a8e32ab32e64d775a17bb705a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5282edc3c7afa8b6281b182576c679d19b20f87a8e32ab32e64d775a17bb705a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:25:14Z\\\",\\\"message\\\":\\\"pods:v4/a13607449821398607916) with []\\\\nI1125 09:25:14.558154 6499 factory.go:1336] Added *v1.Node event handler 7\\\\nI1125 09:25:14.558204 6499 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1125 09:25:14.558314 6499 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 09:25:14.558339 6499 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 09:25:14.558375 6499 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 09:25:14.558427 6499 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 09:25:14.558441 6499 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 09:25:14.558440 6499 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 09:25:14.558467 6499 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 09:25:14.558469 6499 factory.go:656] Stopping watch factory\\\\nI1125 09:25:14.558484 6499 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 09:25:14.558510 6499 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 09:25:14.558633 6499 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 09:25:14.558670 6499 ovnkube.go:599] Stopped ovnkube\\\\nI1125 09:25:14.558702 6499 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 09:25:14.558785 6499 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:25:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-r7x57_openshift-ovn-kubernetes(34502968-eb8d-47fa-83c2-7f6f6735ce2d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:16Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.325292 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:16Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.333054 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.333106 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.333115 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.333127 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.333136 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:16Z","lastTransitionTime":"2025-11-25T09:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.342011 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:16Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.351953 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nvc68" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nvc68\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:16Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.435705 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.435737 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.435748 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.435763 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.435774 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:16Z","lastTransitionTime":"2025-11-25T09:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.538738 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.538782 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.538798 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.538822 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.538840 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:16Z","lastTransitionTime":"2025-11-25T09:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.641773 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.641843 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.641863 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.641893 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.641916 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:16Z","lastTransitionTime":"2025-11-25T09:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.744856 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.744920 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.744939 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.745170 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.745189 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:16Z","lastTransitionTime":"2025-11-25T09:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.848143 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.848246 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.848269 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.848296 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.848316 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:16Z","lastTransitionTime":"2025-11-25T09:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.951644 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.951683 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.951691 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.951705 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:16 crc kubenswrapper[4776]: I1125 09:25:16.951716 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:16Z","lastTransitionTime":"2025-11-25T09:25:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.054198 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.054279 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.054298 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.054328 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.054348 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:17Z","lastTransitionTime":"2025-11-25T09:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.158088 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.158139 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.158155 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.158177 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.158192 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:17Z","lastTransitionTime":"2025-11-25T09:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.261757 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.261798 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.261810 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.261827 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.261840 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:17Z","lastTransitionTime":"2025-11-25T09:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.364619 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.364675 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.364695 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.364727 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.364740 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:17Z","lastTransitionTime":"2025-11-25T09:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.468316 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.468363 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.468375 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.468392 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.468406 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:17Z","lastTransitionTime":"2025-11-25T09:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.570454 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.570501 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.570515 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.570532 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.570547 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:17Z","lastTransitionTime":"2025-11-25T09:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.663347 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.663366 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.663368 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.663403 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:17 crc kubenswrapper[4776]: E1125 09:25:17.663470 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:17 crc kubenswrapper[4776]: E1125 09:25:17.663580 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:17 crc kubenswrapper[4776]: E1125 09:25:17.663647 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:17 crc kubenswrapper[4776]: E1125 09:25:17.663836 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.673646 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.673948 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.674034 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.674187 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.674282 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:17Z","lastTransitionTime":"2025-11-25T09:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.777233 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.777291 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.777315 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.777340 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.777364 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:17Z","lastTransitionTime":"2025-11-25T09:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.880203 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.881213 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.881257 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.881295 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.881315 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:17Z","lastTransitionTime":"2025-11-25T09:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.984167 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.984237 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.984256 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.984281 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:17 crc kubenswrapper[4776]: I1125 09:25:17.984299 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:17Z","lastTransitionTime":"2025-11-25T09:25:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.087237 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.088129 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.088740 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.088938 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.089111 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:18Z","lastTransitionTime":"2025-11-25T09:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.191777 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.191808 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.191820 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.191836 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.191847 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:18Z","lastTransitionTime":"2025-11-25T09:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.294669 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.295012 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.295045 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.295115 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.295139 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:18Z","lastTransitionTime":"2025-11-25T09:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.397854 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.397899 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.397911 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.397929 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.397941 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:18Z","lastTransitionTime":"2025-11-25T09:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.501375 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.501420 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.501431 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.501446 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.501460 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:18Z","lastTransitionTime":"2025-11-25T09:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.604151 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.604193 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.604204 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.604222 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.604234 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:18Z","lastTransitionTime":"2025-11-25T09:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.706302 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.706356 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.706372 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.706394 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.706414 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:18Z","lastTransitionTime":"2025-11-25T09:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.809900 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.809979 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.810003 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.810028 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.810046 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:18Z","lastTransitionTime":"2025-11-25T09:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.913193 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.913242 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.913254 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.913272 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:18 crc kubenswrapper[4776]: I1125 09:25:18.913284 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:18Z","lastTransitionTime":"2025-11-25T09:25:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.016699 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.016990 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.017192 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.017347 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.017486 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:19Z","lastTransitionTime":"2025-11-25T09:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.120244 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.120309 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.120328 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.120353 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.120369 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:19Z","lastTransitionTime":"2025-11-25T09:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.222663 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.222726 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.222752 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.222769 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.222778 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:19Z","lastTransitionTime":"2025-11-25T09:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.324882 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.325172 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.325282 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.325376 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.325477 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:19Z","lastTransitionTime":"2025-11-25T09:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.429307 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.429845 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.430000 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.430194 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.430463 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:19Z","lastTransitionTime":"2025-11-25T09:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.533252 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.533307 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.533322 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.533343 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.533360 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:19Z","lastTransitionTime":"2025-11-25T09:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.636722 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.636758 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.636766 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.636780 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.636788 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:19Z","lastTransitionTime":"2025-11-25T09:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.662200 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.662271 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:19 crc kubenswrapper[4776]: E1125 09:25:19.662295 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.662368 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.662203 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:19 crc kubenswrapper[4776]: E1125 09:25:19.662416 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:19 crc kubenswrapper[4776]: E1125 09:25:19.662441 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:19 crc kubenswrapper[4776]: E1125 09:25:19.662487 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.739472 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.739520 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.739532 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.739550 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.739561 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:19Z","lastTransitionTime":"2025-11-25T09:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.841981 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.842295 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.842322 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.842345 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.842362 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:19Z","lastTransitionTime":"2025-11-25T09:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.945733 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.946041 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.946332 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.946519 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:19 crc kubenswrapper[4776]: I1125 09:25:19.946661 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:19Z","lastTransitionTime":"2025-11-25T09:25:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.049573 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.049610 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.049625 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.049643 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.049653 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:20Z","lastTransitionTime":"2025-11-25T09:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.152832 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.152897 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.152909 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.152924 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.152935 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:20Z","lastTransitionTime":"2025-11-25T09:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.255377 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.255406 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.255417 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.255434 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.255444 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:20Z","lastTransitionTime":"2025-11-25T09:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.358046 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.358102 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.358115 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.358134 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.358151 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:20Z","lastTransitionTime":"2025-11-25T09:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.460534 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.460647 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.460674 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.460719 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.460743 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:20Z","lastTransitionTime":"2025-11-25T09:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.563460 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.563500 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.563514 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.563572 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.563586 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:20Z","lastTransitionTime":"2025-11-25T09:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.666760 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.666812 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.666823 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.666839 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.666855 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:20Z","lastTransitionTime":"2025-11-25T09:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.769229 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.769275 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.769285 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.769301 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.769312 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:20Z","lastTransitionTime":"2025-11-25T09:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.871319 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.871348 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.871358 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.871371 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.871381 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:20Z","lastTransitionTime":"2025-11-25T09:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.974355 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.974676 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.974798 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.974895 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:20 crc kubenswrapper[4776]: I1125 09:25:20.974961 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:20Z","lastTransitionTime":"2025-11-25T09:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.077360 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.077646 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.077770 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.077876 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.077938 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:21Z","lastTransitionTime":"2025-11-25T09:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.180903 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.180948 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.180957 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.180975 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.180987 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:21Z","lastTransitionTime":"2025-11-25T09:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.291578 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.291617 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.291626 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.291644 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.291654 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:21Z","lastTransitionTime":"2025-11-25T09:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.393946 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.394230 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.394268 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.394293 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.394313 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:21Z","lastTransitionTime":"2025-11-25T09:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.496706 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.497026 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.497211 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.497389 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.497570 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:21Z","lastTransitionTime":"2025-11-25T09:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.600399 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.600699 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.600817 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.600934 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.601052 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:21Z","lastTransitionTime":"2025-11-25T09:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.662294 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:21 crc kubenswrapper[4776]: E1125 09:25:21.662438 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.662889 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.663018 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.663103 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:21 crc kubenswrapper[4776]: E1125 09:25:21.663123 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:21 crc kubenswrapper[4776]: E1125 09:25:21.664346 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:21 crc kubenswrapper[4776]: E1125 09:25:21.664446 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.704343 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.704382 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.704394 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.704409 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.704421 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:21Z","lastTransitionTime":"2025-11-25T09:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.807642 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.807685 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.807697 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.807739 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.807750 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:21Z","lastTransitionTime":"2025-11-25T09:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.910272 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.911144 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.911165 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.911197 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:21 crc kubenswrapper[4776]: I1125 09:25:21.911207 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:21Z","lastTransitionTime":"2025-11-25T09:25:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.013451 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.013502 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.013514 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.013532 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.013544 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:22Z","lastTransitionTime":"2025-11-25T09:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.115639 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.115680 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.115691 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.115711 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.115730 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:22Z","lastTransitionTime":"2025-11-25T09:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.218954 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.219001 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.219012 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.219028 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.219037 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:22Z","lastTransitionTime":"2025-11-25T09:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.322338 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.322656 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.322756 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.322873 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.322960 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:22Z","lastTransitionTime":"2025-11-25T09:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.425766 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.425845 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.425869 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.425896 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.425913 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:22Z","lastTransitionTime":"2025-11-25T09:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.461302 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.461337 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.461345 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.461359 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.461369 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:22Z","lastTransitionTime":"2025-11-25T09:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:22 crc kubenswrapper[4776]: E1125 09:25:22.480478 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:22Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.485744 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.485805 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.485823 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.485845 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.485862 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:22Z","lastTransitionTime":"2025-11-25T09:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:22 crc kubenswrapper[4776]: E1125 09:25:22.505749 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:22Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.513910 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.513972 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.513991 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.514018 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.514043 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:22Z","lastTransitionTime":"2025-11-25T09:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:22 crc kubenswrapper[4776]: E1125 09:25:22.527818 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:22Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.531340 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.531409 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.531421 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.531436 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.531446 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:22Z","lastTransitionTime":"2025-11-25T09:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:22 crc kubenswrapper[4776]: E1125 09:25:22.547149 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:22Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.550984 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.551016 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.551028 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.551059 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.551101 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:22Z","lastTransitionTime":"2025-11-25T09:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:22 crc kubenswrapper[4776]: E1125 09:25:22.564974 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:22Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:22 crc kubenswrapper[4776]: E1125 09:25:22.565163 4776 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.566554 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.566579 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.566591 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.566624 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.566636 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:22Z","lastTransitionTime":"2025-11-25T09:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.668475 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.668505 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.668513 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.668540 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.668549 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:22Z","lastTransitionTime":"2025-11-25T09:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.770135 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.770169 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.770179 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.770190 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.770200 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:22Z","lastTransitionTime":"2025-11-25T09:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.872782 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.872812 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.872823 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.872837 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.872846 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:22Z","lastTransitionTime":"2025-11-25T09:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.975503 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.975546 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.975560 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.975579 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:22 crc kubenswrapper[4776]: I1125 09:25:22.975591 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:22Z","lastTransitionTime":"2025-11-25T09:25:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.077445 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.077481 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.077490 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.077503 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.077513 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:23Z","lastTransitionTime":"2025-11-25T09:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.180542 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.180583 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.180591 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.180606 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.180617 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:23Z","lastTransitionTime":"2025-11-25T09:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.283153 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.283227 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.283246 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.283278 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.283298 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:23Z","lastTransitionTime":"2025-11-25T09:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.386791 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.386831 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.386841 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.386857 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.386867 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:23Z","lastTransitionTime":"2025-11-25T09:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.489577 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.489640 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.489659 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.489684 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.489701 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:23Z","lastTransitionTime":"2025-11-25T09:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.592516 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.592562 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.592578 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.592600 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.592619 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:23Z","lastTransitionTime":"2025-11-25T09:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.661764 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.661864 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.661864 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.661785 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:23 crc kubenswrapper[4776]: E1125 09:25:23.661991 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:23 crc kubenswrapper[4776]: E1125 09:25:23.662148 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:23 crc kubenswrapper[4776]: E1125 09:25:23.662454 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:23 crc kubenswrapper[4776]: E1125 09:25:23.662550 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.695120 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.695160 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.695171 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.695183 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.695192 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:23Z","lastTransitionTime":"2025-11-25T09:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.797206 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.797264 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.797281 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.797305 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.797322 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:23Z","lastTransitionTime":"2025-11-25T09:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.900432 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.900473 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.900484 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.900499 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:23 crc kubenswrapper[4776]: I1125 09:25:23.900510 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:23Z","lastTransitionTime":"2025-11-25T09:25:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.002729 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.003258 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.003324 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.003391 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.003490 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:24Z","lastTransitionTime":"2025-11-25T09:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.106365 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.106628 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.106694 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.106759 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.106818 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:24Z","lastTransitionTime":"2025-11-25T09:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.210173 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.210210 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.210219 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.210234 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.210242 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:24Z","lastTransitionTime":"2025-11-25T09:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.312105 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.312158 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.312172 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.312189 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.312201 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:24Z","lastTransitionTime":"2025-11-25T09:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.414665 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.414699 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.414708 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.414720 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.414729 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:24Z","lastTransitionTime":"2025-11-25T09:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.516889 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.516935 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.516944 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.516961 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.516970 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:24Z","lastTransitionTime":"2025-11-25T09:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.619632 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.619678 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.619689 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.619707 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.619722 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:24Z","lastTransitionTime":"2025-11-25T09:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.722169 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.722224 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.722238 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.722254 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.722265 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:24Z","lastTransitionTime":"2025-11-25T09:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.825600 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.825650 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.825662 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.825681 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.825695 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:24Z","lastTransitionTime":"2025-11-25T09:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.927715 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.927758 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.927771 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.927787 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:24 crc kubenswrapper[4776]: I1125 09:25:24.927800 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:24Z","lastTransitionTime":"2025-11-25T09:25:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.030377 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.030415 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.030424 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.030436 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.030445 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:25Z","lastTransitionTime":"2025-11-25T09:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.132894 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.132944 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.132961 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.132985 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.133001 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:25Z","lastTransitionTime":"2025-11-25T09:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.235170 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.235375 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.235394 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.235412 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.235427 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:25Z","lastTransitionTime":"2025-11-25T09:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.338136 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.338193 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.338210 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.338235 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.338252 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:25Z","lastTransitionTime":"2025-11-25T09:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.440905 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.440955 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.440967 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.440986 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.440998 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:25Z","lastTransitionTime":"2025-11-25T09:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.543184 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.543255 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.543280 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.543311 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.543332 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:25Z","lastTransitionTime":"2025-11-25T09:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.646682 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.646730 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.646745 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.646764 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.646779 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:25Z","lastTransitionTime":"2025-11-25T09:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.661331 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.661370 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.661394 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.661333 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:25 crc kubenswrapper[4776]: E1125 09:25:25.661527 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:25 crc kubenswrapper[4776]: E1125 09:25:25.661646 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:25 crc kubenswrapper[4776]: E1125 09:25:25.661747 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:25 crc kubenswrapper[4776]: E1125 09:25:25.661909 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.675987 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db22de3ee20f72a81d958f5fa9b10982a4ec446e092817b1880f669656e7a1ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:25Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.690452 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:25Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.703586 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:25Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.717494 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:25Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.737896 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5282edc3c7afa8b6281b182576c679d19b20f87a8e32ab32e64d775a17bb705a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5282edc3c7afa8b6281b182576c679d19b20f87a8e32ab32e64d775a17bb705a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:25:14Z\\\",\\\"message\\\":\\\"pods:v4/a13607449821398607916) with []\\\\nI1125 09:25:14.558154 6499 factory.go:1336] Added *v1.Node event handler 7\\\\nI1125 09:25:14.558204 6499 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1125 09:25:14.558314 6499 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 09:25:14.558339 6499 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 09:25:14.558375 6499 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 09:25:14.558427 6499 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 09:25:14.558441 6499 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 09:25:14.558440 6499 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 09:25:14.558467 6499 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 09:25:14.558469 6499 factory.go:656] Stopping watch factory\\\\nI1125 09:25:14.558484 6499 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 09:25:14.558510 6499 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 09:25:14.558633 6499 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 09:25:14.558670 6499 ovnkube.go:599] Stopped ovnkube\\\\nI1125 09:25:14.558702 6499 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 09:25:14.558785 6499 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:25:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-r7x57_openshift-ovn-kubernetes(34502968-eb8d-47fa-83c2-7f6f6735ce2d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:25Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.747780 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:25Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.749039 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.749092 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.749111 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.749133 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.749148 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:25Z","lastTransitionTime":"2025-11-25T09:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.758522 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:25Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.768185 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:25Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.777634 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nvc68" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nvc68\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:25Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.793967 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ad63967f82781f8cf6d6c9512713cc8927666707203370dc38b031c201475a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:25Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.805911 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26759602-0b52-410a-939d-c2df88f2b9fb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655c0dff64b7dbe451d7c61bf1e1e63051f02f7587bbb65e71b5ee7d2bec0a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284857ebafc167c12fe70604afa6702c071316774d5976a5e95042968dafe0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7b5ddb321a7c0a66d83ecb015a42c43b13ae4af284f4806d8fcf5a5f3ba005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583bd3f6f361dbdc9c02a6eeeee4a445e2d2d08f605adf2d3c8912cb7d416abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://583bd3f6f361dbdc9c02a6eeeee4a445e2d2d08f605adf2d3c8912cb7d416abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:25Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.820175 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:25Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.831310 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:25Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.840330 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:25Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.851815 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.851855 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.851865 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.851879 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.851888 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:25Z","lastTransitionTime":"2025-11-25T09:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.852685 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:25Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.870476 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:25Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.881687 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:25Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.892802 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a506265-8f64-456c-995f-1bc468cb87dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ead16f9d75513fe9ab6959e2ba0dca4f4b88df8bc65e3b0e6a21f28f2d0d30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb2778606c71db6545b88fd0648a8ab08a66863579467e49a5ece3b36d0d21f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rzwmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:25Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.953927 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.954209 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.954305 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.954395 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:25 crc kubenswrapper[4776]: I1125 09:25:25.954476 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:25Z","lastTransitionTime":"2025-11-25T09:25:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.057293 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.057807 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.057885 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.057952 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.058010 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:26Z","lastTransitionTime":"2025-11-25T09:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.159905 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.160206 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.160523 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.160792 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.161028 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:26Z","lastTransitionTime":"2025-11-25T09:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.263841 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.263873 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.263884 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.263898 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.263909 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:26Z","lastTransitionTime":"2025-11-25T09:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.365884 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.365910 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.365919 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.365934 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.365944 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:26Z","lastTransitionTime":"2025-11-25T09:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.468304 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.468736 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.468859 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.468957 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.469040 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:26Z","lastTransitionTime":"2025-11-25T09:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.572044 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.572339 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.572416 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.572481 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.572575 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:26Z","lastTransitionTime":"2025-11-25T09:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.674504 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.674731 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.674803 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.674896 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.675002 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:26Z","lastTransitionTime":"2025-11-25T09:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.778488 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.778534 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.778546 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.778564 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.778574 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:26Z","lastTransitionTime":"2025-11-25T09:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.881698 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.882351 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.882520 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.882678 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.882856 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:26Z","lastTransitionTime":"2025-11-25T09:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.986480 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.986513 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.986522 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.986536 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:26 crc kubenswrapper[4776]: I1125 09:25:26.986545 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:26Z","lastTransitionTime":"2025-11-25T09:25:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.089353 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.089389 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.089400 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.089419 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.089431 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:27Z","lastTransitionTime":"2025-11-25T09:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.197026 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.197123 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.197185 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.197220 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.197243 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:27Z","lastTransitionTime":"2025-11-25T09:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.300885 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.300937 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.300949 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.300969 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.300983 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:27Z","lastTransitionTime":"2025-11-25T09:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.402997 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.403036 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.403048 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.403085 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.403099 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:27Z","lastTransitionTime":"2025-11-25T09:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.505554 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.505599 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.505609 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.505625 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.505634 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:27Z","lastTransitionTime":"2025-11-25T09:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.608096 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.608140 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.608174 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.608190 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.608200 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:27Z","lastTransitionTime":"2025-11-25T09:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.662267 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.662351 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.662378 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:27 crc kubenswrapper[4776]: E1125 09:25:27.662453 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.662519 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:27 crc kubenswrapper[4776]: E1125 09:25:27.662615 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:27 crc kubenswrapper[4776]: E1125 09:25:27.662761 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:27 crc kubenswrapper[4776]: E1125 09:25:27.662866 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.710449 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.710507 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.710527 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.710551 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.710568 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:27Z","lastTransitionTime":"2025-11-25T09:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.813021 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.813115 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.813132 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.813162 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.813179 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:27Z","lastTransitionTime":"2025-11-25T09:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.858468 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-metrics-certs\") pod \"network-metrics-daemon-nvc68\" (UID: \"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\") " pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:27 crc kubenswrapper[4776]: E1125 09:25:27.858614 4776 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:25:27 crc kubenswrapper[4776]: E1125 09:25:27.858673 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-metrics-certs podName:4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087 nodeName:}" failed. No retries permitted until 2025-11-25 09:25:59.858655395 +0000 UTC m=+104.899714958 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-metrics-certs") pod "network-metrics-daemon-nvc68" (UID: "4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.915880 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.915973 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.916041 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.916133 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:27 crc kubenswrapper[4776]: I1125 09:25:27.916163 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:27Z","lastTransitionTime":"2025-11-25T09:25:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.019122 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.019158 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.019169 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.019186 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.019199 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:28Z","lastTransitionTime":"2025-11-25T09:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.122509 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.122575 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.122598 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.122625 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.122646 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:28Z","lastTransitionTime":"2025-11-25T09:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.224617 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.224647 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.224673 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.224688 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.224699 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:28Z","lastTransitionTime":"2025-11-25T09:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.327007 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.327060 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.327165 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.327197 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.327222 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:28Z","lastTransitionTime":"2025-11-25T09:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.433247 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.433317 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.433336 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.433361 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.433380 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:28Z","lastTransitionTime":"2025-11-25T09:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.536394 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.536437 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.536447 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.536461 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.536470 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:28Z","lastTransitionTime":"2025-11-25T09:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.641285 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.641368 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.641393 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.641421 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.641442 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:28Z","lastTransitionTime":"2025-11-25T09:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.744243 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.744279 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.744290 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.744306 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.744316 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:28Z","lastTransitionTime":"2025-11-25T09:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.847028 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.847078 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.847088 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.847102 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.847110 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:28Z","lastTransitionTime":"2025-11-25T09:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.949762 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.949816 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.949833 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.949854 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:28 crc kubenswrapper[4776]: I1125 09:25:28.949870 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:28Z","lastTransitionTime":"2025-11-25T09:25:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.052576 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.052623 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.052635 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.052652 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.052665 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:29Z","lastTransitionTime":"2025-11-25T09:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.155150 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.155206 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.155223 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.155246 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.155266 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:29Z","lastTransitionTime":"2025-11-25T09:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.258463 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.258539 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.258561 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.258589 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.258610 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:29Z","lastTransitionTime":"2025-11-25T09:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.361891 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.362059 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.362105 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.362130 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.362146 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:29Z","lastTransitionTime":"2025-11-25T09:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.464836 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.464895 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.464912 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.464935 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.464953 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:29Z","lastTransitionTime":"2025-11-25T09:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.567659 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.567744 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.567766 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.567796 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.567817 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:29Z","lastTransitionTime":"2025-11-25T09:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.662323 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.662389 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:29 crc kubenswrapper[4776]: E1125 09:25:29.662478 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.662536 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.662560 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:29 crc kubenswrapper[4776]: E1125 09:25:29.662712 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:29 crc kubenswrapper[4776]: E1125 09:25:29.662774 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.665249 4776 scope.go:117] "RemoveContainer" containerID="5282edc3c7afa8b6281b182576c679d19b20f87a8e32ab32e64d775a17bb705a" Nov 25 09:25:29 crc kubenswrapper[4776]: E1125 09:25:29.665530 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-r7x57_openshift-ovn-kubernetes(34502968-eb8d-47fa-83c2-7f6f6735ce2d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" Nov 25 09:25:29 crc kubenswrapper[4776]: E1125 09:25:29.669505 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.671094 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.671130 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.671143 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.671167 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.671179 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:29Z","lastTransitionTime":"2025-11-25T09:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.774460 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.774516 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.774532 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.774555 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.774571 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:29Z","lastTransitionTime":"2025-11-25T09:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.877863 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.877925 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.877941 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.877965 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.878015 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:29Z","lastTransitionTime":"2025-11-25T09:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.981324 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.981364 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.981376 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.981395 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:29 crc kubenswrapper[4776]: I1125 09:25:29.981411 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:29Z","lastTransitionTime":"2025-11-25T09:25:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.084293 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.084345 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.084361 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.084385 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.084402 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:30Z","lastTransitionTime":"2025-11-25T09:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.131203 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5q5c6_d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c/kube-multus/0.log" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.131242 4776 generic.go:334] "Generic (PLEG): container finished" podID="d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c" containerID="c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9" exitCode=1 Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.131268 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5q5c6" event={"ID":"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c","Type":"ContainerDied","Data":"c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9"} Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.131592 4776 scope.go:117] "RemoveContainer" containerID="c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.155374 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:30Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.176086 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:30Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.187925 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.187956 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.187967 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.187983 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.187994 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:30Z","lastTransitionTime":"2025-11-25T09:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.191535 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:30Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.215611 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5282edc3c7afa8b6281b182576c679d19b20f87a8e32ab32e64d775a17bb705a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5282edc3c7afa8b6281b182576c679d19b20f87a8e32ab32e64d775a17bb705a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:25:14Z\\\",\\\"message\\\":\\\"pods:v4/a13607449821398607916) with []\\\\nI1125 09:25:14.558154 6499 factory.go:1336] Added *v1.Node event handler 7\\\\nI1125 09:25:14.558204 6499 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1125 09:25:14.558314 6499 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 09:25:14.558339 6499 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 09:25:14.558375 6499 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 09:25:14.558427 6499 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 09:25:14.558441 6499 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 09:25:14.558440 6499 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 09:25:14.558467 6499 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 09:25:14.558469 6499 factory.go:656] Stopping watch factory\\\\nI1125 09:25:14.558484 6499 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 09:25:14.558510 6499 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 09:25:14.558633 6499 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 09:25:14.558670 6499 ovnkube.go:599] Stopped ovnkube\\\\nI1125 09:25:14.558702 6499 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 09:25:14.558785 6499 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:25:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-r7x57_openshift-ovn-kubernetes(34502968-eb8d-47fa-83c2-7f6f6735ce2d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:30Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.226327 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:30Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.241907 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db22de3ee20f72a81d958f5fa9b10982a4ec446e092817b1880f669656e7a1ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:30Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.253843 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:30Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.265193 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nvc68" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nvc68\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:30Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.279567 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:30Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.291402 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.291461 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.291474 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.291490 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.291502 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:30Z","lastTransitionTime":"2025-11-25T09:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.295031 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:30Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.309432 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:30Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.321931 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:30Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.338820 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:25:29Z\\\",\\\"message\\\":\\\"2025-11-25T09:24:43+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c6e651f3-f403-47a2-b38e-0d19d75e1f16\\\\n2025-11-25T09:24:43+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c6e651f3-f403-47a2-b38e-0d19d75e1f16 to /host/opt/cni/bin/\\\\n2025-11-25T09:24:44Z [verbose] multus-daemon started\\\\n2025-11-25T09:24:44Z [verbose] Readiness Indicator file check\\\\n2025-11-25T09:25:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:30Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.357221 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ad63967f82781f8cf6d6c9512713cc8927666707203370dc38b031c201475a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:30Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.371754 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26759602-0b52-410a-939d-c2df88f2b9fb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655c0dff64b7dbe451d7c61bf1e1e63051f02f7587bbb65e71b5ee7d2bec0a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284857ebafc167c12fe70604afa6702c071316774d5976a5e95042968dafe0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7b5ddb321a7c0a66d83ecb015a42c43b13ae4af284f4806d8fcf5a5f3ba005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583bd3f6f361dbdc9c02a6eeeee4a445e2d2d08f605adf2d3c8912cb7d416abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://583bd3f6f361dbdc9c02a6eeeee4a445e2d2d08f605adf2d3c8912cb7d416abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:30Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.387035 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:30Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.395479 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.395714 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.395869 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.396008 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.396178 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:30Z","lastTransitionTime":"2025-11-25T09:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.404524 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a506265-8f64-456c-995f-1bc468cb87dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ead16f9d75513fe9ab6959e2ba0dca4f4b88df8bc65e3b0e6a21f28f2d0d30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb2778606c71db6545b88fd0648a8ab08a66863579467e49a5ece3b36d0d21f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rzwmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:30Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.434651 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:30Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.498880 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.499305 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.499443 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.499615 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.499736 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:30Z","lastTransitionTime":"2025-11-25T09:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.602402 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.602433 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.602441 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.602455 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.602464 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:30Z","lastTransitionTime":"2025-11-25T09:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.677375 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.706536 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.706608 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.706629 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.706658 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.706688 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:30Z","lastTransitionTime":"2025-11-25T09:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.809755 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.809829 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.809847 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.809870 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.809886 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:30Z","lastTransitionTime":"2025-11-25T09:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.913435 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.913492 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.913504 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.913523 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:30 crc kubenswrapper[4776]: I1125 09:25:30.913533 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:30Z","lastTransitionTime":"2025-11-25T09:25:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.016903 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.016958 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.016975 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.016996 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.017013 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:31Z","lastTransitionTime":"2025-11-25T09:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.119742 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.119781 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.119793 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.119806 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.119816 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:31Z","lastTransitionTime":"2025-11-25T09:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.137507 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5q5c6_d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c/kube-multus/0.log" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.138042 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5q5c6" event={"ID":"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c","Type":"ContainerStarted","Data":"d7b1ea649a64d7a158675b397137593f6b901e762b67261d2e2872eeedc0d895"} Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.156159 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:31Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.171331 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:31Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.189229 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nvc68" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nvc68\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:31Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.202853 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:31Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.216174 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:31Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.222220 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.222268 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.222285 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.222309 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.222330 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:31Z","lastTransitionTime":"2025-11-25T09:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.232792 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b1ea649a64d7a158675b397137593f6b901e762b67261d2e2872eeedc0d895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:25:29Z\\\",\\\"message\\\":\\\"2025-11-25T09:24:43+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c6e651f3-f403-47a2-b38e-0d19d75e1f16\\\\n2025-11-25T09:24:43+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c6e651f3-f403-47a2-b38e-0d19d75e1f16 to /host/opt/cni/bin/\\\\n2025-11-25T09:24:44Z [verbose] multus-daemon started\\\\n2025-11-25T09:24:44Z [verbose] Readiness Indicator file check\\\\n2025-11-25T09:25:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:25:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:31Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.257444 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ad63967f82781f8cf6d6c9512713cc8927666707203370dc38b031c201475a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:31Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.282795 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26759602-0b52-410a-939d-c2df88f2b9fb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655c0dff64b7dbe451d7c61bf1e1e63051f02f7587bbb65e71b5ee7d2bec0a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284857ebafc167c12fe70604afa6702c071316774d5976a5e95042968dafe0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7b5ddb321a7c0a66d83ecb015a42c43b13ae4af284f4806d8fcf5a5f3ba005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583bd3f6f361dbdc9c02a6eeeee4a445e2d2d08f605adf2d3c8912cb7d416abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://583bd3f6f361dbdc9c02a6eeeee4a445e2d2d08f605adf2d3c8912cb7d416abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:31Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.296304 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51b8c180-5c19-4df6-b088-f1656ef0946e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b232d5cd319131183cefb0df8c537e099bfa1743d7fab3748858d3873a4802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9c968afbab8870d4105cc2b407cb7b2392919716c0a9b5de3ed47f3d12e5a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f9c968afbab8870d4105cc2b407cb7b2392919716c0a9b5de3ed47f3d12e5a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:31Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.316279 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:31Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.325877 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.325939 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.325963 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.325993 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.326015 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:31Z","lastTransitionTime":"2025-11-25T09:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.353408 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:31Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.365100 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:31Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.377185 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a506265-8f64-456c-995f-1bc468cb87dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ead16f9d75513fe9ab6959e2ba0dca4f4b88df8bc65e3b0e6a21f28f2d0d30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb2778606c71db6545b88fd0648a8ab08a66863579467e49a5ece3b36d0d21f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rzwmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:31Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.394955 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:31Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.427591 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5282edc3c7afa8b6281b182576c679d19b20f87a8e32ab32e64d775a17bb705a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5282edc3c7afa8b6281b182576c679d19b20f87a8e32ab32e64d775a17bb705a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:25:14Z\\\",\\\"message\\\":\\\"pods:v4/a13607449821398607916) with []\\\\nI1125 09:25:14.558154 6499 factory.go:1336] Added *v1.Node event handler 7\\\\nI1125 09:25:14.558204 6499 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1125 09:25:14.558314 6499 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 09:25:14.558339 6499 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 09:25:14.558375 6499 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 09:25:14.558427 6499 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 09:25:14.558441 6499 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 09:25:14.558440 6499 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 09:25:14.558467 6499 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 09:25:14.558469 6499 factory.go:656] Stopping watch factory\\\\nI1125 09:25:14.558484 6499 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 09:25:14.558510 6499 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 09:25:14.558633 6499 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 09:25:14.558670 6499 ovnkube.go:599] Stopped ovnkube\\\\nI1125 09:25:14.558702 6499 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 09:25:14.558785 6499 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:25:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-r7x57_openshift-ovn-kubernetes(34502968-eb8d-47fa-83c2-7f6f6735ce2d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:31Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.430162 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.430211 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.430229 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.430252 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.430268 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:31Z","lastTransitionTime":"2025-11-25T09:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.445192 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:31Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.470752 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db22de3ee20f72a81d958f5fa9b10982a4ec446e092817b1880f669656e7a1ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:31Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.492587 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:31Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.514811 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:31Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.532807 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.532874 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.532895 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.532926 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.532947 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:31Z","lastTransitionTime":"2025-11-25T09:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.636502 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.636981 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.637236 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.637435 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.637611 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:31Z","lastTransitionTime":"2025-11-25T09:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.662219 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.662220 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.662285 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.662397 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:31 crc kubenswrapper[4776]: E1125 09:25:31.663126 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:31 crc kubenswrapper[4776]: E1125 09:25:31.663616 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:31 crc kubenswrapper[4776]: E1125 09:25:31.663434 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:31 crc kubenswrapper[4776]: E1125 09:25:31.663294 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.745504 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.745567 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.745589 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.745612 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.745627 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:31Z","lastTransitionTime":"2025-11-25T09:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.890128 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.890159 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.890170 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.890185 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.890196 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:31Z","lastTransitionTime":"2025-11-25T09:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.992887 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.992961 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.992978 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.993002 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:31 crc kubenswrapper[4776]: I1125 09:25:31.993020 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:31Z","lastTransitionTime":"2025-11-25T09:25:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.095723 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.095777 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.095795 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.095819 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.095836 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:32Z","lastTransitionTime":"2025-11-25T09:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.198818 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.198882 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.198900 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.198924 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.198943 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:32Z","lastTransitionTime":"2025-11-25T09:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.302285 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.302359 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.302377 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.302405 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.302426 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:32Z","lastTransitionTime":"2025-11-25T09:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.405328 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.405373 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.405390 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.405417 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.405435 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:32Z","lastTransitionTime":"2025-11-25T09:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.507726 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.507804 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.507825 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.507854 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.507876 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:32Z","lastTransitionTime":"2025-11-25T09:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.611240 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.611336 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.611355 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.611380 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.611398 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:32Z","lastTransitionTime":"2025-11-25T09:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.714595 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.714657 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.714675 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.714732 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.714759 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:32Z","lastTransitionTime":"2025-11-25T09:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.817449 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.817488 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.817499 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.817514 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.817523 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:32Z","lastTransitionTime":"2025-11-25T09:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.820287 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.820313 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.820324 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.820336 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.820346 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:32Z","lastTransitionTime":"2025-11-25T09:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:32 crc kubenswrapper[4776]: E1125 09:25:32.846520 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:32Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.851382 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.851418 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.851430 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.851457 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.851479 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:32Z","lastTransitionTime":"2025-11-25T09:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:32 crc kubenswrapper[4776]: E1125 09:25:32.867845 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:32Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.871705 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.871751 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.871763 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.871780 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.871793 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:32Z","lastTransitionTime":"2025-11-25T09:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:32 crc kubenswrapper[4776]: E1125 09:25:32.885806 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:32Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.889206 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.889242 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.889255 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.889270 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.889283 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:32Z","lastTransitionTime":"2025-11-25T09:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:32 crc kubenswrapper[4776]: E1125 09:25:32.902625 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:32Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.905985 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.906028 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.906102 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.906151 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.906168 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:32Z","lastTransitionTime":"2025-11-25T09:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:32 crc kubenswrapper[4776]: E1125 09:25:32.921309 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:32Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:32 crc kubenswrapper[4776]: E1125 09:25:32.921522 4776 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.923140 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.923184 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.923197 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.923218 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:32 crc kubenswrapper[4776]: I1125 09:25:32.923235 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:32Z","lastTransitionTime":"2025-11-25T09:25:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.025892 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.025943 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.025960 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.025981 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.025997 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:33Z","lastTransitionTime":"2025-11-25T09:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.129384 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.129444 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.129467 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.129496 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.129514 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:33Z","lastTransitionTime":"2025-11-25T09:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.232357 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.232448 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.232476 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.232505 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.232522 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:33Z","lastTransitionTime":"2025-11-25T09:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.335450 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.335506 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.335525 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.335548 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.335566 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:33Z","lastTransitionTime":"2025-11-25T09:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.437670 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.437743 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.437766 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.437797 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.437816 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:33Z","lastTransitionTime":"2025-11-25T09:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.540767 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.540836 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.540856 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.540882 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.540901 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:33Z","lastTransitionTime":"2025-11-25T09:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.643497 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.643555 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.643577 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.643611 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.643633 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:33Z","lastTransitionTime":"2025-11-25T09:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.662238 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.662271 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:33 crc kubenswrapper[4776]: E1125 09:25:33.662388 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.662402 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.662545 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:33 crc kubenswrapper[4776]: E1125 09:25:33.662644 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:33 crc kubenswrapper[4776]: E1125 09:25:33.662721 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:33 crc kubenswrapper[4776]: E1125 09:25:33.662966 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.748228 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.748297 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.748314 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.748339 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.748356 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:33Z","lastTransitionTime":"2025-11-25T09:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.851571 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.851620 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.851638 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.851662 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.851683 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:33Z","lastTransitionTime":"2025-11-25T09:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.955561 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.955641 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.955664 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.955694 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:33 crc kubenswrapper[4776]: I1125 09:25:33.955712 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:33Z","lastTransitionTime":"2025-11-25T09:25:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.059419 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.059501 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.059528 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.059560 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.059582 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:34Z","lastTransitionTime":"2025-11-25T09:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.162522 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.162595 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.162619 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.162651 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.162673 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:34Z","lastTransitionTime":"2025-11-25T09:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.265644 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.265718 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.265738 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.265766 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.265821 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:34Z","lastTransitionTime":"2025-11-25T09:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.368765 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.368811 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.368823 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.368841 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.368853 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:34Z","lastTransitionTime":"2025-11-25T09:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.471838 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.471982 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.472006 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.472036 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.472054 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:34Z","lastTransitionTime":"2025-11-25T09:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.575896 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.575979 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.576000 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.576027 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.576046 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:34Z","lastTransitionTime":"2025-11-25T09:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.679281 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.679351 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.679374 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.679398 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.679420 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:34Z","lastTransitionTime":"2025-11-25T09:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.782843 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.782919 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.782942 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.782974 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.782995 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:34Z","lastTransitionTime":"2025-11-25T09:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.885863 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.885915 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.885933 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.885956 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.885974 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:34Z","lastTransitionTime":"2025-11-25T09:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.989940 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.990035 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.990059 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.990160 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:34 crc kubenswrapper[4776]: I1125 09:25:34.990227 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:34Z","lastTransitionTime":"2025-11-25T09:25:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.093487 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.093561 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.093584 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.093606 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.093623 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:35Z","lastTransitionTime":"2025-11-25T09:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.196333 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.196395 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.196412 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.196438 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.196456 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:35Z","lastTransitionTime":"2025-11-25T09:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.298977 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.299036 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.299052 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.299101 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.299118 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:35Z","lastTransitionTime":"2025-11-25T09:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.400905 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.401305 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.401421 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.401520 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.401617 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:35Z","lastTransitionTime":"2025-11-25T09:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.504649 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.504726 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.504750 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.504782 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.504800 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:35Z","lastTransitionTime":"2025-11-25T09:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.607507 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.607582 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.607604 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.607636 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.607657 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:35Z","lastTransitionTime":"2025-11-25T09:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.661769 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.661859 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.661886 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:35 crc kubenswrapper[4776]: E1125 09:25:35.662035 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.662363 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:35 crc kubenswrapper[4776]: E1125 09:25:35.662489 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:35 crc kubenswrapper[4776]: E1125 09:25:35.662738 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:35 crc kubenswrapper[4776]: E1125 09:25:35.663060 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.688681 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:35Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.709244 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:35Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.710846 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.710899 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.710918 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.710943 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.710956 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:35Z","lastTransitionTime":"2025-11-25T09:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.729800 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:35Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.754494 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5282edc3c7afa8b6281b182576c679d19b20f87a8e32ab32e64d775a17bb705a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5282edc3c7afa8b6281b182576c679d19b20f87a8e32ab32e64d775a17bb705a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:25:14Z\\\",\\\"message\\\":\\\"pods:v4/a13607449821398607916) with []\\\\nI1125 09:25:14.558154 6499 factory.go:1336] Added *v1.Node event handler 7\\\\nI1125 09:25:14.558204 6499 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1125 09:25:14.558314 6499 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 09:25:14.558339 6499 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 09:25:14.558375 6499 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 09:25:14.558427 6499 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 09:25:14.558441 6499 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 09:25:14.558440 6499 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 09:25:14.558467 6499 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 09:25:14.558469 6499 factory.go:656] Stopping watch factory\\\\nI1125 09:25:14.558484 6499 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 09:25:14.558510 6499 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 09:25:14.558633 6499 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 09:25:14.558670 6499 ovnkube.go:599] Stopped ovnkube\\\\nI1125 09:25:14.558702 6499 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 09:25:14.558785 6499 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:25:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-r7x57_openshift-ovn-kubernetes(34502968-eb8d-47fa-83c2-7f6f6735ce2d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:35Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.767461 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:35Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.787682 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db22de3ee20f72a81d958f5fa9b10982a4ec446e092817b1880f669656e7a1ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:35Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.807953 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:35Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.813884 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.813925 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.813938 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.813956 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.813968 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:35Z","lastTransitionTime":"2025-11-25T09:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.823522 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nvc68" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nvc68\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:35Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.843286 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:35Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.859671 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51b8c180-5c19-4df6-b088-f1656ef0946e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b232d5cd319131183cefb0df8c537e099bfa1743d7fab3748858d3873a4802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9c968afbab8870d4105cc2b407cb7b2392919716c0a9b5de3ed47f3d12e5a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f9c968afbab8870d4105cc2b407cb7b2392919716c0a9b5de3ed47f3d12e5a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:35Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.880646 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:35Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.897310 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:35Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.912564 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:35Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.917362 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.917442 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.917466 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.917500 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.917518 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:35Z","lastTransitionTime":"2025-11-25T09:25:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.935242 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b1ea649a64d7a158675b397137593f6b901e762b67261d2e2872eeedc0d895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:25:29Z\\\",\\\"message\\\":\\\"2025-11-25T09:24:43+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c6e651f3-f403-47a2-b38e-0d19d75e1f16\\\\n2025-11-25T09:24:43+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c6e651f3-f403-47a2-b38e-0d19d75e1f16 to /host/opt/cni/bin/\\\\n2025-11-25T09:24:44Z [verbose] multus-daemon started\\\\n2025-11-25T09:24:44Z [verbose] Readiness Indicator file check\\\\n2025-11-25T09:25:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:25:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:35Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.961259 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ad63967f82781f8cf6d6c9512713cc8927666707203370dc38b031c201475a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:35Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:35 crc kubenswrapper[4776]: I1125 09:25:35.980286 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26759602-0b52-410a-939d-c2df88f2b9fb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655c0dff64b7dbe451d7c61bf1e1e63051f02f7587bbb65e71b5ee7d2bec0a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284857ebafc167c12fe70604afa6702c071316774d5976a5e95042968dafe0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7b5ddb321a7c0a66d83ecb015a42c43b13ae4af284f4806d8fcf5a5f3ba005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583bd3f6f361dbdc9c02a6eeeee4a445e2d2d08f605adf2d3c8912cb7d416abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://583bd3f6f361dbdc9c02a6eeeee4a445e2d2d08f605adf2d3c8912cb7d416abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:35Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.000807 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:35Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.020905 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a506265-8f64-456c-995f-1bc468cb87dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ead16f9d75513fe9ab6959e2ba0dca4f4b88df8bc65e3b0e6a21f28f2d0d30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb2778606c71db6545b88fd0648a8ab08a66863579467e49a5ece3b36d0d21f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rzwmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:36Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.022202 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.022273 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.022287 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.022314 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.022330 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:36Z","lastTransitionTime":"2025-11-25T09:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.054801 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:36Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.125284 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.125573 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.125674 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.125796 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.125896 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:36Z","lastTransitionTime":"2025-11-25T09:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.229142 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.229228 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.229266 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.229289 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.229363 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:36Z","lastTransitionTime":"2025-11-25T09:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.333318 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.334232 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.334416 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.334579 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.334718 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:36Z","lastTransitionTime":"2025-11-25T09:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.437883 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.437962 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.437989 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.438024 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.438045 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:36Z","lastTransitionTime":"2025-11-25T09:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.541617 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.541683 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.541702 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.541734 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.541756 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:36Z","lastTransitionTime":"2025-11-25T09:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.644884 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.644940 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.644963 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.644991 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.645013 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:36Z","lastTransitionTime":"2025-11-25T09:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.748768 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.749149 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.749450 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.749615 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.749959 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:36Z","lastTransitionTime":"2025-11-25T09:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.853319 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.853386 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.853401 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.853425 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.853441 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:36Z","lastTransitionTime":"2025-11-25T09:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.956641 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.956694 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.956708 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.956730 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:36 crc kubenswrapper[4776]: I1125 09:25:36.956745 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:36Z","lastTransitionTime":"2025-11-25T09:25:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.059486 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.059540 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.059556 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.059579 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.059596 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:37Z","lastTransitionTime":"2025-11-25T09:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.162148 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.162215 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.162230 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.162256 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.162275 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:37Z","lastTransitionTime":"2025-11-25T09:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.265883 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.265926 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.265935 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.265951 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.265962 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:37Z","lastTransitionTime":"2025-11-25T09:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.368483 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.368522 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.368536 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.368553 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.368565 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:37Z","lastTransitionTime":"2025-11-25T09:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.472120 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.472183 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.472203 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.472233 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.472254 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:37Z","lastTransitionTime":"2025-11-25T09:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.574864 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.574957 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.574973 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.575034 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.575048 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:37Z","lastTransitionTime":"2025-11-25T09:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.662202 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.662336 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.662336 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.662524 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:37 crc kubenswrapper[4776]: E1125 09:25:37.662509 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:37 crc kubenswrapper[4776]: E1125 09:25:37.662634 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:37 crc kubenswrapper[4776]: E1125 09:25:37.662777 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:37 crc kubenswrapper[4776]: E1125 09:25:37.662909 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.678364 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.678464 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.678483 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.678548 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.678568 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:37Z","lastTransitionTime":"2025-11-25T09:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.782014 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.782115 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.782133 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.782163 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.782182 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:37Z","lastTransitionTime":"2025-11-25T09:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.885665 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.885728 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.885810 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.885847 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.885870 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:37Z","lastTransitionTime":"2025-11-25T09:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.990132 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.990220 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.990244 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.990284 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:37 crc kubenswrapper[4776]: I1125 09:25:37.990309 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:37Z","lastTransitionTime":"2025-11-25T09:25:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.093694 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.093751 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.093762 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.093783 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.093795 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:38Z","lastTransitionTime":"2025-11-25T09:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.196631 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.196737 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.196763 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.196805 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.196832 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:38Z","lastTransitionTime":"2025-11-25T09:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.300818 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.300907 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.300933 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.300966 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.300992 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:38Z","lastTransitionTime":"2025-11-25T09:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.403951 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.404017 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.404035 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.404115 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.404135 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:38Z","lastTransitionTime":"2025-11-25T09:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.507701 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.507760 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.507808 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.507834 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.507852 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:38Z","lastTransitionTime":"2025-11-25T09:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.611395 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.611465 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.611489 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.611515 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.611532 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:38Z","lastTransitionTime":"2025-11-25T09:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.714465 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.714529 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.714546 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.714572 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.714589 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:38Z","lastTransitionTime":"2025-11-25T09:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.817697 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.817738 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.817748 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.817765 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.817777 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:38Z","lastTransitionTime":"2025-11-25T09:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.921157 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.921236 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.921253 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.921275 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:38 crc kubenswrapper[4776]: I1125 09:25:38.921288 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:38Z","lastTransitionTime":"2025-11-25T09:25:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.023897 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.023982 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.024007 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.024037 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.024059 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:39Z","lastTransitionTime":"2025-11-25T09:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.127041 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.127382 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.127591 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.127738 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.128053 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:39Z","lastTransitionTime":"2025-11-25T09:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.231327 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.231676 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.231769 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.231868 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.232016 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:39Z","lastTransitionTime":"2025-11-25T09:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.334699 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.334765 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.334777 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.334792 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.334802 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:39Z","lastTransitionTime":"2025-11-25T09:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.437619 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.437673 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.437688 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.437709 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.437726 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:39Z","lastTransitionTime":"2025-11-25T09:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.541162 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.541221 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.541232 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.541247 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.541274 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:39Z","lastTransitionTime":"2025-11-25T09:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.588456 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:25:39 crc kubenswrapper[4776]: E1125 09:25:39.588629 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:43.588607893 +0000 UTC m=+148.629667456 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.588677 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.588721 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.588750 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.588777 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:39 crc kubenswrapper[4776]: E1125 09:25:39.588893 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:25:39 crc kubenswrapper[4776]: E1125 09:25:39.588896 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:25:39 crc kubenswrapper[4776]: E1125 09:25:39.588926 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:25:39 crc kubenswrapper[4776]: E1125 09:25:39.588946 4776 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:25:39 crc kubenswrapper[4776]: E1125 09:25:39.588945 4776 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:25:39 crc kubenswrapper[4776]: E1125 09:25:39.588907 4776 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:25:39 crc kubenswrapper[4776]: E1125 09:25:39.589002 4776 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:25:39 crc kubenswrapper[4776]: E1125 09:25:39.589002 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 09:26:43.588985433 +0000 UTC m=+148.630044996 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:25:39 crc kubenswrapper[4776]: E1125 09:25:39.589020 4776 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:25:39 crc kubenswrapper[4776]: E1125 09:25:39.589052 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 09:26:43.589041134 +0000 UTC m=+148.630100687 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:25:39 crc kubenswrapper[4776]: E1125 09:25:39.589217 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:26:43.589190308 +0000 UTC m=+148.630249901 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:25:39 crc kubenswrapper[4776]: E1125 09:25:39.589250 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:26:43.589232149 +0000 UTC m=+148.630291732 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.643675 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.643739 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.643762 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.643797 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.643820 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:39Z","lastTransitionTime":"2025-11-25T09:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.661428 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.661535 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.661598 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:39 crc kubenswrapper[4776]: E1125 09:25:39.661798 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.661850 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:39 crc kubenswrapper[4776]: E1125 09:25:39.661983 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:39 crc kubenswrapper[4776]: E1125 09:25:39.662090 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:39 crc kubenswrapper[4776]: E1125 09:25:39.662180 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.747224 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.747282 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.747305 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.747332 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.747353 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:39Z","lastTransitionTime":"2025-11-25T09:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.849791 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.849858 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.849875 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.849899 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.849917 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:39Z","lastTransitionTime":"2025-11-25T09:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.953633 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.953706 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.953724 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.953750 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:39 crc kubenswrapper[4776]: I1125 09:25:39.953769 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:39Z","lastTransitionTime":"2025-11-25T09:25:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.056604 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.056688 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.056713 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.056752 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.056777 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:40Z","lastTransitionTime":"2025-11-25T09:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.160543 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.160608 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.160628 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.160656 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.160674 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:40Z","lastTransitionTime":"2025-11-25T09:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.263624 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.263693 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.263715 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.263741 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.263760 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:40Z","lastTransitionTime":"2025-11-25T09:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.367293 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.367377 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.367399 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.367428 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.367449 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:40Z","lastTransitionTime":"2025-11-25T09:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.469403 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.469434 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.469442 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.469454 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.469462 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:40Z","lastTransitionTime":"2025-11-25T09:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.572304 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.572355 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.572366 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.572386 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.572397 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:40Z","lastTransitionTime":"2025-11-25T09:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.675476 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.675541 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.675566 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.675596 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.675620 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:40Z","lastTransitionTime":"2025-11-25T09:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.778309 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.778346 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.778370 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.778392 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.778404 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:40Z","lastTransitionTime":"2025-11-25T09:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.881121 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.881216 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.881236 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.881261 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.881317 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:40Z","lastTransitionTime":"2025-11-25T09:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.984236 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.984297 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.984308 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.984331 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:40 crc kubenswrapper[4776]: I1125 09:25:40.984342 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:40Z","lastTransitionTime":"2025-11-25T09:25:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.087406 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.087467 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.087479 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.087505 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.087517 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:41Z","lastTransitionTime":"2025-11-25T09:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.191402 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.191485 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.191505 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.191539 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.191558 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:41Z","lastTransitionTime":"2025-11-25T09:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.294794 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.294888 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.294914 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.294998 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.295028 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:41Z","lastTransitionTime":"2025-11-25T09:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.398540 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.398612 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.398636 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.398667 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.398694 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:41Z","lastTransitionTime":"2025-11-25T09:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.502107 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.502147 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.502156 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.502172 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.502184 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:41Z","lastTransitionTime":"2025-11-25T09:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.605837 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.605892 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.605906 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.605927 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.605943 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:41Z","lastTransitionTime":"2025-11-25T09:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.662037 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.662125 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.662225 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.662358 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:41 crc kubenswrapper[4776]: E1125 09:25:41.662517 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:41 crc kubenswrapper[4776]: E1125 09:25:41.662757 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:41 crc kubenswrapper[4776]: E1125 09:25:41.662924 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:41 crc kubenswrapper[4776]: E1125 09:25:41.662994 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.709219 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.709298 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.709313 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.709333 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.709347 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:41Z","lastTransitionTime":"2025-11-25T09:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.812834 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.812910 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.812924 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.812944 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.812956 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:41Z","lastTransitionTime":"2025-11-25T09:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.915951 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.916033 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.916056 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.916124 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:41 crc kubenswrapper[4776]: I1125 09:25:41.916149 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:41Z","lastTransitionTime":"2025-11-25T09:25:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.018550 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.018616 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.018633 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.018658 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.018675 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:42Z","lastTransitionTime":"2025-11-25T09:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.122314 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.122360 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.122374 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.122393 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.122406 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:42Z","lastTransitionTime":"2025-11-25T09:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.225386 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.225444 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.225464 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.225488 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.225506 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:42Z","lastTransitionTime":"2025-11-25T09:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.329059 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.329176 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.329201 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.329232 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.329255 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:42Z","lastTransitionTime":"2025-11-25T09:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.432001 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.432096 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.432115 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.432137 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.432158 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:42Z","lastTransitionTime":"2025-11-25T09:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.535148 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.535231 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.535251 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.535277 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.535295 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:42Z","lastTransitionTime":"2025-11-25T09:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.638984 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.639047 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.639106 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.639134 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.639152 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:42Z","lastTransitionTime":"2025-11-25T09:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.742965 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.743038 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.743061 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.743111 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.743129 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:42Z","lastTransitionTime":"2025-11-25T09:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.845351 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.845406 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.845417 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.845434 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.845446 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:42Z","lastTransitionTime":"2025-11-25T09:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.947871 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.947904 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.947911 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.947937 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:42 crc kubenswrapper[4776]: I1125 09:25:42.947946 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:42Z","lastTransitionTime":"2025-11-25T09:25:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.050337 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.050387 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.050399 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.050423 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.050434 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:43Z","lastTransitionTime":"2025-11-25T09:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.152828 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.152859 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.152867 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.152879 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.152893 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:43Z","lastTransitionTime":"2025-11-25T09:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.255960 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.256036 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.256059 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.256126 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.256154 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:43Z","lastTransitionTime":"2025-11-25T09:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.257797 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.257846 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.257865 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.257884 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.257901 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:43Z","lastTransitionTime":"2025-11-25T09:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:43 crc kubenswrapper[4776]: E1125 09:25:43.280256 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.286029 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.286119 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.286136 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.286162 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.286179 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:43Z","lastTransitionTime":"2025-11-25T09:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:43 crc kubenswrapper[4776]: E1125 09:25:43.306243 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.311555 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.311713 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.311745 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.311815 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.311837 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:43Z","lastTransitionTime":"2025-11-25T09:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:43 crc kubenswrapper[4776]: E1125 09:25:43.331969 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.337795 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.337889 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.337947 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.337971 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.338017 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:43Z","lastTransitionTime":"2025-11-25T09:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:43 crc kubenswrapper[4776]: E1125 09:25:43.358839 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.369309 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.369367 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.369392 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.369420 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.369444 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:43Z","lastTransitionTime":"2025-11-25T09:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:43 crc kubenswrapper[4776]: E1125 09:25:43.389510 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:43 crc kubenswrapper[4776]: E1125 09:25:43.389677 4776 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.391671 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.391733 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.391746 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.391764 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.391775 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:43Z","lastTransitionTime":"2025-11-25T09:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.494284 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.494392 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.494427 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.494455 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.494475 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:43Z","lastTransitionTime":"2025-11-25T09:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.597224 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.597301 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.597328 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.597393 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.597417 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:43Z","lastTransitionTime":"2025-11-25T09:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.661723 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.661793 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:43 crc kubenswrapper[4776]: E1125 09:25:43.661925 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.662054 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.662129 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:43 crc kubenswrapper[4776]: E1125 09:25:43.662275 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:43 crc kubenswrapper[4776]: E1125 09:25:43.662444 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:43 crc kubenswrapper[4776]: E1125 09:25:43.662545 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.700606 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.700712 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.700737 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.700823 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.700959 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:43Z","lastTransitionTime":"2025-11-25T09:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.803822 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.803871 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.803887 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.803911 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.803930 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:43Z","lastTransitionTime":"2025-11-25T09:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.906637 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.906697 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.906709 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.906727 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:43 crc kubenswrapper[4776]: I1125 09:25:43.906743 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:43Z","lastTransitionTime":"2025-11-25T09:25:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.009617 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.009679 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.009696 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.009718 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.009735 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:44Z","lastTransitionTime":"2025-11-25T09:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.112875 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.112945 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.112965 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.112989 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.113005 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:44Z","lastTransitionTime":"2025-11-25T09:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.216271 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.216385 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.216402 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.216427 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.216447 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:44Z","lastTransitionTime":"2025-11-25T09:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.319767 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.319851 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.319887 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.319920 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.319941 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:44Z","lastTransitionTime":"2025-11-25T09:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.422727 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.422862 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.422896 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.422926 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.422946 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:44Z","lastTransitionTime":"2025-11-25T09:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.526457 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.526655 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.526680 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.526713 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.526730 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:44Z","lastTransitionTime":"2025-11-25T09:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.629887 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.629935 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.629947 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.629966 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.629979 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:44Z","lastTransitionTime":"2025-11-25T09:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.662872 4776 scope.go:117] "RemoveContainer" containerID="5282edc3c7afa8b6281b182576c679d19b20f87a8e32ab32e64d775a17bb705a" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.734102 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.734289 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.734304 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.734322 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.734335 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:44Z","lastTransitionTime":"2025-11-25T09:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.837409 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.837497 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.837521 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.837553 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.837584 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:44Z","lastTransitionTime":"2025-11-25T09:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.940111 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.940148 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.940157 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.940173 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:44 crc kubenswrapper[4776]: I1125 09:25:44.940183 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:44Z","lastTransitionTime":"2025-11-25T09:25:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.043002 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.043093 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.043118 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.043147 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.043170 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:45Z","lastTransitionTime":"2025-11-25T09:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.145938 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.146005 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.146027 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.146097 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.146126 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:45Z","lastTransitionTime":"2025-11-25T09:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.194656 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-r7x57_34502968-eb8d-47fa-83c2-7f6f6735ce2d/ovnkube-controller/2.log" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.198110 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" event={"ID":"34502968-eb8d-47fa-83c2-7f6f6735ce2d","Type":"ContainerStarted","Data":"3c5b9f801210cc7df25d48a313b74f7a347bfd5cba0855ee054fc48baf53e3b0"} Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.198669 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.215217 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.227189 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.242128 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nvc68" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nvc68\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.248803 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.248878 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.248890 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.248907 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.248918 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:45Z","lastTransitionTime":"2025-11-25T09:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.255141 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26759602-0b52-410a-939d-c2df88f2b9fb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655c0dff64b7dbe451d7c61bf1e1e63051f02f7587bbb65e71b5ee7d2bec0a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284857ebafc167c12fe70604afa6702c071316774d5976a5e95042968dafe0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7b5ddb321a7c0a66d83ecb015a42c43b13ae4af284f4806d8fcf5a5f3ba005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583bd3f6f361dbdc9c02a6eeeee4a445e2d2d08f605adf2d3c8912cb7d416abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://583bd3f6f361dbdc9c02a6eeeee4a445e2d2d08f605adf2d3c8912cb7d416abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.265758 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51b8c180-5c19-4df6-b088-f1656ef0946e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b232d5cd319131183cefb0df8c537e099bfa1743d7fab3748858d3873a4802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9c968afbab8870d4105cc2b407cb7b2392919716c0a9b5de3ed47f3d12e5a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f9c968afbab8870d4105cc2b407cb7b2392919716c0a9b5de3ed47f3d12e5a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.280161 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.291759 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.302465 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.315604 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b1ea649a64d7a158675b397137593f6b901e762b67261d2e2872eeedc0d895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:25:29Z\\\",\\\"message\\\":\\\"2025-11-25T09:24:43+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c6e651f3-f403-47a2-b38e-0d19d75e1f16\\\\n2025-11-25T09:24:43+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c6e651f3-f403-47a2-b38e-0d19d75e1f16 to /host/opt/cni/bin/\\\\n2025-11-25T09:24:44Z [verbose] multus-daemon started\\\\n2025-11-25T09:24:44Z [verbose] Readiness Indicator file check\\\\n2025-11-25T09:25:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:25:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.332264 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ad63967f82781f8cf6d6c9512713cc8927666707203370dc38b031c201475a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.351873 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.351924 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.351936 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.351959 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.351973 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:45Z","lastTransitionTime":"2025-11-25T09:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.360409 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.372158 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.383903 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a506265-8f64-456c-995f-1bc468cb87dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ead16f9d75513fe9ab6959e2ba0dca4f4b88df8bc65e3b0e6a21f28f2d0d30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb2778606c71db6545b88fd0648a8ab08a66863579467e49a5ece3b36d0d21f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rzwmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.397203 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db22de3ee20f72a81d958f5fa9b10982a4ec446e092817b1880f669656e7a1ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.407189 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.417090 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.426671 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.442472 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c5b9f801210cc7df25d48a313b74f7a347bfd5cba0855ee054fc48baf53e3b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5282edc3c7afa8b6281b182576c679d19b20f87a8e32ab32e64d775a17bb705a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:25:14Z\\\",\\\"message\\\":\\\"pods:v4/a13607449821398607916) with []\\\\nI1125 09:25:14.558154 6499 factory.go:1336] Added *v1.Node event handler 7\\\\nI1125 09:25:14.558204 6499 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1125 09:25:14.558314 6499 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 09:25:14.558339 6499 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 09:25:14.558375 6499 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 09:25:14.558427 6499 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 09:25:14.558441 6499 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 09:25:14.558440 6499 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 09:25:14.558467 6499 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 09:25:14.558469 6499 factory.go:656] Stopping watch factory\\\\nI1125 09:25:14.558484 6499 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 09:25:14.558510 6499 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 09:25:14.558633 6499 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 09:25:14.558670 6499 ovnkube.go:599] Stopped ovnkube\\\\nI1125 09:25:14.558702 6499 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 09:25:14.558785 6499 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:25:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.453848 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.453895 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.453872 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.453905 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.453949 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.453964 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:45Z","lastTransitionTime":"2025-11-25T09:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.555883 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.555917 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.555925 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.555937 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.555946 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:45Z","lastTransitionTime":"2025-11-25T09:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.658481 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.658526 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.658537 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.658552 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.658563 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:45Z","lastTransitionTime":"2025-11-25T09:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.661846 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.661873 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.661955 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.661978 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:45 crc kubenswrapper[4776]: E1125 09:25:45.662061 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:45 crc kubenswrapper[4776]: E1125 09:25:45.662138 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:45 crc kubenswrapper[4776]: E1125 09:25:45.662221 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:45 crc kubenswrapper[4776]: E1125 09:25:45.662319 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.675022 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db22de3ee20f72a81d958f5fa9b10982a4ec446e092817b1880f669656e7a1ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.688961 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.699583 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.710884 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.732027 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c5b9f801210cc7df25d48a313b74f7a347bfd5cba0855ee054fc48baf53e3b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5282edc3c7afa8b6281b182576c679d19b20f87a8e32ab32e64d775a17bb705a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:25:14Z\\\",\\\"message\\\":\\\"pods:v4/a13607449821398607916) with []\\\\nI1125 09:25:14.558154 6499 factory.go:1336] Added *v1.Node event handler 7\\\\nI1125 09:25:14.558204 6499 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1125 09:25:14.558314 6499 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 09:25:14.558339 6499 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 09:25:14.558375 6499 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 09:25:14.558427 6499 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 09:25:14.558441 6499 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 09:25:14.558440 6499 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 09:25:14.558467 6499 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 09:25:14.558469 6499 factory.go:656] Stopping watch factory\\\\nI1125 09:25:14.558484 6499 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 09:25:14.558510 6499 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 09:25:14.558633 6499 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 09:25:14.558670 6499 ovnkube.go:599] Stopped ovnkube\\\\nI1125 09:25:14.558702 6499 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 09:25:14.558785 6499 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:25:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.741579 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.751208 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.760900 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.760941 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.760949 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.760963 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.760972 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:45Z","lastTransitionTime":"2025-11-25T09:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.764102 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.781721 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nvc68" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nvc68\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.797569 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ad63967f82781f8cf6d6c9512713cc8927666707203370dc38b031c201475a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.809565 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26759602-0b52-410a-939d-c2df88f2b9fb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655c0dff64b7dbe451d7c61bf1e1e63051f02f7587bbb65e71b5ee7d2bec0a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284857ebafc167c12fe70604afa6702c071316774d5976a5e95042968dafe0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7b5ddb321a7c0a66d83ecb015a42c43b13ae4af284f4806d8fcf5a5f3ba005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583bd3f6f361dbdc9c02a6eeeee4a445e2d2d08f605adf2d3c8912cb7d416abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://583bd3f6f361dbdc9c02a6eeeee4a445e2d2d08f605adf2d3c8912cb7d416abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.822229 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51b8c180-5c19-4df6-b088-f1656ef0946e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b232d5cd319131183cefb0df8c537e099bfa1743d7fab3748858d3873a4802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9c968afbab8870d4105cc2b407cb7b2392919716c0a9b5de3ed47f3d12e5a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f9c968afbab8870d4105cc2b407cb7b2392919716c0a9b5de3ed47f3d12e5a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.834683 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.845888 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.855458 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.863632 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.863680 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.863690 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.863706 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.863717 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:45Z","lastTransitionTime":"2025-11-25T09:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.867314 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b1ea649a64d7a158675b397137593f6b901e762b67261d2e2872eeedc0d895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:25:29Z\\\",\\\"message\\\":\\\"2025-11-25T09:24:43+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c6e651f3-f403-47a2-b38e-0d19d75e1f16\\\\n2025-11-25T09:24:43+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c6e651f3-f403-47a2-b38e-0d19d75e1f16 to /host/opt/cni/bin/\\\\n2025-11-25T09:24:44Z [verbose] multus-daemon started\\\\n2025-11-25T09:24:44Z [verbose] Readiness Indicator file check\\\\n2025-11-25T09:25:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:25:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.885461 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.895214 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.905306 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a506265-8f64-456c-995f-1bc468cb87dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ead16f9d75513fe9ab6959e2ba0dca4f4b88df8bc65e3b0e6a21f28f2d0d30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb2778606c71db6545b88fd0648a8ab08a66863579467e49a5ece3b36d0d21f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rzwmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.966478 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.966518 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.966529 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.966546 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:45 crc kubenswrapper[4776]: I1125 09:25:45.966557 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:45Z","lastTransitionTime":"2025-11-25T09:25:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.069103 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.069131 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.069140 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.069152 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.069161 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:46Z","lastTransitionTime":"2025-11-25T09:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.171504 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.171814 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.171826 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.171844 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.171856 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:46Z","lastTransitionTime":"2025-11-25T09:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.203217 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-r7x57_34502968-eb8d-47fa-83c2-7f6f6735ce2d/ovnkube-controller/3.log" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.204046 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-r7x57_34502968-eb8d-47fa-83c2-7f6f6735ce2d/ovnkube-controller/2.log" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.208951 4776 generic.go:334] "Generic (PLEG): container finished" podID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerID="3c5b9f801210cc7df25d48a313b74f7a347bfd5cba0855ee054fc48baf53e3b0" exitCode=1 Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.209005 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" event={"ID":"34502968-eb8d-47fa-83c2-7f6f6735ce2d","Type":"ContainerDied","Data":"3c5b9f801210cc7df25d48a313b74f7a347bfd5cba0855ee054fc48baf53e3b0"} Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.209062 4776 scope.go:117] "RemoveContainer" containerID="5282edc3c7afa8b6281b182576c679d19b20f87a8e32ab32e64d775a17bb705a" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.209878 4776 scope.go:117] "RemoveContainer" containerID="3c5b9f801210cc7df25d48a313b74f7a347bfd5cba0855ee054fc48baf53e3b0" Nov 25 09:25:46 crc kubenswrapper[4776]: E1125 09:25:46.210118 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-r7x57_openshift-ovn-kubernetes(34502968-eb8d-47fa-83c2-7f6f6735ce2d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.238789 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c5b9f801210cc7df25d48a313b74f7a347bfd5cba0855ee054fc48baf53e3b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5282edc3c7afa8b6281b182576c679d19b20f87a8e32ab32e64d775a17bb705a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:25:14Z\\\",\\\"message\\\":\\\"pods:v4/a13607449821398607916) with []\\\\nI1125 09:25:14.558154 6499 factory.go:1336] Added *v1.Node event handler 7\\\\nI1125 09:25:14.558204 6499 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1125 09:25:14.558314 6499 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 09:25:14.558339 6499 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 09:25:14.558375 6499 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 09:25:14.558427 6499 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 09:25:14.558441 6499 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 09:25:14.558440 6499 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 09:25:14.558467 6499 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 09:25:14.558469 6499 factory.go:656] Stopping watch factory\\\\nI1125 09:25:14.558484 6499 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 09:25:14.558510 6499 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 09:25:14.558633 6499 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 09:25:14.558670 6499 ovnkube.go:599] Stopped ovnkube\\\\nI1125 09:25:14.558702 6499 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 09:25:14.558785 6499 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:25:13Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c5b9f801210cc7df25d48a313b74f7a347bfd5cba0855ee054fc48baf53e3b0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:25:45Z\\\",\\\"message\\\":\\\"ctory.go:656] Stopping watch factory\\\\nI1125 09:25:45.621334 6872 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 09:25:45.621345 6872 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 09:25:45.621360 6872 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 09:25:45.621369 6872 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 09:25:45.621376 6872 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 09:25:45.621395 6872 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1125 09:25:45.621486 6872 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 09:25:45.621573 6872 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 09:25:45.621682 6872 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:25:45.622084 6872 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:25:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.252637 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.271198 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db22de3ee20f72a81d958f5fa9b10982a4ec446e092817b1880f669656e7a1ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.274406 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.274468 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.274487 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.274512 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.274529 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:46Z","lastTransitionTime":"2025-11-25T09:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.291843 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.312257 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.326966 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.338656 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.354845 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.365793 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nvc68" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nvc68\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.376303 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.376903 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.376935 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.376948 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.376968 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.376981 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:46Z","lastTransitionTime":"2025-11-25T09:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.390394 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b1ea649a64d7a158675b397137593f6b901e762b67261d2e2872eeedc0d895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:25:29Z\\\",\\\"message\\\":\\\"2025-11-25T09:24:43+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c6e651f3-f403-47a2-b38e-0d19d75e1f16\\\\n2025-11-25T09:24:43+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c6e651f3-f403-47a2-b38e-0d19d75e1f16 to /host/opt/cni/bin/\\\\n2025-11-25T09:24:44Z [verbose] multus-daemon started\\\\n2025-11-25T09:24:44Z [verbose] Readiness Indicator file check\\\\n2025-11-25T09:25:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:25:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.402557 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ad63967f82781f8cf6d6c9512713cc8927666707203370dc38b031c201475a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.414946 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26759602-0b52-410a-939d-c2df88f2b9fb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655c0dff64b7dbe451d7c61bf1e1e63051f02f7587bbb65e71b5ee7d2bec0a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284857ebafc167c12fe70604afa6702c071316774d5976a5e95042968dafe0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7b5ddb321a7c0a66d83ecb015a42c43b13ae4af284f4806d8fcf5a5f3ba005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583bd3f6f361dbdc9c02a6eeeee4a445e2d2d08f605adf2d3c8912cb7d416abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://583bd3f6f361dbdc9c02a6eeeee4a445e2d2d08f605adf2d3c8912cb7d416abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.425575 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51b8c180-5c19-4df6-b088-f1656ef0946e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b232d5cd319131183cefb0df8c537e099bfa1743d7fab3748858d3873a4802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9c968afbab8870d4105cc2b407cb7b2392919716c0a9b5de3ed47f3d12e5a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f9c968afbab8870d4105cc2b407cb7b2392919716c0a9b5de3ed47f3d12e5a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.438550 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.450420 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.479828 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.479874 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.479887 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.479907 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.479919 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:46Z","lastTransitionTime":"2025-11-25T09:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.480616 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.495571 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.510935 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a506265-8f64-456c-995f-1bc468cb87dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ead16f9d75513fe9ab6959e2ba0dca4f4b88df8bc65e3b0e6a21f28f2d0d30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb2778606c71db6545b88fd0648a8ab08a66863579467e49a5ece3b36d0d21f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rzwmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.582635 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.582673 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.582685 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.582701 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.582714 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:46Z","lastTransitionTime":"2025-11-25T09:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.686127 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.686204 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.686228 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.686261 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.686287 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:46Z","lastTransitionTime":"2025-11-25T09:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.794265 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.794307 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.794325 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.794348 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.794365 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:46Z","lastTransitionTime":"2025-11-25T09:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.897589 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.897663 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.897680 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.897706 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:46 crc kubenswrapper[4776]: I1125 09:25:46.897723 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:46Z","lastTransitionTime":"2025-11-25T09:25:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.001222 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.001261 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.001269 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.001283 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.001292 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:47Z","lastTransitionTime":"2025-11-25T09:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.104096 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.104135 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.104157 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.104182 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.104197 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:47Z","lastTransitionTime":"2025-11-25T09:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.206031 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.206128 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.206151 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.206177 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.206199 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:47Z","lastTransitionTime":"2025-11-25T09:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.213628 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-r7x57_34502968-eb8d-47fa-83c2-7f6f6735ce2d/ovnkube-controller/3.log" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.219112 4776 scope.go:117] "RemoveContainer" containerID="3c5b9f801210cc7df25d48a313b74f7a347bfd5cba0855ee054fc48baf53e3b0" Nov 25 09:25:47 crc kubenswrapper[4776]: E1125 09:25:47.219414 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-r7x57_openshift-ovn-kubernetes(34502968-eb8d-47fa-83c2-7f6f6735ce2d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.251734 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.267685 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.283127 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a506265-8f64-456c-995f-1bc468cb87dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ead16f9d75513fe9ab6959e2ba0dca4f4b88df8bc65e3b0e6a21f28f2d0d30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb2778606c71db6545b88fd0648a8ab08a66863579467e49a5ece3b36d0d21f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rzwmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.299743 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.309902 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.309977 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.309991 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.310009 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.310022 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:47Z","lastTransitionTime":"2025-11-25T09:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.319532 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c5b9f801210cc7df25d48a313b74f7a347bfd5cba0855ee054fc48baf53e3b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c5b9f801210cc7df25d48a313b74f7a347bfd5cba0855ee054fc48baf53e3b0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:25:45Z\\\",\\\"message\\\":\\\"ctory.go:656] Stopping watch factory\\\\nI1125 09:25:45.621334 6872 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 09:25:45.621345 6872 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 09:25:45.621360 6872 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 09:25:45.621369 6872 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 09:25:45.621376 6872 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 09:25:45.621395 6872 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1125 09:25:45.621486 6872 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 09:25:45.621573 6872 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 09:25:45.621682 6872 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:25:45.622084 6872 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:25:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-r7x57_openshift-ovn-kubernetes(34502968-eb8d-47fa-83c2-7f6f6735ce2d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.330955 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.346446 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db22de3ee20f72a81d958f5fa9b10982a4ec446e092817b1880f669656e7a1ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.357172 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.369204 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.380375 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.397815 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.409388 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nvc68" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nvc68\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.411407 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.411432 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.411441 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.411453 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.411461 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:47Z","lastTransitionTime":"2025-11-25T09:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.423468 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.435939 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.452195 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b1ea649a64d7a158675b397137593f6b901e762b67261d2e2872eeedc0d895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:25:29Z\\\",\\\"message\\\":\\\"2025-11-25T09:24:43+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c6e651f3-f403-47a2-b38e-0d19d75e1f16\\\\n2025-11-25T09:24:43+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c6e651f3-f403-47a2-b38e-0d19d75e1f16 to /host/opt/cni/bin/\\\\n2025-11-25T09:24:44Z [verbose] multus-daemon started\\\\n2025-11-25T09:24:44Z [verbose] Readiness Indicator file check\\\\n2025-11-25T09:25:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:25:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.474210 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ad63967f82781f8cf6d6c9512713cc8927666707203370dc38b031c201475a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.491442 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26759602-0b52-410a-939d-c2df88f2b9fb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655c0dff64b7dbe451d7c61bf1e1e63051f02f7587bbb65e71b5ee7d2bec0a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284857ebafc167c12fe70604afa6702c071316774d5976a5e95042968dafe0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7b5ddb321a7c0a66d83ecb015a42c43b13ae4af284f4806d8fcf5a5f3ba005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583bd3f6f361dbdc9c02a6eeeee4a445e2d2d08f605adf2d3c8912cb7d416abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://583bd3f6f361dbdc9c02a6eeeee4a445e2d2d08f605adf2d3c8912cb7d416abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.503364 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51b8c180-5c19-4df6-b088-f1656ef0946e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b232d5cd319131183cefb0df8c537e099bfa1743d7fab3748858d3873a4802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9c968afbab8870d4105cc2b407cb7b2392919716c0a9b5de3ed47f3d12e5a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f9c968afbab8870d4105cc2b407cb7b2392919716c0a9b5de3ed47f3d12e5a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.514521 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.514571 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.514765 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.514844 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.514862 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:47Z","lastTransitionTime":"2025-11-25T09:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.518306 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.617534 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.617561 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.617571 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.617583 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.617591 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:47Z","lastTransitionTime":"2025-11-25T09:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.661678 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.661733 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:47 crc kubenswrapper[4776]: E1125 09:25:47.661804 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.661678 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.661870 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:47 crc kubenswrapper[4776]: E1125 09:25:47.661961 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:47 crc kubenswrapper[4776]: E1125 09:25:47.662163 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:47 crc kubenswrapper[4776]: E1125 09:25:47.662261 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.719512 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.719580 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.719601 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.719624 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.719643 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:47Z","lastTransitionTime":"2025-11-25T09:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.824225 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.824287 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.824311 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.824336 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.824356 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:47Z","lastTransitionTime":"2025-11-25T09:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.927779 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.927843 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.927867 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.927899 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:47 crc kubenswrapper[4776]: I1125 09:25:47.927920 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:47Z","lastTransitionTime":"2025-11-25T09:25:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.030611 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.030680 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.030698 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.030726 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.030746 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:48Z","lastTransitionTime":"2025-11-25T09:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.139745 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.139823 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.139849 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.139880 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.139904 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:48Z","lastTransitionTime":"2025-11-25T09:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.242429 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.242502 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.242524 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.242552 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.242570 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:48Z","lastTransitionTime":"2025-11-25T09:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.346123 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.346226 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.346255 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.346292 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.346309 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:48Z","lastTransitionTime":"2025-11-25T09:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.449600 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.449661 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.449703 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.449731 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.449752 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:48Z","lastTransitionTime":"2025-11-25T09:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.552715 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.552787 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.552805 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.552832 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.552851 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:48Z","lastTransitionTime":"2025-11-25T09:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.656027 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.656121 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.656153 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.656181 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.656215 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:48Z","lastTransitionTime":"2025-11-25T09:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.758833 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.758964 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.758987 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.759057 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.759136 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:48Z","lastTransitionTime":"2025-11-25T09:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.862485 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.862577 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.862595 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.862618 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.862634 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:48Z","lastTransitionTime":"2025-11-25T09:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.965525 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.965592 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.965618 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.965651 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:48 crc kubenswrapper[4776]: I1125 09:25:48.965675 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:48Z","lastTransitionTime":"2025-11-25T09:25:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.069659 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.069728 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.069745 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.069822 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.069848 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:49Z","lastTransitionTime":"2025-11-25T09:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.172870 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.172939 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.172958 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.172986 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.173003 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:49Z","lastTransitionTime":"2025-11-25T09:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.276991 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.277098 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.277120 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.277145 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.277163 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:49Z","lastTransitionTime":"2025-11-25T09:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.380637 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.380711 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.380729 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.380753 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.380775 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:49Z","lastTransitionTime":"2025-11-25T09:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.484156 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.484202 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.484218 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.484241 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.484305 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:49Z","lastTransitionTime":"2025-11-25T09:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.587529 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.587597 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.587625 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.587880 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.587908 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:49Z","lastTransitionTime":"2025-11-25T09:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.662128 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:49 crc kubenswrapper[4776]: E1125 09:25:49.662324 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.662420 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.662528 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.662420 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:49 crc kubenswrapper[4776]: E1125 09:25:49.662741 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:49 crc kubenswrapper[4776]: E1125 09:25:49.662941 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:49 crc kubenswrapper[4776]: E1125 09:25:49.663189 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.690631 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.690690 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.690702 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.690720 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.690732 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:49Z","lastTransitionTime":"2025-11-25T09:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.794676 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.794737 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.794761 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.794790 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.794814 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:49Z","lastTransitionTime":"2025-11-25T09:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.898898 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.898972 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.898992 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.899044 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:49 crc kubenswrapper[4776]: I1125 09:25:49.899104 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:49Z","lastTransitionTime":"2025-11-25T09:25:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.002285 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.002415 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.002434 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.002455 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.002471 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:50Z","lastTransitionTime":"2025-11-25T09:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.105113 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.105142 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.105150 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.105164 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.105172 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:50Z","lastTransitionTime":"2025-11-25T09:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.209445 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.209505 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.209519 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.209539 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.209556 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:50Z","lastTransitionTime":"2025-11-25T09:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.312664 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.312716 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.312734 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.312754 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.312773 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:50Z","lastTransitionTime":"2025-11-25T09:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.417186 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.417256 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.417271 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.417296 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.417314 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:50Z","lastTransitionTime":"2025-11-25T09:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.520541 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.520586 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.520597 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.520612 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.520625 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:50Z","lastTransitionTime":"2025-11-25T09:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.623965 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.624002 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.624011 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.624024 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.624033 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:50Z","lastTransitionTime":"2025-11-25T09:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.726860 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.726926 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.726938 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.726955 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.726964 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:50Z","lastTransitionTime":"2025-11-25T09:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.830117 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.830516 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.830588 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.830629 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.830676 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:50Z","lastTransitionTime":"2025-11-25T09:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.939797 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.939835 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.939846 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.939862 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:50 crc kubenswrapper[4776]: I1125 09:25:50.939875 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:50Z","lastTransitionTime":"2025-11-25T09:25:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.043446 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.043524 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.043548 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.043577 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.043599 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:51Z","lastTransitionTime":"2025-11-25T09:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.146938 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.146993 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.147009 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.147027 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.147042 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:51Z","lastTransitionTime":"2025-11-25T09:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.249284 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.249326 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.249343 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.249363 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.249377 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:51Z","lastTransitionTime":"2025-11-25T09:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.352573 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.352620 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.352631 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.352648 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.352660 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:51Z","lastTransitionTime":"2025-11-25T09:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.455298 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.455396 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.455414 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.455441 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.455461 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:51Z","lastTransitionTime":"2025-11-25T09:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.558662 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.558754 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.558777 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.558802 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.558819 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:51Z","lastTransitionTime":"2025-11-25T09:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.661223 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.661278 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.661295 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.661321 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.661339 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:51Z","lastTransitionTime":"2025-11-25T09:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.661370 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.661410 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.661447 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:51 crc kubenswrapper[4776]: E1125 09:25:51.661651 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:51 crc kubenswrapper[4776]: E1125 09:25:51.661821 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:51 crc kubenswrapper[4776]: E1125 09:25:51.662306 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.662769 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:51 crc kubenswrapper[4776]: E1125 09:25:51.662932 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.764801 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.764875 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.764899 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.764927 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.764952 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:51Z","lastTransitionTime":"2025-11-25T09:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.868277 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.868330 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.868348 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.868371 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.868387 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:51Z","lastTransitionTime":"2025-11-25T09:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.970889 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.970929 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.970940 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.970957 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:51 crc kubenswrapper[4776]: I1125 09:25:51.970970 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:51Z","lastTransitionTime":"2025-11-25T09:25:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.073513 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.073539 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.073547 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.073559 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.073567 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:52Z","lastTransitionTime":"2025-11-25T09:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.176805 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.176862 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.176880 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.176901 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.176917 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:52Z","lastTransitionTime":"2025-11-25T09:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.280112 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.280180 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.280204 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.280233 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.280255 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:52Z","lastTransitionTime":"2025-11-25T09:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.383231 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.383281 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.383293 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.383311 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.383324 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:52Z","lastTransitionTime":"2025-11-25T09:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.486788 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.486830 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.486841 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.486859 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.486875 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:52Z","lastTransitionTime":"2025-11-25T09:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.589490 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.589530 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.589543 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.589559 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.589573 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:52Z","lastTransitionTime":"2025-11-25T09:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.691812 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.691881 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.691918 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.691949 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.691969 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:52Z","lastTransitionTime":"2025-11-25T09:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.794135 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.794214 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.794238 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.794271 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.794292 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:52Z","lastTransitionTime":"2025-11-25T09:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.897105 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.897167 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.897183 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.897522 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:52 crc kubenswrapper[4776]: I1125 09:25:52.897557 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:52Z","lastTransitionTime":"2025-11-25T09:25:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.000975 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.001035 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.001053 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.001101 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.001119 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:53Z","lastTransitionTime":"2025-11-25T09:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.104892 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.104928 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.104937 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.104951 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.104965 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:53Z","lastTransitionTime":"2025-11-25T09:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.207229 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.207272 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.207284 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.207300 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.207313 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:53Z","lastTransitionTime":"2025-11-25T09:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.310026 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.310295 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.310366 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.310443 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.310511 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:53Z","lastTransitionTime":"2025-11-25T09:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.414286 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.414339 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.414355 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.414376 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.414392 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:53Z","lastTransitionTime":"2025-11-25T09:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.517633 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.517715 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.517730 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.517747 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.517756 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:53Z","lastTransitionTime":"2025-11-25T09:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.619825 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.619914 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.619938 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.619969 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.619991 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:53Z","lastTransitionTime":"2025-11-25T09:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.651552 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.651601 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.651612 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.651629 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.651642 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:53Z","lastTransitionTime":"2025-11-25T09:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.662137 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:53 crc kubenswrapper[4776]: E1125 09:25:53.662289 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.662317 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.662341 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:53 crc kubenswrapper[4776]: E1125 09:25:53.662412 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.662430 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:53 crc kubenswrapper[4776]: E1125 09:25:53.662511 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:53 crc kubenswrapper[4776]: E1125 09:25:53.662677 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:53 crc kubenswrapper[4776]: E1125 09:25:53.670252 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.675271 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.675314 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.675324 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.675363 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.675375 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:53Z","lastTransitionTime":"2025-11-25T09:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:53 crc kubenswrapper[4776]: E1125 09:25:53.703708 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.713849 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.713942 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.713969 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.714020 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.714058 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:53Z","lastTransitionTime":"2025-11-25T09:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:53 crc kubenswrapper[4776]: E1125 09:25:53.742306 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.749381 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.749411 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.749420 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.749432 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.749441 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:53Z","lastTransitionTime":"2025-11-25T09:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:53 crc kubenswrapper[4776]: E1125 09:25:53.763359 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.767176 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.767207 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.767215 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.767228 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.767237 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:53Z","lastTransitionTime":"2025-11-25T09:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:53 crc kubenswrapper[4776]: E1125 09:25:53.779305 4776 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a4d2f912-d0fd-42f2-920e-273120324c92\\\",\\\"systemUUID\\\":\\\"1c412267-e4cd-44b2-89dc-7f2cc5766618\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:53Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:53 crc kubenswrapper[4776]: E1125 09:25:53.779447 4776 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.780877 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.780916 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.780928 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.780944 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.780953 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:53Z","lastTransitionTime":"2025-11-25T09:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.884120 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.884195 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.884212 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.884236 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.884255 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:53Z","lastTransitionTime":"2025-11-25T09:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.987345 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.987401 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.987413 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.987429 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:53 crc kubenswrapper[4776]: I1125 09:25:53.987440 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:53Z","lastTransitionTime":"2025-11-25T09:25:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.089825 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.089916 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.089935 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.089959 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.089977 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:54Z","lastTransitionTime":"2025-11-25T09:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.193029 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.193132 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.193155 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.193184 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.193206 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:54Z","lastTransitionTime":"2025-11-25T09:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.295701 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.295753 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.295765 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.295781 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.295793 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:54Z","lastTransitionTime":"2025-11-25T09:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.397854 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.397894 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.397903 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.397921 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.397931 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:54Z","lastTransitionTime":"2025-11-25T09:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.507580 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.507617 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.507625 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.507639 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.507647 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:54Z","lastTransitionTime":"2025-11-25T09:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.609763 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.609850 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.609867 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.609892 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.609910 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:54Z","lastTransitionTime":"2025-11-25T09:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.712996 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.713109 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.713147 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.713179 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.713202 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:54Z","lastTransitionTime":"2025-11-25T09:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.816700 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.816783 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.816805 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.816830 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.816855 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:54Z","lastTransitionTime":"2025-11-25T09:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.919424 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.919487 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.919505 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.919529 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:54 crc kubenswrapper[4776]: I1125 09:25:54.919548 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:54Z","lastTransitionTime":"2025-11-25T09:25:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.023647 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.023711 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.023733 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.023762 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.023783 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:55Z","lastTransitionTime":"2025-11-25T09:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.126405 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.126457 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.126474 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.126494 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.126509 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:55Z","lastTransitionTime":"2025-11-25T09:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.229951 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.230272 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.230369 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.230469 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.230562 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:55Z","lastTransitionTime":"2025-11-25T09:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.333640 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.333712 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.333730 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.333754 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.333772 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:55Z","lastTransitionTime":"2025-11-25T09:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.437153 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.437209 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.437220 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.437240 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.437256 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:55Z","lastTransitionTime":"2025-11-25T09:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.539821 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.539865 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.539877 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.539894 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.539907 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:55Z","lastTransitionTime":"2025-11-25T09:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.643678 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.643739 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.643753 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.643776 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.643791 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:55Z","lastTransitionTime":"2025-11-25T09:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.661402 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:55 crc kubenswrapper[4776]: E1125 09:25:55.661567 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.661652 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.661755 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:55 crc kubenswrapper[4776]: E1125 09:25:55.661830 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.661862 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:55 crc kubenswrapper[4776]: E1125 09:25:55.662023 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:55 crc kubenswrapper[4776]: E1125 09:25:55.662236 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.694029 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"70332448-1918-4624-96cd-206ad2620c7c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ebe095106e95a23fc63001f042197ebc0799fa5126d729fead71c3a72719f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93fd6cfdc2eb470ee3848171040245ebba68f48897b6abc519cd182a3c8f71b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9dc2c08878eb3dfde1483a6c3edba49ec6887c954dae71885331c67bf0f64f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://37f4a13594a301afd075b5fb5dc6c3a3fc820d376ac026d6c2ff13af0d493209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://23ffcbb8606f10820f78446571c9df8f557b8a601a4b741f46a6b77c2b2a518a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://60522a830698be0a7b0cec858edfc5c12fcd7d9cd99b94fda6037e0b57f1fd22\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb4d47bd63c284b4a08f5507ec606fd9bd618b7b731497cd10ee2b97f95aee8e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://483af096f42ecd19e9cba404e8ee16a6650692570f21d48f02570806a3d2b454\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.712152 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fa47ebcc-a95e-4693-876d-7284c28c3ade\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0bd15355bc66e9e954857ec0789bf7e949dbd6aeb23073674f45fb56fc644d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t8ftp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-84dqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.730605 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a506265-8f64-456c-995f-1bc468cb87dd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29ead16f9d75513fe9ab6959e2ba0dca4f4b88df8bc65e3b0e6a21f28f2d0d30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb2778606c71db6545b88fd0648a8ab08a66863579467e49a5ece3b36d0d21f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rcf26\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-rzwmj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.746801 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.747034 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.747084 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.747108 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.747124 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:55Z","lastTransitionTime":"2025-11-25T09:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.753725 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5238060c-61a9-42a7-8eb7-961ba07cf3a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c6056703a9bc52c936e9e202b84cee744f59fd2d1ef42bc537f537678598a68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://90b098cc297ebdbe4e31b61800d691cb0c3c5814c8e60abac4c2ffe36ee7e2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9836d5211ef7b70bab65a7d8b182aa0e02085df02528a19db3b21ff7301d670\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://db22de3ee20f72a81d958f5fa9b10982a4ec446e092817b1880f669656e7a1ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://399a4f10c2249bacf7e517465e5b2240fcfeb62bb027a2cd4a475742d54eef4d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:24:32Z\\\",\\\"message\\\":\\\"le observer\\\\nW1125 09:24:31.728575 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1125 09:24:31.728858 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:24:31.729919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1566869983/tls.crt::/tmp/serving-cert-1566869983/tls.key\\\\\\\"\\\\nI1125 09:24:32.121778 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:24:32.128212 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:24:32.128260 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:24:32.128665 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:24:32.128688 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:24:32.139387 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:24:32.139435 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139447 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:24:32.139459 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:24:32.139468 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:24:32.139476 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:24:32.139485 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:24:32.139644 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:24:32.143619 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:31Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6535cc2cdbad9c5b2ae9bb8ad90345edc1ea77c41e84bf1090efec658d3a1ddf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e4b5cc924a6486c44334a9f96d9969955144e80e546eca8a6e4ae4faa13878f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.776432 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a4dab1da-2c11-436a-b64a-7bb682ea60d9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20ed5099652c62158adf9604f6f51431dfdad65857b15b113fc9196307201f04\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281547d4d4442adc288e329733bf38f863408f495c786584e8cec7aa7a99f7f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://778ecf1e896e4b1314eddb8a4176a722fb22df576ebe53d27f5e05caa5f36ddc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d05eacba75b3f2a211ee6696ab4872805cba9707a2b86e3e1c5bc8bf67fcf33\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.796337 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19e15b0b8586f2319da82ffb64ea63a0ffa6df756e246f9b85e41b8a0734fbf2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8aa5b16235d427aed6d28dc20e367cb47b930cccf5433b6270ce8de3e3c94f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.815336 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.841514 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34502968-eb8d-47fa-83c2-7f6f6735ce2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c5b9f801210cc7df25d48a313b74f7a347bfd5cba0855ee054fc48baf53e3b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c5b9f801210cc7df25d48a313b74f7a347bfd5cba0855ee054fc48baf53e3b0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:25:45Z\\\",\\\"message\\\":\\\"ctory.go:656] Stopping watch factory\\\\nI1125 09:25:45.621334 6872 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 09:25:45.621345 6872 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 09:25:45.621360 6872 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 09:25:45.621369 6872 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 09:25:45.621376 6872 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 09:25:45.621395 6872 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1125 09:25:45.621486 6872 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 09:25:45.621573 6872 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1125 09:25:45.621682 6872 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:25:45.622084 6872 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:25:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-r7x57_openshift-ovn-kubernetes(34502968-eb8d-47fa-83c2-7f6f6735ce2d)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmdxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-r7x57\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.849753 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.849799 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.849811 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.849829 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.849841 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:55Z","lastTransitionTime":"2025-11-25T09:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.856448 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6gp8r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0d194bcf-a8c4-4dd9-8239-d4172c5480e3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9b8aaef5feb786547066568f3daf234a2388c3d0e820b354928ea355b467279\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4cx8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:44Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6gp8r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.868880 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.880463 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.891050 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-nvc68" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5n95\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-nvc68\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.900732 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"26759602-0b52-410a-939d-c2df88f2b9fb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://655c0dff64b7dbe451d7c61bf1e1e63051f02f7587bbb65e71b5ee7d2bec0a2a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284857ebafc167c12fe70604afa6702c071316774d5976a5e95042968dafe0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf7b5ddb321a7c0a66d83ecb015a42c43b13ae4af284f4806d8fcf5a5f3ba005\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://583bd3f6f361dbdc9c02a6eeeee4a445e2d2d08f605adf2d3c8912cb7d416abb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://583bd3f6f361dbdc9c02a6eeeee4a445e2d2d08f605adf2d3c8912cb7d416abb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.908849 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"51b8c180-5c19-4df6-b088-f1656ef0946e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21b232d5cd319131183cefb0df8c537e099bfa1743d7fab3748858d3873a4802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f9c968afbab8870d4105cc2b407cb7b2392919716c0a9b5de3ed47f3d12e5a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9f9c968afbab8870d4105cc2b407cb7b2392919716c0a9b5de3ed47f3d12e5a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:15Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.920897 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a99d58727a5ca0de397ee391d6fcd2584bdb84eb21fef6197c9925bdf59d8a2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.930825 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a2b4582be3a99e3963f5790b01fd0d392bcc5ca94914698f0fd7460da1fcf32b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.940038 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hgvxs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9adab68-3143-470e-9988-a0190c8e1bc2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23309005a821e290cd840657ccda118a14a9c712972dc66466626b89d641a342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pjv57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hgvxs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.951581 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5q5c6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:25:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b1ea649a64d7a158675b397137593f6b901e762b67261d2e2872eeedc0d895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:25:29Z\\\",\\\"message\\\":\\\"2025-11-25T09:24:43+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c6e651f3-f403-47a2-b38e-0d19d75e1f16\\\\n2025-11-25T09:24:43+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c6e651f3-f403-47a2-b38e-0d19d75e1f16 to /host/opt/cni/bin/\\\\n2025-11-25T09:24:44Z [verbose] multus-daemon started\\\\n2025-11-25T09:24:44Z [verbose] Readiness Indicator file check\\\\n2025-11-25T09:25:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:25:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9d5qr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5q5c6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.952413 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.952464 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.952476 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.952492 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.952503 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:55Z","lastTransitionTime":"2025-11-25T09:25:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:55 crc kubenswrapper[4776]: I1125 09:25:55.966556 4776 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hl78" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5a1bccd-873a-4af1-8217-c60998b613f2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:24:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2ad63967f82781f8cf6d6c9512713cc8927666707203370dc38b031c201475a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:24:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://08b625beb6756a64a8a50d9c4fdab9635606051212ea3d633e800ad63b52c127\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c1132d6ae5645727a1e8620d979862b7223564198c99c022b269709f0cdc860\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://415eb6be2f5759ade4b8096f7d0d9b1a9b7fc535e834d435d2f84c47eba445c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7adb6fc5ef5a329f5c8d4dafafac77027121c5187b0e9348f161d90d3c7f9c76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62268bf2f3e1f6b399805baa104d456054a7135802c985dda16232b9b43610fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b69520c85c007f847a4a11f785629865a9d7ef612ce3ec581ba301dbb73cd25\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:24:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:24:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kk9n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:24:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hl78\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:25:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.055138 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.055183 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.055197 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.055214 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.055226 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:56Z","lastTransitionTime":"2025-11-25T09:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.158817 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.158900 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.158926 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.158960 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.158982 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:56Z","lastTransitionTime":"2025-11-25T09:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.261922 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.261993 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.262015 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.262042 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.262099 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:56Z","lastTransitionTime":"2025-11-25T09:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.364423 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.364462 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.364476 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.364496 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.364510 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:56Z","lastTransitionTime":"2025-11-25T09:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.466797 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.467207 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.467399 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.467647 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.467868 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:56Z","lastTransitionTime":"2025-11-25T09:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.571416 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.571481 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.571507 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.571535 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.571556 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:56Z","lastTransitionTime":"2025-11-25T09:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.674036 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.674137 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.674161 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.674187 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.674209 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:56Z","lastTransitionTime":"2025-11-25T09:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.777923 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.777990 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.778000 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.778022 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.778037 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:56Z","lastTransitionTime":"2025-11-25T09:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.880781 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.880835 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.880852 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.880874 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.880890 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:56Z","lastTransitionTime":"2025-11-25T09:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.983680 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.983737 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.983754 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.983782 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:56 crc kubenswrapper[4776]: I1125 09:25:56.983800 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:56Z","lastTransitionTime":"2025-11-25T09:25:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.087168 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.087252 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.087278 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.087470 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.087594 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:57Z","lastTransitionTime":"2025-11-25T09:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.190632 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.190690 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.190709 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.190733 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.190751 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:57Z","lastTransitionTime":"2025-11-25T09:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.293019 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.293126 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.293144 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.293174 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.293192 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:57Z","lastTransitionTime":"2025-11-25T09:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.395937 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.396007 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.396030 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.396109 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.396134 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:57Z","lastTransitionTime":"2025-11-25T09:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.499871 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.499921 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.499930 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.499948 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.499959 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:57Z","lastTransitionTime":"2025-11-25T09:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.602785 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.602826 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.602838 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.602857 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.602874 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:57Z","lastTransitionTime":"2025-11-25T09:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.662325 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.662588 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.662653 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.662594 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:57 crc kubenswrapper[4776]: E1125 09:25:57.662857 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:57 crc kubenswrapper[4776]: E1125 09:25:57.662966 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:57 crc kubenswrapper[4776]: E1125 09:25:57.663104 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:57 crc kubenswrapper[4776]: E1125 09:25:57.663181 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.706766 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.706836 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.706851 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.706890 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.706912 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:57Z","lastTransitionTime":"2025-11-25T09:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.809380 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.809470 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.809487 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.809513 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.809532 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:57Z","lastTransitionTime":"2025-11-25T09:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.911998 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.912100 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.912123 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.912146 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:57 crc kubenswrapper[4776]: I1125 09:25:57.912162 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:57Z","lastTransitionTime":"2025-11-25T09:25:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.015479 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.015522 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.015539 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.015561 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.015577 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:58Z","lastTransitionTime":"2025-11-25T09:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.117847 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.117906 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.117929 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.117960 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.117982 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:58Z","lastTransitionTime":"2025-11-25T09:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.220166 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.220207 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.220215 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.220227 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.220236 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:58Z","lastTransitionTime":"2025-11-25T09:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.323226 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.323303 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.323330 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.323357 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.323376 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:58Z","lastTransitionTime":"2025-11-25T09:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.426481 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.426545 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.426558 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.426586 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.426605 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:58Z","lastTransitionTime":"2025-11-25T09:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.529364 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.529408 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.529416 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.529437 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.529449 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:58Z","lastTransitionTime":"2025-11-25T09:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.631987 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.632044 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.632055 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.632093 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.632105 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:58Z","lastTransitionTime":"2025-11-25T09:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.741706 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.741774 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.741793 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.741819 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.741838 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:58Z","lastTransitionTime":"2025-11-25T09:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.844689 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.844753 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.844770 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.844795 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.844812 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:58Z","lastTransitionTime":"2025-11-25T09:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.947546 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.947580 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.947591 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.947607 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:58 crc kubenswrapper[4776]: I1125 09:25:58.947618 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:58Z","lastTransitionTime":"2025-11-25T09:25:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.049988 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.050045 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.050096 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.050122 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.050140 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:59Z","lastTransitionTime":"2025-11-25T09:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.152798 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.152864 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.152887 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.152916 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.152940 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:59Z","lastTransitionTime":"2025-11-25T09:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.255734 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.255805 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.255822 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.255848 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.255867 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:59Z","lastTransitionTime":"2025-11-25T09:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.358518 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.358585 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.358609 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.358644 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.358667 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:59Z","lastTransitionTime":"2025-11-25T09:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.461118 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.461172 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.461189 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.461213 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.461230 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:59Z","lastTransitionTime":"2025-11-25T09:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.564920 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.564994 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.565014 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.565041 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.565100 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:59Z","lastTransitionTime":"2025-11-25T09:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.661792 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.661880 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.661803 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:25:59 crc kubenswrapper[4776]: E1125 09:25:59.661995 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:25:59 crc kubenswrapper[4776]: E1125 09:25:59.662213 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:25:59 crc kubenswrapper[4776]: E1125 09:25:59.662350 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.662472 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:25:59 crc kubenswrapper[4776]: E1125 09:25:59.662592 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.668163 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.668241 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.669024 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.669053 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.669110 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:59Z","lastTransitionTime":"2025-11-25T09:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.771626 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.771666 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.771677 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.771695 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.771706 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:59Z","lastTransitionTime":"2025-11-25T09:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.874999 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.875113 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.875152 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.875184 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.875245 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:25:59Z","lastTransitionTime":"2025-11-25T09:25:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:25:59 crc kubenswrapper[4776]: I1125 09:25:59.916109 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-metrics-certs\") pod \"network-metrics-daemon-nvc68\" (UID: \"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\") " pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:25:59 crc kubenswrapper[4776]: E1125 09:25:59.916358 4776 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:25:59 crc kubenswrapper[4776]: E1125 09:25:59.916686 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-metrics-certs podName:4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087 nodeName:}" failed. No retries permitted until 2025-11-25 09:27:03.916487217 +0000 UTC m=+168.957546820 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-metrics-certs") pod "network-metrics-daemon-nvc68" (UID: "4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.355895 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.355957 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.355980 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.356010 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.356033 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:00Z","lastTransitionTime":"2025-11-25T09:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.458363 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.458413 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.458424 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.458440 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.458450 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:00Z","lastTransitionTime":"2025-11-25T09:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.561502 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.561558 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.561574 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.561596 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.561613 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:00Z","lastTransitionTime":"2025-11-25T09:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.664245 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.664328 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.664352 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.664382 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.664407 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:00Z","lastTransitionTime":"2025-11-25T09:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.767592 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.767661 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.767690 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.767717 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.767740 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:00Z","lastTransitionTime":"2025-11-25T09:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.872119 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.872180 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.872197 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.872223 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.872243 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:00Z","lastTransitionTime":"2025-11-25T09:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.976141 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.976611 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.976792 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.976959 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:00 crc kubenswrapper[4776]: I1125 09:26:00.977173 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:00Z","lastTransitionTime":"2025-11-25T09:26:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.079953 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.079999 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.080009 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.080025 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.080035 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:01Z","lastTransitionTime":"2025-11-25T09:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.183130 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.183178 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.183249 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.183269 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.183279 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:01Z","lastTransitionTime":"2025-11-25T09:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.286743 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.286802 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.286826 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.286853 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.286870 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:01Z","lastTransitionTime":"2025-11-25T09:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.389621 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.389683 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.389696 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.389714 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.389725 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:01Z","lastTransitionTime":"2025-11-25T09:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.493387 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.493453 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.493476 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.493510 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.493535 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:01Z","lastTransitionTime":"2025-11-25T09:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.596927 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.597357 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.597539 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.597714 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.597873 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:01Z","lastTransitionTime":"2025-11-25T09:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.662809 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.662897 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.663309 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.663660 4776 scope.go:117] "RemoveContainer" containerID="3c5b9f801210cc7df25d48a313b74f7a347bfd5cba0855ee054fc48baf53e3b0" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.663867 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:26:01 crc kubenswrapper[4776]: E1125 09:26:01.664011 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:26:01 crc kubenswrapper[4776]: E1125 09:26:01.664015 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-r7x57_openshift-ovn-kubernetes(34502968-eb8d-47fa-83c2-7f6f6735ce2d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" Nov 25 09:26:01 crc kubenswrapper[4776]: E1125 09:26:01.664157 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:26:01 crc kubenswrapper[4776]: E1125 09:26:01.664249 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:26:01 crc kubenswrapper[4776]: E1125 09:26:01.664341 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.700970 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.701106 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.701135 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.701173 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.701198 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:01Z","lastTransitionTime":"2025-11-25T09:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.804395 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.804447 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.804466 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.804489 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.804506 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:01Z","lastTransitionTime":"2025-11-25T09:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.908506 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.908558 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.908566 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.908584 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:01 crc kubenswrapper[4776]: I1125 09:26:01.908598 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:01Z","lastTransitionTime":"2025-11-25T09:26:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.011804 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.011869 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.011888 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.011911 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.011940 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:02Z","lastTransitionTime":"2025-11-25T09:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.115225 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.115286 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.115296 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.115314 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.115331 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:02Z","lastTransitionTime":"2025-11-25T09:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.218982 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.219037 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.219051 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.219088 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.219103 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:02Z","lastTransitionTime":"2025-11-25T09:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.321812 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.321888 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.321908 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.321937 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.321957 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:02Z","lastTransitionTime":"2025-11-25T09:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.424885 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.424965 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.424978 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.424999 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.425013 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:02Z","lastTransitionTime":"2025-11-25T09:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.528413 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.528493 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.528513 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.528538 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.528553 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:02Z","lastTransitionTime":"2025-11-25T09:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.631723 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.631771 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.631784 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.631801 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.631816 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:02Z","lastTransitionTime":"2025-11-25T09:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.735278 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.735351 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.735370 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.735399 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.735418 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:02Z","lastTransitionTime":"2025-11-25T09:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.838671 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.838729 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.838746 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.838800 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.838821 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:02Z","lastTransitionTime":"2025-11-25T09:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.942701 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.942761 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.942773 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.942792 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:02 crc kubenswrapper[4776]: I1125 09:26:02.942806 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:02Z","lastTransitionTime":"2025-11-25T09:26:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.045803 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.045871 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.045895 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.045924 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.045948 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:03Z","lastTransitionTime":"2025-11-25T09:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.149392 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.149456 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.149473 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.149502 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.149521 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:03Z","lastTransitionTime":"2025-11-25T09:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.253167 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.253213 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.253223 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.253238 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.253247 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:03Z","lastTransitionTime":"2025-11-25T09:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.356051 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.356114 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.356126 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.356150 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.356162 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:03Z","lastTransitionTime":"2025-11-25T09:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.458597 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.458658 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.458678 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.458704 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.458723 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:03Z","lastTransitionTime":"2025-11-25T09:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.560732 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.560782 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.560791 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.560807 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.560818 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:03Z","lastTransitionTime":"2025-11-25T09:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.661917 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.661944 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:26:03 crc kubenswrapper[4776]: E1125 09:26:03.662032 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.662097 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:26:03 crc kubenswrapper[4776]: E1125 09:26:03.662195 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.662327 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:26:03 crc kubenswrapper[4776]: E1125 09:26:03.662489 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:26:03 crc kubenswrapper[4776]: E1125 09:26:03.662604 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.663425 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.663473 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.663492 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.663883 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.663935 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:03Z","lastTransitionTime":"2025-11-25T09:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.766513 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.766566 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.766583 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.766605 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.766622 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:03Z","lastTransitionTime":"2025-11-25T09:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.869390 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.869441 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.869452 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.869468 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.869481 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:03Z","lastTransitionTime":"2025-11-25T09:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.895719 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.895779 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.895795 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.895816 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.895839 4776 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:26:03Z","lastTransitionTime":"2025-11-25T09:26:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.959595 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-tbrhl"] Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.960207 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tbrhl" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.967031 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.967478 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.968522 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.970054 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 25 09:26:03 crc kubenswrapper[4776]: I1125 09:26:03.996975 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=87.996957669 podStartE2EDuration="1m27.996957669s" podCreationTimestamp="2025-11-25 09:24:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:03.996355074 +0000 UTC m=+109.037414627" watchObservedRunningTime="2025-11-25 09:26:03.996957669 +0000 UTC m=+109.038017222" Nov 25 09:26:04 crc kubenswrapper[4776]: I1125 09:26:04.027427 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podStartSLOduration=83.02740308 podStartE2EDuration="1m23.02740308s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:04.017481022 +0000 UTC m=+109.058540605" watchObservedRunningTime="2025-11-25 09:26:04.02740308 +0000 UTC m=+109.068462663" Nov 25 09:26:04 crc kubenswrapper[4776]: I1125 09:26:04.053850 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-rzwmj" podStartSLOduration=83.053834921 podStartE2EDuration="1m23.053834921s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:04.028140139 +0000 UTC m=+109.069199732" watchObservedRunningTime="2025-11-25 09:26:04.053834921 +0000 UTC m=+109.094894474" Nov 25 09:26:04 crc kubenswrapper[4776]: I1125 09:26:04.066047 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7872a565-95a1-42ff-8ac7-356972474e48-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-tbrhl\" (UID: \"7872a565-95a1-42ff-8ac7-356972474e48\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tbrhl" Nov 25 09:26:04 crc kubenswrapper[4776]: I1125 09:26:04.066190 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7872a565-95a1-42ff-8ac7-356972474e48-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-tbrhl\" (UID: \"7872a565-95a1-42ff-8ac7-356972474e48\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tbrhl" Nov 25 09:26:04 crc kubenswrapper[4776]: I1125 09:26:04.066337 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7872a565-95a1-42ff-8ac7-356972474e48-service-ca\") pod \"cluster-version-operator-5c965bbfc6-tbrhl\" (UID: \"7872a565-95a1-42ff-8ac7-356972474e48\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tbrhl" Nov 25 09:26:04 crc kubenswrapper[4776]: I1125 09:26:04.066409 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7872a565-95a1-42ff-8ac7-356972474e48-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-tbrhl\" (UID: \"7872a565-95a1-42ff-8ac7-356972474e48\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tbrhl" Nov 25 09:26:04 crc kubenswrapper[4776]: I1125 09:26:04.066501 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7872a565-95a1-42ff-8ac7-356972474e48-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-tbrhl\" (UID: \"7872a565-95a1-42ff-8ac7-356972474e48\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tbrhl" Nov 25 09:26:04 crc kubenswrapper[4776]: I1125 09:26:04.067575 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-6gp8r" podStartSLOduration=83.067554474 podStartE2EDuration="1m23.067554474s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:04.066533709 +0000 UTC m=+109.107593262" watchObservedRunningTime="2025-11-25 09:26:04.067554474 +0000 UTC m=+109.108614067" Nov 25 09:26:04 crc kubenswrapper[4776]: I1125 09:26:04.081290 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=89.081280137 podStartE2EDuration="1m29.081280137s" podCreationTimestamp="2025-11-25 09:24:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:04.08058307 +0000 UTC m=+109.121642643" watchObservedRunningTime="2025-11-25 09:26:04.081280137 +0000 UTC m=+109.122339690" Nov 25 09:26:04 crc kubenswrapper[4776]: I1125 09:26:04.097998 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=82.097974154 podStartE2EDuration="1m22.097974154s" podCreationTimestamp="2025-11-25 09:24:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:04.097616525 +0000 UTC m=+109.138676108" watchObservedRunningTime="2025-11-25 09:26:04.097974154 +0000 UTC m=+109.139033717" Nov 25 09:26:04 crc kubenswrapper[4776]: I1125 09:26:04.167390 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7872a565-95a1-42ff-8ac7-356972474e48-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-tbrhl\" (UID: \"7872a565-95a1-42ff-8ac7-356972474e48\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tbrhl" Nov 25 09:26:04 crc kubenswrapper[4776]: I1125 09:26:04.167435 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7872a565-95a1-42ff-8ac7-356972474e48-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-tbrhl\" (UID: \"7872a565-95a1-42ff-8ac7-356972474e48\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tbrhl" Nov 25 09:26:04 crc kubenswrapper[4776]: I1125 09:26:04.167465 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7872a565-95a1-42ff-8ac7-356972474e48-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-tbrhl\" (UID: \"7872a565-95a1-42ff-8ac7-356972474e48\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tbrhl" Nov 25 09:26:04 crc kubenswrapper[4776]: I1125 09:26:04.167571 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7872a565-95a1-42ff-8ac7-356972474e48-service-ca\") pod \"cluster-version-operator-5c965bbfc6-tbrhl\" (UID: \"7872a565-95a1-42ff-8ac7-356972474e48\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tbrhl" Nov 25 09:26:04 crc kubenswrapper[4776]: I1125 09:26:04.167574 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7872a565-95a1-42ff-8ac7-356972474e48-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-tbrhl\" (UID: \"7872a565-95a1-42ff-8ac7-356972474e48\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tbrhl" Nov 25 09:26:04 crc kubenswrapper[4776]: I1125 09:26:04.167619 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7872a565-95a1-42ff-8ac7-356972474e48-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-tbrhl\" (UID: \"7872a565-95a1-42ff-8ac7-356972474e48\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tbrhl" Nov 25 09:26:04 crc kubenswrapper[4776]: I1125 09:26:04.167652 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7872a565-95a1-42ff-8ac7-356972474e48-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-tbrhl\" (UID: \"7872a565-95a1-42ff-8ac7-356972474e48\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tbrhl" Nov 25 09:26:04 crc kubenswrapper[4776]: I1125 09:26:04.168883 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7872a565-95a1-42ff-8ac7-356972474e48-service-ca\") pod \"cluster-version-operator-5c965bbfc6-tbrhl\" (UID: \"7872a565-95a1-42ff-8ac7-356972474e48\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tbrhl" Nov 25 09:26:04 crc kubenswrapper[4776]: I1125 09:26:04.177688 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7872a565-95a1-42ff-8ac7-356972474e48-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-tbrhl\" (UID: \"7872a565-95a1-42ff-8ac7-356972474e48\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tbrhl" Nov 25 09:26:04 crc kubenswrapper[4776]: I1125 09:26:04.189257 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7872a565-95a1-42ff-8ac7-356972474e48-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-tbrhl\" (UID: \"7872a565-95a1-42ff-8ac7-356972474e48\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tbrhl" Nov 25 09:26:04 crc kubenswrapper[4776]: I1125 09:26:04.197956 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-hgvxs" podStartSLOduration=83.197939063 podStartE2EDuration="1m23.197939063s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:04.184605 +0000 UTC m=+109.225664553" watchObservedRunningTime="2025-11-25 09:26:04.197939063 +0000 UTC m=+109.238998616" Nov 25 09:26:04 crc kubenswrapper[4776]: I1125 09:26:04.214826 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-5q5c6" podStartSLOduration=83.214803725 podStartE2EDuration="1m23.214803725s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:04.198498597 +0000 UTC m=+109.239558150" watchObservedRunningTime="2025-11-25 09:26:04.214803725 +0000 UTC m=+109.255863288" Nov 25 09:26:04 crc kubenswrapper[4776]: I1125 09:26:04.215453 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-4hl78" podStartSLOduration=83.215444321 podStartE2EDuration="1m23.215444321s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:04.21542904 +0000 UTC m=+109.256488663" watchObservedRunningTime="2025-11-25 09:26:04.215444321 +0000 UTC m=+109.256503874" Nov 25 09:26:04 crc kubenswrapper[4776]: I1125 09:26:04.248099 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=58.248081556 podStartE2EDuration="58.248081556s" podCreationTimestamp="2025-11-25 09:25:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:04.247087412 +0000 UTC m=+109.288146965" watchObservedRunningTime="2025-11-25 09:26:04.248081556 +0000 UTC m=+109.289141109" Nov 25 09:26:04 crc kubenswrapper[4776]: I1125 09:26:04.273121 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=34.273094872 podStartE2EDuration="34.273094872s" podCreationTimestamp="2025-11-25 09:25:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:04.259110982 +0000 UTC m=+109.300170555" watchObservedRunningTime="2025-11-25 09:26:04.273094872 +0000 UTC m=+109.314154425" Nov 25 09:26:04 crc kubenswrapper[4776]: I1125 09:26:04.279310 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tbrhl" Nov 25 09:26:04 crc kubenswrapper[4776]: I1125 09:26:04.372313 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tbrhl" event={"ID":"7872a565-95a1-42ff-8ac7-356972474e48","Type":"ContainerStarted","Data":"4f39c9dd77ec01db11d39bdcf3402668c595f311536b4d95634c52264e774789"} Nov 25 09:26:05 crc kubenswrapper[4776]: I1125 09:26:05.378217 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tbrhl" event={"ID":"7872a565-95a1-42ff-8ac7-356972474e48","Type":"ContainerStarted","Data":"82a6ed4c39fd69b36875fce0abc6ad2515495e82f6c7b789ddf570e0f5aeaee4"} Nov 25 09:26:05 crc kubenswrapper[4776]: I1125 09:26:05.394891 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-tbrhl" podStartSLOduration=84.39485651 podStartE2EDuration="1m24.39485651s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:05.3944594 +0000 UTC m=+110.435518953" watchObservedRunningTime="2025-11-25 09:26:05.39485651 +0000 UTC m=+110.435916103" Nov 25 09:26:05 crc kubenswrapper[4776]: I1125 09:26:05.662110 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:26:05 crc kubenswrapper[4776]: I1125 09:26:05.662213 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:26:05 crc kubenswrapper[4776]: I1125 09:26:05.664083 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:26:05 crc kubenswrapper[4776]: E1125 09:26:05.664053 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:26:05 crc kubenswrapper[4776]: I1125 09:26:05.664172 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:26:05 crc kubenswrapper[4776]: E1125 09:26:05.664226 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:26:05 crc kubenswrapper[4776]: E1125 09:26:05.664356 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:26:05 crc kubenswrapper[4776]: E1125 09:26:05.664496 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:26:07 crc kubenswrapper[4776]: I1125 09:26:07.661857 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:26:07 crc kubenswrapper[4776]: I1125 09:26:07.661874 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:26:07 crc kubenswrapper[4776]: I1125 09:26:07.661944 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:26:07 crc kubenswrapper[4776]: I1125 09:26:07.662039 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:26:07 crc kubenswrapper[4776]: E1125 09:26:07.662145 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:26:07 crc kubenswrapper[4776]: E1125 09:26:07.662293 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:26:07 crc kubenswrapper[4776]: E1125 09:26:07.662354 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:26:07 crc kubenswrapper[4776]: E1125 09:26:07.662397 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:26:09 crc kubenswrapper[4776]: I1125 09:26:09.662325 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:26:09 crc kubenswrapper[4776]: I1125 09:26:09.662376 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:26:09 crc kubenswrapper[4776]: I1125 09:26:09.662393 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:26:09 crc kubenswrapper[4776]: E1125 09:26:09.662467 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:26:09 crc kubenswrapper[4776]: I1125 09:26:09.662478 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:26:09 crc kubenswrapper[4776]: E1125 09:26:09.662578 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:26:09 crc kubenswrapper[4776]: E1125 09:26:09.662638 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:26:09 crc kubenswrapper[4776]: E1125 09:26:09.662697 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:26:11 crc kubenswrapper[4776]: I1125 09:26:11.661802 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:26:11 crc kubenswrapper[4776]: I1125 09:26:11.661862 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:26:11 crc kubenswrapper[4776]: I1125 09:26:11.661962 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:26:11 crc kubenswrapper[4776]: E1125 09:26:11.661958 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:26:11 crc kubenswrapper[4776]: I1125 09:26:11.662136 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:26:11 crc kubenswrapper[4776]: E1125 09:26:11.662144 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:26:11 crc kubenswrapper[4776]: E1125 09:26:11.662207 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:26:11 crc kubenswrapper[4776]: E1125 09:26:11.662303 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:26:13 crc kubenswrapper[4776]: I1125 09:26:13.661272 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:26:13 crc kubenswrapper[4776]: I1125 09:26:13.661315 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:26:13 crc kubenswrapper[4776]: I1125 09:26:13.661397 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:26:13 crc kubenswrapper[4776]: E1125 09:26:13.662132 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:26:13 crc kubenswrapper[4776]: E1125 09:26:13.662205 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:26:13 crc kubenswrapper[4776]: E1125 09:26:13.662277 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:26:13 crc kubenswrapper[4776]: I1125 09:26:13.662338 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:26:13 crc kubenswrapper[4776]: E1125 09:26:13.662620 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:26:15 crc kubenswrapper[4776]: E1125 09:26:15.658599 4776 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 25 09:26:15 crc kubenswrapper[4776]: I1125 09:26:15.661333 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:26:15 crc kubenswrapper[4776]: E1125 09:26:15.661661 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:26:15 crc kubenswrapper[4776]: I1125 09:26:15.661864 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:26:15 crc kubenswrapper[4776]: E1125 09:26:15.663791 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:26:15 crc kubenswrapper[4776]: I1125 09:26:15.663824 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:26:15 crc kubenswrapper[4776]: I1125 09:26:15.663868 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:26:15 crc kubenswrapper[4776]: E1125 09:26:15.663955 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:26:15 crc kubenswrapper[4776]: E1125 09:26:15.664059 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:26:15 crc kubenswrapper[4776]: E1125 09:26:15.786557 4776 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 09:26:16 crc kubenswrapper[4776]: I1125 09:26:16.428223 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5q5c6_d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c/kube-multus/1.log" Nov 25 09:26:16 crc kubenswrapper[4776]: I1125 09:26:16.428814 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5q5c6_d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c/kube-multus/0.log" Nov 25 09:26:16 crc kubenswrapper[4776]: I1125 09:26:16.428867 4776 generic.go:334] "Generic (PLEG): container finished" podID="d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c" containerID="d7b1ea649a64d7a158675b397137593f6b901e762b67261d2e2872eeedc0d895" exitCode=1 Nov 25 09:26:16 crc kubenswrapper[4776]: I1125 09:26:16.428906 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5q5c6" event={"ID":"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c","Type":"ContainerDied","Data":"d7b1ea649a64d7a158675b397137593f6b901e762b67261d2e2872eeedc0d895"} Nov 25 09:26:16 crc kubenswrapper[4776]: I1125 09:26:16.428943 4776 scope.go:117] "RemoveContainer" containerID="c2436f929435ab26f55c9a44330802a5174d82c6d42db13481fe871cd0b4e8a9" Nov 25 09:26:16 crc kubenswrapper[4776]: I1125 09:26:16.429500 4776 scope.go:117] "RemoveContainer" containerID="d7b1ea649a64d7a158675b397137593f6b901e762b67261d2e2872eeedc0d895" Nov 25 09:26:16 crc kubenswrapper[4776]: E1125 09:26:16.429732 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-5q5c6_openshift-multus(d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c)\"" pod="openshift-multus/multus-5q5c6" podUID="d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c" Nov 25 09:26:16 crc kubenswrapper[4776]: I1125 09:26:16.662456 4776 scope.go:117] "RemoveContainer" containerID="3c5b9f801210cc7df25d48a313b74f7a347bfd5cba0855ee054fc48baf53e3b0" Nov 25 09:26:16 crc kubenswrapper[4776]: E1125 09:26:16.662722 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-r7x57_openshift-ovn-kubernetes(34502968-eb8d-47fa-83c2-7f6f6735ce2d)\"" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" Nov 25 09:26:17 crc kubenswrapper[4776]: I1125 09:26:17.434261 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5q5c6_d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c/kube-multus/1.log" Nov 25 09:26:17 crc kubenswrapper[4776]: I1125 09:26:17.661386 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:26:17 crc kubenswrapper[4776]: I1125 09:26:17.661493 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:26:17 crc kubenswrapper[4776]: I1125 09:26:17.661614 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:26:17 crc kubenswrapper[4776]: E1125 09:26:17.661610 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:26:17 crc kubenswrapper[4776]: E1125 09:26:17.661762 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:26:17 crc kubenswrapper[4776]: E1125 09:26:17.661898 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:26:17 crc kubenswrapper[4776]: I1125 09:26:17.661988 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:26:17 crc kubenswrapper[4776]: E1125 09:26:17.662125 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:26:19 crc kubenswrapper[4776]: I1125 09:26:19.661569 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:26:19 crc kubenswrapper[4776]: I1125 09:26:19.661736 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:26:19 crc kubenswrapper[4776]: E1125 09:26:19.661785 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:26:19 crc kubenswrapper[4776]: I1125 09:26:19.661888 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:26:19 crc kubenswrapper[4776]: E1125 09:26:19.662107 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:26:19 crc kubenswrapper[4776]: I1125 09:26:19.662453 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:26:19 crc kubenswrapper[4776]: E1125 09:26:19.662579 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:26:19 crc kubenswrapper[4776]: E1125 09:26:19.662893 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:26:20 crc kubenswrapper[4776]: E1125 09:26:20.787710 4776 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 09:26:21 crc kubenswrapper[4776]: I1125 09:26:21.661332 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:26:21 crc kubenswrapper[4776]: I1125 09:26:21.661366 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:26:21 crc kubenswrapper[4776]: E1125 09:26:21.661441 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:26:21 crc kubenswrapper[4776]: E1125 09:26:21.661532 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:26:21 crc kubenswrapper[4776]: I1125 09:26:21.661366 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:26:21 crc kubenswrapper[4776]: I1125 09:26:21.661569 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:26:21 crc kubenswrapper[4776]: E1125 09:26:21.661720 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:26:21 crc kubenswrapper[4776]: E1125 09:26:21.661743 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:26:23 crc kubenswrapper[4776]: I1125 09:26:23.661306 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:26:23 crc kubenswrapper[4776]: I1125 09:26:23.661374 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:26:23 crc kubenswrapper[4776]: I1125 09:26:23.661452 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:26:23 crc kubenswrapper[4776]: E1125 09:26:23.661457 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:26:23 crc kubenswrapper[4776]: E1125 09:26:23.661542 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:26:23 crc kubenswrapper[4776]: E1125 09:26:23.661607 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:26:23 crc kubenswrapper[4776]: I1125 09:26:23.661603 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:26:23 crc kubenswrapper[4776]: E1125 09:26:23.661669 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:26:25 crc kubenswrapper[4776]: I1125 09:26:25.661373 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:26:25 crc kubenswrapper[4776]: I1125 09:26:25.661373 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:26:25 crc kubenswrapper[4776]: E1125 09:26:25.662655 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:26:25 crc kubenswrapper[4776]: I1125 09:26:25.662673 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:26:25 crc kubenswrapper[4776]: I1125 09:26:25.662714 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:26:25 crc kubenswrapper[4776]: E1125 09:26:25.662802 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:26:25 crc kubenswrapper[4776]: E1125 09:26:25.662892 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:26:25 crc kubenswrapper[4776]: E1125 09:26:25.662968 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:26:25 crc kubenswrapper[4776]: E1125 09:26:25.789492 4776 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 09:26:27 crc kubenswrapper[4776]: I1125 09:26:27.662238 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:26:27 crc kubenswrapper[4776]: I1125 09:26:27.662312 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:26:27 crc kubenswrapper[4776]: I1125 09:26:27.662356 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:26:27 crc kubenswrapper[4776]: E1125 09:26:27.662399 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:26:27 crc kubenswrapper[4776]: E1125 09:26:27.662523 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:26:27 crc kubenswrapper[4776]: I1125 09:26:27.662549 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:26:27 crc kubenswrapper[4776]: E1125 09:26:27.662787 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:26:27 crc kubenswrapper[4776]: E1125 09:26:27.662935 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:26:28 crc kubenswrapper[4776]: I1125 09:26:28.662600 4776 scope.go:117] "RemoveContainer" containerID="3c5b9f801210cc7df25d48a313b74f7a347bfd5cba0855ee054fc48baf53e3b0" Nov 25 09:26:29 crc kubenswrapper[4776]: I1125 09:26:29.482549 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-r7x57_34502968-eb8d-47fa-83c2-7f6f6735ce2d/ovnkube-controller/3.log" Nov 25 09:26:29 crc kubenswrapper[4776]: I1125 09:26:29.486947 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" event={"ID":"34502968-eb8d-47fa-83c2-7f6f6735ce2d","Type":"ContainerStarted","Data":"fdec62f7fcca2fe281153267dffee23264c5d60c8d63fe9d451cd8699b962ae7"} Nov 25 09:26:29 crc kubenswrapper[4776]: I1125 09:26:29.487568 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:26:29 crc kubenswrapper[4776]: I1125 09:26:29.523257 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" podStartSLOduration=108.523227617 podStartE2EDuration="1m48.523227617s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:29.522686333 +0000 UTC m=+134.563745896" watchObservedRunningTime="2025-11-25 09:26:29.523227617 +0000 UTC m=+134.564287180" Nov 25 09:26:29 crc kubenswrapper[4776]: I1125 09:26:29.639147 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-nvc68"] Nov 25 09:26:29 crc kubenswrapper[4776]: I1125 09:26:29.639269 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:26:29 crc kubenswrapper[4776]: E1125 09:26:29.639390 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:26:29 crc kubenswrapper[4776]: I1125 09:26:29.661657 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:26:29 crc kubenswrapper[4776]: I1125 09:26:29.661679 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:26:29 crc kubenswrapper[4776]: I1125 09:26:29.661677 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:26:29 crc kubenswrapper[4776]: E1125 09:26:29.661784 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:26:29 crc kubenswrapper[4776]: E1125 09:26:29.661863 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:26:29 crc kubenswrapper[4776]: E1125 09:26:29.662104 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:26:30 crc kubenswrapper[4776]: E1125 09:26:30.791586 4776 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 09:26:31 crc kubenswrapper[4776]: I1125 09:26:31.662100 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:26:31 crc kubenswrapper[4776]: I1125 09:26:31.662109 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:26:31 crc kubenswrapper[4776]: I1125 09:26:31.662056 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:26:31 crc kubenswrapper[4776]: I1125 09:26:31.662247 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:26:31 crc kubenswrapper[4776]: E1125 09:26:31.662287 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:26:31 crc kubenswrapper[4776]: E1125 09:26:31.662397 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:26:31 crc kubenswrapper[4776]: I1125 09:26:31.662767 4776 scope.go:117] "RemoveContainer" containerID="d7b1ea649a64d7a158675b397137593f6b901e762b67261d2e2872eeedc0d895" Nov 25 09:26:31 crc kubenswrapper[4776]: E1125 09:26:31.662814 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:26:31 crc kubenswrapper[4776]: E1125 09:26:31.662863 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:26:32 crc kubenswrapper[4776]: I1125 09:26:32.501393 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5q5c6_d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c/kube-multus/1.log" Nov 25 09:26:32 crc kubenswrapper[4776]: I1125 09:26:32.501844 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5q5c6" event={"ID":"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c","Type":"ContainerStarted","Data":"ea0a18f856c9c659938df76dc703106bc870a8067586cfcda259e8dfe225e2ab"} Nov 25 09:26:33 crc kubenswrapper[4776]: I1125 09:26:33.661667 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:26:33 crc kubenswrapper[4776]: I1125 09:26:33.661743 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:26:33 crc kubenswrapper[4776]: E1125 09:26:33.661846 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:26:33 crc kubenswrapper[4776]: I1125 09:26:33.661910 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:26:33 crc kubenswrapper[4776]: E1125 09:26:33.662005 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:26:33 crc kubenswrapper[4776]: I1125 09:26:33.662252 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:26:33 crc kubenswrapper[4776]: E1125 09:26:33.662352 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:26:33 crc kubenswrapper[4776]: E1125 09:26:33.662591 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:26:35 crc kubenswrapper[4776]: I1125 09:26:35.661477 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:26:35 crc kubenswrapper[4776]: I1125 09:26:35.661595 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:26:35 crc kubenswrapper[4776]: I1125 09:26:35.661601 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:26:35 crc kubenswrapper[4776]: I1125 09:26:35.661611 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:26:35 crc kubenswrapper[4776]: E1125 09:26:35.666209 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:26:35 crc kubenswrapper[4776]: E1125 09:26:35.666541 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:26:35 crc kubenswrapper[4776]: E1125 09:26:35.666646 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:26:35 crc kubenswrapper[4776]: E1125 09:26:35.666705 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-nvc68" podUID="4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087" Nov 25 09:26:37 crc kubenswrapper[4776]: I1125 09:26:37.661880 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:26:37 crc kubenswrapper[4776]: I1125 09:26:37.661923 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:26:37 crc kubenswrapper[4776]: I1125 09:26:37.661957 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:26:37 crc kubenswrapper[4776]: I1125 09:26:37.662044 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:26:37 crc kubenswrapper[4776]: I1125 09:26:37.666581 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 25 09:26:37 crc kubenswrapper[4776]: I1125 09:26:37.666934 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 25 09:26:37 crc kubenswrapper[4776]: I1125 09:26:37.667242 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 25 09:26:37 crc kubenswrapper[4776]: I1125 09:26:37.667300 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 25 09:26:37 crc kubenswrapper[4776]: I1125 09:26:37.668147 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 25 09:26:37 crc kubenswrapper[4776]: I1125 09:26:37.668217 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 25 09:26:43 crc kubenswrapper[4776]: I1125 09:26:43.656914 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:43 crc kubenswrapper[4776]: E1125 09:26:43.657187 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:28:45.657135445 +0000 UTC m=+270.698195038 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:43 crc kubenswrapper[4776]: I1125 09:26:43.657466 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:26:43 crc kubenswrapper[4776]: I1125 09:26:43.657700 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:26:43 crc kubenswrapper[4776]: I1125 09:26:43.657773 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:26:43 crc kubenswrapper[4776]: I1125 09:26:43.657827 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:26:43 crc kubenswrapper[4776]: I1125 09:26:43.658872 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:26:43 crc kubenswrapper[4776]: I1125 09:26:43.664760 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:26:43 crc kubenswrapper[4776]: I1125 09:26:43.664890 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:26:43 crc kubenswrapper[4776]: I1125 09:26:43.664937 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:26:43 crc kubenswrapper[4776]: I1125 09:26:43.678649 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:26:43 crc kubenswrapper[4776]: I1125 09:26:43.687830 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:26:43 crc kubenswrapper[4776]: I1125 09:26:43.695305 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:26:43 crc kubenswrapper[4776]: W1125 09:26:43.893840 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-20b33f244badf3bb1a4911462ec5fbdc9bb3ac8d7799a29abb162639c9ce5f0c WatchSource:0}: Error finding container 20b33f244badf3bb1a4911462ec5fbdc9bb3ac8d7799a29abb162639c9ce5f0c: Status 404 returned error can't find the container with id 20b33f244badf3bb1a4911462ec5fbdc9bb3ac8d7799a29abb162639c9ce5f0c Nov 25 09:26:43 crc kubenswrapper[4776]: W1125 09:26:43.949740 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-e9cc8a6304800c8fa3323357d6c3000b728726eb4f0eb6ac8e1fb7de67472cfa WatchSource:0}: Error finding container e9cc8a6304800c8fa3323357d6c3000b728726eb4f0eb6ac8e1fb7de67472cfa: Status 404 returned error can't find the container with id e9cc8a6304800c8fa3323357d6c3000b728726eb4f0eb6ac8e1fb7de67472cfa Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.541557 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"d769119a0a2775247958efa35bbc209df8b7db5c551ddc1ab19141e8b0429239"} Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.541647 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"e9cc8a6304800c8fa3323357d6c3000b728726eb4f0eb6ac8e1fb7de67472cfa"} Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.543995 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"d61a3d49dbe4ea1b30a643d717fe7c79d75ba18ffaa1cafd206fe71bc358c6ce"} Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.544101 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"20b33f244badf3bb1a4911462ec5fbdc9bb3ac8d7799a29abb162639c9ce5f0c"} Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.546801 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"e4fc137eaecd5840a8614962905eb18411d41b9009dfe8e0b61596e836bf807f"} Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.546887 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"7c534c2a73741a85cfd1b65f414e5b6b27cbf90f7367f595488cc0b23ef4f36c"} Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.547191 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.715591 4776 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.760684 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xh8xm"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.761165 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xh8xm" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.762685 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-srdjj"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.763186 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.764504 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-fjr5v"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.764876 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-fjr5v" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.766438 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fq5rl"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.766771 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.768342 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.768810 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.769631 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.769686 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.769686 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.770301 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.770711 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.770984 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.771230 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.771585 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.771967 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.773123 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.773261 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.773385 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.773548 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.774144 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8rhfr"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.774223 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.774799 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.774991 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.775392 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.776217 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-sb2sq"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.777161 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-sb2sq" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.778317 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-mt26t"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.778765 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-mt26t" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.779707 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.780995 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-pxpfx"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.781934 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pxpfx" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.784640 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-r979m"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.805690 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-bhgv2"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.805806 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-r979m" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.832109 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.833564 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.833595 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.833818 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.834706 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.834814 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.834949 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vpp8l"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.835879 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.835957 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.836023 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.836297 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.836373 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.836486 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.836533 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.836576 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.836719 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-x9k2l"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.836754 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.836772 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.836842 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.836962 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.837219 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-tjf22"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.837388 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.837513 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.837567 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vpp8l" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.837589 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qqkgv"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.837538 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.837846 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x9k2l" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.837916 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-twg6n"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.838313 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.838496 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-qqkgv" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.838631 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dlv77"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.838787 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-twg6n" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.838967 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pgjrp"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.839168 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.839272 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.839338 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.839451 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7r6h5"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.839461 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.839494 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.839526 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dlv77" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.839580 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.839613 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.839629 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pgjrp" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.839723 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.839737 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.839831 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.839909 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.839948 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.839964 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7r6h5" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.839980 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.840027 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.840085 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.840112 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.840164 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.840189 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.840249 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.840277 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.840311 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.840326 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.840348 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.840380 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.840400 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.840421 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.839984 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.840469 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.840497 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.840577 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.840648 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.850457 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wrbxz"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.851005 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wrbxz" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.851735 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rbfl6"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.852342 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rbfl6" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.855540 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.855699 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.855999 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.856134 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.856475 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.856645 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.856773 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.857846 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.863372 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.863406 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.863497 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-wzwq4"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.863558 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.864285 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-wzwq4" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.866957 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.867169 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.868890 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.869045 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.870809 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.874987 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/fffe85b6-294d-4f43-ad27-65a5d093c076-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fjr5v\" (UID: \"fffe85b6-294d-4f43-ad27-65a5d093c076\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fjr5v" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.875019 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-oauth-serving-cert\") pod \"console-f9d7485db-bhgv2\" (UID: \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\") " pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.875121 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/be1426b0-cf02-44d3-87b8-aa0394af1afa-audit-dir\") pod \"apiserver-7bbb656c7d-hffrq\" (UID: \"be1426b0-cf02-44d3-87b8-aa0394af1afa\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.875234 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-audit-policies\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.875258 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2jsz\" (UniqueName: \"kubernetes.io/projected/490fba04-edd0-4edc-9f1e-4e08debf284f-kube-api-access-x2jsz\") pod \"controller-manager-879f6c89f-fq5rl\" (UID: \"490fba04-edd0-4edc-9f1e-4e08debf284f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.875377 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm7sd\" (UniqueName: \"kubernetes.io/projected/c7ba5522-7755-4e0a-96e8-6d818e130fa2-kube-api-access-lm7sd\") pod \"route-controller-manager-6576b87f9c-n6sd4\" (UID: \"c7ba5522-7755-4e0a-96e8-6d818e130fa2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.875398 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/be1426b0-cf02-44d3-87b8-aa0394af1afa-audit-policies\") pod \"apiserver-7bbb656c7d-hffrq\" (UID: \"be1426b0-cf02-44d3-87b8-aa0394af1afa\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.875457 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.875504 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nkvq\" (UniqueName: \"kubernetes.io/projected/b858f43c-8ede-490b-866e-56a8839c8d8b-kube-api-access-8nkvq\") pod \"console-operator-58897d9998-mt26t\" (UID: \"b858f43c-8ede-490b-866e-56a8839c8d8b\") " pod="openshift-console-operator/console-operator-58897d9998-mt26t" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.875522 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/10d2d0a3-eaba-486f-a833-0da30ad1f18e-machine-approver-tls\") pod \"machine-approver-56656f9798-pxpfx\" (UID: \"10d2d0a3-eaba-486f-a833-0da30ad1f18e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pxpfx" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.875646 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/974a7107-64ce-4f92-842b-7927d3ba006d-config\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.875670 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.875689 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/490fba04-edd0-4edc-9f1e-4e08debf284f-config\") pod \"controller-manager-879f6c89f-fq5rl\" (UID: \"490fba04-edd0-4edc-9f1e-4e08debf284f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.875707 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-console-oauth-config\") pod \"console-f9d7485db-bhgv2\" (UID: \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\") " pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.875724 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/974a7107-64ce-4f92-842b-7927d3ba006d-audit-dir\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.875742 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/be1426b0-cf02-44d3-87b8-aa0394af1afa-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hffrq\" (UID: \"be1426b0-cf02-44d3-87b8-aa0394af1afa\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.875760 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/10d2d0a3-eaba-486f-a833-0da30ad1f18e-auth-proxy-config\") pod \"machine-approver-56656f9798-pxpfx\" (UID: \"10d2d0a3-eaba-486f-a833-0da30ad1f18e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pxpfx" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.875778 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c7ba5522-7755-4e0a-96e8-6d818e130fa2-client-ca\") pod \"route-controller-manager-6576b87f9c-n6sd4\" (UID: \"c7ba5522-7755-4e0a-96e8-6d818e130fa2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.875910 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.868862 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.875936 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qk9t\" (UniqueName: \"kubernetes.io/projected/1d06dda7-d25a-4d6d-93b3-9a8579876bee-kube-api-access-9qk9t\") pod \"openshift-apiserver-operator-796bbdcf4f-xh8xm\" (UID: \"1d06dda7-d25a-4d6d-93b3-9a8579876bee\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xh8xm" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.879971 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b858f43c-8ede-490b-866e-56a8839c8d8b-config\") pod \"console-operator-58897d9998-mt26t\" (UID: \"b858f43c-8ede-490b-866e-56a8839c8d8b\") " pod="openshift-console-operator/console-operator-58897d9998-mt26t" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.880010 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.880041 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10d2d0a3-eaba-486f-a833-0da30ad1f18e-config\") pod \"machine-approver-56656f9798-pxpfx\" (UID: \"10d2d0a3-eaba-486f-a833-0da30ad1f18e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pxpfx" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.880086 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/974a7107-64ce-4f92-842b-7927d3ba006d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.880117 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be1426b0-cf02-44d3-87b8-aa0394af1afa-serving-cert\") pod \"apiserver-7bbb656c7d-hffrq\" (UID: \"be1426b0-cf02-44d3-87b8-aa0394af1afa\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.880146 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/be1426b0-cf02-44d3-87b8-aa0394af1afa-encryption-config\") pod \"apiserver-7bbb656c7d-hffrq\" (UID: \"be1426b0-cf02-44d3-87b8-aa0394af1afa\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.880183 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b858f43c-8ede-490b-866e-56a8839c8d8b-serving-cert\") pod \"console-operator-58897d9998-mt26t\" (UID: \"b858f43c-8ede-490b-866e-56a8839c8d8b\") " pod="openshift-console-operator/console-operator-58897d9998-mt26t" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.880214 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e67eb5b-2b9f-4553-861e-d687a07a1d12-serving-cert\") pod \"authentication-operator-69f744f599-sb2sq\" (UID: \"5e67eb5b-2b9f-4553-861e-d687a07a1d12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sb2sq" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.880501 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.880530 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.880558 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dj8fq\" (UniqueName: \"kubernetes.io/projected/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-kube-api-access-dj8fq\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.880582 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-trusted-ca-bundle\") pod \"console-f9d7485db-bhgv2\" (UID: \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\") " pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.880613 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvkbd\" (UniqueName: \"kubernetes.io/projected/5e67eb5b-2b9f-4553-861e-d687a07a1d12-kube-api-access-mvkbd\") pod \"authentication-operator-69f744f599-sb2sq\" (UID: \"5e67eb5b-2b9f-4553-861e-d687a07a1d12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sb2sq" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.880640 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d06dda7-d25a-4d6d-93b3-9a8579876bee-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xh8xm\" (UID: \"1d06dda7-d25a-4d6d-93b3-9a8579876bee\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xh8xm" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.880674 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/490fba04-edd0-4edc-9f1e-4e08debf284f-serving-cert\") pod \"controller-manager-879f6c89f-fq5rl\" (UID: \"490fba04-edd0-4edc-9f1e-4e08debf284f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.880701 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e67eb5b-2b9f-4553-861e-d687a07a1d12-service-ca-bundle\") pod \"authentication-operator-69f744f599-sb2sq\" (UID: \"5e67eb5b-2b9f-4553-861e-d687a07a1d12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sb2sq" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.880728 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7ba5522-7755-4e0a-96e8-6d818e130fa2-config\") pod \"route-controller-manager-6576b87f9c-n6sd4\" (UID: \"c7ba5522-7755-4e0a-96e8-6d818e130fa2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.880752 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/974a7107-64ce-4f92-842b-7927d3ba006d-etcd-serving-ca\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.880774 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/974a7107-64ce-4f92-842b-7927d3ba006d-serving-cert\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.880811 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.880842 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/490fba04-edd0-4edc-9f1e-4e08debf284f-client-ca\") pod \"controller-manager-879f6c89f-fq5rl\" (UID: \"490fba04-edd0-4edc-9f1e-4e08debf284f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.880874 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/974a7107-64ce-4f92-842b-7927d3ba006d-encryption-config\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.880903 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7ba5522-7755-4e0a-96e8-6d818e130fa2-serving-cert\") pod \"route-controller-manager-6576b87f9c-n6sd4\" (UID: \"c7ba5522-7755-4e0a-96e8-6d818e130fa2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.880935 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksp8p\" (UniqueName: \"kubernetes.io/projected/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-kube-api-access-ksp8p\") pod \"console-f9d7485db-bhgv2\" (UID: \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\") " pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.880967 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-console-config\") pod \"console-f9d7485db-bhgv2\" (UID: \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\") " pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.881140 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qp554\" (UniqueName: \"kubernetes.io/projected/10d2d0a3-eaba-486f-a833-0da30ad1f18e-kube-api-access-qp554\") pod \"machine-approver-56656f9798-pxpfx\" (UID: \"10d2d0a3-eaba-486f-a833-0da30ad1f18e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pxpfx" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.881175 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/974a7107-64ce-4f92-842b-7927d3ba006d-audit\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.882516 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.883359 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.884088 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.884618 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.884885 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.885397 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-rc22s"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.886149 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/974a7107-64ce-4f92-842b-7927d3ba006d-image-import-ca\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.886187 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/be1426b0-cf02-44d3-87b8-aa0394af1afa-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hffrq\" (UID: \"be1426b0-cf02-44d3-87b8-aa0394af1afa\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.886228 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-console-serving-cert\") pod \"console-f9d7485db-bhgv2\" (UID: \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\") " pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.886259 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.886298 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/be1426b0-cf02-44d3-87b8-aa0394af1afa-etcd-client\") pod \"apiserver-7bbb656c7d-hffrq\" (UID: \"be1426b0-cf02-44d3-87b8-aa0394af1afa\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.886822 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkcvv\" (UniqueName: \"kubernetes.io/projected/529b4e1d-e99c-4342-8cff-68c1e31a776f-kube-api-access-hkcvv\") pod \"downloads-7954f5f757-r979m\" (UID: \"529b4e1d-e99c-4342-8cff-68c1e31a776f\") " pod="openshift-console/downloads-7954f5f757-r979m" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.886855 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.886882 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b858f43c-8ede-490b-866e-56a8839c8d8b-trusted-ca\") pod \"console-operator-58897d9998-mt26t\" (UID: \"b858f43c-8ede-490b-866e-56a8839c8d8b\") " pod="openshift-console-operator/console-operator-58897d9998-mt26t" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.886908 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e67eb5b-2b9f-4553-861e-d687a07a1d12-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-sb2sq\" (UID: \"5e67eb5b-2b9f-4553-861e-d687a07a1d12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sb2sq" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.886990 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppr85\" (UniqueName: \"kubernetes.io/projected/fffe85b6-294d-4f43-ad27-65a5d093c076-kube-api-access-ppr85\") pod \"machine-api-operator-5694c8668f-fjr5v\" (UID: \"fffe85b6-294d-4f43-ad27-65a5d093c076\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fjr5v" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.887105 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csbqw\" (UniqueName: \"kubernetes.io/projected/be1426b0-cf02-44d3-87b8-aa0394af1afa-kube-api-access-csbqw\") pod \"apiserver-7bbb656c7d-hffrq\" (UID: \"be1426b0-cf02-44d3-87b8-aa0394af1afa\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.887355 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-service-ca\") pod \"console-f9d7485db-bhgv2\" (UID: \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\") " pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.887473 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.887521 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/fffe85b6-294d-4f43-ad27-65a5d093c076-images\") pod \"machine-api-operator-5694c8668f-fjr5v\" (UID: \"fffe85b6-294d-4f43-ad27-65a5d093c076\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fjr5v" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.887568 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/490fba04-edd0-4edc-9f1e-4e08debf284f-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-fq5rl\" (UID: \"490fba04-edd0-4edc-9f1e-4e08debf284f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.887599 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e67eb5b-2b9f-4553-861e-d687a07a1d12-config\") pod \"authentication-operator-69f744f599-sb2sq\" (UID: \"5e67eb5b-2b9f-4553-861e-d687a07a1d12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sb2sq" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.887626 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zmmg\" (UniqueName: \"kubernetes.io/projected/974a7107-64ce-4f92-842b-7927d3ba006d-kube-api-access-8zmmg\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.887650 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-audit-dir\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.887676 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.887719 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/974a7107-64ce-4f92-842b-7927d3ba006d-node-pullsecrets\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.897303 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.897493 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.899281 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.899779 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.900180 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.900389 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rc22s" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.904109 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.904508 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.904702 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.905389 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.906106 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-f4747"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.906702 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.906941 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f4747" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.908026 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/974a7107-64ce-4f92-842b-7927d3ba006d-etcd-client\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.908100 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fffe85b6-294d-4f43-ad27-65a5d093c076-config\") pod \"machine-api-operator-5694c8668f-fjr5v\" (UID: \"fffe85b6-294d-4f43-ad27-65a5d093c076\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fjr5v" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.908127 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d06dda7-d25a-4d6d-93b3-9a8579876bee-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xh8xm\" (UID: \"1d06dda7-d25a-4d6d-93b3-9a8579876bee\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xh8xm" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.914935 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.915772 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.916437 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-fjr5v"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.916660 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.916874 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.918335 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.918495 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.927285 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.932160 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xh8xm"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.935660 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.945239 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.945772 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-8nrlj"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.946549 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-8nrlj" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.947321 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwxqt"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.948025 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwxqt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.949298 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nzcz7"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.950739 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-m9xnc"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.951341 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pxtkj"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.951713 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-mv7lp"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.952172 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mv7lp" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.952328 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tbbm7"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.952453 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nzcz7" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.952649 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-m9xnc" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.952835 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pxtkj" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.953964 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tbbm7" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.955559 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h8t6d"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.955951 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wxwlt"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.956372 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wxwlt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.956514 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h8t6d" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.959595 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5n2sp"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.960229 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5n2sp" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.961130 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401035-dht2z"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.961883 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401035-dht2z" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.963022 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.963196 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-tckf4"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.963766 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-tckf4" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.964189 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-g4wgk"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.964706 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g4wgk" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.965623 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-sb2sq"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.967477 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.968911 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-7vqhc"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.970565 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-7vqhc" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.978131 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-x9k2l"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.978258 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-srdjj"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.981260 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-mt26t"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.983234 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.983757 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vpp8l"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.991918 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-tjf22"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.993612 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.995922 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qqkgv"] Nov 25 09:26:44 crc kubenswrapper[4776]: I1125 09:26:44.998397 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fq5rl"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.000745 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-r979m"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.001362 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-wzwq4"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.002864 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.003165 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-f4747"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.003431 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wrbxz"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.008628 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/490fba04-edd0-4edc-9f1e-4e08debf284f-serving-cert\") pod \"controller-manager-879f6c89f-fq5rl\" (UID: \"490fba04-edd0-4edc-9f1e-4e08debf284f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.008656 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e67eb5b-2b9f-4553-861e-d687a07a1d12-service-ca-bundle\") pod \"authentication-operator-69f744f599-sb2sq\" (UID: \"5e67eb5b-2b9f-4553-861e-d687a07a1d12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sb2sq" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.008675 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7ba5522-7755-4e0a-96e8-6d818e130fa2-config\") pod \"route-controller-manager-6576b87f9c-n6sd4\" (UID: \"c7ba5522-7755-4e0a-96e8-6d818e130fa2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.008690 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/974a7107-64ce-4f92-842b-7927d3ba006d-etcd-serving-ca\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.008705 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/974a7107-64ce-4f92-842b-7927d3ba006d-serving-cert\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.008727 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.008748 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/490fba04-edd0-4edc-9f1e-4e08debf284f-client-ca\") pod \"controller-manager-879f6c89f-fq5rl\" (UID: \"490fba04-edd0-4edc-9f1e-4e08debf284f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.008764 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/974a7107-64ce-4f92-842b-7927d3ba006d-encryption-config\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.008780 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7ba5522-7755-4e0a-96e8-6d818e130fa2-serving-cert\") pod \"route-controller-manager-6576b87f9c-n6sd4\" (UID: \"c7ba5522-7755-4e0a-96e8-6d818e130fa2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.008794 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksp8p\" (UniqueName: \"kubernetes.io/projected/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-kube-api-access-ksp8p\") pod \"console-f9d7485db-bhgv2\" (UID: \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\") " pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.008810 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-console-config\") pod \"console-f9d7485db-bhgv2\" (UID: \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\") " pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.008826 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qp554\" (UniqueName: \"kubernetes.io/projected/10d2d0a3-eaba-486f-a833-0da30ad1f18e-kube-api-access-qp554\") pod \"machine-approver-56656f9798-pxpfx\" (UID: \"10d2d0a3-eaba-486f-a833-0da30ad1f18e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pxpfx" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.008841 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/974a7107-64ce-4f92-842b-7927d3ba006d-audit\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.008855 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/974a7107-64ce-4f92-842b-7927d3ba006d-image-import-ca\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.008871 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/be1426b0-cf02-44d3-87b8-aa0394af1afa-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hffrq\" (UID: \"be1426b0-cf02-44d3-87b8-aa0394af1afa\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.008885 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-console-serving-cert\") pod \"console-f9d7485db-bhgv2\" (UID: \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\") " pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.008901 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.008916 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/be1426b0-cf02-44d3-87b8-aa0394af1afa-etcd-client\") pod \"apiserver-7bbb656c7d-hffrq\" (UID: \"be1426b0-cf02-44d3-87b8-aa0394af1afa\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.008931 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkcvv\" (UniqueName: \"kubernetes.io/projected/529b4e1d-e99c-4342-8cff-68c1e31a776f-kube-api-access-hkcvv\") pod \"downloads-7954f5f757-r979m\" (UID: \"529b4e1d-e99c-4342-8cff-68c1e31a776f\") " pod="openshift-console/downloads-7954f5f757-r979m" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.008948 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.008962 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b858f43c-8ede-490b-866e-56a8839c8d8b-trusted-ca\") pod \"console-operator-58897d9998-mt26t\" (UID: \"b858f43c-8ede-490b-866e-56a8839c8d8b\") " pod="openshift-console-operator/console-operator-58897d9998-mt26t" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.008977 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e67eb5b-2b9f-4553-861e-d687a07a1d12-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-sb2sq\" (UID: \"5e67eb5b-2b9f-4553-861e-d687a07a1d12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sb2sq" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.008994 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppr85\" (UniqueName: \"kubernetes.io/projected/fffe85b6-294d-4f43-ad27-65a5d093c076-kube-api-access-ppr85\") pod \"machine-api-operator-5694c8668f-fjr5v\" (UID: \"fffe85b6-294d-4f43-ad27-65a5d093c076\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fjr5v" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009017 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csbqw\" (UniqueName: \"kubernetes.io/projected/be1426b0-cf02-44d3-87b8-aa0394af1afa-kube-api-access-csbqw\") pod \"apiserver-7bbb656c7d-hffrq\" (UID: \"be1426b0-cf02-44d3-87b8-aa0394af1afa\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009033 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-service-ca\") pod \"console-f9d7485db-bhgv2\" (UID: \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\") " pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009054 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/50f49648-47d6-45d8-9563-f76b7547cad6-etcd-ca\") pod \"etcd-operator-b45778765-qqkgv\" (UID: \"50f49648-47d6-45d8-9563-f76b7547cad6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qqkgv" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009087 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009103 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/fffe85b6-294d-4f43-ad27-65a5d093c076-images\") pod \"machine-api-operator-5694c8668f-fjr5v\" (UID: \"fffe85b6-294d-4f43-ad27-65a5d093c076\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fjr5v" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009119 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gvxl\" (UniqueName: \"kubernetes.io/projected/50f49648-47d6-45d8-9563-f76b7547cad6-kube-api-access-8gvxl\") pod \"etcd-operator-b45778765-qqkgv\" (UID: \"50f49648-47d6-45d8-9563-f76b7547cad6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qqkgv" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009139 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/490fba04-edd0-4edc-9f1e-4e08debf284f-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-fq5rl\" (UID: \"490fba04-edd0-4edc-9f1e-4e08debf284f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009155 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e67eb5b-2b9f-4553-861e-d687a07a1d12-config\") pod \"authentication-operator-69f744f599-sb2sq\" (UID: \"5e67eb5b-2b9f-4553-861e-d687a07a1d12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sb2sq" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009172 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zmmg\" (UniqueName: \"kubernetes.io/projected/974a7107-64ce-4f92-842b-7927d3ba006d-kube-api-access-8zmmg\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009188 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-audit-dir\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009203 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009219 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/974a7107-64ce-4f92-842b-7927d3ba006d-node-pullsecrets\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009234 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/50f49648-47d6-45d8-9563-f76b7547cad6-etcd-client\") pod \"etcd-operator-b45778765-qqkgv\" (UID: \"50f49648-47d6-45d8-9563-f76b7547cad6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qqkgv" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009251 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/974a7107-64ce-4f92-842b-7927d3ba006d-etcd-client\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009267 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fffe85b6-294d-4f43-ad27-65a5d093c076-config\") pod \"machine-api-operator-5694c8668f-fjr5v\" (UID: \"fffe85b6-294d-4f43-ad27-65a5d093c076\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fjr5v" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009285 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d06dda7-d25a-4d6d-93b3-9a8579876bee-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xh8xm\" (UID: \"1d06dda7-d25a-4d6d-93b3-9a8579876bee\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xh8xm" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009302 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/fffe85b6-294d-4f43-ad27-65a5d093c076-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fjr5v\" (UID: \"fffe85b6-294d-4f43-ad27-65a5d093c076\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fjr5v" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009320 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-oauth-serving-cert\") pod \"console-f9d7485db-bhgv2\" (UID: \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\") " pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009336 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/be1426b0-cf02-44d3-87b8-aa0394af1afa-audit-dir\") pod \"apiserver-7bbb656c7d-hffrq\" (UID: \"be1426b0-cf02-44d3-87b8-aa0394af1afa\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009352 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50f49648-47d6-45d8-9563-f76b7547cad6-config\") pod \"etcd-operator-b45778765-qqkgv\" (UID: \"50f49648-47d6-45d8-9563-f76b7547cad6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qqkgv" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009369 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-audit-policies\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009385 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2jsz\" (UniqueName: \"kubernetes.io/projected/490fba04-edd0-4edc-9f1e-4e08debf284f-kube-api-access-x2jsz\") pod \"controller-manager-879f6c89f-fq5rl\" (UID: \"490fba04-edd0-4edc-9f1e-4e08debf284f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009401 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm7sd\" (UniqueName: \"kubernetes.io/projected/c7ba5522-7755-4e0a-96e8-6d818e130fa2-kube-api-access-lm7sd\") pod \"route-controller-manager-6576b87f9c-n6sd4\" (UID: \"c7ba5522-7755-4e0a-96e8-6d818e130fa2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009416 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/be1426b0-cf02-44d3-87b8-aa0394af1afa-audit-policies\") pod \"apiserver-7bbb656c7d-hffrq\" (UID: \"be1426b0-cf02-44d3-87b8-aa0394af1afa\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009431 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009446 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nkvq\" (UniqueName: \"kubernetes.io/projected/b858f43c-8ede-490b-866e-56a8839c8d8b-kube-api-access-8nkvq\") pod \"console-operator-58897d9998-mt26t\" (UID: \"b858f43c-8ede-490b-866e-56a8839c8d8b\") " pod="openshift-console-operator/console-operator-58897d9998-mt26t" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009461 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/10d2d0a3-eaba-486f-a833-0da30ad1f18e-machine-approver-tls\") pod \"machine-approver-56656f9798-pxpfx\" (UID: \"10d2d0a3-eaba-486f-a833-0da30ad1f18e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pxpfx" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009476 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/974a7107-64ce-4f92-842b-7927d3ba006d-config\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009491 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009506 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/490fba04-edd0-4edc-9f1e-4e08debf284f-config\") pod \"controller-manager-879f6c89f-fq5rl\" (UID: \"490fba04-edd0-4edc-9f1e-4e08debf284f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009521 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-console-oauth-config\") pod \"console-f9d7485db-bhgv2\" (UID: \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\") " pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009538 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/974a7107-64ce-4f92-842b-7927d3ba006d-audit-dir\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009552 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/be1426b0-cf02-44d3-87b8-aa0394af1afa-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hffrq\" (UID: \"be1426b0-cf02-44d3-87b8-aa0394af1afa\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009568 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/10d2d0a3-eaba-486f-a833-0da30ad1f18e-auth-proxy-config\") pod \"machine-approver-56656f9798-pxpfx\" (UID: \"10d2d0a3-eaba-486f-a833-0da30ad1f18e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pxpfx" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009594 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c7ba5522-7755-4e0a-96e8-6d818e130fa2-client-ca\") pod \"route-controller-manager-6576b87f9c-n6sd4\" (UID: \"c7ba5522-7755-4e0a-96e8-6d818e130fa2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009610 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009625 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qk9t\" (UniqueName: \"kubernetes.io/projected/1d06dda7-d25a-4d6d-93b3-9a8579876bee-kube-api-access-9qk9t\") pod \"openshift-apiserver-operator-796bbdcf4f-xh8xm\" (UID: \"1d06dda7-d25a-4d6d-93b3-9a8579876bee\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xh8xm" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009641 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b858f43c-8ede-490b-866e-56a8839c8d8b-config\") pod \"console-operator-58897d9998-mt26t\" (UID: \"b858f43c-8ede-490b-866e-56a8839c8d8b\") " pod="openshift-console-operator/console-operator-58897d9998-mt26t" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009656 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50f49648-47d6-45d8-9563-f76b7547cad6-serving-cert\") pod \"etcd-operator-b45778765-qqkgv\" (UID: \"50f49648-47d6-45d8-9563-f76b7547cad6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qqkgv" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009679 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009696 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10d2d0a3-eaba-486f-a833-0da30ad1f18e-config\") pod \"machine-approver-56656f9798-pxpfx\" (UID: \"10d2d0a3-eaba-486f-a833-0da30ad1f18e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pxpfx" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009710 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/974a7107-64ce-4f92-842b-7927d3ba006d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009727 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be1426b0-cf02-44d3-87b8-aa0394af1afa-serving-cert\") pod \"apiserver-7bbb656c7d-hffrq\" (UID: \"be1426b0-cf02-44d3-87b8-aa0394af1afa\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.009742 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/be1426b0-cf02-44d3-87b8-aa0394af1afa-encryption-config\") pod \"apiserver-7bbb656c7d-hffrq\" (UID: \"be1426b0-cf02-44d3-87b8-aa0394af1afa\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.008674 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-bhgv2"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.010438 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-rc22s"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.010462 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dlv77"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.010768 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b858f43c-8ede-490b-866e-56a8839c8d8b-serving-cert\") pod \"console-operator-58897d9998-mt26t\" (UID: \"b858f43c-8ede-490b-866e-56a8839c8d8b\") " pod="openshift-console-operator/console-operator-58897d9998-mt26t" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.010795 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e67eb5b-2b9f-4553-861e-d687a07a1d12-serving-cert\") pod \"authentication-operator-69f744f599-sb2sq\" (UID: \"5e67eb5b-2b9f-4553-861e-d687a07a1d12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sb2sq" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.010816 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.010838 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/50f49648-47d6-45d8-9563-f76b7547cad6-etcd-service-ca\") pod \"etcd-operator-b45778765-qqkgv\" (UID: \"50f49648-47d6-45d8-9563-f76b7547cad6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qqkgv" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.010839 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e67eb5b-2b9f-4553-861e-d687a07a1d12-service-ca-bundle\") pod \"authentication-operator-69f744f599-sb2sq\" (UID: \"5e67eb5b-2b9f-4553-861e-d687a07a1d12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sb2sq" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.010855 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.010883 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj8fq\" (UniqueName: \"kubernetes.io/projected/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-kube-api-access-dj8fq\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.010904 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-trusted-ca-bundle\") pod \"console-f9d7485db-bhgv2\" (UID: \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\") " pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.010922 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvkbd\" (UniqueName: \"kubernetes.io/projected/5e67eb5b-2b9f-4553-861e-d687a07a1d12-kube-api-access-mvkbd\") pod \"authentication-operator-69f744f599-sb2sq\" (UID: \"5e67eb5b-2b9f-4553-861e-d687a07a1d12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sb2sq" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.010939 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d06dda7-d25a-4d6d-93b3-9a8579876bee-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xh8xm\" (UID: \"1d06dda7-d25a-4d6d-93b3-9a8579876bee\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xh8xm" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.011317 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/974a7107-64ce-4f92-842b-7927d3ba006d-audit-dir\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.012433 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-console-config\") pod \"console-f9d7485db-bhgv2\" (UID: \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\") " pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.012967 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d06dda7-d25a-4d6d-93b3-9a8579876bee-config\") pod \"openshift-apiserver-operator-796bbdcf4f-xh8xm\" (UID: \"1d06dda7-d25a-4d6d-93b3-9a8579876bee\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xh8xm" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.014285 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8rhfr"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.014315 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pgjrp"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.014325 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7r6h5"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.014897 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/be1426b0-cf02-44d3-87b8-aa0394af1afa-audit-policies\") pod \"apiserver-7bbb656c7d-hffrq\" (UID: \"be1426b0-cf02-44d3-87b8-aa0394af1afa\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.015094 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/974a7107-64ce-4f92-842b-7927d3ba006d-audit\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.015552 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-oauth-serving-cert\") pod \"console-f9d7485db-bhgv2\" (UID: \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\") " pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.015565 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-trusted-ca-bundle\") pod \"console-f9d7485db-bhgv2\" (UID: \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\") " pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.015598 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/be1426b0-cf02-44d3-87b8-aa0394af1afa-audit-dir\") pod \"apiserver-7bbb656c7d-hffrq\" (UID: \"be1426b0-cf02-44d3-87b8-aa0394af1afa\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.015611 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/974a7107-64ce-4f92-842b-7927d3ba006d-config\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.016158 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.016174 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-audit-policies\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.016259 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/be1426b0-cf02-44d3-87b8-aa0394af1afa-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hffrq\" (UID: \"be1426b0-cf02-44d3-87b8-aa0394af1afa\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.016991 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/974a7107-64ce-4f92-842b-7927d3ba006d-serving-cert\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.017275 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.017331 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b858f43c-8ede-490b-866e-56a8839c8d8b-trusted-ca\") pod \"console-operator-58897d9998-mt26t\" (UID: \"b858f43c-8ede-490b-866e-56a8839c8d8b\") " pod="openshift-console-operator/console-operator-58897d9998-mt26t" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.018666 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-xj5z4"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.019039 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/974a7107-64ce-4f92-842b-7927d3ba006d-image-import-ca\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.019158 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/974a7107-64ce-4f92-842b-7927d3ba006d-encryption-config\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.019366 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-m9xnc"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.019390 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nzcz7"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.019479 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-xj5z4" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.019779 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/490fba04-edd0-4edc-9f1e-4e08debf284f-client-ca\") pod \"controller-manager-879f6c89f-fq5rl\" (UID: \"490fba04-edd0-4edc-9f1e-4e08debf284f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.020441 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d06dda7-d25a-4d6d-93b3-9a8579876bee-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-xh8xm\" (UID: \"1d06dda7-d25a-4d6d-93b3-9a8579876bee\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xh8xm" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.020598 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.020676 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/be1426b0-cf02-44d3-87b8-aa0394af1afa-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hffrq\" (UID: \"be1426b0-cf02-44d3-87b8-aa0394af1afa\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.020771 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/490fba04-edd0-4edc-9f1e-4e08debf284f-config\") pod \"controller-manager-879f6c89f-fq5rl\" (UID: \"490fba04-edd0-4edc-9f1e-4e08debf284f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.021308 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/fffe85b6-294d-4f43-ad27-65a5d093c076-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-fjr5v\" (UID: \"fffe85b6-294d-4f43-ad27-65a5d093c076\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fjr5v" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.021362 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/10d2d0a3-eaba-486f-a833-0da30ad1f18e-machine-approver-tls\") pod \"machine-approver-56656f9798-pxpfx\" (UID: \"10d2d0a3-eaba-486f-a833-0da30ad1f18e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pxpfx" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.021501 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b858f43c-8ede-490b-866e-56a8839c8d8b-config\") pod \"console-operator-58897d9998-mt26t\" (UID: \"b858f43c-8ede-490b-866e-56a8839c8d8b\") " pod="openshift-console-operator/console-operator-58897d9998-mt26t" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.021944 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/fffe85b6-294d-4f43-ad27-65a5d093c076-images\") pod \"machine-api-operator-5694c8668f-fjr5v\" (UID: \"fffe85b6-294d-4f43-ad27-65a5d093c076\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fjr5v" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.022001 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.022026 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-tckf4"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.022354 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-console-serving-cert\") pod \"console-f9d7485db-bhgv2\" (UID: \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\") " pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.022279 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.022458 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pxtkj"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.023025 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fffe85b6-294d-4f43-ad27-65a5d093c076-config\") pod \"machine-api-operator-5694c8668f-fjr5v\" (UID: \"fffe85b6-294d-4f43-ad27-65a5d093c076\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fjr5v" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.023116 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e67eb5b-2b9f-4553-861e-d687a07a1d12-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-sb2sq\" (UID: \"5e67eb5b-2b9f-4553-861e-d687a07a1d12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sb2sq" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.023487 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rbfl6"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.023580 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10d2d0a3-eaba-486f-a833-0da30ad1f18e-config\") pod \"machine-approver-56656f9798-pxpfx\" (UID: \"10d2d0a3-eaba-486f-a833-0da30ad1f18e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pxpfx" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.023710 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/be1426b0-cf02-44d3-87b8-aa0394af1afa-etcd-client\") pod \"apiserver-7bbb656c7d-hffrq\" (UID: \"be1426b0-cf02-44d3-87b8-aa0394af1afa\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.023954 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-service-ca\") pod \"console-f9d7485db-bhgv2\" (UID: \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\") " pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.024164 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.024333 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/974a7107-64ce-4f92-842b-7927d3ba006d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.024499 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.024689 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/974a7107-64ce-4f92-842b-7927d3ba006d-node-pullsecrets\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.024908 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.025322 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7ba5522-7755-4e0a-96e8-6d818e130fa2-config\") pod \"route-controller-manager-6576b87f9c-n6sd4\" (UID: \"c7ba5522-7755-4e0a-96e8-6d818e130fa2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.025430 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7ba5522-7755-4e0a-96e8-6d818e130fa2-serving-cert\") pod \"route-controller-manager-6576b87f9c-n6sd4\" (UID: \"c7ba5522-7755-4e0a-96e8-6d818e130fa2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.025483 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-console-oauth-config\") pod \"console-f9d7485db-bhgv2\" (UID: \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\") " pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.025542 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/974a7107-64ce-4f92-842b-7927d3ba006d-etcd-serving-ca\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.025579 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-audit-dir\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.025791 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-7vqhc"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.025870 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e67eb5b-2b9f-4553-861e-d687a07a1d12-config\") pod \"authentication-operator-69f744f599-sb2sq\" (UID: \"5e67eb5b-2b9f-4553-861e-d687a07a1d12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sb2sq" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.025882 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwxqt"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.026248 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/10d2d0a3-eaba-486f-a833-0da30ad1f18e-auth-proxy-config\") pod \"machine-approver-56656f9798-pxpfx\" (UID: \"10d2d0a3-eaba-486f-a833-0da30ad1f18e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pxpfx" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.026617 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.026735 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/490fba04-edd0-4edc-9f1e-4e08debf284f-serving-cert\") pod \"controller-manager-879f6c89f-fq5rl\" (UID: \"490fba04-edd0-4edc-9f1e-4e08debf284f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.027003 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401035-dht2z"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.027506 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c7ba5522-7755-4e0a-96e8-6d818e130fa2-client-ca\") pod \"route-controller-manager-6576b87f9c-n6sd4\" (UID: \"c7ba5522-7755-4e0a-96e8-6d818e130fa2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.027883 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.028959 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/be1426b0-cf02-44d3-87b8-aa0394af1afa-encryption-config\") pod \"apiserver-7bbb656c7d-hffrq\" (UID: \"be1426b0-cf02-44d3-87b8-aa0394af1afa\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.029113 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-twg6n"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.029461 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b858f43c-8ede-490b-866e-56a8839c8d8b-serving-cert\") pod \"console-operator-58897d9998-mt26t\" (UID: \"b858f43c-8ede-490b-866e-56a8839c8d8b\") " pod="openshift-console-operator/console-operator-58897d9998-mt26t" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.029770 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.029810 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wxwlt"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.030182 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/490fba04-edd0-4edc-9f1e-4e08debf284f-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-fq5rl\" (UID: \"490fba04-edd0-4edc-9f1e-4e08debf284f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.030756 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.030954 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be1426b0-cf02-44d3-87b8-aa0394af1afa-serving-cert\") pod \"apiserver-7bbb656c7d-hffrq\" (UID: \"be1426b0-cf02-44d3-87b8-aa0394af1afa\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.031000 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5n2sp"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.031800 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/974a7107-64ce-4f92-842b-7927d3ba006d-etcd-client\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.032021 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-mv7lp"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.033030 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tbbm7"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.034089 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h8t6d"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.035059 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-g4wgk"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.036201 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-twswh"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.037420 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-hkhrr"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.037784 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-hkhrr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.038159 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e67eb5b-2b9f-4553-861e-d687a07a1d12-serving-cert\") pod \"authentication-operator-69f744f599-sb2sq\" (UID: \"5e67eb5b-2b9f-4553-861e-d687a07a1d12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sb2sq" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.038167 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-hkhrr"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.038377 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-twswh" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.039151 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-twswh"] Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.043454 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.062916 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.083282 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.102915 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.111950 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/50f49648-47d6-45d8-9563-f76b7547cad6-etcd-ca\") pod \"etcd-operator-b45778765-qqkgv\" (UID: \"50f49648-47d6-45d8-9563-f76b7547cad6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qqkgv" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.112117 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gvxl\" (UniqueName: \"kubernetes.io/projected/50f49648-47d6-45d8-9563-f76b7547cad6-kube-api-access-8gvxl\") pod \"etcd-operator-b45778765-qqkgv\" (UID: \"50f49648-47d6-45d8-9563-f76b7547cad6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qqkgv" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.112234 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/50f49648-47d6-45d8-9563-f76b7547cad6-etcd-client\") pod \"etcd-operator-b45778765-qqkgv\" (UID: \"50f49648-47d6-45d8-9563-f76b7547cad6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qqkgv" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.112347 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50f49648-47d6-45d8-9563-f76b7547cad6-config\") pod \"etcd-operator-b45778765-qqkgv\" (UID: \"50f49648-47d6-45d8-9563-f76b7547cad6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qqkgv" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.112460 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50f49648-47d6-45d8-9563-f76b7547cad6-serving-cert\") pod \"etcd-operator-b45778765-qqkgv\" (UID: \"50f49648-47d6-45d8-9563-f76b7547cad6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qqkgv" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.112654 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/50f49648-47d6-45d8-9563-f76b7547cad6-etcd-service-ca\") pod \"etcd-operator-b45778765-qqkgv\" (UID: \"50f49648-47d6-45d8-9563-f76b7547cad6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qqkgv" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.113736 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50f49648-47d6-45d8-9563-f76b7547cad6-config\") pod \"etcd-operator-b45778765-qqkgv\" (UID: \"50f49648-47d6-45d8-9563-f76b7547cad6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qqkgv" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.113817 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/50f49648-47d6-45d8-9563-f76b7547cad6-etcd-ca\") pod \"etcd-operator-b45778765-qqkgv\" (UID: \"50f49648-47d6-45d8-9563-f76b7547cad6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qqkgv" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.114343 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/50f49648-47d6-45d8-9563-f76b7547cad6-etcd-service-ca\") pod \"etcd-operator-b45778765-qqkgv\" (UID: \"50f49648-47d6-45d8-9563-f76b7547cad6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qqkgv" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.117000 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/50f49648-47d6-45d8-9563-f76b7547cad6-etcd-client\") pod \"etcd-operator-b45778765-qqkgv\" (UID: \"50f49648-47d6-45d8-9563-f76b7547cad6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qqkgv" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.118352 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50f49648-47d6-45d8-9563-f76b7547cad6-serving-cert\") pod \"etcd-operator-b45778765-qqkgv\" (UID: \"50f49648-47d6-45d8-9563-f76b7547cad6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qqkgv" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.124580 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.145302 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.163554 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.183412 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.203202 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.223098 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.243153 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.263588 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.283245 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.302969 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.323837 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.342977 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.363357 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.384247 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.403910 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.443887 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.463823 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.503426 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.506034 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.524166 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.544665 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.563836 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.582890 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.622618 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.643290 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.662932 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.682994 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.703629 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.722844 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.743749 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.763032 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.783152 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.805484 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.823510 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.843190 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.864004 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.884127 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.903257 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.923007 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.944173 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.962301 4776 request.go:700] Waited for 1.009155615s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca-operator/configmaps?fieldSelector=metadata.name%3Dservice-ca-operator-config&limit=500&resourceVersion=0 Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.964254 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 25 09:26:45 crc kubenswrapper[4776]: I1125 09:26:45.983946 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.004037 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.023665 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.043688 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.063758 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.082549 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.104209 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.124420 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.143994 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.163386 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.183852 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.203922 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.223954 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.243795 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.264219 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.284398 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.318891 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.323385 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.343022 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.363334 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.384813 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.403871 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.423639 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.443402 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.463267 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.483862 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.512734 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.523471 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.545049 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.563843 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.583555 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.603696 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.641081 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qp554\" (UniqueName: \"kubernetes.io/projected/10d2d0a3-eaba-486f-a833-0da30ad1f18e-kube-api-access-qp554\") pod \"machine-approver-56656f9798-pxpfx\" (UID: \"10d2d0a3-eaba-486f-a833-0da30ad1f18e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pxpfx" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.663893 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dj8fq\" (UniqueName: \"kubernetes.io/projected/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-kube-api-access-dj8fq\") pod \"oauth-openshift-558db77b4-8rhfr\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.686729 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nkvq\" (UniqueName: \"kubernetes.io/projected/b858f43c-8ede-490b-866e-56a8839c8d8b-kube-api-access-8nkvq\") pod \"console-operator-58897d9998-mt26t\" (UID: \"b858f43c-8ede-490b-866e-56a8839c8d8b\") " pod="openshift-console-operator/console-operator-58897d9998-mt26t" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.708989 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2jsz\" (UniqueName: \"kubernetes.io/projected/490fba04-edd0-4edc-9f1e-4e08debf284f-kube-api-access-x2jsz\") pod \"controller-manager-879f6c89f-fq5rl\" (UID: \"490fba04-edd0-4edc-9f1e-4e08debf284f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.721034 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm7sd\" (UniqueName: \"kubernetes.io/projected/c7ba5522-7755-4e0a-96e8-6d818e130fa2-kube-api-access-lm7sd\") pod \"route-controller-manager-6576b87f9c-n6sd4\" (UID: \"c7ba5522-7755-4e0a-96e8-6d818e130fa2\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.723996 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.737140 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.744340 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.747830 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.764201 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.788446 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-mt26t" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.796652 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvkbd\" (UniqueName: \"kubernetes.io/projected/5e67eb5b-2b9f-4553-861e-d687a07a1d12-kube-api-access-mvkbd\") pod \"authentication-operator-69f744f599-sb2sq\" (UID: \"5e67eb5b-2b9f-4553-861e-d687a07a1d12\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-sb2sq" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.808318 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pxpfx" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.820179 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkcvv\" (UniqueName: \"kubernetes.io/projected/529b4e1d-e99c-4342-8cff-68c1e31a776f-kube-api-access-hkcvv\") pod \"downloads-7954f5f757-r979m\" (UID: \"529b4e1d-e99c-4342-8cff-68c1e31a776f\") " pod="openshift-console/downloads-7954f5f757-r979m" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.853305 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppr85\" (UniqueName: \"kubernetes.io/projected/fffe85b6-294d-4f43-ad27-65a5d093c076-kube-api-access-ppr85\") pod \"machine-api-operator-5694c8668f-fjr5v\" (UID: \"fffe85b6-294d-4f43-ad27-65a5d093c076\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-fjr5v" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.858038 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-r979m" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.862103 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csbqw\" (UniqueName: \"kubernetes.io/projected/be1426b0-cf02-44d3-87b8-aa0394af1afa-kube-api-access-csbqw\") pod \"apiserver-7bbb656c7d-hffrq\" (UID: \"be1426b0-cf02-44d3-87b8-aa0394af1afa\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.881700 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qk9t\" (UniqueName: \"kubernetes.io/projected/1d06dda7-d25a-4d6d-93b3-9a8579876bee-kube-api-access-9qk9t\") pod \"openshift-apiserver-operator-796bbdcf4f-xh8xm\" (UID: \"1d06dda7-d25a-4d6d-93b3-9a8579876bee\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xh8xm" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.897634 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zmmg\" (UniqueName: \"kubernetes.io/projected/974a7107-64ce-4f92-842b-7927d3ba006d-kube-api-access-8zmmg\") pod \"apiserver-76f77b778f-srdjj\" (UID: \"974a7107-64ce-4f92-842b-7927d3ba006d\") " pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.922665 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksp8p\" (UniqueName: \"kubernetes.io/projected/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-kube-api-access-ksp8p\") pod \"console-f9d7485db-bhgv2\" (UID: \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\") " pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.924946 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.945150 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.945836 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8rhfr"] Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.952095 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xh8xm" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.952625 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.963183 4776 request.go:700] Waited for 1.924865012s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-canary/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.964860 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.966052 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-fjr5v" Nov 25 09:26:46 crc kubenswrapper[4776]: W1125 09:26:46.971034 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5ae11bc_db7c_4665_88e4_ef3b0eb83581.slice/crio-c7be2d6b7d9f8dacd74ac864092f2282a5ad82be9eda5a52b5c5b3c23e02f46a WatchSource:0}: Error finding container c7be2d6b7d9f8dacd74ac864092f2282a5ad82be9eda5a52b5c5b3c23e02f46a: Status 404 returned error can't find the container with id c7be2d6b7d9f8dacd74ac864092f2282a5ad82be9eda5a52b5c5b3c23e02f46a Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.976659 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4"] Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.983872 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 25 09:26:46 crc kubenswrapper[4776]: I1125 09:26:46.985421 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" Nov 25 09:26:46 crc kubenswrapper[4776]: W1125 09:26:46.994966 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7ba5522_7755_4e0a_96e8_6d818e130fa2.slice/crio-ac4a185653aa1a8f1db2726538a9079d4f34d1b3503b632a3403bee2e488996e WatchSource:0}: Error finding container ac4a185653aa1a8f1db2726538a9079d4f34d1b3503b632a3403bee2e488996e: Status 404 returned error can't find the container with id ac4a185653aa1a8f1db2726538a9079d4f34d1b3503b632a3403bee2e488996e Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.003919 4776 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.013616 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-mt26t"] Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.020399 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.023996 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.043884 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.064153 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-sb2sq" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.069057 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-r979m"] Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.083005 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gvxl\" (UniqueName: \"kubernetes.io/projected/50f49648-47d6-45d8-9563-f76b7547cad6-kube-api-access-8gvxl\") pod \"etcd-operator-b45778765-qqkgv\" (UID: \"50f49648-47d6-45d8-9563-f76b7547cad6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-qqkgv" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.138495 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8m9tr\" (UniqueName: \"kubernetes.io/projected/e88ac546-8758-4cce-ae60-e49a1dc7c5ea-kube-api-access-8m9tr\") pod \"dns-operator-744455d44c-twg6n\" (UID: \"e88ac546-8758-4cce-ae60-e49a1dc7c5ea\") " pod="openshift-dns-operator/dns-operator-744455d44c-twg6n" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.138534 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/471ebf6d-cd02-461d-96c9-d2b503d65cb8-config\") pod \"kube-apiserver-operator-766d6c64bb-wrbxz\" (UID: \"471ebf6d-cd02-461d-96c9-d2b503d65cb8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wrbxz" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.138554 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/52e1736f-df9e-4444-bb28-4d89911000de-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-pgjrp\" (UID: \"52e1736f-df9e-4444-bb28-4d89911000de\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pgjrp" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.138572 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d2aa408d-c0e1-4697-b9e8-dee9276e8b47-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-7r6h5\" (UID: \"d2aa408d-c0e1-4697-b9e8-dee9276e8b47\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7r6h5" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.138591 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2441a0f3-b420-4a64-803b-81082cc15874-trusted-ca\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.138619 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8003bc91-3e3b-4e88-8942-3ec54fc30aef-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-rc22s\" (UID: \"8003bc91-3e3b-4e88-8942-3ec54fc30aef\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rc22s" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.138657 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e308a810-6256-4459-a587-d93ed4899fa8-images\") pod \"machine-config-operator-74547568cd-f4747\" (UID: \"e308a810-6256-4459-a587-d93ed4899fa8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f4747" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.138673 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/471ebf6d-cd02-461d-96c9-d2b503d65cb8-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-wrbxz\" (UID: \"471ebf6d-cd02-461d-96c9-d2b503d65cb8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wrbxz" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.138689 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1411e81d-100d-44a3-957f-5cc5a89e7748-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-dlv77\" (UID: \"1411e81d-100d-44a3-957f-5cc5a89e7748\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dlv77" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.138750 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1411e81d-100d-44a3-957f-5cc5a89e7748-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-dlv77\" (UID: \"1411e81d-100d-44a3-957f-5cc5a89e7748\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dlv77" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.138781 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwv8h\" (UniqueName: \"kubernetes.io/projected/2441a0f3-b420-4a64-803b-81082cc15874-kube-api-access-mwv8h\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.138798 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/049123ea-7dfd-43a3-b65c-9fb0c7b6c40b-available-featuregates\") pod \"openshift-config-operator-7777fb866f-x9k2l\" (UID: \"049123ea-7dfd-43a3-b65c-9fb0c7b6c40b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x9k2l" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.138821 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e1c3eef-d1b8-40c8-bbad-6e92ea15732e-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-vpp8l\" (UID: \"2e1c3eef-d1b8-40c8-bbad-6e92ea15732e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vpp8l" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.138847 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qvvq\" (UniqueName: \"kubernetes.io/projected/049123ea-7dfd-43a3-b65c-9fb0c7b6c40b-kube-api-access-2qvvq\") pod \"openshift-config-operator-7777fb866f-x9k2l\" (UID: \"049123ea-7dfd-43a3-b65c-9fb0c7b6c40b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x9k2l" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.138862 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8003bc91-3e3b-4e88-8942-3ec54fc30aef-proxy-tls\") pod \"machine-config-controller-84d6567774-rc22s\" (UID: \"8003bc91-3e3b-4e88-8942-3ec54fc30aef\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rc22s" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.138895 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2441a0f3-b420-4a64-803b-81082cc15874-installation-pull-secrets\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.138929 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9qz5\" (UniqueName: \"kubernetes.io/projected/1411e81d-100d-44a3-957f-5cc5a89e7748-kube-api-access-p9qz5\") pod \"cluster-image-registry-operator-dc59b4c8b-dlv77\" (UID: \"1411e81d-100d-44a3-957f-5cc5a89e7748\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dlv77" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.139000 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2441a0f3-b420-4a64-803b-81082cc15874-bound-sa-token\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.139028 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n68rq\" (UniqueName: \"kubernetes.io/projected/d2aa408d-c0e1-4697-b9e8-dee9276e8b47-kube-api-access-n68rq\") pod \"kube-storage-version-migrator-operator-b67b599dd-7r6h5\" (UID: \"d2aa408d-c0e1-4697-b9e8-dee9276e8b47\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7r6h5" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.139096 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w28t\" (UniqueName: \"kubernetes.io/projected/2e1c3eef-d1b8-40c8-bbad-6e92ea15732e-kube-api-access-4w28t\") pod \"openshift-controller-manager-operator-756b6f6bc6-vpp8l\" (UID: \"2e1c3eef-d1b8-40c8-bbad-6e92ea15732e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vpp8l" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.139111 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fggj\" (UniqueName: \"kubernetes.io/projected/52e1736f-df9e-4444-bb28-4d89911000de-kube-api-access-9fggj\") pod \"cluster-samples-operator-665b6dd947-pgjrp\" (UID: \"52e1736f-df9e-4444-bb28-4d89911000de\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pgjrp" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.139127 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/1411e81d-100d-44a3-957f-5cc5a89e7748-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-dlv77\" (UID: \"1411e81d-100d-44a3-957f-5cc5a89e7748\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dlv77" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.139143 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/049123ea-7dfd-43a3-b65c-9fb0c7b6c40b-serving-cert\") pod \"openshift-config-operator-7777fb866f-x9k2l\" (UID: \"049123ea-7dfd-43a3-b65c-9fb0c7b6c40b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x9k2l" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.139158 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2aa408d-c0e1-4697-b9e8-dee9276e8b47-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-7r6h5\" (UID: \"d2aa408d-c0e1-4697-b9e8-dee9276e8b47\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7r6h5" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.139189 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.139207 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv9q2\" (UniqueName: \"kubernetes.io/projected/a229c7a5-8a0d-451b-9b28-da9c8762044a-kube-api-access-hv9q2\") pod \"control-plane-machine-set-operator-78cbb6b69f-rbfl6\" (UID: \"a229c7a5-8a0d-451b-9b28-da9c8762044a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rbfl6" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.139223 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jlrd\" (UniqueName: \"kubernetes.io/projected/8003bc91-3e3b-4e88-8942-3ec54fc30aef-kube-api-access-2jlrd\") pod \"machine-config-controller-84d6567774-rc22s\" (UID: \"8003bc91-3e3b-4e88-8942-3ec54fc30aef\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rc22s" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.139240 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a229c7a5-8a0d-451b-9b28-da9c8762044a-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-rbfl6\" (UID: \"a229c7a5-8a0d-451b-9b28-da9c8762044a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rbfl6" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.139255 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e88ac546-8758-4cce-ae60-e49a1dc7c5ea-metrics-tls\") pod \"dns-operator-744455d44c-twg6n\" (UID: \"e88ac546-8758-4cce-ae60-e49a1dc7c5ea\") " pod="openshift-dns-operator/dns-operator-744455d44c-twg6n" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.139268 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e308a810-6256-4459-a587-d93ed4899fa8-proxy-tls\") pod \"machine-config-operator-74547568cd-f4747\" (UID: \"e308a810-6256-4459-a587-d93ed4899fa8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f4747" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.139283 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2e1c3eef-d1b8-40c8-bbad-6e92ea15732e-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-vpp8l\" (UID: \"2e1c3eef-d1b8-40c8-bbad-6e92ea15732e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vpp8l" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.139298 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qs4bf\" (UniqueName: \"kubernetes.io/projected/b80070a8-0eaa-4fe8-92ca-7c9afe029e04-kube-api-access-qs4bf\") pod \"migrator-59844c95c7-wzwq4\" (UID: \"b80070a8-0eaa-4fe8-92ca-7c9afe029e04\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-wzwq4" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.139324 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmv87\" (UniqueName: \"kubernetes.io/projected/e308a810-6256-4459-a587-d93ed4899fa8-kube-api-access-bmv87\") pod \"machine-config-operator-74547568cd-f4747\" (UID: \"e308a810-6256-4459-a587-d93ed4899fa8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f4747" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.139367 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e308a810-6256-4459-a587-d93ed4899fa8-auth-proxy-config\") pod \"machine-config-operator-74547568cd-f4747\" (UID: \"e308a810-6256-4459-a587-d93ed4899fa8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f4747" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.139388 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2441a0f3-b420-4a64-803b-81082cc15874-registry-tls\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.139402 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/471ebf6d-cd02-461d-96c9-d2b503d65cb8-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-wrbxz\" (UID: \"471ebf6d-cd02-461d-96c9-d2b503d65cb8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wrbxz" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.139435 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2441a0f3-b420-4a64-803b-81082cc15874-ca-trust-extracted\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.139464 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2441a0f3-b420-4a64-803b-81082cc15874-registry-certificates\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:47 crc kubenswrapper[4776]: E1125 09:26:47.152983 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:47.652966147 +0000 UTC m=+152.694025700 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.174562 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.197466 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-qqkgv" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.241527 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.241815 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/52e1736f-df9e-4444-bb28-4d89911000de-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-pgjrp\" (UID: \"52e1736f-df9e-4444-bb28-4d89911000de\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pgjrp" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.241839 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8m9tr\" (UniqueName: \"kubernetes.io/projected/e88ac546-8758-4cce-ae60-e49a1dc7c5ea-kube-api-access-8m9tr\") pod \"dns-operator-744455d44c-twg6n\" (UID: \"e88ac546-8758-4cce-ae60-e49a1dc7c5ea\") " pod="openshift-dns-operator/dns-operator-744455d44c-twg6n" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242004 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2441a0f3-b420-4a64-803b-81082cc15874-trusted-ca\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242024 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/df8b1df7-103a-49ba-855e-face92da97dd-node-bootstrap-token\") pod \"machine-config-server-xj5z4\" (UID: \"df8b1df7-103a-49ba-855e-face92da97dd\") " pod="openshift-machine-config-operator/machine-config-server-xj5z4" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242108 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8003bc91-3e3b-4e88-8942-3ec54fc30aef-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-rc22s\" (UID: \"8003bc91-3e3b-4e88-8942-3ec54fc30aef\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rc22s" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242128 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e308a810-6256-4459-a587-d93ed4899fa8-images\") pod \"machine-config-operator-74547568cd-f4747\" (UID: \"e308a810-6256-4459-a587-d93ed4899fa8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f4747" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242143 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n78j9\" (UniqueName: \"kubernetes.io/projected/8ba97f9b-a0a1-4514-8577-0b36eb65bbd3-kube-api-access-n78j9\") pod \"multus-admission-controller-857f4d67dd-m9xnc\" (UID: \"8ba97f9b-a0a1-4514-8577-0b36eb65bbd3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-m9xnc" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242160 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a01b028-a86d-40d3-9633-3f849f6d0840-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-h8t6d\" (UID: \"5a01b028-a86d-40d3-9633-3f849f6d0840\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h8t6d" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242175 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/337ebdb5-efc9-4bcb-8630-ab09a36c5bec-signing-key\") pod \"service-ca-9c57cc56f-tckf4\" (UID: \"337ebdb5-efc9-4bcb-8630-ab09a36c5bec\") " pod="openshift-service-ca/service-ca-9c57cc56f-tckf4" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242201 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/471ebf6d-cd02-461d-96c9-d2b503d65cb8-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-wrbxz\" (UID: \"471ebf6d-cd02-461d-96c9-d2b503d65cb8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wrbxz" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242220 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb6w2\" (UniqueName: \"kubernetes.io/projected/ae2cfb63-dc76-40c2-9070-d5c29a6a272d-kube-api-access-sb6w2\") pod \"csi-hostpathplugin-twswh\" (UID: \"ae2cfb63-dc76-40c2-9070-d5c29a6a272d\") " pod="hostpath-provisioner/csi-hostpathplugin-twswh" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242234 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4738a95e-6584-4c14-b545-d0e96d678ab0-trusted-ca\") pod \"ingress-operator-5b745b69d9-g4wgk\" (UID: \"4738a95e-6584-4c14-b545-d0e96d678ab0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g4wgk" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242256 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwv8h\" (UniqueName: \"kubernetes.io/projected/2441a0f3-b420-4a64-803b-81082cc15874-kube-api-access-mwv8h\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242269 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/09269fbc-f59c-42cc-a832-2285987526db-cert\") pod \"ingress-canary-hkhrr\" (UID: \"09269fbc-f59c-42cc-a832-2285987526db\") " pod="openshift-ingress-canary/ingress-canary-hkhrr" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242285 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bb9001d-7d85-4cad-8587-e47e6f1c7aa9-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-tbbm7\" (UID: \"4bb9001d-7d85-4cad-8587-e47e6f1c7aa9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tbbm7" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242302 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5a01b028-a86d-40d3-9633-3f849f6d0840-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-h8t6d\" (UID: \"5a01b028-a86d-40d3-9633-3f849f6d0840\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h8t6d" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242317 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e8765bc6-c8e3-40d7-a061-211b1cad5df2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5n2sp\" (UID: \"e8765bc6-c8e3-40d7-a061-211b1cad5df2\") " pod="openshift-marketplace/marketplace-operator-79b997595-5n2sp" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242333 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qvvq\" (UniqueName: \"kubernetes.io/projected/049123ea-7dfd-43a3-b65c-9fb0c7b6c40b-kube-api-access-2qvvq\") pod \"openshift-config-operator-7777fb866f-x9k2l\" (UID: \"049123ea-7dfd-43a3-b65c-9fb0c7b6c40b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x9k2l" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242348 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z627v\" (UniqueName: \"kubernetes.io/projected/df8b1df7-103a-49ba-855e-face92da97dd-kube-api-access-z627v\") pod \"machine-config-server-xj5z4\" (UID: \"df8b1df7-103a-49ba-855e-face92da97dd\") " pod="openshift-machine-config-operator/machine-config-server-xj5z4" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242365 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2441a0f3-b420-4a64-803b-81082cc15874-installation-pull-secrets\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242379 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8003bc91-3e3b-4e88-8942-3ec54fc30aef-proxy-tls\") pod \"machine-config-controller-84d6567774-rc22s\" (UID: \"8003bc91-3e3b-4e88-8942-3ec54fc30aef\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rc22s" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242398 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e932b24b-a48e-47c4-bb74-779fd2ec8ea9-config\") pod \"service-ca-operator-777779d784-mv7lp\" (UID: \"e932b24b-a48e-47c4-bb74-779fd2ec8ea9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mv7lp" Nov 25 09:26:47 crc kubenswrapper[4776]: E1125 09:26:47.242433 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:47.742407956 +0000 UTC m=+152.783467519 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242476 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ae2cfb63-dc76-40c2-9070-d5c29a6a272d-plugins-dir\") pod \"csi-hostpathplugin-twswh\" (UID: \"ae2cfb63-dc76-40c2-9070-d5c29a6a272d\") " pod="hostpath-provisioner/csi-hostpathplugin-twswh" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242511 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9qz5\" (UniqueName: \"kubernetes.io/projected/1411e81d-100d-44a3-957f-5cc5a89e7748-kube-api-access-p9qz5\") pod \"cluster-image-registry-operator-dc59b4c8b-dlv77\" (UID: \"1411e81d-100d-44a3-957f-5cc5a89e7748\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dlv77" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242536 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a9a85890-83ba-4e99-bdbe-386e0bbbecd6-webhook-cert\") pod \"packageserver-d55dfcdfc-pxtkj\" (UID: \"a9a85890-83ba-4e99-bdbe-386e0bbbecd6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pxtkj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242568 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjvk6\" (UniqueName: \"kubernetes.io/projected/a9a85890-83ba-4e99-bdbe-386e0bbbecd6-kube-api-access-bjvk6\") pod \"packageserver-d55dfcdfc-pxtkj\" (UID: \"a9a85890-83ba-4e99-bdbe-386e0bbbecd6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pxtkj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242597 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e932b24b-a48e-47c4-bb74-779fd2ec8ea9-serving-cert\") pod \"service-ca-operator-777779d784-mv7lp\" (UID: \"e932b24b-a48e-47c4-bb74-779fd2ec8ea9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mv7lp" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242630 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4738a95e-6584-4c14-b545-d0e96d678ab0-bound-sa-token\") pod \"ingress-operator-5b745b69d9-g4wgk\" (UID: \"4738a95e-6584-4c14-b545-d0e96d678ab0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g4wgk" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242655 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e8765bc6-c8e3-40d7-a061-211b1cad5df2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5n2sp\" (UID: \"e8765bc6-c8e3-40d7-a061-211b1cad5df2\") " pod="openshift-marketplace/marketplace-operator-79b997595-5n2sp" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242687 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n68rq\" (UniqueName: \"kubernetes.io/projected/d2aa408d-c0e1-4697-b9e8-dee9276e8b47-kube-api-access-n68rq\") pod \"kube-storage-version-migrator-operator-b67b599dd-7r6h5\" (UID: \"d2aa408d-c0e1-4697-b9e8-dee9276e8b47\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7r6h5" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242708 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/02eb8668-3740-4714-98ae-e993184549aa-default-certificate\") pod \"router-default-5444994796-8nrlj\" (UID: \"02eb8668-3740-4714-98ae-e993184549aa\") " pod="openshift-ingress/router-default-5444994796-8nrlj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242731 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8ba97f9b-a0a1-4514-8577-0b36eb65bbd3-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-m9xnc\" (UID: \"8ba97f9b-a0a1-4514-8577-0b36eb65bbd3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-m9xnc" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242770 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w28t\" (UniqueName: \"kubernetes.io/projected/2e1c3eef-d1b8-40c8-bbad-6e92ea15732e-kube-api-access-4w28t\") pod \"openshift-controller-manager-operator-756b6f6bc6-vpp8l\" (UID: \"2e1c3eef-d1b8-40c8-bbad-6e92ea15732e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vpp8l" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242791 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fggj\" (UniqueName: \"kubernetes.io/projected/52e1736f-df9e-4444-bb28-4d89911000de-kube-api-access-9fggj\") pod \"cluster-samples-operator-665b6dd947-pgjrp\" (UID: \"52e1736f-df9e-4444-bb28-4d89911000de\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pgjrp" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242815 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ae2cfb63-dc76-40c2-9070-d5c29a6a272d-csi-data-dir\") pod \"csi-hostpathplugin-twswh\" (UID: \"ae2cfb63-dc76-40c2-9070-d5c29a6a272d\") " pod="hostpath-provisioner/csi-hostpathplugin-twswh" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242835 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a9a85890-83ba-4e99-bdbe-386e0bbbecd6-apiservice-cert\") pod \"packageserver-d55dfcdfc-pxtkj\" (UID: \"a9a85890-83ba-4e99-bdbe-386e0bbbecd6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pxtkj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242859 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/049123ea-7dfd-43a3-b65c-9fb0c7b6c40b-serving-cert\") pod \"openshift-config-operator-7777fb866f-x9k2l\" (UID: \"049123ea-7dfd-43a3-b65c-9fb0c7b6c40b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x9k2l" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242887 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv9q2\" (UniqueName: \"kubernetes.io/projected/a229c7a5-8a0d-451b-9b28-da9c8762044a-kube-api-access-hv9q2\") pod \"control-plane-machine-set-operator-78cbb6b69f-rbfl6\" (UID: \"a229c7a5-8a0d-451b-9b28-da9c8762044a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rbfl6" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242912 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kx5ld\" (UniqueName: \"kubernetes.io/projected/8b743ab8-67de-493a-bdf2-c0682a192869-kube-api-access-kx5ld\") pod \"package-server-manager-789f6589d5-hwxqt\" (UID: \"8b743ab8-67de-493a-bdf2-c0682a192869\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwxqt" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242937 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jlrd\" (UniqueName: \"kubernetes.io/projected/8003bc91-3e3b-4e88-8942-3ec54fc30aef-kube-api-access-2jlrd\") pod \"machine-config-controller-84d6567774-rc22s\" (UID: \"8003bc91-3e3b-4e88-8942-3ec54fc30aef\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rc22s" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242958 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e88ac546-8758-4cce-ae60-e49a1dc7c5ea-metrics-tls\") pod \"dns-operator-744455d44c-twg6n\" (UID: \"e88ac546-8758-4cce-ae60-e49a1dc7c5ea\") " pod="openshift-dns-operator/dns-operator-744455d44c-twg6n" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.242981 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5slr6\" (UniqueName: \"kubernetes.io/projected/e932b24b-a48e-47c4-bb74-779fd2ec8ea9-kube-api-access-5slr6\") pod \"service-ca-operator-777779d784-mv7lp\" (UID: \"e932b24b-a48e-47c4-bb74-779fd2ec8ea9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mv7lp" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.245347 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e308a810-6256-4459-a587-d93ed4899fa8-images\") pod \"machine-config-operator-74547568cd-f4747\" (UID: \"e308a810-6256-4459-a587-d93ed4899fa8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f4747" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.248434 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2441a0f3-b420-4a64-803b-81082cc15874-trusted-ca\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.248509 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmv87\" (UniqueName: \"kubernetes.io/projected/e308a810-6256-4459-a587-d93ed4899fa8-kube-api-access-bmv87\") pod \"machine-config-operator-74547568cd-f4747\" (UID: \"e308a810-6256-4459-a587-d93ed4899fa8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f4747" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.248526 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/049123ea-7dfd-43a3-b65c-9fb0c7b6c40b-serving-cert\") pod \"openshift-config-operator-7777fb866f-x9k2l\" (UID: \"049123ea-7dfd-43a3-b65c-9fb0c7b6c40b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x9k2l" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.248553 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2441a0f3-b420-4a64-803b-81082cc15874-installation-pull-secrets\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.248563 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/025d75af-9ed0-4122-b97e-82dd1abd66fe-profile-collector-cert\") pod \"olm-operator-6b444d44fb-wxwlt\" (UID: \"025d75af-9ed0-4122-b97e-82dd1abd66fe\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wxwlt" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.248765 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79jht\" (UniqueName: \"kubernetes.io/projected/cf21b2c5-1b6a-458b-9abd-96b309d1626d-kube-api-access-79jht\") pod \"catalog-operator-68c6474976-nzcz7\" (UID: \"cf21b2c5-1b6a-458b-9abd-96b309d1626d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nzcz7" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.248799 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2441a0f3-b420-4a64-803b-81082cc15874-registry-certificates\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.248824 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ae2cfb63-dc76-40c2-9070-d5c29a6a272d-socket-dir\") pod \"csi-hostpathplugin-twswh\" (UID: \"ae2cfb63-dc76-40c2-9070-d5c29a6a272d\") " pod="hostpath-provisioner/csi-hostpathplugin-twswh" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.248852 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ae2cfb63-dc76-40c2-9070-d5c29a6a272d-mountpoint-dir\") pod \"csi-hostpathplugin-twswh\" (UID: \"ae2cfb63-dc76-40c2-9070-d5c29a6a272d\") " pod="hostpath-provisioner/csi-hostpathplugin-twswh" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.248879 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/471ebf6d-cd02-461d-96c9-d2b503d65cb8-config\") pod \"kube-apiserver-operator-766d6c64bb-wrbxz\" (UID: \"471ebf6d-cd02-461d-96c9-d2b503d65cb8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wrbxz" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.248905 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d2aa408d-c0e1-4697-b9e8-dee9276e8b47-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-7r6h5\" (UID: \"d2aa408d-c0e1-4697-b9e8-dee9276e8b47\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7r6h5" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.248946 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/02eb8668-3740-4714-98ae-e993184549aa-stats-auth\") pod \"router-default-5444994796-8nrlj\" (UID: \"02eb8668-3740-4714-98ae-e993184549aa\") " pod="openshift-ingress/router-default-5444994796-8nrlj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.248969 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzcwt\" (UniqueName: \"kubernetes.io/projected/4738a95e-6584-4c14-b545-d0e96d678ab0-kube-api-access-xzcwt\") pod \"ingress-operator-5b745b69d9-g4wgk\" (UID: \"4738a95e-6584-4c14-b545-d0e96d678ab0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g4wgk" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.249010 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1411e81d-100d-44a3-957f-5cc5a89e7748-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-dlv77\" (UID: \"1411e81d-100d-44a3-957f-5cc5a89e7748\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dlv77" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.249046 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/02eb8668-3740-4714-98ae-e993184549aa-service-ca-bundle\") pod \"router-default-5444994796-8nrlj\" (UID: \"02eb8668-3740-4714-98ae-e993184549aa\") " pod="openshift-ingress/router-default-5444994796-8nrlj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.249084 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsx5t\" (UniqueName: \"kubernetes.io/projected/337ebdb5-efc9-4bcb-8630-ab09a36c5bec-kube-api-access-rsx5t\") pod \"service-ca-9c57cc56f-tckf4\" (UID: \"337ebdb5-efc9-4bcb-8630-ab09a36c5bec\") " pod="openshift-service-ca/service-ca-9c57cc56f-tckf4" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.249110 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqd5x\" (UniqueName: \"kubernetes.io/projected/09269fbc-f59c-42cc-a832-2285987526db-kube-api-access-fqd5x\") pod \"ingress-canary-hkhrr\" (UID: \"09269fbc-f59c-42cc-a832-2285987526db\") " pod="openshift-ingress-canary/ingress-canary-hkhrr" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.249138 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1411e81d-100d-44a3-957f-5cc5a89e7748-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-dlv77\" (UID: \"1411e81d-100d-44a3-957f-5cc5a89e7748\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dlv77" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.249158 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/025d75af-9ed0-4122-b97e-82dd1abd66fe-srv-cert\") pod \"olm-operator-6b444d44fb-wxwlt\" (UID: \"025d75af-9ed0-4122-b97e-82dd1abd66fe\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wxwlt" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.249179 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fdwr\" (UniqueName: \"kubernetes.io/projected/4b51fc9b-c461-4e80-9a56-4f0e43e02399-kube-api-access-8fdwr\") pod \"collect-profiles-29401035-dht2z\" (UID: \"4b51fc9b-c461-4e80-9a56-4f0e43e02399\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401035-dht2z" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.249205 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f0ed4ec3-141e-413a-8229-86baf2265ece-metrics-tls\") pod \"dns-default-7vqhc\" (UID: \"f0ed4ec3-141e-413a-8229-86baf2265ece\") " pod="openshift-dns/dns-default-7vqhc" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.249230 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/049123ea-7dfd-43a3-b65c-9fb0c7b6c40b-available-featuregates\") pod \"openshift-config-operator-7777fb866f-x9k2l\" (UID: \"049123ea-7dfd-43a3-b65c-9fb0c7b6c40b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x9k2l" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.249254 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e1c3eef-d1b8-40c8-bbad-6e92ea15732e-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-vpp8l\" (UID: \"2e1c3eef-d1b8-40c8-bbad-6e92ea15732e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vpp8l" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.249279 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm6k2\" (UniqueName: \"kubernetes.io/projected/02eb8668-3740-4714-98ae-e993184549aa-kube-api-access-lm6k2\") pod \"router-default-5444994796-8nrlj\" (UID: \"02eb8668-3740-4714-98ae-e993184549aa\") " pod="openshift-ingress/router-default-5444994796-8nrlj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.249302 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/337ebdb5-efc9-4bcb-8630-ab09a36c5bec-signing-cabundle\") pod \"service-ca-9c57cc56f-tckf4\" (UID: \"337ebdb5-efc9-4bcb-8630-ab09a36c5bec\") " pod="openshift-service-ca/service-ca-9c57cc56f-tckf4" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.249343 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8b743ab8-67de-493a-bdf2-c0682a192869-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hwxqt\" (UID: \"8b743ab8-67de-493a-bdf2-c0682a192869\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwxqt" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.249364 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ae2cfb63-dc76-40c2-9070-d5c29a6a272d-registration-dir\") pod \"csi-hostpathplugin-twswh\" (UID: \"ae2cfb63-dc76-40c2-9070-d5c29a6a272d\") " pod="hostpath-provisioner/csi-hostpathplugin-twswh" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.249398 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/df8b1df7-103a-49ba-855e-face92da97dd-certs\") pod \"machine-config-server-xj5z4\" (UID: \"df8b1df7-103a-49ba-855e-face92da97dd\") " pod="openshift-machine-config-operator/machine-config-server-xj5z4" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.249419 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f0ed4ec3-141e-413a-8229-86baf2265ece-config-volume\") pod \"dns-default-7vqhc\" (UID: \"f0ed4ec3-141e-413a-8229-86baf2265ece\") " pod="openshift-dns/dns-default-7vqhc" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.249443 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2441a0f3-b420-4a64-803b-81082cc15874-bound-sa-token\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.249465 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qm8bb\" (UniqueName: \"kubernetes.io/projected/e8765bc6-c8e3-40d7-a061-211b1cad5df2-kube-api-access-qm8bb\") pod \"marketplace-operator-79b997595-5n2sp\" (UID: \"e8765bc6-c8e3-40d7-a061-211b1cad5df2\") " pod="openshift-marketplace/marketplace-operator-79b997595-5n2sp" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.251252 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/1411e81d-100d-44a3-957f-5cc5a89e7748-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-dlv77\" (UID: \"1411e81d-100d-44a3-957f-5cc5a89e7748\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dlv77" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.251480 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bb9001d-7d85-4cad-8587-e47e6f1c7aa9-config\") pod \"kube-controller-manager-operator-78b949d7b-tbbm7\" (UID: \"4bb9001d-7d85-4cad-8587-e47e6f1c7aa9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tbbm7" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.251523 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2aa408d-c0e1-4697-b9e8-dee9276e8b47-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-7r6h5\" (UID: \"d2aa408d-c0e1-4697-b9e8-dee9276e8b47\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7r6h5" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.251550 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.251570 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a229c7a5-8a0d-451b-9b28-da9c8762044a-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-rbfl6\" (UID: \"a229c7a5-8a0d-451b-9b28-da9c8762044a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rbfl6" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.251588 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e308a810-6256-4459-a587-d93ed4899fa8-proxy-tls\") pod \"machine-config-operator-74547568cd-f4747\" (UID: \"e308a810-6256-4459-a587-d93ed4899fa8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f4747" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.251607 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/02eb8668-3740-4714-98ae-e993184549aa-metrics-certs\") pod \"router-default-5444994796-8nrlj\" (UID: \"02eb8668-3740-4714-98ae-e993184549aa\") " pod="openshift-ingress/router-default-5444994796-8nrlj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.251625 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2e1c3eef-d1b8-40c8-bbad-6e92ea15732e-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-vpp8l\" (UID: \"2e1c3eef-d1b8-40c8-bbad-6e92ea15732e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vpp8l" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.251642 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qs4bf\" (UniqueName: \"kubernetes.io/projected/b80070a8-0eaa-4fe8-92ca-7c9afe029e04-kube-api-access-qs4bf\") pod \"migrator-59844c95c7-wzwq4\" (UID: \"b80070a8-0eaa-4fe8-92ca-7c9afe029e04\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-wzwq4" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.251674 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cf21b2c5-1b6a-458b-9abd-96b309d1626d-profile-collector-cert\") pod \"catalog-operator-68c6474976-nzcz7\" (UID: \"cf21b2c5-1b6a-458b-9abd-96b309d1626d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nzcz7" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.251694 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b51fc9b-c461-4e80-9a56-4f0e43e02399-secret-volume\") pod \"collect-profiles-29401035-dht2z\" (UID: \"4b51fc9b-c461-4e80-9a56-4f0e43e02399\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401035-dht2z" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.251826 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2441a0f3-b420-4a64-803b-81082cc15874-registry-certificates\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.251854 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e1c3eef-d1b8-40c8-bbad-6e92ea15732e-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-vpp8l\" (UID: \"2e1c3eef-d1b8-40c8-bbad-6e92ea15732e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vpp8l" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.252058 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/52e1736f-df9e-4444-bb28-4d89911000de-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-pgjrp\" (UID: \"52e1736f-df9e-4444-bb28-4d89911000de\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pgjrp" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.253173 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2aa408d-c0e1-4697-b9e8-dee9276e8b47-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-7r6h5\" (UID: \"d2aa408d-c0e1-4697-b9e8-dee9276e8b47\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7r6h5" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.253216 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1411e81d-100d-44a3-957f-5cc5a89e7748-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-dlv77\" (UID: \"1411e81d-100d-44a3-957f-5cc5a89e7748\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dlv77" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.253257 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/471ebf6d-cd02-461d-96c9-d2b503d65cb8-config\") pod \"kube-apiserver-operator-766d6c64bb-wrbxz\" (UID: \"471ebf6d-cd02-461d-96c9-d2b503d65cb8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wrbxz" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.253520 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e308a810-6256-4459-a587-d93ed4899fa8-auth-proxy-config\") pod \"machine-config-operator-74547568cd-f4747\" (UID: \"e308a810-6256-4459-a587-d93ed4899fa8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f4747" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.253548 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvpxx\" (UniqueName: \"kubernetes.io/projected/f0ed4ec3-141e-413a-8229-86baf2265ece-kube-api-access-dvpxx\") pod \"dns-default-7vqhc\" (UID: \"f0ed4ec3-141e-413a-8229-86baf2265ece\") " pod="openshift-dns/dns-default-7vqhc" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.253680 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/049123ea-7dfd-43a3-b65c-9fb0c7b6c40b-available-featuregates\") pod \"openshift-config-operator-7777fb866f-x9k2l\" (UID: \"049123ea-7dfd-43a3-b65c-9fb0c7b6c40b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x9k2l" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.253831 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-fjr5v"] Nov 25 09:26:47 crc kubenswrapper[4776]: E1125 09:26:47.254080 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:47.754048194 +0000 UTC m=+152.795107747 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.254239 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8003bc91-3e3b-4e88-8942-3ec54fc30aef-proxy-tls\") pod \"machine-config-controller-84d6567774-rc22s\" (UID: \"8003bc91-3e3b-4e88-8942-3ec54fc30aef\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rc22s" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.254308 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2441a0f3-b420-4a64-803b-81082cc15874-registry-tls\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.254545 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/471ebf6d-cd02-461d-96c9-d2b503d65cb8-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-wrbxz\" (UID: \"471ebf6d-cd02-461d-96c9-d2b503d65cb8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wrbxz" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.255638 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2441a0f3-b420-4a64-803b-81082cc15874-ca-trust-extracted\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.255673 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbg6n\" (UniqueName: \"kubernetes.io/projected/025d75af-9ed0-4122-b97e-82dd1abd66fe-kube-api-access-gbg6n\") pod \"olm-operator-6b444d44fb-wxwlt\" (UID: \"025d75af-9ed0-4122-b97e-82dd1abd66fe\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wxwlt" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.255706 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a01b028-a86d-40d3-9633-3f849f6d0840-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-h8t6d\" (UID: \"5a01b028-a86d-40d3-9633-3f849f6d0840\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h8t6d" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.255727 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b51fc9b-c461-4e80-9a56-4f0e43e02399-config-volume\") pod \"collect-profiles-29401035-dht2z\" (UID: \"4b51fc9b-c461-4e80-9a56-4f0e43e02399\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401035-dht2z" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.255749 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4738a95e-6584-4c14-b545-d0e96d678ab0-metrics-tls\") pod \"ingress-operator-5b745b69d9-g4wgk\" (UID: \"4738a95e-6584-4c14-b545-d0e96d678ab0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g4wgk" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.255772 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a9a85890-83ba-4e99-bdbe-386e0bbbecd6-tmpfs\") pod \"packageserver-d55dfcdfc-pxtkj\" (UID: \"a9a85890-83ba-4e99-bdbe-386e0bbbecd6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pxtkj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.255793 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cf21b2c5-1b6a-458b-9abd-96b309d1626d-srv-cert\") pod \"catalog-operator-68c6474976-nzcz7\" (UID: \"cf21b2c5-1b6a-458b-9abd-96b309d1626d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nzcz7" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.255817 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4bb9001d-7d85-4cad-8587-e47e6f1c7aa9-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-tbbm7\" (UID: \"4bb9001d-7d85-4cad-8587-e47e6f1c7aa9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tbbm7" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.254575 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e308a810-6256-4459-a587-d93ed4899fa8-auth-proxy-config\") pod \"machine-config-operator-74547568cd-f4747\" (UID: \"e308a810-6256-4459-a587-d93ed4899fa8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f4747" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.256570 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2e1c3eef-d1b8-40c8-bbad-6e92ea15732e-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-vpp8l\" (UID: \"2e1c3eef-d1b8-40c8-bbad-6e92ea15732e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vpp8l" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.254630 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e88ac546-8758-4cce-ae60-e49a1dc7c5ea-metrics-tls\") pod \"dns-operator-744455d44c-twg6n\" (UID: \"e88ac546-8758-4cce-ae60-e49a1dc7c5ea\") " pod="openshift-dns-operator/dns-operator-744455d44c-twg6n" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.257227 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2441a0f3-b420-4a64-803b-81082cc15874-ca-trust-extracted\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.258690 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/1411e81d-100d-44a3-957f-5cc5a89e7748-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-dlv77\" (UID: \"1411e81d-100d-44a3-957f-5cc5a89e7748\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dlv77" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.258689 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8003bc91-3e3b-4e88-8942-3ec54fc30aef-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-rc22s\" (UID: \"8003bc91-3e3b-4e88-8942-3ec54fc30aef\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rc22s" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.259747 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d2aa408d-c0e1-4697-b9e8-dee9276e8b47-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-7r6h5\" (UID: \"d2aa408d-c0e1-4697-b9e8-dee9276e8b47\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7r6h5" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.261489 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/471ebf6d-cd02-461d-96c9-d2b503d65cb8-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-wrbxz\" (UID: \"471ebf6d-cd02-461d-96c9-d2b503d65cb8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wrbxz" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.261561 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e308a810-6256-4459-a587-d93ed4899fa8-proxy-tls\") pod \"machine-config-operator-74547568cd-f4747\" (UID: \"e308a810-6256-4459-a587-d93ed4899fa8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f4747" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.261643 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2441a0f3-b420-4a64-803b-81082cc15874-registry-tls\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.261794 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a229c7a5-8a0d-451b-9b28-da9c8762044a-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-rbfl6\" (UID: \"a229c7a5-8a0d-451b-9b28-da9c8762044a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rbfl6" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.262503 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/471ebf6d-cd02-461d-96c9-d2b503d65cb8-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-wrbxz\" (UID: \"471ebf6d-cd02-461d-96c9-d2b503d65cb8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wrbxz" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.309039 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8m9tr\" (UniqueName: \"kubernetes.io/projected/e88ac546-8758-4cce-ae60-e49a1dc7c5ea-kube-api-access-8m9tr\") pod \"dns-operator-744455d44c-twg6n\" (UID: \"e88ac546-8758-4cce-ae60-e49a1dc7c5ea\") " pod="openshift-dns-operator/dns-operator-744455d44c-twg6n" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.320882 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w28t\" (UniqueName: \"kubernetes.io/projected/2e1c3eef-d1b8-40c8-bbad-6e92ea15732e-kube-api-access-4w28t\") pod \"openshift-controller-manager-operator-756b6f6bc6-vpp8l\" (UID: \"2e1c3eef-d1b8-40c8-bbad-6e92ea15732e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vpp8l" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.345227 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n68rq\" (UniqueName: \"kubernetes.io/projected/d2aa408d-c0e1-4697-b9e8-dee9276e8b47-kube-api-access-n68rq\") pod \"kube-storage-version-migrator-operator-b67b599dd-7r6h5\" (UID: \"d2aa408d-c0e1-4697-b9e8-dee9276e8b47\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7r6h5" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.357737 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:47 crc kubenswrapper[4776]: E1125 09:26:47.358057 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:47.858030692 +0000 UTC m=+152.899090245 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.358201 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a9a85890-83ba-4e99-bdbe-386e0bbbecd6-tmpfs\") pod \"packageserver-d55dfcdfc-pxtkj\" (UID: \"a9a85890-83ba-4e99-bdbe-386e0bbbecd6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pxtkj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.358249 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cf21b2c5-1b6a-458b-9abd-96b309d1626d-srv-cert\") pod \"catalog-operator-68c6474976-nzcz7\" (UID: \"cf21b2c5-1b6a-458b-9abd-96b309d1626d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nzcz7" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.358275 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4bb9001d-7d85-4cad-8587-e47e6f1c7aa9-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-tbbm7\" (UID: \"4bb9001d-7d85-4cad-8587-e47e6f1c7aa9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tbbm7" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.358326 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/df8b1df7-103a-49ba-855e-face92da97dd-node-bootstrap-token\") pod \"machine-config-server-xj5z4\" (UID: \"df8b1df7-103a-49ba-855e-face92da97dd\") " pod="openshift-machine-config-operator/machine-config-server-xj5z4" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.358353 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n78j9\" (UniqueName: \"kubernetes.io/projected/8ba97f9b-a0a1-4514-8577-0b36eb65bbd3-kube-api-access-n78j9\") pod \"multus-admission-controller-857f4d67dd-m9xnc\" (UID: \"8ba97f9b-a0a1-4514-8577-0b36eb65bbd3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-m9xnc" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.358382 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a01b028-a86d-40d3-9633-3f849f6d0840-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-h8t6d\" (UID: \"5a01b028-a86d-40d3-9633-3f849f6d0840\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h8t6d" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.358405 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/337ebdb5-efc9-4bcb-8630-ab09a36c5bec-signing-key\") pod \"service-ca-9c57cc56f-tckf4\" (UID: \"337ebdb5-efc9-4bcb-8630-ab09a36c5bec\") " pod="openshift-service-ca/service-ca-9c57cc56f-tckf4" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.358430 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb6w2\" (UniqueName: \"kubernetes.io/projected/ae2cfb63-dc76-40c2-9070-d5c29a6a272d-kube-api-access-sb6w2\") pod \"csi-hostpathplugin-twswh\" (UID: \"ae2cfb63-dc76-40c2-9070-d5c29a6a272d\") " pod="hostpath-provisioner/csi-hostpathplugin-twswh" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.358453 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4738a95e-6584-4c14-b545-d0e96d678ab0-trusted-ca\") pod \"ingress-operator-5b745b69d9-g4wgk\" (UID: \"4738a95e-6584-4c14-b545-d0e96d678ab0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g4wgk" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.358486 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/09269fbc-f59c-42cc-a832-2285987526db-cert\") pod \"ingress-canary-hkhrr\" (UID: \"09269fbc-f59c-42cc-a832-2285987526db\") " pod="openshift-ingress-canary/ingress-canary-hkhrr" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.358511 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bb9001d-7d85-4cad-8587-e47e6f1c7aa9-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-tbbm7\" (UID: \"4bb9001d-7d85-4cad-8587-e47e6f1c7aa9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tbbm7" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.358533 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5a01b028-a86d-40d3-9633-3f849f6d0840-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-h8t6d\" (UID: \"5a01b028-a86d-40d3-9633-3f849f6d0840\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h8t6d" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.358558 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e8765bc6-c8e3-40d7-a061-211b1cad5df2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5n2sp\" (UID: \"e8765bc6-c8e3-40d7-a061-211b1cad5df2\") " pod="openshift-marketplace/marketplace-operator-79b997595-5n2sp" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.358591 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z627v\" (UniqueName: \"kubernetes.io/projected/df8b1df7-103a-49ba-855e-face92da97dd-kube-api-access-z627v\") pod \"machine-config-server-xj5z4\" (UID: \"df8b1df7-103a-49ba-855e-face92da97dd\") " pod="openshift-machine-config-operator/machine-config-server-xj5z4" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.358620 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e932b24b-a48e-47c4-bb74-779fd2ec8ea9-config\") pod \"service-ca-operator-777779d784-mv7lp\" (UID: \"e932b24b-a48e-47c4-bb74-779fd2ec8ea9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mv7lp" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.358650 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a9a85890-83ba-4e99-bdbe-386e0bbbecd6-webhook-cert\") pod \"packageserver-d55dfcdfc-pxtkj\" (UID: \"a9a85890-83ba-4e99-bdbe-386e0bbbecd6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pxtkj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.358674 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ae2cfb63-dc76-40c2-9070-d5c29a6a272d-plugins-dir\") pod \"csi-hostpathplugin-twswh\" (UID: \"ae2cfb63-dc76-40c2-9070-d5c29a6a272d\") " pod="hostpath-provisioner/csi-hostpathplugin-twswh" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.358703 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjvk6\" (UniqueName: \"kubernetes.io/projected/a9a85890-83ba-4e99-bdbe-386e0bbbecd6-kube-api-access-bjvk6\") pod \"packageserver-d55dfcdfc-pxtkj\" (UID: \"a9a85890-83ba-4e99-bdbe-386e0bbbecd6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pxtkj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.358726 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4738a95e-6584-4c14-b545-d0e96d678ab0-bound-sa-token\") pod \"ingress-operator-5b745b69d9-g4wgk\" (UID: \"4738a95e-6584-4c14-b545-d0e96d678ab0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g4wgk" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.358747 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e932b24b-a48e-47c4-bb74-779fd2ec8ea9-serving-cert\") pod \"service-ca-operator-777779d784-mv7lp\" (UID: \"e932b24b-a48e-47c4-bb74-779fd2ec8ea9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mv7lp" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.358770 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e8765bc6-c8e3-40d7-a061-211b1cad5df2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5n2sp\" (UID: \"e8765bc6-c8e3-40d7-a061-211b1cad5df2\") " pod="openshift-marketplace/marketplace-operator-79b997595-5n2sp" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.358793 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/02eb8668-3740-4714-98ae-e993184549aa-default-certificate\") pod \"router-default-5444994796-8nrlj\" (UID: \"02eb8668-3740-4714-98ae-e993184549aa\") " pod="openshift-ingress/router-default-5444994796-8nrlj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.358816 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8ba97f9b-a0a1-4514-8577-0b36eb65bbd3-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-m9xnc\" (UID: \"8ba97f9b-a0a1-4514-8577-0b36eb65bbd3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-m9xnc" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.358854 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ae2cfb63-dc76-40c2-9070-d5c29a6a272d-csi-data-dir\") pod \"csi-hostpathplugin-twswh\" (UID: \"ae2cfb63-dc76-40c2-9070-d5c29a6a272d\") " pod="hostpath-provisioner/csi-hostpathplugin-twswh" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.358881 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a9a85890-83ba-4e99-bdbe-386e0bbbecd6-apiservice-cert\") pod \"packageserver-d55dfcdfc-pxtkj\" (UID: \"a9a85890-83ba-4e99-bdbe-386e0bbbecd6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pxtkj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.358913 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kx5ld\" (UniqueName: \"kubernetes.io/projected/8b743ab8-67de-493a-bdf2-c0682a192869-kube-api-access-kx5ld\") pod \"package-server-manager-789f6589d5-hwxqt\" (UID: \"8b743ab8-67de-493a-bdf2-c0682a192869\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwxqt" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.358946 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5slr6\" (UniqueName: \"kubernetes.io/projected/e932b24b-a48e-47c4-bb74-779fd2ec8ea9-kube-api-access-5slr6\") pod \"service-ca-operator-777779d784-mv7lp\" (UID: \"e932b24b-a48e-47c4-bb74-779fd2ec8ea9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mv7lp" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.358986 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/025d75af-9ed0-4122-b97e-82dd1abd66fe-profile-collector-cert\") pod \"olm-operator-6b444d44fb-wxwlt\" (UID: \"025d75af-9ed0-4122-b97e-82dd1abd66fe\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wxwlt" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359013 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79jht\" (UniqueName: \"kubernetes.io/projected/cf21b2c5-1b6a-458b-9abd-96b309d1626d-kube-api-access-79jht\") pod \"catalog-operator-68c6474976-nzcz7\" (UID: \"cf21b2c5-1b6a-458b-9abd-96b309d1626d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nzcz7" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359043 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ae2cfb63-dc76-40c2-9070-d5c29a6a272d-socket-dir\") pod \"csi-hostpathplugin-twswh\" (UID: \"ae2cfb63-dc76-40c2-9070-d5c29a6a272d\") " pod="hostpath-provisioner/csi-hostpathplugin-twswh" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359092 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ae2cfb63-dc76-40c2-9070-d5c29a6a272d-mountpoint-dir\") pod \"csi-hostpathplugin-twswh\" (UID: \"ae2cfb63-dc76-40c2-9070-d5c29a6a272d\") " pod="hostpath-provisioner/csi-hostpathplugin-twswh" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359123 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/02eb8668-3740-4714-98ae-e993184549aa-stats-auth\") pod \"router-default-5444994796-8nrlj\" (UID: \"02eb8668-3740-4714-98ae-e993184549aa\") " pod="openshift-ingress/router-default-5444994796-8nrlj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359144 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzcwt\" (UniqueName: \"kubernetes.io/projected/4738a95e-6584-4c14-b545-d0e96d678ab0-kube-api-access-xzcwt\") pod \"ingress-operator-5b745b69d9-g4wgk\" (UID: \"4738a95e-6584-4c14-b545-d0e96d678ab0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g4wgk" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359172 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/02eb8668-3740-4714-98ae-e993184549aa-service-ca-bundle\") pod \"router-default-5444994796-8nrlj\" (UID: \"02eb8668-3740-4714-98ae-e993184549aa\") " pod="openshift-ingress/router-default-5444994796-8nrlj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359205 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsx5t\" (UniqueName: \"kubernetes.io/projected/337ebdb5-efc9-4bcb-8630-ab09a36c5bec-kube-api-access-rsx5t\") pod \"service-ca-9c57cc56f-tckf4\" (UID: \"337ebdb5-efc9-4bcb-8630-ab09a36c5bec\") " pod="openshift-service-ca/service-ca-9c57cc56f-tckf4" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359236 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqd5x\" (UniqueName: \"kubernetes.io/projected/09269fbc-f59c-42cc-a832-2285987526db-kube-api-access-fqd5x\") pod \"ingress-canary-hkhrr\" (UID: \"09269fbc-f59c-42cc-a832-2285987526db\") " pod="openshift-ingress-canary/ingress-canary-hkhrr" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359245 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a9a85890-83ba-4e99-bdbe-386e0bbbecd6-tmpfs\") pod \"packageserver-d55dfcdfc-pxtkj\" (UID: \"a9a85890-83ba-4e99-bdbe-386e0bbbecd6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pxtkj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359259 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/025d75af-9ed0-4122-b97e-82dd1abd66fe-srv-cert\") pod \"olm-operator-6b444d44fb-wxwlt\" (UID: \"025d75af-9ed0-4122-b97e-82dd1abd66fe\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wxwlt" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359282 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fdwr\" (UniqueName: \"kubernetes.io/projected/4b51fc9b-c461-4e80-9a56-4f0e43e02399-kube-api-access-8fdwr\") pod \"collect-profiles-29401035-dht2z\" (UID: \"4b51fc9b-c461-4e80-9a56-4f0e43e02399\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401035-dht2z" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359308 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f0ed4ec3-141e-413a-8229-86baf2265ece-metrics-tls\") pod \"dns-default-7vqhc\" (UID: \"f0ed4ec3-141e-413a-8229-86baf2265ece\") " pod="openshift-dns/dns-default-7vqhc" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359447 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm6k2\" (UniqueName: \"kubernetes.io/projected/02eb8668-3740-4714-98ae-e993184549aa-kube-api-access-lm6k2\") pod \"router-default-5444994796-8nrlj\" (UID: \"02eb8668-3740-4714-98ae-e993184549aa\") " pod="openshift-ingress/router-default-5444994796-8nrlj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359479 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/337ebdb5-efc9-4bcb-8630-ab09a36c5bec-signing-cabundle\") pod \"service-ca-9c57cc56f-tckf4\" (UID: \"337ebdb5-efc9-4bcb-8630-ab09a36c5bec\") " pod="openshift-service-ca/service-ca-9c57cc56f-tckf4" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359509 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8b743ab8-67de-493a-bdf2-c0682a192869-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hwxqt\" (UID: \"8b743ab8-67de-493a-bdf2-c0682a192869\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwxqt" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359532 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/df8b1df7-103a-49ba-855e-face92da97dd-certs\") pod \"machine-config-server-xj5z4\" (UID: \"df8b1df7-103a-49ba-855e-face92da97dd\") " pod="openshift-machine-config-operator/machine-config-server-xj5z4" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359551 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f0ed4ec3-141e-413a-8229-86baf2265ece-config-volume\") pod \"dns-default-7vqhc\" (UID: \"f0ed4ec3-141e-413a-8229-86baf2265ece\") " pod="openshift-dns/dns-default-7vqhc" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359572 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ae2cfb63-dc76-40c2-9070-d5c29a6a272d-registration-dir\") pod \"csi-hostpathplugin-twswh\" (UID: \"ae2cfb63-dc76-40c2-9070-d5c29a6a272d\") " pod="hostpath-provisioner/csi-hostpathplugin-twswh" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359602 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qm8bb\" (UniqueName: \"kubernetes.io/projected/e8765bc6-c8e3-40d7-a061-211b1cad5df2-kube-api-access-qm8bb\") pod \"marketplace-operator-79b997595-5n2sp\" (UID: \"e8765bc6-c8e3-40d7-a061-211b1cad5df2\") " pod="openshift-marketplace/marketplace-operator-79b997595-5n2sp" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359631 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bb9001d-7d85-4cad-8587-e47e6f1c7aa9-config\") pod \"kube-controller-manager-operator-78b949d7b-tbbm7\" (UID: \"4bb9001d-7d85-4cad-8587-e47e6f1c7aa9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tbbm7" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359666 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359690 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/02eb8668-3740-4714-98ae-e993184549aa-metrics-certs\") pod \"router-default-5444994796-8nrlj\" (UID: \"02eb8668-3740-4714-98ae-e993184549aa\") " pod="openshift-ingress/router-default-5444994796-8nrlj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359721 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cf21b2c5-1b6a-458b-9abd-96b309d1626d-profile-collector-cert\") pod \"catalog-operator-68c6474976-nzcz7\" (UID: \"cf21b2c5-1b6a-458b-9abd-96b309d1626d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nzcz7" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359763 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b51fc9b-c461-4e80-9a56-4f0e43e02399-secret-volume\") pod \"collect-profiles-29401035-dht2z\" (UID: \"4b51fc9b-c461-4e80-9a56-4f0e43e02399\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401035-dht2z" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359799 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvpxx\" (UniqueName: \"kubernetes.io/projected/f0ed4ec3-141e-413a-8229-86baf2265ece-kube-api-access-dvpxx\") pod \"dns-default-7vqhc\" (UID: \"f0ed4ec3-141e-413a-8229-86baf2265ece\") " pod="openshift-dns/dns-default-7vqhc" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359827 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbg6n\" (UniqueName: \"kubernetes.io/projected/025d75af-9ed0-4122-b97e-82dd1abd66fe-kube-api-access-gbg6n\") pod \"olm-operator-6b444d44fb-wxwlt\" (UID: \"025d75af-9ed0-4122-b97e-82dd1abd66fe\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wxwlt" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359849 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b51fc9b-c461-4e80-9a56-4f0e43e02399-config-volume\") pod \"collect-profiles-29401035-dht2z\" (UID: \"4b51fc9b-c461-4e80-9a56-4f0e43e02399\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401035-dht2z" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359872 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a01b028-a86d-40d3-9633-3f849f6d0840-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-h8t6d\" (UID: \"5a01b028-a86d-40d3-9633-3f849f6d0840\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h8t6d" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.359898 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4738a95e-6584-4c14-b545-d0e96d678ab0-metrics-tls\") pod \"ingress-operator-5b745b69d9-g4wgk\" (UID: \"4738a95e-6584-4c14-b545-d0e96d678ab0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g4wgk" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.361609 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cf21b2c5-1b6a-458b-9abd-96b309d1626d-srv-cert\") pod \"catalog-operator-68c6474976-nzcz7\" (UID: \"cf21b2c5-1b6a-458b-9abd-96b309d1626d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nzcz7" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.362325 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ae2cfb63-dc76-40c2-9070-d5c29a6a272d-socket-dir\") pod \"csi-hostpathplugin-twswh\" (UID: \"ae2cfb63-dc76-40c2-9070-d5c29a6a272d\") " pod="hostpath-provisioner/csi-hostpathplugin-twswh" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.362394 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ae2cfb63-dc76-40c2-9070-d5c29a6a272d-mountpoint-dir\") pod \"csi-hostpathplugin-twswh\" (UID: \"ae2cfb63-dc76-40c2-9070-d5c29a6a272d\") " pod="hostpath-provisioner/csi-hostpathplugin-twswh" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.362933 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e932b24b-a48e-47c4-bb74-779fd2ec8ea9-serving-cert\") pod \"service-ca-operator-777779d784-mv7lp\" (UID: \"e932b24b-a48e-47c4-bb74-779fd2ec8ea9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mv7lp" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.365850 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bb9001d-7d85-4cad-8587-e47e6f1c7aa9-config\") pod \"kube-controller-manager-operator-78b949d7b-tbbm7\" (UID: \"4bb9001d-7d85-4cad-8587-e47e6f1c7aa9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tbbm7" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.367212 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/02eb8668-3740-4714-98ae-e993184549aa-service-ca-bundle\") pod \"router-default-5444994796-8nrlj\" (UID: \"02eb8668-3740-4714-98ae-e993184549aa\") " pod="openshift-ingress/router-default-5444994796-8nrlj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.367611 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4738a95e-6584-4c14-b545-d0e96d678ab0-metrics-tls\") pod \"ingress-operator-5b745b69d9-g4wgk\" (UID: \"4738a95e-6584-4c14-b545-d0e96d678ab0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g4wgk" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.368645 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f0ed4ec3-141e-413a-8229-86baf2265ece-metrics-tls\") pod \"dns-default-7vqhc\" (UID: \"f0ed4ec3-141e-413a-8229-86baf2265ece\") " pod="openshift-dns/dns-default-7vqhc" Nov 25 09:26:47 crc kubenswrapper[4776]: E1125 09:26:47.368942 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:47.868921131 +0000 UTC m=+152.909980784 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.369291 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b51fc9b-c461-4e80-9a56-4f0e43e02399-config-volume\") pod \"collect-profiles-29401035-dht2z\" (UID: \"4b51fc9b-c461-4e80-9a56-4f0e43e02399\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401035-dht2z" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.369731 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a01b028-a86d-40d3-9633-3f849f6d0840-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-h8t6d\" (UID: \"5a01b028-a86d-40d3-9633-3f849f6d0840\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h8t6d" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.370348 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f0ed4ec3-141e-413a-8229-86baf2265ece-config-volume\") pod \"dns-default-7vqhc\" (UID: \"f0ed4ec3-141e-413a-8229-86baf2265ece\") " pod="openshift-dns/dns-default-7vqhc" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.371866 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e8765bc6-c8e3-40d7-a061-211b1cad5df2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5n2sp\" (UID: \"e8765bc6-c8e3-40d7-a061-211b1cad5df2\") " pod="openshift-marketplace/marketplace-operator-79b997595-5n2sp" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.372490 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-srdjj"] Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.373687 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ae2cfb63-dc76-40c2-9070-d5c29a6a272d-registration-dir\") pod \"csi-hostpathplugin-twswh\" (UID: \"ae2cfb63-dc76-40c2-9070-d5c29a6a272d\") " pod="hostpath-provisioner/csi-hostpathplugin-twswh" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.373817 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ae2cfb63-dc76-40c2-9070-d5c29a6a272d-csi-data-dir\") pod \"csi-hostpathplugin-twswh\" (UID: \"ae2cfb63-dc76-40c2-9070-d5c29a6a272d\") " pod="hostpath-provisioner/csi-hostpathplugin-twswh" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.373929 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/025d75af-9ed0-4122-b97e-82dd1abd66fe-srv-cert\") pod \"olm-operator-6b444d44fb-wxwlt\" (UID: \"025d75af-9ed0-4122-b97e-82dd1abd66fe\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wxwlt" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.376294 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-sb2sq"] Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.376841 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b51fc9b-c461-4e80-9a56-4f0e43e02399-secret-volume\") pod \"collect-profiles-29401035-dht2z\" (UID: \"4b51fc9b-c461-4e80-9a56-4f0e43e02399\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401035-dht2z" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.378746 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/09269fbc-f59c-42cc-a832-2285987526db-cert\") pod \"ingress-canary-hkhrr\" (UID: \"09269fbc-f59c-42cc-a832-2285987526db\") " pod="openshift-ingress-canary/ingress-canary-hkhrr" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.379863 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4bb9001d-7d85-4cad-8587-e47e6f1c7aa9-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-tbbm7\" (UID: \"4bb9001d-7d85-4cad-8587-e47e6f1c7aa9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tbbm7" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.380348 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ae2cfb63-dc76-40c2-9070-d5c29a6a272d-plugins-dir\") pod \"csi-hostpathplugin-twswh\" (UID: \"ae2cfb63-dc76-40c2-9070-d5c29a6a272d\") " pod="hostpath-provisioner/csi-hostpathplugin-twswh" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.382265 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4738a95e-6584-4c14-b545-d0e96d678ab0-trusted-ca\") pod \"ingress-operator-5b745b69d9-g4wgk\" (UID: \"4738a95e-6584-4c14-b545-d0e96d678ab0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g4wgk" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.387853 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e932b24b-a48e-47c4-bb74-779fd2ec8ea9-config\") pod \"service-ca-operator-777779d784-mv7lp\" (UID: \"e932b24b-a48e-47c4-bb74-779fd2ec8ea9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mv7lp" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.388159 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8b743ab8-67de-493a-bdf2-c0682a192869-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hwxqt\" (UID: \"8b743ab8-67de-493a-bdf2-c0682a192869\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwxqt" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.391260 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv9q2\" (UniqueName: \"kubernetes.io/projected/a229c7a5-8a0d-451b-9b28-da9c8762044a-kube-api-access-hv9q2\") pod \"control-plane-machine-set-operator-78cbb6b69f-rbfl6\" (UID: \"a229c7a5-8a0d-451b-9b28-da9c8762044a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rbfl6" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.391624 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/02eb8668-3740-4714-98ae-e993184549aa-stats-auth\") pod \"router-default-5444994796-8nrlj\" (UID: \"02eb8668-3740-4714-98ae-e993184549aa\") " pod="openshift-ingress/router-default-5444994796-8nrlj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.397169 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/02eb8668-3740-4714-98ae-e993184549aa-metrics-certs\") pod \"router-default-5444994796-8nrlj\" (UID: \"02eb8668-3740-4714-98ae-e993184549aa\") " pod="openshift-ingress/router-default-5444994796-8nrlj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.397323 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a9a85890-83ba-4e99-bdbe-386e0bbbecd6-apiservice-cert\") pod \"packageserver-d55dfcdfc-pxtkj\" (UID: \"a9a85890-83ba-4e99-bdbe-386e0bbbecd6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pxtkj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.397503 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a01b028-a86d-40d3-9633-3f849f6d0840-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-h8t6d\" (UID: \"5a01b028-a86d-40d3-9633-3f849f6d0840\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h8t6d" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.398149 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jlrd\" (UniqueName: \"kubernetes.io/projected/8003bc91-3e3b-4e88-8942-3ec54fc30aef-kube-api-access-2jlrd\") pod \"machine-config-controller-84d6567774-rc22s\" (UID: \"8003bc91-3e3b-4e88-8942-3ec54fc30aef\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rc22s" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.398588 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cf21b2c5-1b6a-458b-9abd-96b309d1626d-profile-collector-cert\") pod \"catalog-operator-68c6474976-nzcz7\" (UID: \"cf21b2c5-1b6a-458b-9abd-96b309d1626d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nzcz7" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.399498 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/02eb8668-3740-4714-98ae-e993184549aa-default-certificate\") pod \"router-default-5444994796-8nrlj\" (UID: \"02eb8668-3740-4714-98ae-e993184549aa\") " pod="openshift-ingress/router-default-5444994796-8nrlj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.399602 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e8765bc6-c8e3-40d7-a061-211b1cad5df2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5n2sp\" (UID: \"e8765bc6-c8e3-40d7-a061-211b1cad5df2\") " pod="openshift-marketplace/marketplace-operator-79b997595-5n2sp" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.399696 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/025d75af-9ed0-4122-b97e-82dd1abd66fe-profile-collector-cert\") pod \"olm-operator-6b444d44fb-wxwlt\" (UID: \"025d75af-9ed0-4122-b97e-82dd1abd66fe\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wxwlt" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.401535 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/df8b1df7-103a-49ba-855e-face92da97dd-certs\") pod \"machine-config-server-xj5z4\" (UID: \"df8b1df7-103a-49ba-855e-face92da97dd\") " pod="openshift-machine-config-operator/machine-config-server-xj5z4" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.401725 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/df8b1df7-103a-49ba-855e-face92da97dd-node-bootstrap-token\") pod \"machine-config-server-xj5z4\" (UID: \"df8b1df7-103a-49ba-855e-face92da97dd\") " pod="openshift-machine-config-operator/machine-config-server-xj5z4" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.411800 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a9a85890-83ba-4e99-bdbe-386e0bbbecd6-webhook-cert\") pod \"packageserver-d55dfcdfc-pxtkj\" (UID: \"a9a85890-83ba-4e99-bdbe-386e0bbbecd6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pxtkj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.412008 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9qz5\" (UniqueName: \"kubernetes.io/projected/1411e81d-100d-44a3-957f-5cc5a89e7748-kube-api-access-p9qz5\") pod \"cluster-image-registry-operator-dc59b4c8b-dlv77\" (UID: \"1411e81d-100d-44a3-957f-5cc5a89e7748\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dlv77" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.413904 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8ba97f9b-a0a1-4514-8577-0b36eb65bbd3-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-m9xnc\" (UID: \"8ba97f9b-a0a1-4514-8577-0b36eb65bbd3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-m9xnc" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.414297 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/337ebdb5-efc9-4bcb-8630-ab09a36c5bec-signing-cabundle\") pod \"service-ca-9c57cc56f-tckf4\" (UID: \"337ebdb5-efc9-4bcb-8630-ab09a36c5bec\") " pod="openshift-service-ca/service-ca-9c57cc56f-tckf4" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.416729 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/337ebdb5-efc9-4bcb-8630-ab09a36c5bec-signing-key\") pod \"service-ca-9c57cc56f-tckf4\" (UID: \"337ebdb5-efc9-4bcb-8630-ab09a36c5bec\") " pod="openshift-service-ca/service-ca-9c57cc56f-tckf4" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.426822 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwv8h\" (UniqueName: \"kubernetes.io/projected/2441a0f3-b420-4a64-803b-81082cc15874-kube-api-access-mwv8h\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.441436 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qvvq\" (UniqueName: \"kubernetes.io/projected/049123ea-7dfd-43a3-b65c-9fb0c7b6c40b-kube-api-access-2qvvq\") pod \"openshift-config-operator-7777fb866f-x9k2l\" (UID: \"049123ea-7dfd-43a3-b65c-9fb0c7b6c40b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-x9k2l" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.461415 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:47 crc kubenswrapper[4776]: E1125 09:26:47.462301 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:47.962279187 +0000 UTC m=+153.003338740 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.464048 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fggj\" (UniqueName: \"kubernetes.io/projected/52e1736f-df9e-4444-bb28-4d89911000de-kube-api-access-9fggj\") pod \"cluster-samples-operator-665b6dd947-pgjrp\" (UID: \"52e1736f-df9e-4444-bb28-4d89911000de\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pgjrp" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.469844 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vpp8l" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.480439 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x9k2l" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.485155 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmv87\" (UniqueName: \"kubernetes.io/projected/e308a810-6256-4459-a587-d93ed4899fa8-kube-api-access-bmv87\") pod \"machine-config-operator-74547568cd-f4747\" (UID: \"e308a810-6256-4459-a587-d93ed4899fa8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f4747" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.486467 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-qqkgv"] Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.498092 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1411e81d-100d-44a3-957f-5cc5a89e7748-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-dlv77\" (UID: \"1411e81d-100d-44a3-957f-5cc5a89e7748\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dlv77" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.504176 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-twg6n" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.505347 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fq5rl"] Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.508810 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dlv77" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.509437 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xh8xm"] Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.510554 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq"] Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.516838 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pgjrp" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.518024 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-bhgv2"] Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.524619 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2441a0f3-b420-4a64-803b-81082cc15874-bound-sa-token\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.532545 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7r6h5" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.538004 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wrbxz" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.545889 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qs4bf\" (UniqueName: \"kubernetes.io/projected/b80070a8-0eaa-4fe8-92ca-7c9afe029e04-kube-api-access-qs4bf\") pod \"migrator-59844c95c7-wzwq4\" (UID: \"b80070a8-0eaa-4fe8-92ca-7c9afe029e04\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-wzwq4" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.548029 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rbfl6" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.553914 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-wzwq4" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.557725 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rc22s" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.563215 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:47 crc kubenswrapper[4776]: E1125 09:26:47.564047 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:48.06403003 +0000 UTC m=+153.105089583 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.564970 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pxpfx" event={"ID":"10d2d0a3-eaba-486f-a833-0da30ad1f18e","Type":"ContainerStarted","Data":"319ead6eace837f5e4fb6bec9c1aeb87e705dd2153404eaf0248803e2c7bc52b"} Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.565010 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pxpfx" event={"ID":"10d2d0a3-eaba-486f-a833-0da30ad1f18e","Type":"ContainerStarted","Data":"ca5c5c73d66985211b52960e5d8d273600a1fa374bc849df8daf3e06a9e7ad51"} Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.565945 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f4747" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.572587 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" event={"ID":"490fba04-edd0-4edc-9f1e-4e08debf284f","Type":"ContainerStarted","Data":"26614fee53750b1a6612a048fa02b3f83264c3b8996d41e6b69181ad31742c45"} Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.591533 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-r979m" event={"ID":"529b4e1d-e99c-4342-8cff-68c1e31a776f","Type":"ContainerStarted","Data":"7e439c271c65b4ad711496e4cff0e6a49ce9ecc2f65cec74ef7a45281f7d53c2"} Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.591597 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-r979m" event={"ID":"529b4e1d-e99c-4342-8cff-68c1e31a776f","Type":"ContainerStarted","Data":"289c12db5f1491c26811c7ff52ec762e10eeea71eb6d2cbfbde9f950568beb82"} Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.591933 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-r979m" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.593949 4776 patch_prober.go:28] interesting pod/downloads-7954f5f757-r979m container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.593997 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r979m" podUID="529b4e1d-e99c-4342-8cff-68c1e31a776f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.599576 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fdwr\" (UniqueName: \"kubernetes.io/projected/4b51fc9b-c461-4e80-9a56-4f0e43e02399-kube-api-access-8fdwr\") pod \"collect-profiles-29401035-dht2z\" (UID: \"4b51fc9b-c461-4e80-9a56-4f0e43e02399\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401035-dht2z" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.604792 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79jht\" (UniqueName: \"kubernetes.io/projected/cf21b2c5-1b6a-458b-9abd-96b309d1626d-kube-api-access-79jht\") pod \"catalog-operator-68c6474976-nzcz7\" (UID: \"cf21b2c5-1b6a-458b-9abd-96b309d1626d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nzcz7" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.606504 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" event={"ID":"be1426b0-cf02-44d3-87b8-aa0394af1afa","Type":"ContainerStarted","Data":"b3d06ce39030d17ed88b1f53ebcabf8ed828139cd41a7018921a42ed83eaeb57"} Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.624477 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4bb9001d-7d85-4cad-8587-e47e6f1c7aa9-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-tbbm7\" (UID: \"4bb9001d-7d85-4cad-8587-e47e6f1c7aa9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tbbm7" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.624819 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-fjr5v" event={"ID":"fffe85b6-294d-4f43-ad27-65a5d093c076","Type":"ContainerStarted","Data":"9181a33f077330faa5755ffc932d1cd510d27a8a934bbc7adc37cb4f72f4d9c1"} Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.627507 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-fjr5v" event={"ID":"fffe85b6-294d-4f43-ad27-65a5d093c076","Type":"ContainerStarted","Data":"478c0c4eeedeb7bceb772d12da13e38abc93caec178a353ba4646167001bfbbb"} Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.627543 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-srdjj" event={"ID":"974a7107-64ce-4f92-842b-7927d3ba006d","Type":"ContainerStarted","Data":"d892a91db43ec3036b4be6d3ead177189b1f8fdd61621f16936d20a3ebc7d01e"} Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.632263 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401035-dht2z" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.639274 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-mt26t" event={"ID":"b858f43c-8ede-490b-866e-56a8839c8d8b","Type":"ContainerStarted","Data":"1fcd0a5fd842cbd90520bb3c8d18c4d59a11de1fccf72aedb1c4d07c867ff1d8"} Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.639314 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-mt26t" event={"ID":"b858f43c-8ede-490b-866e-56a8839c8d8b","Type":"ContainerStarted","Data":"236a979ceeee70b6f296ab7df2fa5f52f6bcfb3d7622e23c8c453f5ad493722e"} Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.639993 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-mt26t" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.641452 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xh8xm" event={"ID":"1d06dda7-d25a-4d6d-93b3-9a8579876bee","Type":"ContainerStarted","Data":"5194c250d612158f49d3dc2a8ef47c2aa064ab7eaee11ceee29c43ad7c5f3091"} Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.643364 4776 patch_prober.go:28] interesting pod/console-operator-58897d9998-mt26t container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.643409 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-mt26t" podUID="b858f43c-8ede-490b-866e-56a8839c8d8b" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.646249 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzcwt\" (UniqueName: \"kubernetes.io/projected/4738a95e-6584-4c14-b545-d0e96d678ab0-kube-api-access-xzcwt\") pod \"ingress-operator-5b745b69d9-g4wgk\" (UID: \"4738a95e-6584-4c14-b545-d0e96d678ab0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g4wgk" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.649828 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4" event={"ID":"c7ba5522-7755-4e0a-96e8-6d818e130fa2","Type":"ContainerStarted","Data":"b3e28a8d029b15ccffe0ab864f1506542921325f7da5dfe877595c48db93c082"} Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.649900 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4" event={"ID":"c7ba5522-7755-4e0a-96e8-6d818e130fa2","Type":"ContainerStarted","Data":"ac4a185653aa1a8f1db2726538a9079d4f34d1b3503b632a3403bee2e488996e"} Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.651372 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.653180 4776 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-n6sd4 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.653212 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4" podUID="c7ba5522-7755-4e0a-96e8-6d818e130fa2" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.656747 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-sb2sq" event={"ID":"5e67eb5b-2b9f-4553-861e-d687a07a1d12","Type":"ContainerStarted","Data":"1259a373610fb960b7caa2550d36ff480fe70c036b32bcb3dae44c58cf7d10b2"} Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.656795 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-sb2sq" event={"ID":"5e67eb5b-2b9f-4553-861e-d687a07a1d12","Type":"ContainerStarted","Data":"f91795605a8cfa1e8b7e943739fbdb3ae52f31735eff8654dc6881309d2aeda2"} Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.660475 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-qqkgv" event={"ID":"50f49648-47d6-45d8-9563-f76b7547cad6","Type":"ContainerStarted","Data":"c02363e4d0fbf764d3900dc38f4a1770cfae17a0a2c42dc9dcd394a665d59109"} Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.662119 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsx5t\" (UniqueName: \"kubernetes.io/projected/337ebdb5-efc9-4bcb-8630-ab09a36c5bec-kube-api-access-rsx5t\") pod \"service-ca-9c57cc56f-tckf4\" (UID: \"337ebdb5-efc9-4bcb-8630-ab09a36c5bec\") " pod="openshift-service-ca/service-ca-9c57cc56f-tckf4" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.666874 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:47 crc kubenswrapper[4776]: E1125 09:26:47.668283 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:48.168266994 +0000 UTC m=+153.209326547 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.674747 4776 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-8rhfr container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" start-of-body= Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.674788 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" podUID="c5ae11bc-db7c-4665-88e4-ef3b0eb83581" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.676647 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-bhgv2" event={"ID":"04fe29b5-d8c6-46c5-9d1f-76005e225fd6","Type":"ContainerStarted","Data":"7822ef037f518dffb97dd42c557b0f98dbd1ff31fc53a6a167d11c58d44cdd27"} Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.676685 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.676698 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" event={"ID":"c5ae11bc-db7c-4665-88e4-ef3b0eb83581","Type":"ContainerStarted","Data":"d884031de43de28107048c44a1903141869dd6318f0e8ec733b6b2eb1ce4ac13"} Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.676707 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" event={"ID":"c5ae11bc-db7c-4665-88e4-ef3b0eb83581","Type":"ContainerStarted","Data":"c7be2d6b7d9f8dacd74ac864092f2282a5ad82be9eda5a52b5c5b3c23e02f46a"} Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.697087 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqd5x\" (UniqueName: \"kubernetes.io/projected/09269fbc-f59c-42cc-a832-2285987526db-kube-api-access-fqd5x\") pod \"ingress-canary-hkhrr\" (UID: \"09269fbc-f59c-42cc-a832-2285987526db\") " pod="openshift-ingress-canary/ingress-canary-hkhrr" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.705180 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm6k2\" (UniqueName: \"kubernetes.io/projected/02eb8668-3740-4714-98ae-e993184549aa-kube-api-access-lm6k2\") pod \"router-default-5444994796-8nrlj\" (UID: \"02eb8668-3740-4714-98ae-e993184549aa\") " pod="openshift-ingress/router-default-5444994796-8nrlj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.716928 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vpp8l"] Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.722900 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n78j9\" (UniqueName: \"kubernetes.io/projected/8ba97f9b-a0a1-4514-8577-0b36eb65bbd3-kube-api-access-n78j9\") pod \"multus-admission-controller-857f4d67dd-m9xnc\" (UID: \"8ba97f9b-a0a1-4514-8577-0b36eb65bbd3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-m9xnc" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.745818 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-x9k2l"] Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.751399 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbg6n\" (UniqueName: \"kubernetes.io/projected/025d75af-9ed0-4122-b97e-82dd1abd66fe-kube-api-access-gbg6n\") pod \"olm-operator-6b444d44fb-wxwlt\" (UID: \"025d75af-9ed0-4122-b97e-82dd1abd66fe\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wxwlt" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.764572 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvpxx\" (UniqueName: \"kubernetes.io/projected/f0ed4ec3-141e-413a-8229-86baf2265ece-kube-api-access-dvpxx\") pod \"dns-default-7vqhc\" (UID: \"f0ed4ec3-141e-413a-8229-86baf2265ece\") " pod="openshift-dns/dns-default-7vqhc" Nov 25 09:26:47 crc kubenswrapper[4776]: W1125 09:26:47.772309 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e1c3eef_d1b8_40c8_bbad_6e92ea15732e.slice/crio-87da58d67a8260695a83915818ab4cd5c59bdab6e9aafc9be01de5d91ef17f46 WatchSource:0}: Error finding container 87da58d67a8260695a83915818ab4cd5c59bdab6e9aafc9be01de5d91ef17f46: Status 404 returned error can't find the container with id 87da58d67a8260695a83915818ab4cd5c59bdab6e9aafc9be01de5d91ef17f46 Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.772990 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:47 crc kubenswrapper[4776]: E1125 09:26:47.778368 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:48.278350153 +0000 UTC m=+153.319409766 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.798407 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kx5ld\" (UniqueName: \"kubernetes.io/projected/8b743ab8-67de-493a-bdf2-c0682a192869-kube-api-access-kx5ld\") pod \"package-server-manager-789f6589d5-hwxqt\" (UID: \"8b743ab8-67de-493a-bdf2-c0682a192869\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwxqt" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.806544 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5slr6\" (UniqueName: \"kubernetes.io/projected/e932b24b-a48e-47c4-bb74-779fd2ec8ea9-kube-api-access-5slr6\") pod \"service-ca-operator-777779d784-mv7lp\" (UID: \"e932b24b-a48e-47c4-bb74-779fd2ec8ea9\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mv7lp" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.815394 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dlv77"] Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.819850 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.819908 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:26:47 crc kubenswrapper[4776]: W1125 09:26:47.820212 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod049123ea_7dfd_43a3_b65c_9fb0c7b6c40b.slice/crio-0b0f4cf88a7df1ce1ce9b27da6eb32eed694fd0395ef6ef9b343723b6a4ef82e WatchSource:0}: Error finding container 0b0f4cf88a7df1ce1ce9b27da6eb32eed694fd0395ef6ef9b343723b6a4ef82e: Status 404 returned error can't find the container with id 0b0f4cf88a7df1ce1ce9b27da6eb32eed694fd0395ef6ef9b343723b6a4ef82e Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.825895 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb6w2\" (UniqueName: \"kubernetes.io/projected/ae2cfb63-dc76-40c2-9070-d5c29a6a272d-kube-api-access-sb6w2\") pod \"csi-hostpathplugin-twswh\" (UID: \"ae2cfb63-dc76-40c2-9070-d5c29a6a272d\") " pod="hostpath-provisioner/csi-hostpathplugin-twswh" Nov 25 09:26:47 crc kubenswrapper[4776]: W1125 09:26:47.828961 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1411e81d_100d_44a3_957f_5cc5a89e7748.slice/crio-2cef61c12bd6086b63d509b326bab97cba0b321e77902b1903710c33004deb71 WatchSource:0}: Error finding container 2cef61c12bd6086b63d509b326bab97cba0b321e77902b1903710c33004deb71: Status 404 returned error can't find the container with id 2cef61c12bd6086b63d509b326bab97cba0b321e77902b1903710c33004deb71 Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.856966 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qm8bb\" (UniqueName: \"kubernetes.io/projected/e8765bc6-c8e3-40d7-a061-211b1cad5df2-kube-api-access-qm8bb\") pod \"marketplace-operator-79b997595-5n2sp\" (UID: \"e8765bc6-c8e3-40d7-a061-211b1cad5df2\") " pod="openshift-marketplace/marketplace-operator-79b997595-5n2sp" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.871036 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-8nrlj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.874255 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:47 crc kubenswrapper[4776]: E1125 09:26:47.874585 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:48.37457013 +0000 UTC m=+153.415629683 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.874779 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5a01b028-a86d-40d3-9633-3f849f6d0840-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-h8t6d\" (UID: \"5a01b028-a86d-40d3-9633-3f849f6d0840\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h8t6d" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.877060 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwxqt" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.885030 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mv7lp" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.886746 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z627v\" (UniqueName: \"kubernetes.io/projected/df8b1df7-103a-49ba-855e-face92da97dd-kube-api-access-z627v\") pod \"machine-config-server-xj5z4\" (UID: \"df8b1df7-103a-49ba-855e-face92da97dd\") " pod="openshift-machine-config-operator/machine-config-server-xj5z4" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.890504 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nzcz7" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.895297 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-m9xnc" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.898025 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjvk6\" (UniqueName: \"kubernetes.io/projected/a9a85890-83ba-4e99-bdbe-386e0bbbecd6-kube-api-access-bjvk6\") pod \"packageserver-d55dfcdfc-pxtkj\" (UID: \"a9a85890-83ba-4e99-bdbe-386e0bbbecd6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pxtkj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.900385 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pxtkj" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.907401 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tbbm7" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.912888 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wxwlt" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.919319 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5n2sp" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.923169 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4738a95e-6584-4c14-b545-d0e96d678ab0-bound-sa-token\") pod \"ingress-operator-5b745b69d9-g4wgk\" (UID: \"4738a95e-6584-4c14-b545-d0e96d678ab0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g4wgk" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.925773 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h8t6d" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.939884 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-tckf4" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.946246 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-7vqhc" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.952127 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-xj5z4" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.958775 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-hkhrr" Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.976180 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:47 crc kubenswrapper[4776]: E1125 09:26:47.976473 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:48.476462447 +0000 UTC m=+153.517522000 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:47 crc kubenswrapper[4776]: I1125 09:26:47.978892 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-twswh" Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.001214 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g4wgk" Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.078237 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:48 crc kubenswrapper[4776]: E1125 09:26:48.078824 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:48.578808394 +0000 UTC m=+153.619867947 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:48 crc kubenswrapper[4776]: W1125 09:26:48.109044 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02eb8668_3740_4714_98ae_e993184549aa.slice/crio-077aecff6bdc899c15b8dc6b38d5fa31bc521d3a5618d29e2541e29b1de5f665 WatchSource:0}: Error finding container 077aecff6bdc899c15b8dc6b38d5fa31bc521d3a5618d29e2541e29b1de5f665: Status 404 returned error can't find the container with id 077aecff6bdc899c15b8dc6b38d5fa31bc521d3a5618d29e2541e29b1de5f665 Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.184048 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:48 crc kubenswrapper[4776]: E1125 09:26:48.184515 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:48.684499675 +0000 UTC m=+153.725559228 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.216897 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-wzwq4"] Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.285025 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:48 crc kubenswrapper[4776]: E1125 09:26:48.285286 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:48.785257134 +0000 UTC m=+153.826316687 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.285487 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:48 crc kubenswrapper[4776]: E1125 09:26:48.285835 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:48.785827338 +0000 UTC m=+153.826886891 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.296133 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7r6h5"] Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.312237 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-twg6n"] Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.312577 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wrbxz"] Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.375769 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401035-dht2z"] Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.386630 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:48 crc kubenswrapper[4776]: E1125 09:26:48.386895 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:48.886880504 +0000 UTC m=+153.927940047 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.404854 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-f4747"] Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.416931 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pgjrp"] Nov 25 09:26:48 crc kubenswrapper[4776]: W1125 09:26:48.432915 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod471ebf6d_cd02_461d_96c9_d2b503d65cb8.slice/crio-9f1dfe913018da9c71634da03832b563d852a6b23aa564cc205e45448893a406 WatchSource:0}: Error finding container 9f1dfe913018da9c71634da03832b563d852a6b23aa564cc205e45448893a406: Status 404 returned error can't find the container with id 9f1dfe913018da9c71634da03832b563d852a6b23aa564cc205e45448893a406 Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.444925 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-rc22s"] Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.476849 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rbfl6"] Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.489758 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:48 crc kubenswrapper[4776]: E1125 09:26:48.490968 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:48.990949484 +0000 UTC m=+154.032009037 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.500865 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pxtkj"] Nov 25 09:26:48 crc kubenswrapper[4776]: W1125 09:26:48.533193 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode308a810_6256_4459_a587_d93ed4899fa8.slice/crio-ae2da9a35c9874d06a8bb2dcb123353ba03c7985ef1d657620cdf973b1149aed WatchSource:0}: Error finding container ae2da9a35c9874d06a8bb2dcb123353ba03c7985ef1d657620cdf973b1149aed: Status 404 returned error can't find the container with id ae2da9a35c9874d06a8bb2dcb123353ba03c7985ef1d657620cdf973b1149aed Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.595488 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:48 crc kubenswrapper[4776]: E1125 09:26:48.596142 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:49.096128532 +0000 UTC m=+154.137188075 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:48 crc kubenswrapper[4776]: W1125 09:26:48.606838 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda229c7a5_8a0d_451b_9b28_da9c8762044a.slice/crio-ba8adc65243bef9657edfb1459a1bde4153c5f369c5a0407bc925e968a43cc6e WatchSource:0}: Error finding container ba8adc65243bef9657edfb1459a1bde4153c5f369c5a0407bc925e968a43cc6e: Status 404 returned error can't find the container with id ba8adc65243bef9657edfb1459a1bde4153c5f369c5a0407bc925e968a43cc6e Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.689357 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4" podStartSLOduration=126.689336654 podStartE2EDuration="2m6.689336654s" podCreationTimestamp="2025-11-25 09:24:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:48.642389815 +0000 UTC m=+153.683449368" watchObservedRunningTime="2025-11-25 09:26:48.689336654 +0000 UTC m=+153.730396207" Nov 25 09:26:48 crc kubenswrapper[4776]: W1125 09:26:48.694942 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9a85890_83ba_4e99_bdbe_386e0bbbecd6.slice/crio-9a0505c9d1e919d8498f5a9694f3216d1f14b055f4758a1843274d74438aa3e9 WatchSource:0}: Error finding container 9a0505c9d1e919d8498f5a9694f3216d1f14b055f4758a1843274d74438aa3e9: Status 404 returned error can't find the container with id 9a0505c9d1e919d8498f5a9694f3216d1f14b055f4758a1843274d74438aa3e9 Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.696690 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:48 crc kubenswrapper[4776]: E1125 09:26:48.697037 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:49.197021534 +0000 UTC m=+154.238081087 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.699957 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-twg6n" event={"ID":"e88ac546-8758-4cce-ae60-e49a1dc7c5ea","Type":"ContainerStarted","Data":"72d6455dbce71e31524e43a80300b9c22519f16bcc2f873a7ec79559e4a0c21c"} Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.704914 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-qqkgv" event={"ID":"50f49648-47d6-45d8-9563-f76b7547cad6","Type":"ContainerStarted","Data":"bd734b4ce88a8856c1dd7a80a30765762626d723250ebaf448cfdf119fbb7a9f"} Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.719612 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pgjrp" event={"ID":"52e1736f-df9e-4444-bb28-4d89911000de","Type":"ContainerStarted","Data":"24d403af29208a2100dc49c7c6639091958a981b5dcb0ada1146fe9aa69c1804"} Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.723422 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-wzwq4" event={"ID":"b80070a8-0eaa-4fe8-92ca-7c9afe029e04","Type":"ContainerStarted","Data":"88b402c898ac11fd5e88c07961b5a249f1e9ba88f1a94168b5a48058ac5873f6"} Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.768490 4776 generic.go:334] "Generic (PLEG): container finished" podID="974a7107-64ce-4f92-842b-7927d3ba006d" containerID="47f51ded7071180fc82da5a05a3eac896517c301108cad3760eb20a68ab3b3a5" exitCode=0 Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.768910 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-srdjj" event={"ID":"974a7107-64ce-4f92-842b-7927d3ba006d","Type":"ContainerDied","Data":"47f51ded7071180fc82da5a05a3eac896517c301108cad3760eb20a68ab3b3a5"} Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.780376 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vpp8l" event={"ID":"2e1c3eef-d1b8-40c8-bbad-6e92ea15732e","Type":"ContainerStarted","Data":"e8cd4256541e23155119a5a32e3897e4958fc040701ebafc65950d267a4746b6"} Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.780421 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vpp8l" event={"ID":"2e1c3eef-d1b8-40c8-bbad-6e92ea15732e","Type":"ContainerStarted","Data":"87da58d67a8260695a83915818ab4cd5c59bdab6e9aafc9be01de5d91ef17f46"} Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.782514 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xh8xm" event={"ID":"1d06dda7-d25a-4d6d-93b3-9a8579876bee","Type":"ContainerStarted","Data":"51ab5b87920fb240792d3843aa2953344bae81710221f386a9609043400fbeea"} Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.786443 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-8nrlj" event={"ID":"02eb8668-3740-4714-98ae-e993184549aa","Type":"ContainerStarted","Data":"82ee6b3029378a5526cd9d3cc6134ed7a6de2b4d94d616620566b460cff49420"} Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.786475 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-8nrlj" event={"ID":"02eb8668-3740-4714-98ae-e993184549aa","Type":"ContainerStarted","Data":"077aecff6bdc899c15b8dc6b38d5fa31bc521d3a5618d29e2541e29b1de5f665"} Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.788336 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" event={"ID":"490fba04-edd0-4edc-9f1e-4e08debf284f","Type":"ContainerStarted","Data":"6ec4c7094e26a33c9942383e806dbf27649e9a5b3c66dc847760cd562f8587d7"} Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.789490 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.790448 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wrbxz" event={"ID":"471ebf6d-cd02-461d-96c9-d2b503d65cb8","Type":"ContainerStarted","Data":"9f1dfe913018da9c71634da03832b563d852a6b23aa564cc205e45448893a406"} Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.790700 4776 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-fq5rl container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.790731 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" podUID="490fba04-edd0-4edc-9f1e-4e08debf284f" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.791807 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-bhgv2" event={"ID":"04fe29b5-d8c6-46c5-9d1f-76005e225fd6","Type":"ContainerStarted","Data":"cf7f59f46490d2b9142f335ae37337163611607d48c7ed1849cc2986f0939532"} Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.797949 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:48 crc kubenswrapper[4776]: E1125 09:26:48.799238 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:49.299220928 +0000 UTC m=+154.340280481 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.802041 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-fjr5v" event={"ID":"fffe85b6-294d-4f43-ad27-65a5d093c076","Type":"ContainerStarted","Data":"ec0791af7e649845923f3e2700f3c60e7fa4f5e694f074c1f3fdd7f7bc8d4335"} Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.804674 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f4747" event={"ID":"e308a810-6256-4459-a587-d93ed4899fa8","Type":"ContainerStarted","Data":"ae2da9a35c9874d06a8bb2dcb123353ba03c7985ef1d657620cdf973b1149aed"} Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.811459 4776 generic.go:334] "Generic (PLEG): container finished" podID="be1426b0-cf02-44d3-87b8-aa0394af1afa" containerID="cb5b31dcadef8b66da36b9e292724b0893c039ed8e3c414e0890a58da0756fef" exitCode=0 Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.811524 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" event={"ID":"be1426b0-cf02-44d3-87b8-aa0394af1afa","Type":"ContainerDied","Data":"cb5b31dcadef8b66da36b9e292724b0893c039ed8e3c414e0890a58da0756fef"} Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.846835 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-xj5z4" event={"ID":"df8b1df7-103a-49ba-855e-face92da97dd","Type":"ContainerStarted","Data":"06ef83ba6b7dd8cc48faa2baa5fe6ef7823b38c12cc3421bb452ae876b5fe35b"} Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.853007 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pxpfx" event={"ID":"10d2d0a3-eaba-486f-a833-0da30ad1f18e","Type":"ContainerStarted","Data":"47f0d3a96137743903ade716e23ac1a87cab18b5488537d6903bcea42c286047"} Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.868770 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401035-dht2z" event={"ID":"4b51fc9b-c461-4e80-9a56-4f0e43e02399","Type":"ContainerStarted","Data":"8f34a2a36c30d5862e517fa9c193c6f488947582dd60db9420d739b1aff92ede"} Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.876151 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-8nrlj" Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.878172 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7r6h5" event={"ID":"d2aa408d-c0e1-4697-b9e8-dee9276e8b47","Type":"ContainerStarted","Data":"e825bc89451c5edac6d2eb29c4a446bf5d91ef67da4ea921ca99d65a09693f93"} Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.880184 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rbfl6" event={"ID":"a229c7a5-8a0d-451b-9b28-da9c8762044a","Type":"ContainerStarted","Data":"ba8adc65243bef9657edfb1459a1bde4153c5f369c5a0407bc925e968a43cc6e"} Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.891094 4776 generic.go:334] "Generic (PLEG): container finished" podID="049123ea-7dfd-43a3-b65c-9fb0c7b6c40b" containerID="fe8df474d3859122e1f63602f4f906307d894d002a82a08e710829bf918d8408" exitCode=0 Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.891197 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x9k2l" event={"ID":"049123ea-7dfd-43a3-b65c-9fb0c7b6c40b","Type":"ContainerDied","Data":"fe8df474d3859122e1f63602f4f906307d894d002a82a08e710829bf918d8408"} Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.891272 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x9k2l" event={"ID":"049123ea-7dfd-43a3-b65c-9fb0c7b6c40b","Type":"ContainerStarted","Data":"0b0f4cf88a7df1ce1ce9b27da6eb32eed694fd0395ef6ef9b343723b6a4ef82e"} Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.900804 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dlv77" event={"ID":"1411e81d-100d-44a3-957f-5cc5a89e7748","Type":"ContainerStarted","Data":"4664b6c0fd2120a5547f5629421cc7fa8856846f09e021ad19d63e3f4c7c2813"} Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.900855 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dlv77" event={"ID":"1411e81d-100d-44a3-957f-5cc5a89e7748","Type":"ContainerStarted","Data":"2cef61c12bd6086b63d509b326bab97cba0b321e77902b1903710c33004deb71"} Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.901752 4776 patch_prober.go:28] interesting pod/downloads-7954f5f757-r979m container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.901786 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r979m" podUID="529b4e1d-e99c-4342-8cff-68c1e31a776f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.902824 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:48 crc kubenswrapper[4776]: E1125 09:26:48.904187 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:49.40417124 +0000 UTC m=+154.445230793 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.911471 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4" Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.954319 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-r979m" podStartSLOduration=127.954304759 podStartE2EDuration="2m7.954304759s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:48.918852853 +0000 UTC m=+153.959912406" watchObservedRunningTime="2025-11-25 09:26:48.954304759 +0000 UTC m=+153.995364312" Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.991317 4776 patch_prober.go:28] interesting pod/router-default-5444994796-8nrlj container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Nov 25 09:26:48 crc kubenswrapper[4776]: I1125 09:26:48.991569 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nrlj" podUID="02eb8668-3740-4714-98ae-e993184549aa" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.007108 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.008216 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" podStartSLOduration=128.00820112 podStartE2EDuration="2m8.00820112s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:49.007530513 +0000 UTC m=+154.048590066" watchObservedRunningTime="2025-11-25 09:26:49.00820112 +0000 UTC m=+154.049260673" Nov 25 09:26:49 crc kubenswrapper[4776]: E1125 09:26:49.008853 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:49.508829115 +0000 UTC m=+154.549888668 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.012481 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:49 crc kubenswrapper[4776]: E1125 09:26:49.018159 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:49.518141925 +0000 UTC m=+154.559201578 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.050608 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-mt26t" podStartSLOduration=128.050590587 podStartE2EDuration="2m8.050590587s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:49.048597088 +0000 UTC m=+154.089656641" watchObservedRunningTime="2025-11-25 09:26:49.050590587 +0000 UTC m=+154.091650140" Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.114115 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:49 crc kubenswrapper[4776]: E1125 09:26:49.114400 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:49.614386132 +0000 UTC m=+154.655445685 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.116698 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-mt26t" Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.217967 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:49 crc kubenswrapper[4776]: E1125 09:26:49.218668 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:49.718655198 +0000 UTC m=+154.759714751 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.244883 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-sb2sq" podStartSLOduration=128.244863065 podStartE2EDuration="2m8.244863065s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:49.244119377 +0000 UTC m=+154.285178930" watchObservedRunningTime="2025-11-25 09:26:49.244863065 +0000 UTC m=+154.285922618" Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.319435 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:49 crc kubenswrapper[4776]: E1125 09:26:49.320511 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:49.820478143 +0000 UTC m=+154.861537696 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.327830 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h8t6d"] Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.330163 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-7vqhc"] Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.350130 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-mv7lp"] Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.365902 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-g4wgk"] Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.421714 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:49 crc kubenswrapper[4776]: E1125 09:26:49.422031 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:49.922020201 +0000 UTC m=+154.963079754 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.443785 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tbbm7"] Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.468541 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-tckf4"] Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.475914 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-hkhrr"] Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.477899 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5n2sp"] Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.525541 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:49 crc kubenswrapper[4776]: E1125 09:26:49.525845 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:50.025830705 +0000 UTC m=+155.066890258 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:49 crc kubenswrapper[4776]: W1125 09:26:49.539227 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4bb9001d_7d85_4cad_8587_e47e6f1c7aa9.slice/crio-b2951288685bb4a3cf08b9306bd6a10a9211cf17edacd9e220966528e0655b09 WatchSource:0}: Error finding container b2951288685bb4a3cf08b9306bd6a10a9211cf17edacd9e220966528e0655b09: Status 404 returned error can't find the container with id b2951288685bb4a3cf08b9306bd6a10a9211cf17edacd9e220966528e0655b09 Nov 25 09:26:49 crc kubenswrapper[4776]: W1125 09:26:49.540878 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09269fbc_f59c_42cc_a832_2285987526db.slice/crio-9f1fbfe05e8503eb42352566cd4500fded4bcc21bda373b56826fb58b105a510 WatchSource:0}: Error finding container 9f1fbfe05e8503eb42352566cd4500fded4bcc21bda373b56826fb58b105a510: Status 404 returned error can't find the container with id 9f1fbfe05e8503eb42352566cd4500fded4bcc21bda373b56826fb58b105a510 Nov 25 09:26:49 crc kubenswrapper[4776]: W1125 09:26:49.546325 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod337ebdb5_efc9_4bcb_8630_ab09a36c5bec.slice/crio-5300e50d34e2a52f1dd3a3e81c3985a051137185d999a958ab7467776ccfff71 WatchSource:0}: Error finding container 5300e50d34e2a52f1dd3a3e81c3985a051137185d999a958ab7467776ccfff71: Status 404 returned error can't find the container with id 5300e50d34e2a52f1dd3a3e81c3985a051137185d999a958ab7467776ccfff71 Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.589262 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-m9xnc"] Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.606848 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-xh8xm" podStartSLOduration=128.606828044 podStartE2EDuration="2m8.606828044s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:49.597198876 +0000 UTC m=+154.638258429" watchObservedRunningTime="2025-11-25 09:26:49.606828044 +0000 UTC m=+154.647887597" Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.608630 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwxqt"] Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.617507 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-twswh"] Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.619278 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nzcz7"] Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.623326 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wxwlt"] Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.627114 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:49 crc kubenswrapper[4776]: E1125 09:26:49.630170 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:50.1301537 +0000 UTC m=+155.171213253 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.648603 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:26:49 crc kubenswrapper[4776]: W1125 09:26:49.698243 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod025d75af_9ed0_4122_b97e_82dd1abd66fe.slice/crio-6c2930a28eed833b3a4017584a055f701ab7826f09b8c63e8b5b1b5d8fe7c6b1 WatchSource:0}: Error finding container 6c2930a28eed833b3a4017584a055f701ab7826f09b8c63e8b5b1b5d8fe7c6b1: Status 404 returned error can't find the container with id 6c2930a28eed833b3a4017584a055f701ab7826f09b8c63e8b5b1b5d8fe7c6b1 Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.720593 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" podStartSLOduration=128.720578164 podStartE2EDuration="2m8.720578164s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:49.718514653 +0000 UTC m=+154.759574206" watchObservedRunningTime="2025-11-25 09:26:49.720578164 +0000 UTC m=+154.761637717" Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.727609 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:49 crc kubenswrapper[4776]: E1125 09:26:49.727944 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:50.227927795 +0000 UTC m=+155.268987348 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:49 crc kubenswrapper[4776]: W1125 09:26:49.728685 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcf21b2c5_1b6a_458b_9abd_96b309d1626d.slice/crio-6c15ce6994da0cb2c64b7aeead0f27b68b3f004f7494879713af02470a9b6689 WatchSource:0}: Error finding container 6c15ce6994da0cb2c64b7aeead0f27b68b3f004f7494879713af02470a9b6689: Status 404 returned error can't find the container with id 6c15ce6994da0cb2c64b7aeead0f27b68b3f004f7494879713af02470a9b6689 Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.761290 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-bhgv2" podStartSLOduration=128.761273249 podStartE2EDuration="2m8.761273249s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:49.7588853 +0000 UTC m=+154.799944853" watchObservedRunningTime="2025-11-25 09:26:49.761273249 +0000 UTC m=+154.802332802" Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.831643 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:49 crc kubenswrapper[4776]: E1125 09:26:49.831930 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:50.331918254 +0000 UTC m=+155.372977807 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.873607 4776 patch_prober.go:28] interesting pod/router-default-5444994796-8nrlj container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.873876 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nrlj" podUID="02eb8668-3740-4714-98ae-e993184549aa" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.933748 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:49 crc kubenswrapper[4776]: E1125 09:26:49.934209 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:50.43418886 +0000 UTC m=+155.475248413 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.949092 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-8nrlj" podStartSLOduration=128.949077027 podStartE2EDuration="2m8.949077027s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:49.947677183 +0000 UTC m=+154.988736736" watchObservedRunningTime="2025-11-25 09:26:49.949077027 +0000 UTC m=+154.990136580" Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.949718 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rbfl6" event={"ID":"a229c7a5-8a0d-451b-9b28-da9c8762044a","Type":"ContainerStarted","Data":"342330d5b39b81bf7d4909e95abbc9456ae6b9e0b20bcac20ab888027ad1a6c9"} Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.954879 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rc22s" event={"ID":"8003bc91-3e3b-4e88-8942-3ec54fc30aef","Type":"ContainerStarted","Data":"9519aae80067951fbe0323e1c80d2fd343e7a5f0d9225679ff96938aaa3132f4"} Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.954905 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rc22s" event={"ID":"8003bc91-3e3b-4e88-8942-3ec54fc30aef","Type":"ContainerStarted","Data":"6c07441cd14673d762c5bf055db2b018fc2f220b413459ca8744e771d21d7e1f"} Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.956118 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wxwlt" event={"ID":"025d75af-9ed0-4122-b97e-82dd1abd66fe","Type":"ContainerStarted","Data":"6c2930a28eed833b3a4017584a055f701ab7826f09b8c63e8b5b1b5d8fe7c6b1"} Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.957579 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tbbm7" event={"ID":"4bb9001d-7d85-4cad-8587-e47e6f1c7aa9","Type":"ContainerStarted","Data":"b2951288685bb4a3cf08b9306bd6a10a9211cf17edacd9e220966528e0655b09"} Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.979531 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mv7lp" event={"ID":"e932b24b-a48e-47c4-bb74-779fd2ec8ea9","Type":"ContainerStarted","Data":"4f24ba11cc556182eb49535babe02e0e97e1a26c21dc05960988a7870b9b668d"} Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.979582 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mv7lp" event={"ID":"e932b24b-a48e-47c4-bb74-779fd2ec8ea9","Type":"ContainerStarted","Data":"5a2806d1a18121944d6d0e2525eff616a274f0668ef167104c6bee2e858febe9"} Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.981425 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pgjrp" event={"ID":"52e1736f-df9e-4444-bb28-4d89911000de","Type":"ContainerStarted","Data":"c80f295ef743b0bf65bba762a507bec100d4ddf92eca1b6d43f03f94ae591fea"} Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.982831 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h8t6d" event={"ID":"5a01b028-a86d-40d3-9633-3f849f6d0840","Type":"ContainerStarted","Data":"4eb4afcd392af30d05813b5456bac073e777fbdcfca771783ff7b9a40f447524"} Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.985744 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x9k2l" event={"ID":"049123ea-7dfd-43a3-b65c-9fb0c7b6c40b","Type":"ContainerStarted","Data":"6b518d7d8963a3c242a5efccd5740f5d5289ca50e666cbcf24fc254a5b3c8013"} Nov 25 09:26:49 crc kubenswrapper[4776]: I1125 09:26:49.985931 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x9k2l" Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.019787 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pxtkj" event={"ID":"a9a85890-83ba-4e99-bdbe-386e0bbbecd6","Type":"ContainerStarted","Data":"75e6319ec31a09433df8cf649f2e0f79f1a304c1975e72f7202429268e1b805a"} Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.019834 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pxtkj" event={"ID":"a9a85890-83ba-4e99-bdbe-386e0bbbecd6","Type":"ContainerStarted","Data":"9a0505c9d1e919d8498f5a9694f3216d1f14b055f4758a1843274d74438aa3e9"} Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.020407 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pxtkj" Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.021894 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwxqt" event={"ID":"8b743ab8-67de-493a-bdf2-c0682a192869","Type":"ContainerStarted","Data":"e27d3c24d9b9f613d41df16709fbc5bff98e63a361331ea5007ccb649b3a3244"} Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.023324 4776 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-pxtkj container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" start-of-body= Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.023384 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pxtkj" podUID="a9a85890-83ba-4e99-bdbe-386e0bbbecd6" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.023552 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-vpp8l" podStartSLOduration=129.023532476 podStartE2EDuration="2m9.023532476s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:50.021395934 +0000 UTC m=+155.062455487" watchObservedRunningTime="2025-11-25 09:26:50.023532476 +0000 UTC m=+155.064592029" Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.023604 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5n2sp" event={"ID":"e8765bc6-c8e3-40d7-a061-211b1cad5df2","Type":"ContainerStarted","Data":"c03b771376224415a70b8bf2e6d4b61d05e29b7ba7909208777791fbe3b54f5c"} Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.033683 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-hkhrr" event={"ID":"09269fbc-f59c-42cc-a832-2285987526db","Type":"ContainerStarted","Data":"9f1fbfe05e8503eb42352566cd4500fded4bcc21bda373b56826fb58b105a510"} Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.034741 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:50 crc kubenswrapper[4776]: E1125 09:26:50.035840 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:50.5358248 +0000 UTC m=+155.576884353 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.062054 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nzcz7" event={"ID":"cf21b2c5-1b6a-458b-9abd-96b309d1626d","Type":"ContainerStarted","Data":"6c15ce6994da0cb2c64b7aeead0f27b68b3f004f7494879713af02470a9b6689"} Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.063519 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-tckf4" event={"ID":"337ebdb5-efc9-4bcb-8630-ab09a36c5bec","Type":"ContainerStarted","Data":"5300e50d34e2a52f1dd3a3e81c3985a051137185d999a958ab7467776ccfff71"} Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.065968 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-wzwq4" event={"ID":"b80070a8-0eaa-4fe8-92ca-7c9afe029e04","Type":"ContainerStarted","Data":"e2a6d82eb570021140be2ac4dcc4fc469c84a64c6d8c0345b12b2bd954abeaaa"} Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.065997 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-wzwq4" event={"ID":"b80070a8-0eaa-4fe8-92ca-7c9afe029e04","Type":"ContainerStarted","Data":"8aacd7230108c80e4962dba0bc5731d10ef656d1525b5517006e06a3b6104db4"} Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.072872 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7r6h5" event={"ID":"d2aa408d-c0e1-4697-b9e8-dee9276e8b47","Type":"ContainerStarted","Data":"ed9f3551af686f9d1b628d91480249110f87bbbdaefe9bd5e7c6c298323b0d4f"} Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.075251 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f4747" event={"ID":"e308a810-6256-4459-a587-d93ed4899fa8","Type":"ContainerStarted","Data":"82d9960ffc068beb1e30daa00c6cec53ce217d38c56608112d95d88f4b1a7580"} Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.077549 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wrbxz" event={"ID":"471ebf6d-cd02-461d-96c9-d2b503d65cb8","Type":"ContainerStarted","Data":"0f8d2e2e4a4dbcfa56a13e7f706fb64b0d96399227e6f0734cd1d45f58162cdd"} Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.085257 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-twswh" event={"ID":"ae2cfb63-dc76-40c2-9070-d5c29a6a272d","Type":"ContainerStarted","Data":"983dc68d7955085dd30a4055277977c563b60c38548b4b242be978a8ecfad675"} Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.088499 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-7vqhc" event={"ID":"f0ed4ec3-141e-413a-8229-86baf2265ece","Type":"ContainerStarted","Data":"9e7058d50d6296ec788bf2b81415e54ac0684a8d25e403e562391930fa49da26"} Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.098720 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-xj5z4" event={"ID":"df8b1df7-103a-49ba-855e-face92da97dd","Type":"ContainerStarted","Data":"b1d2ad261921e72a782e73fdd0df07581f387dc73605ef13cb1eafd00e52bcbd"} Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.102727 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-srdjj" event={"ID":"974a7107-64ce-4f92-842b-7927d3ba006d","Type":"ContainerStarted","Data":"38f8dac1a196c74f050300a2405d3f33d99e165a20284590a05e4939376f615a"} Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.109316 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g4wgk" event={"ID":"4738a95e-6584-4c14-b545-d0e96d678ab0","Type":"ContainerStarted","Data":"c395f905b448cd3ab7c8bb6f8b7faf08c25c6fd9a8abd8c7950e215dd39af664"} Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.113414 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" event={"ID":"be1426b0-cf02-44d3-87b8-aa0394af1afa","Type":"ContainerStarted","Data":"04a9b1c60b93d309fa7156c913e590e67ff5dbf30f0e4db01df858b9ec8f0d03"} Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.123331 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401035-dht2z" event={"ID":"4b51fc9b-c461-4e80-9a56-4f0e43e02399","Type":"ContainerStarted","Data":"3f95652774809f6ee88562122ca9c3788000edd112955de44fc0382bad0e8c84"} Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.129819 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-twg6n" event={"ID":"e88ac546-8758-4cce-ae60-e49a1dc7c5ea","Type":"ContainerStarted","Data":"8684aa1f4f570ca3f1430feea6abea4f7778bd3b22f4c9f2af15ecc4fc18f14e"} Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.136306 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-m9xnc" event={"ID":"8ba97f9b-a0a1-4514-8577-0b36eb65bbd3","Type":"ContainerStarted","Data":"168cd4ef80a41a0612dea7090c2d4cf78543bf9aa9bf1c215416dfdf1e711143"} Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.139559 4776 patch_prober.go:28] interesting pod/downloads-7954f5f757-r979m container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.139593 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r979m" podUID="529b4e1d-e99c-4342-8cff-68c1e31a776f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.139613 4776 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-fq5rl container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.139662 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" podUID="490fba04-edd0-4edc-9f1e-4e08debf284f" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Nov 25 09:26:50 crc kubenswrapper[4776]: E1125 09:26:50.140942 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:50.640925176 +0000 UTC m=+155.681984729 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.141035 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.141243 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:50 crc kubenswrapper[4776]: E1125 09:26:50.142609 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:50.642597687 +0000 UTC m=+155.683657240 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.241752 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:50 crc kubenswrapper[4776]: E1125 09:26:50.245354 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:50.745339214 +0000 UTC m=+155.786398767 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.305134 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-qqkgv" podStartSLOduration=129.305119081 podStartE2EDuration="2m9.305119081s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:50.301272466 +0000 UTC m=+155.342332019" watchObservedRunningTime="2025-11-25 09:26:50.305119081 +0000 UTC m=+155.346178634" Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.348497 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:50 crc kubenswrapper[4776]: E1125 09:26:50.348809 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:50.8487971 +0000 UTC m=+155.889856653 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.382660 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pxpfx" podStartSLOduration=129.382636776 podStartE2EDuration="2m9.382636776s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:50.322623173 +0000 UTC m=+155.363682736" watchObservedRunningTime="2025-11-25 09:26:50.382636776 +0000 UTC m=+155.423696329" Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.383190 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dlv77" podStartSLOduration=129.383184469 podStartE2EDuration="2m9.383184469s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:50.361151245 +0000 UTC m=+155.402210838" watchObservedRunningTime="2025-11-25 09:26:50.383184469 +0000 UTC m=+155.424244022" Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.440578 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-fjr5v" podStartSLOduration=129.440562276 podStartE2EDuration="2m9.440562276s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:50.427895773 +0000 UTC m=+155.468955326" watchObservedRunningTime="2025-11-25 09:26:50.440562276 +0000 UTC m=+155.481621829" Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.448922 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:50 crc kubenswrapper[4776]: E1125 09:26:50.449282 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:50.949269311 +0000 UTC m=+155.990328864 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.544360 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pxtkj" podStartSLOduration=129.544321519 podStartE2EDuration="2m9.544321519s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:50.541834907 +0000 UTC m=+155.582894470" watchObservedRunningTime="2025-11-25 09:26:50.544321519 +0000 UTC m=+155.585381072" Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.550556 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:50 crc kubenswrapper[4776]: E1125 09:26:50.550907 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:51.050893951 +0000 UTC m=+156.091953504 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.635655 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-7r6h5" podStartSLOduration=129.635634144 podStartE2EDuration="2m9.635634144s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:50.580293087 +0000 UTC m=+155.621352640" watchObservedRunningTime="2025-11-25 09:26:50.635634144 +0000 UTC m=+155.676693697" Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.651633 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:50 crc kubenswrapper[4776]: E1125 09:26:50.652006 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:51.151991238 +0000 UTC m=+156.193050791 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.661639 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rbfl6" podStartSLOduration=129.661621326 podStartE2EDuration="2m9.661621326s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:50.640264699 +0000 UTC m=+155.681324252" watchObservedRunningTime="2025-11-25 09:26:50.661621326 +0000 UTC m=+155.702680889" Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.662885 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x9k2l" podStartSLOduration=129.662876767 podStartE2EDuration="2m9.662876767s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:50.66136623 +0000 UTC m=+155.702425783" watchObservedRunningTime="2025-11-25 09:26:50.662876767 +0000 UTC m=+155.703936320" Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.697489 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29401035-dht2z" podStartSLOduration=129.697471892 podStartE2EDuration="2m9.697471892s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:50.696269112 +0000 UTC m=+155.737328675" watchObservedRunningTime="2025-11-25 09:26:50.697471892 +0000 UTC m=+155.738531445" Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.753650 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:50 crc kubenswrapper[4776]: E1125 09:26:50.754037 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:51.254023008 +0000 UTC m=+156.295082571 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.856491 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:50 crc kubenswrapper[4776]: E1125 09:26:50.856937 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:51.35691926 +0000 UTC m=+156.397978813 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.877497 4776 patch_prober.go:28] interesting pod/router-default-5444994796-8nrlj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:26:50 crc kubenswrapper[4776]: [-]has-synced failed: reason withheld Nov 25 09:26:50 crc kubenswrapper[4776]: [+]process-running ok Nov 25 09:26:50 crc kubenswrapper[4776]: healthz check failed Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.877573 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nrlj" podUID="02eb8668-3740-4714-98ae-e993184549aa" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:26:50 crc kubenswrapper[4776]: I1125 09:26:50.958089 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:50 crc kubenswrapper[4776]: E1125 09:26:50.958474 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:51.458439477 +0000 UTC m=+156.499499030 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.058826 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:51 crc kubenswrapper[4776]: E1125 09:26:51.059684 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:51.559663057 +0000 UTC m=+156.600722610 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.159554 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-twg6n" event={"ID":"e88ac546-8758-4cce-ae60-e49a1dc7c5ea","Type":"ContainerStarted","Data":"91877d149352d3c211e81e6a4ebd2d28859fc0c6779c2c435055bc9ea97a1cd8"} Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.160703 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:51 crc kubenswrapper[4776]: E1125 09:26:51.161197 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:51.661182685 +0000 UTC m=+156.702242238 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.174921 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rc22s" event={"ID":"8003bc91-3e3b-4e88-8942-3ec54fc30aef","Type":"ContainerStarted","Data":"5e98cd09ed2b8c1f711216597f50144b86855e1b66dea7a0a7123f1602d2f8be"} Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.179924 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g4wgk" event={"ID":"4738a95e-6584-4c14-b545-d0e96d678ab0","Type":"ContainerStarted","Data":"a290bffbe2583c16f3e3f4aa5a572c6d14e251fbc181780af7b9404e740438d4"} Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.179978 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g4wgk" event={"ID":"4738a95e-6584-4c14-b545-d0e96d678ab0","Type":"ContainerStarted","Data":"084a164606467571164707f63246f2f6f5f38f92350c6213e533a2d0371751ba"} Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.190681 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-twg6n" podStartSLOduration=130.190661863 podStartE2EDuration="2m10.190661863s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:51.188938271 +0000 UTC m=+156.229997824" watchObservedRunningTime="2025-11-25 09:26:51.190661863 +0000 UTC m=+156.231721436" Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.194765 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-srdjj" event={"ID":"974a7107-64ce-4f92-842b-7927d3ba006d","Type":"ContainerStarted","Data":"8577e53fa42483be5c96d6edd87fa7e5616f91a43be56c0684671d0181ce99db"} Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.210488 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rc22s" podStartSLOduration=130.210469402 podStartE2EDuration="2m10.210469402s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:51.209358645 +0000 UTC m=+156.250418198" watchObservedRunningTime="2025-11-25 09:26:51.210469402 +0000 UTC m=+156.251528955" Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.227809 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pgjrp" event={"ID":"52e1736f-df9e-4444-bb28-4d89911000de","Type":"ContainerStarted","Data":"f504954628b4f573cc9d01532a08e2b8380cb9f76a2c05dba7dd3ab21b2cbbc2"} Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.228829 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g4wgk" podStartSLOduration=130.228815196 podStartE2EDuration="2m10.228815196s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:51.226316924 +0000 UTC m=+156.267376467" watchObservedRunningTime="2025-11-25 09:26:51.228815196 +0000 UTC m=+156.269874749" Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.242423 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h8t6d" event={"ID":"5a01b028-a86d-40d3-9633-3f849f6d0840","Type":"ContainerStarted","Data":"42dbf7c948ab011a0e15a5e798f3a8cbffdfc1286fb1a8d5ebdf7f01e2e9cdfd"} Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.253101 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-tckf4" event={"ID":"337ebdb5-efc9-4bcb-8630-ab09a36c5bec","Type":"ContainerStarted","Data":"0701b2335ff2ef41ddc21164e1e6b2729dab5100787409d980db9a52dd0bb77e"} Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.262586 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:51 crc kubenswrapper[4776]: E1125 09:26:51.264102 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:51.764083947 +0000 UTC m=+156.805143500 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.277318 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wxwlt" event={"ID":"025d75af-9ed0-4122-b97e-82dd1abd66fe","Type":"ContainerStarted","Data":"1fd4d055d7cb881233d5d053fe916cf05db44ec079c339dc3aa8000bdf4706ec"} Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.278095 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wxwlt" Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.283167 4776 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-wxwlt container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" start-of-body= Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.283218 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wxwlt" podUID="025d75af-9ed0-4122-b97e-82dd1abd66fe" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.308258 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tbbm7" event={"ID":"4bb9001d-7d85-4cad-8587-e47e6f1c7aa9","Type":"ContainerStarted","Data":"24ba53b2d49e2646d1fe20137369e775fe127307ef226eba1b8dae84282c50f5"} Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.333170 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-tckf4" podStartSLOduration=129.333151602 podStartE2EDuration="2m9.333151602s" podCreationTimestamp="2025-11-25 09:24:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:51.332465106 +0000 UTC m=+156.373524659" watchObservedRunningTime="2025-11-25 09:26:51.333151602 +0000 UTC m=+156.374211155" Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.333920 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-srdjj" podStartSLOduration=130.333911591 podStartE2EDuration="2m10.333911591s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:51.29982844 +0000 UTC m=+156.340887993" watchObservedRunningTime="2025-11-25 09:26:51.333911591 +0000 UTC m=+156.374971144" Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.350304 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-m9xnc" event={"ID":"8ba97f9b-a0a1-4514-8577-0b36eb65bbd3","Type":"ContainerStarted","Data":"ea3235bbc628212df1fcb5b59fc828c0f10e55ec8ef31a735ea92e478aa03654"} Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.358999 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wxwlt" podStartSLOduration=130.358986011 podStartE2EDuration="2m10.358986011s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:51.355293879 +0000 UTC m=+156.396353432" watchObservedRunningTime="2025-11-25 09:26:51.358986011 +0000 UTC m=+156.400045564" Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.370881 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:51 crc kubenswrapper[4776]: E1125 09:26:51.372458 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:51.872446883 +0000 UTC m=+156.913506436 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.375629 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwxqt" event={"ID":"8b743ab8-67de-493a-bdf2-c0682a192869","Type":"ContainerStarted","Data":"7a378b5d4a6ef2b11601de3fd6d15a502628a399197bf62b53910d4db4b126e3"} Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.375670 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwxqt" event={"ID":"8b743ab8-67de-493a-bdf2-c0682a192869","Type":"ContainerStarted","Data":"0aeefacafe6fcc62146a21ef433946d8605c63ef7976d42095142813aadccfd4"} Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.376235 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwxqt" Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.389149 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5n2sp" event={"ID":"e8765bc6-c8e3-40d7-a061-211b1cad5df2","Type":"ContainerStarted","Data":"aa1237b8018ef898ff3df28567501bf0a4db0a98d91104b753dc8e81f603467a"} Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.389941 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-5n2sp" Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.391634 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-pgjrp" podStartSLOduration=130.391625107 podStartE2EDuration="2m10.391625107s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:51.390264383 +0000 UTC m=+156.431323936" watchObservedRunningTime="2025-11-25 09:26:51.391625107 +0000 UTC m=+156.432684660" Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.399535 4776 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5n2sp container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.399575 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5n2sp" podUID="e8765bc6-c8e3-40d7-a061-211b1cad5df2" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.404632 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f4747" event={"ID":"e308a810-6256-4459-a587-d93ed4899fa8","Type":"ContainerStarted","Data":"39210e4cd2ddbd76083bfe02c31847cb79ebd38256f59c9f94ddb9a74e5d31ab"} Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.417768 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-hkhrr" event={"ID":"09269fbc-f59c-42cc-a832-2285987526db","Type":"ContainerStarted","Data":"e29bcfdf53e3dce33675c6a649e4e00056077b61748be11ca9c4f6b01da007c9"} Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.441728 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-7vqhc" event={"ID":"f0ed4ec3-141e-413a-8229-86baf2265ece","Type":"ContainerStarted","Data":"4b593c1ef1d6b5aa48dad34d4c571a6c9ba45dd5542c554d54cf124d4f046a5a"} Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.441779 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-7vqhc" event={"ID":"f0ed4ec3-141e-413a-8229-86baf2265ece","Type":"ContainerStarted","Data":"5b6dddebaa782d76a00b9f226b74fd9dbde56063472af3345cabc05bed292c7b"} Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.442180 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-7vqhc" Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.462930 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nzcz7" event={"ID":"cf21b2c5-1b6a-458b-9abd-96b309d1626d","Type":"ContainerStarted","Data":"03c616103210ff531f847014c5c1848a9884f263b6bf109ad663c8c5dcb5a33c"} Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.463687 4776 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-pxtkj container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" start-of-body= Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.463723 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pxtkj" podUID="a9a85890-83ba-4e99-bdbe-386e0bbbecd6" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.464794 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-h8t6d" podStartSLOduration=130.464774103 podStartE2EDuration="2m10.464774103s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:51.42656859 +0000 UTC m=+156.467628143" watchObservedRunningTime="2025-11-25 09:26:51.464774103 +0000 UTC m=+156.505833656" Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.479845 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:51 crc kubenswrapper[4776]: E1125 09:26:51.481401 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:51.981379834 +0000 UTC m=+157.022439377 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.509778 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwxqt" podStartSLOduration=130.509755604 podStartE2EDuration="2m10.509755604s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:51.465412229 +0000 UTC m=+156.506471912" watchObservedRunningTime="2025-11-25 09:26:51.509755604 +0000 UTC m=+156.550815157" Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.511716 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-hkhrr" podStartSLOduration=7.511704373 podStartE2EDuration="7.511704373s" podCreationTimestamp="2025-11-25 09:26:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:51.510145584 +0000 UTC m=+156.551205147" watchObservedRunningTime="2025-11-25 09:26:51.511704373 +0000 UTC m=+156.552763926" Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.588077 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:51 crc kubenswrapper[4776]: E1125 09:26:51.591077 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:52.091043722 +0000 UTC m=+157.132103275 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.623038 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-m9xnc" podStartSLOduration=130.623019942 podStartE2EDuration="2m10.623019942s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:51.569392727 +0000 UTC m=+156.610452280" watchObservedRunningTime="2025-11-25 09:26:51.623019942 +0000 UTC m=+156.664079495" Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.692548 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:51 crc kubenswrapper[4776]: E1125 09:26:51.693204 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:52.193184055 +0000 UTC m=+157.234243608 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.748006 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f4747" podStartSLOduration=130.747990659 podStartE2EDuration="2m10.747990659s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:51.624373715 +0000 UTC m=+156.665433278" watchObservedRunningTime="2025-11-25 09:26:51.747990659 +0000 UTC m=+156.789050212" Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.748207 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tbbm7" podStartSLOduration=130.748202634 podStartE2EDuration="2m10.748202634s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:51.739436077 +0000 UTC m=+156.780495630" watchObservedRunningTime="2025-11-25 09:26:51.748202634 +0000 UTC m=+156.789262187" Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.794675 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:51 crc kubenswrapper[4776]: E1125 09:26:51.795096 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:52.295059061 +0000 UTC m=+157.336118614 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.840994 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-5n2sp" podStartSLOduration=130.840972955 podStartE2EDuration="2m10.840972955s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:51.837428228 +0000 UTC m=+156.878487781" watchObservedRunningTime="2025-11-25 09:26:51.840972955 +0000 UTC m=+156.882032508" Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.875193 4776 patch_prober.go:28] interesting pod/router-default-5444994796-8nrlj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:26:51 crc kubenswrapper[4776]: [-]has-synced failed: reason withheld Nov 25 09:26:51 crc kubenswrapper[4776]: [+]process-running ok Nov 25 09:26:51 crc kubenswrapper[4776]: healthz check failed Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.875275 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nrlj" podUID="02eb8668-3740-4714-98ae-e993184549aa" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.895368 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:51 crc kubenswrapper[4776]: E1125 09:26:51.895525 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:52.395503632 +0000 UTC m=+157.436563185 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.895673 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:51 crc kubenswrapper[4776]: E1125 09:26:51.895942 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:52.395930233 +0000 UTC m=+157.436989776 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.944749 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-xj5z4" podStartSLOduration=7.944732388 podStartE2EDuration="7.944732388s" podCreationTimestamp="2025-11-25 09:26:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:51.942566835 +0000 UTC m=+156.983626388" watchObservedRunningTime="2025-11-25 09:26:51.944732388 +0000 UTC m=+156.985791941" Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.945389 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-wzwq4" podStartSLOduration=130.945380884 podStartE2EDuration="2m10.945380884s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:51.890307634 +0000 UTC m=+156.931367187" watchObservedRunningTime="2025-11-25 09:26:51.945380884 +0000 UTC m=+156.986440437" Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.953437 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.953493 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.955025 4776 patch_prober.go:28] interesting pod/apiserver-76f77b778f-srdjj container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.11:8443/livez\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.955103 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-srdjj" podUID="974a7107-64ce-4f92-842b-7927d3ba006d" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.11:8443/livez\": dial tcp 10.217.0.11:8443: connect: connection refused" Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.976938 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" podStartSLOduration=129.976852861 podStartE2EDuration="2m9.976852861s" podCreationTimestamp="2025-11-25 09:24:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:51.976290118 +0000 UTC m=+157.017349671" watchObservedRunningTime="2025-11-25 09:26:51.976852861 +0000 UTC m=+157.017912414" Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.996437 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:51 crc kubenswrapper[4776]: E1125 09:26:51.996616 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:52.496592429 +0000 UTC m=+157.537651982 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:51 crc kubenswrapper[4776]: I1125 09:26:51.996665 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:51 crc kubenswrapper[4776]: E1125 09:26:51.996961 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:52.496950408 +0000 UTC m=+157.538009961 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.015187 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mv7lp" podStartSLOduration=130.015169298 podStartE2EDuration="2m10.015169298s" podCreationTimestamp="2025-11-25 09:24:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:52.013558848 +0000 UTC m=+157.054618401" watchObservedRunningTime="2025-11-25 09:26:52.015169298 +0000 UTC m=+157.056228851" Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.021761 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.021824 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.074751 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-7vqhc" podStartSLOduration=8.074734379 podStartE2EDuration="8.074734379s" podCreationTimestamp="2025-11-25 09:26:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:52.073763975 +0000 UTC m=+157.114823528" watchObservedRunningTime="2025-11-25 09:26:52.074734379 +0000 UTC m=+157.115793932" Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.097739 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:52 crc kubenswrapper[4776]: E1125 09:26:52.097947 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:52.597918132 +0000 UTC m=+157.638977695 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.097990 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:52 crc kubenswrapper[4776]: E1125 09:26:52.098376 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:52.598365853 +0000 UTC m=+157.639425486 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.128697 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-wrbxz" podStartSLOduration=131.128679241 podStartE2EDuration="2m11.128679241s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:52.126255931 +0000 UTC m=+157.167315484" watchObservedRunningTime="2025-11-25 09:26:52.128679241 +0000 UTC m=+157.169738794" Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.129954 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nzcz7" podStartSLOduration=131.129948873 podStartE2EDuration="2m11.129948873s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:52.100916856 +0000 UTC m=+157.141976409" watchObservedRunningTime="2025-11-25 09:26:52.129948873 +0000 UTC m=+157.171008426" Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.199312 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:52 crc kubenswrapper[4776]: E1125 09:26:52.199525 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:52.69949125 +0000 UTC m=+157.740550803 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.199609 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:52 crc kubenswrapper[4776]: E1125 09:26:52.199997 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:52.699978122 +0000 UTC m=+157.741038105 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.300737 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:52 crc kubenswrapper[4776]: E1125 09:26:52.300899 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:52.800874534 +0000 UTC m=+157.841934087 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.300970 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:52 crc kubenswrapper[4776]: E1125 09:26:52.301349 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:52.801338156 +0000 UTC m=+157.842397779 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.402501 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:52 crc kubenswrapper[4776]: E1125 09:26:52.402785 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:52.902759971 +0000 UTC m=+157.943819524 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.403014 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:52 crc kubenswrapper[4776]: E1125 09:26:52.403346 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:52.903334175 +0000 UTC m=+157.944393728 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.472428 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-twswh" event={"ID":"ae2cfb63-dc76-40c2-9070-d5c29a6a272d","Type":"ContainerStarted","Data":"b2514b054034f57866e9b3d60f182814a3e4d5b7c64dd766868e019bf5f5c751"} Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.476405 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-m9xnc" event={"ID":"8ba97f9b-a0a1-4514-8577-0b36eb65bbd3","Type":"ContainerStarted","Data":"ba297f56de2cb4e1e57e7cb95cde75af389041cfeaf6a00c33137462613aa27c"} Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.476454 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nzcz7" Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.477778 4776 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5n2sp container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.477819 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5n2sp" podUID="e8765bc6-c8e3-40d7-a061-211b1cad5df2" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.504123 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:52 crc kubenswrapper[4776]: E1125 09:26:52.504419 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:53.004403381 +0000 UTC m=+158.045462934 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.555498 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nzcz7" Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.605805 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:52 crc kubenswrapper[4776]: E1125 09:26:52.608157 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:53.108141804 +0000 UTC m=+158.149201357 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.674441 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-wxwlt" Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.706743 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:52 crc kubenswrapper[4776]: E1125 09:26:52.706872 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:53.206844151 +0000 UTC m=+158.247903704 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.707052 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:52 crc kubenswrapper[4776]: E1125 09:26:52.707360 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:53.207352744 +0000 UTC m=+158.248412287 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.808700 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:52 crc kubenswrapper[4776]: E1125 09:26:52.809135 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:53.309116067 +0000 UTC m=+158.350175620 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.875228 4776 patch_prober.go:28] interesting pod/router-default-5444994796-8nrlj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:26:52 crc kubenswrapper[4776]: [-]has-synced failed: reason withheld Nov 25 09:26:52 crc kubenswrapper[4776]: [+]process-running ok Nov 25 09:26:52 crc kubenswrapper[4776]: healthz check failed Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.875462 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nrlj" podUID="02eb8668-3740-4714-98ae-e993184549aa" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:26:52 crc kubenswrapper[4776]: I1125 09:26:52.910341 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:52 crc kubenswrapper[4776]: E1125 09:26:52.910630 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:53.410619294 +0000 UTC m=+158.451678847 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.011744 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:53 crc kubenswrapper[4776]: E1125 09:26:53.011930 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:53.511903306 +0000 UTC m=+158.552962859 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.011988 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:53 crc kubenswrapper[4776]: E1125 09:26:53.012294 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:53.512279416 +0000 UTC m=+158.553339029 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.112906 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:53 crc kubenswrapper[4776]: E1125 09:26:53.113101 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:53.613078045 +0000 UTC m=+158.654137598 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.113138 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:53 crc kubenswrapper[4776]: E1125 09:26:53.113448 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:53.613440894 +0000 UTC m=+158.654500447 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.126508 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.214455 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:53 crc kubenswrapper[4776]: E1125 09:26:53.214621 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:53.714596892 +0000 UTC m=+158.755656445 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.214727 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:53 crc kubenswrapper[4776]: E1125 09:26:53.215015 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:53.715002822 +0000 UTC m=+158.756062375 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.315768 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:53 crc kubenswrapper[4776]: E1125 09:26:53.315965 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:53.815936045 +0000 UTC m=+158.856995598 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.417859 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:53 crc kubenswrapper[4776]: E1125 09:26:53.418278 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:53.918260612 +0000 UTC m=+158.959320165 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.468858 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pxtkj" Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.480348 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-twswh" event={"ID":"ae2cfb63-dc76-40c2-9070-d5c29a6a272d","Type":"ContainerStarted","Data":"0cf6a8b7335287424f4d2968bba01e5d27c9ece96840bc5fe7b1abf306879844"} Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.480882 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-twswh" event={"ID":"ae2cfb63-dc76-40c2-9070-d5c29a6a272d","Type":"ContainerStarted","Data":"709ecf91c0c2c6df7001031198201011fe0a34104bd0aaa5ca93044371f201ad"} Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.480822 4776 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5n2sp container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.480951 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5n2sp" podUID="e8765bc6-c8e3-40d7-a061-211b1cad5df2" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.503366 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hffrq" Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.519679 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:53 crc kubenswrapper[4776]: E1125 09:26:53.519838 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:54.019820311 +0000 UTC m=+159.060879864 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.520166 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:53 crc kubenswrapper[4776]: E1125 09:26:53.521453 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:54.021437221 +0000 UTC m=+159.062496774 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.621008 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:53 crc kubenswrapper[4776]: E1125 09:26:53.621204 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:54.121166804 +0000 UTC m=+159.162226357 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.621265 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:53 crc kubenswrapper[4776]: E1125 09:26:53.621778 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:54.121770119 +0000 UTC m=+159.162829672 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.722620 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:53 crc kubenswrapper[4776]: E1125 09:26:53.722860 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:26:54.222834265 +0000 UTC m=+159.263893818 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.823867 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:53 crc kubenswrapper[4776]: E1125 09:26:53.824281 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:26:54.32426491 +0000 UTC m=+159.365324463 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-tjf22" (UID: "2441a0f3-b420-4a64-803b-81082cc15874") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.836490 4776 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.843155 4776 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-25T09:26:53.836530233Z","Handler":null,"Name":""} Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.846282 4776 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.846324 4776 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.863033 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zd2l7"] Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.863959 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zd2l7" Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.871259 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.874484 4776 patch_prober.go:28] interesting pod/router-default-5444994796-8nrlj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:26:53 crc kubenswrapper[4776]: [-]has-synced failed: reason withheld Nov 25 09:26:53 crc kubenswrapper[4776]: [+]process-running ok Nov 25 09:26:53 crc kubenswrapper[4776]: healthz check failed Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.874546 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nrlj" podUID="02eb8668-3740-4714-98ae-e993184549aa" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.901096 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zd2l7"] Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.924778 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.924971 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txb6p\" (UniqueName: \"kubernetes.io/projected/2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6-kube-api-access-txb6p\") pod \"certified-operators-zd2l7\" (UID: \"2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6\") " pod="openshift-marketplace/certified-operators-zd2l7" Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.925005 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6-catalog-content\") pod \"certified-operators-zd2l7\" (UID: \"2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6\") " pod="openshift-marketplace/certified-operators-zd2l7" Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.925041 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6-utilities\") pod \"certified-operators-zd2l7\" (UID: \"2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6\") " pod="openshift-marketplace/certified-operators-zd2l7" Nov 25 09:26:53 crc kubenswrapper[4776]: I1125 09:26:53.947444 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.026420 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.026780 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txb6p\" (UniqueName: \"kubernetes.io/projected/2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6-kube-api-access-txb6p\") pod \"certified-operators-zd2l7\" (UID: \"2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6\") " pod="openshift-marketplace/certified-operators-zd2l7" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.026822 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6-catalog-content\") pod \"certified-operators-zd2l7\" (UID: \"2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6\") " pod="openshift-marketplace/certified-operators-zd2l7" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.026874 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6-utilities\") pod \"certified-operators-zd2l7\" (UID: \"2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6\") " pod="openshift-marketplace/certified-operators-zd2l7" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.027544 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6-catalog-content\") pod \"certified-operators-zd2l7\" (UID: \"2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6\") " pod="openshift-marketplace/certified-operators-zd2l7" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.027624 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6-utilities\") pod \"certified-operators-zd2l7\" (UID: \"2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6\") " pod="openshift-marketplace/certified-operators-zd2l7" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.029298 4776 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.029332 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.062272 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xz58l"] Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.063329 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xz58l" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.065570 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.071125 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txb6p\" (UniqueName: \"kubernetes.io/projected/2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6-kube-api-access-txb6p\") pod \"certified-operators-zd2l7\" (UID: \"2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6\") " pod="openshift-marketplace/certified-operators-zd2l7" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.071721 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-tjf22\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.088587 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.100929 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xz58l"] Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.127630 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e057eef9-0e1f-4272-8525-3a4350250d34-catalog-content\") pod \"community-operators-xz58l\" (UID: \"e057eef9-0e1f-4272-8525-3a4350250d34\") " pod="openshift-marketplace/community-operators-xz58l" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.127680 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66kjv\" (UniqueName: \"kubernetes.io/projected/e057eef9-0e1f-4272-8525-3a4350250d34-kube-api-access-66kjv\") pod \"community-operators-xz58l\" (UID: \"e057eef9-0e1f-4272-8525-3a4350250d34\") " pod="openshift-marketplace/community-operators-xz58l" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.127699 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e057eef9-0e1f-4272-8525-3a4350250d34-utilities\") pod \"community-operators-xz58l\" (UID: \"e057eef9-0e1f-4272-8525-3a4350250d34\") " pod="openshift-marketplace/community-operators-xz58l" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.162576 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-x9k2l" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.180330 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zd2l7" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.230604 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e057eef9-0e1f-4272-8525-3a4350250d34-catalog-content\") pod \"community-operators-xz58l\" (UID: \"e057eef9-0e1f-4272-8525-3a4350250d34\") " pod="openshift-marketplace/community-operators-xz58l" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.230664 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66kjv\" (UniqueName: \"kubernetes.io/projected/e057eef9-0e1f-4272-8525-3a4350250d34-kube-api-access-66kjv\") pod \"community-operators-xz58l\" (UID: \"e057eef9-0e1f-4272-8525-3a4350250d34\") " pod="openshift-marketplace/community-operators-xz58l" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.230685 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e057eef9-0e1f-4272-8525-3a4350250d34-utilities\") pod \"community-operators-xz58l\" (UID: \"e057eef9-0e1f-4272-8525-3a4350250d34\") " pod="openshift-marketplace/community-operators-xz58l" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.231777 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e057eef9-0e1f-4272-8525-3a4350250d34-catalog-content\") pod \"community-operators-xz58l\" (UID: \"e057eef9-0e1f-4272-8525-3a4350250d34\") " pod="openshift-marketplace/community-operators-xz58l" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.234681 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e057eef9-0e1f-4272-8525-3a4350250d34-utilities\") pod \"community-operators-xz58l\" (UID: \"e057eef9-0e1f-4272-8525-3a4350250d34\") " pod="openshift-marketplace/community-operators-xz58l" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.270930 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66kjv\" (UniqueName: \"kubernetes.io/projected/e057eef9-0e1f-4272-8525-3a4350250d34-kube-api-access-66kjv\") pod \"community-operators-xz58l\" (UID: \"e057eef9-0e1f-4272-8525-3a4350250d34\") " pod="openshift-marketplace/community-operators-xz58l" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.293182 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4hbp8"] Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.294133 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4hbp8" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.335506 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4hbp8"] Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.335785 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9a42349-e392-40c2-a6e8-32917240a4ad-catalog-content\") pod \"certified-operators-4hbp8\" (UID: \"b9a42349-e392-40c2-a6e8-32917240a4ad\") " pod="openshift-marketplace/certified-operators-4hbp8" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.335849 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nqbw\" (UniqueName: \"kubernetes.io/projected/b9a42349-e392-40c2-a6e8-32917240a4ad-kube-api-access-8nqbw\") pod \"certified-operators-4hbp8\" (UID: \"b9a42349-e392-40c2-a6e8-32917240a4ad\") " pod="openshift-marketplace/certified-operators-4hbp8" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.335928 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9a42349-e392-40c2-a6e8-32917240a4ad-utilities\") pod \"certified-operators-4hbp8\" (UID: \"b9a42349-e392-40c2-a6e8-32917240a4ad\") " pod="openshift-marketplace/certified-operators-4hbp8" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.401873 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xz58l" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.439160 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9a42349-e392-40c2-a6e8-32917240a4ad-utilities\") pod \"certified-operators-4hbp8\" (UID: \"b9a42349-e392-40c2-a6e8-32917240a4ad\") " pod="openshift-marketplace/certified-operators-4hbp8" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.439217 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9a42349-e392-40c2-a6e8-32917240a4ad-catalog-content\") pod \"certified-operators-4hbp8\" (UID: \"b9a42349-e392-40c2-a6e8-32917240a4ad\") " pod="openshift-marketplace/certified-operators-4hbp8" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.439256 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nqbw\" (UniqueName: \"kubernetes.io/projected/b9a42349-e392-40c2-a6e8-32917240a4ad-kube-api-access-8nqbw\") pod \"certified-operators-4hbp8\" (UID: \"b9a42349-e392-40c2-a6e8-32917240a4ad\") " pod="openshift-marketplace/certified-operators-4hbp8" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.439885 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9a42349-e392-40c2-a6e8-32917240a4ad-utilities\") pod \"certified-operators-4hbp8\" (UID: \"b9a42349-e392-40c2-a6e8-32917240a4ad\") " pod="openshift-marketplace/certified-operators-4hbp8" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.440109 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9a42349-e392-40c2-a6e8-32917240a4ad-catalog-content\") pod \"certified-operators-4hbp8\" (UID: \"b9a42349-e392-40c2-a6e8-32917240a4ad\") " pod="openshift-marketplace/certified-operators-4hbp8" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.459018 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rpmxl"] Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.463690 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rpmxl" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.470991 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nqbw\" (UniqueName: \"kubernetes.io/projected/b9a42349-e392-40c2-a6e8-32917240a4ad-kube-api-access-8nqbw\") pod \"certified-operators-4hbp8\" (UID: \"b9a42349-e392-40c2-a6e8-32917240a4ad\") " pod="openshift-marketplace/certified-operators-4hbp8" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.485415 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rpmxl"] Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.514390 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-twswh" event={"ID":"ae2cfb63-dc76-40c2-9070-d5c29a6a272d","Type":"ContainerStarted","Data":"a3da8e094983b44f17ea7ca00af48cdbeff8d8e1f158c9ae6c5c0e8f617c9730"} Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.540059 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa898cba-424e-4a13-93de-0e869ccc6780-catalog-content\") pod \"community-operators-rpmxl\" (UID: \"fa898cba-424e-4a13-93de-0e869ccc6780\") " pod="openshift-marketplace/community-operators-rpmxl" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.540130 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa898cba-424e-4a13-93de-0e869ccc6780-utilities\") pod \"community-operators-rpmxl\" (UID: \"fa898cba-424e-4a13-93de-0e869ccc6780\") " pod="openshift-marketplace/community-operators-rpmxl" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.540181 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfs2q\" (UniqueName: \"kubernetes.io/projected/fa898cba-424e-4a13-93de-0e869ccc6780-kube-api-access-rfs2q\") pod \"community-operators-rpmxl\" (UID: \"fa898cba-424e-4a13-93de-0e869ccc6780\") " pod="openshift-marketplace/community-operators-rpmxl" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.561048 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-twswh" podStartSLOduration=10.561030157 podStartE2EDuration="10.561030157s" podCreationTimestamp="2025-11-25 09:26:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:54.558759141 +0000 UTC m=+159.599818694" watchObservedRunningTime="2025-11-25 09:26:54.561030157 +0000 UTC m=+159.602089700" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.607006 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zd2l7"] Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.610336 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4hbp8" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.644505 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa898cba-424e-4a13-93de-0e869ccc6780-catalog-content\") pod \"community-operators-rpmxl\" (UID: \"fa898cba-424e-4a13-93de-0e869ccc6780\") " pod="openshift-marketplace/community-operators-rpmxl" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.645765 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa898cba-424e-4a13-93de-0e869ccc6780-catalog-content\") pod \"community-operators-rpmxl\" (UID: \"fa898cba-424e-4a13-93de-0e869ccc6780\") " pod="openshift-marketplace/community-operators-rpmxl" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.645876 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa898cba-424e-4a13-93de-0e869ccc6780-utilities\") pod \"community-operators-rpmxl\" (UID: \"fa898cba-424e-4a13-93de-0e869ccc6780\") " pod="openshift-marketplace/community-operators-rpmxl" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.646003 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfs2q\" (UniqueName: \"kubernetes.io/projected/fa898cba-424e-4a13-93de-0e869ccc6780-kube-api-access-rfs2q\") pod \"community-operators-rpmxl\" (UID: \"fa898cba-424e-4a13-93de-0e869ccc6780\") " pod="openshift-marketplace/community-operators-rpmxl" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.647089 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa898cba-424e-4a13-93de-0e869ccc6780-utilities\") pod \"community-operators-rpmxl\" (UID: \"fa898cba-424e-4a13-93de-0e869ccc6780\") " pod="openshift-marketplace/community-operators-rpmxl" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.666806 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfs2q\" (UniqueName: \"kubernetes.io/projected/fa898cba-424e-4a13-93de-0e869ccc6780-kube-api-access-rfs2q\") pod \"community-operators-rpmxl\" (UID: \"fa898cba-424e-4a13-93de-0e869ccc6780\") " pod="openshift-marketplace/community-operators-rpmxl" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.787688 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xz58l"] Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.799402 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rpmxl" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.845044 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4hbp8"] Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.878307 4776 patch_prober.go:28] interesting pod/router-default-5444994796-8nrlj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:26:54 crc kubenswrapper[4776]: [-]has-synced failed: reason withheld Nov 25 09:26:54 crc kubenswrapper[4776]: [+]process-running ok Nov 25 09:26:54 crc kubenswrapper[4776]: healthz check failed Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.878368 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nrlj" podUID="02eb8668-3740-4714-98ae-e993184549aa" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:26:54 crc kubenswrapper[4776]: I1125 09:26:54.960343 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-tjf22"] Nov 25 09:26:55 crc kubenswrapper[4776]: I1125 09:26:55.010197 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rpmxl"] Nov 25 09:26:55 crc kubenswrapper[4776]: W1125 09:26:55.023578 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa898cba_424e_4a13_93de_0e869ccc6780.slice/crio-a558cbb45ddf4d73617df1e89c9d52a41fd0c17bf70b9bfd8408284a23af4162 WatchSource:0}: Error finding container a558cbb45ddf4d73617df1e89c9d52a41fd0c17bf70b9bfd8408284a23af4162: Status 404 returned error can't find the container with id a558cbb45ddf4d73617df1e89c9d52a41fd0c17bf70b9bfd8408284a23af4162 Nov 25 09:26:55 crc kubenswrapper[4776]: I1125 09:26:55.520388 4776 generic.go:334] "Generic (PLEG): container finished" podID="b9a42349-e392-40c2-a6e8-32917240a4ad" containerID="f8c42243e16782a0ddbc885783ad5d47d190bd89b4e9f78a77fa89969136cf8a" exitCode=0 Nov 25 09:26:55 crc kubenswrapper[4776]: I1125 09:26:55.520469 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4hbp8" event={"ID":"b9a42349-e392-40c2-a6e8-32917240a4ad","Type":"ContainerDied","Data":"f8c42243e16782a0ddbc885783ad5d47d190bd89b4e9f78a77fa89969136cf8a"} Nov 25 09:26:55 crc kubenswrapper[4776]: I1125 09:26:55.520797 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4hbp8" event={"ID":"b9a42349-e392-40c2-a6e8-32917240a4ad","Type":"ContainerStarted","Data":"82f4bfa4b3677700320ef1f2ff5b571fe3d8fc39e6f88f04be2c994030940ed5"} Nov 25 09:26:55 crc kubenswrapper[4776]: I1125 09:26:55.522089 4776 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 09:26:55 crc kubenswrapper[4776]: I1125 09:26:55.523501 4776 generic.go:334] "Generic (PLEG): container finished" podID="e057eef9-0e1f-4272-8525-3a4350250d34" containerID="7caabe31431087b47cbe5e03f5d129421ae23d1e4a00404793e64b1b43c78160" exitCode=0 Nov 25 09:26:55 crc kubenswrapper[4776]: I1125 09:26:55.523545 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xz58l" event={"ID":"e057eef9-0e1f-4272-8525-3a4350250d34","Type":"ContainerDied","Data":"7caabe31431087b47cbe5e03f5d129421ae23d1e4a00404793e64b1b43c78160"} Nov 25 09:26:55 crc kubenswrapper[4776]: I1125 09:26:55.523566 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xz58l" event={"ID":"e057eef9-0e1f-4272-8525-3a4350250d34","Type":"ContainerStarted","Data":"f1b8f0f731be5211e5ca88944ed7d31c2f317b2b939e3f627c0162fb7ac9201a"} Nov 25 09:26:55 crc kubenswrapper[4776]: I1125 09:26:55.530044 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa898cba-424e-4a13-93de-0e869ccc6780" containerID="65ffd35e733c330afec288d913f6b319055816eb0c048d59c0bfa50eb662e9cd" exitCode=0 Nov 25 09:26:55 crc kubenswrapper[4776]: I1125 09:26:55.530100 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpmxl" event={"ID":"fa898cba-424e-4a13-93de-0e869ccc6780","Type":"ContainerDied","Data":"65ffd35e733c330afec288d913f6b319055816eb0c048d59c0bfa50eb662e9cd"} Nov 25 09:26:55 crc kubenswrapper[4776]: I1125 09:26:55.530149 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpmxl" event={"ID":"fa898cba-424e-4a13-93de-0e869ccc6780","Type":"ContainerStarted","Data":"a558cbb45ddf4d73617df1e89c9d52a41fd0c17bf70b9bfd8408284a23af4162"} Nov 25 09:26:55 crc kubenswrapper[4776]: I1125 09:26:55.531871 4776 generic.go:334] "Generic (PLEG): container finished" podID="4b51fc9b-c461-4e80-9a56-4f0e43e02399" containerID="3f95652774809f6ee88562122ca9c3788000edd112955de44fc0382bad0e8c84" exitCode=0 Nov 25 09:26:55 crc kubenswrapper[4776]: I1125 09:26:55.531942 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401035-dht2z" event={"ID":"4b51fc9b-c461-4e80-9a56-4f0e43e02399","Type":"ContainerDied","Data":"3f95652774809f6ee88562122ca9c3788000edd112955de44fc0382bad0e8c84"} Nov 25 09:26:55 crc kubenswrapper[4776]: I1125 09:26:55.534481 4776 generic.go:334] "Generic (PLEG): container finished" podID="2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6" containerID="69d6eade4eba1db1723bfebcd0379bb921a12d7d5d045d56597e365d50ea6184" exitCode=0 Nov 25 09:26:55 crc kubenswrapper[4776]: I1125 09:26:55.534587 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zd2l7" event={"ID":"2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6","Type":"ContainerDied","Data":"69d6eade4eba1db1723bfebcd0379bb921a12d7d5d045d56597e365d50ea6184"} Nov 25 09:26:55 crc kubenswrapper[4776]: I1125 09:26:55.534634 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zd2l7" event={"ID":"2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6","Type":"ContainerStarted","Data":"1000b0616d2704456e7d987010b9bd00f6990de2c2a44dd67d4fa61f3ff8536a"} Nov 25 09:26:55 crc kubenswrapper[4776]: I1125 09:26:55.537346 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" event={"ID":"2441a0f3-b420-4a64-803b-81082cc15874","Type":"ContainerStarted","Data":"1426561e5140b314f69ecb69456d3fdf76ef4d0201b681d26b5d01bc154eaedd"} Nov 25 09:26:55 crc kubenswrapper[4776]: I1125 09:26:55.537392 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" event={"ID":"2441a0f3-b420-4a64-803b-81082cc15874","Type":"ContainerStarted","Data":"c760fdf52209f8d7e7374c21df5d5a74e223b27cc07cb83155583ea1c8e24fcb"} Nov 25 09:26:55 crc kubenswrapper[4776]: I1125 09:26:55.632328 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" podStartSLOduration=134.632311647 podStartE2EDuration="2m14.632311647s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:55.624539525 +0000 UTC m=+160.665599088" watchObservedRunningTime="2025-11-25 09:26:55.632311647 +0000 UTC m=+160.673371200" Nov 25 09:26:55 crc kubenswrapper[4776]: I1125 09:26:55.670476 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 25 09:26:55 crc kubenswrapper[4776]: I1125 09:26:55.874625 4776 patch_prober.go:28] interesting pod/router-default-5444994796-8nrlj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:26:55 crc kubenswrapper[4776]: [-]has-synced failed: reason withheld Nov 25 09:26:55 crc kubenswrapper[4776]: [+]process-running ok Nov 25 09:26:55 crc kubenswrapper[4776]: healthz check failed Nov 25 09:26:55 crc kubenswrapper[4776]: I1125 09:26:55.874678 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nrlj" podUID="02eb8668-3740-4714-98ae-e993184549aa" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.062148 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-424gh"] Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.063437 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-424gh" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.065521 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.068543 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-424gh"] Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.175102 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4-catalog-content\") pod \"redhat-marketplace-424gh\" (UID: \"a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4\") " pod="openshift-marketplace/redhat-marketplace-424gh" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.175184 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4-utilities\") pod \"redhat-marketplace-424gh\" (UID: \"a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4\") " pod="openshift-marketplace/redhat-marketplace-424gh" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.175296 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdjdp\" (UniqueName: \"kubernetes.io/projected/a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4-kube-api-access-bdjdp\") pod \"redhat-marketplace-424gh\" (UID: \"a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4\") " pod="openshift-marketplace/redhat-marketplace-424gh" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.276689 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdjdp\" (UniqueName: \"kubernetes.io/projected/a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4-kube-api-access-bdjdp\") pod \"redhat-marketplace-424gh\" (UID: \"a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4\") " pod="openshift-marketplace/redhat-marketplace-424gh" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.276765 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4-catalog-content\") pod \"redhat-marketplace-424gh\" (UID: \"a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4\") " pod="openshift-marketplace/redhat-marketplace-424gh" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.276794 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4-utilities\") pod \"redhat-marketplace-424gh\" (UID: \"a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4\") " pod="openshift-marketplace/redhat-marketplace-424gh" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.277280 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4-utilities\") pod \"redhat-marketplace-424gh\" (UID: \"a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4\") " pod="openshift-marketplace/redhat-marketplace-424gh" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.280395 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4-catalog-content\") pod \"redhat-marketplace-424gh\" (UID: \"a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4\") " pod="openshift-marketplace/redhat-marketplace-424gh" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.310056 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdjdp\" (UniqueName: \"kubernetes.io/projected/a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4-kube-api-access-bdjdp\") pod \"redhat-marketplace-424gh\" (UID: \"a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4\") " pod="openshift-marketplace/redhat-marketplace-424gh" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.415240 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-424gh" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.475491 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lvw5w"] Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.478203 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.478813 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.479246 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lvw5w" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.481574 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.481818 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.484613 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lvw5w"] Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.492354 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.542786 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.582572 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d03c6574-9476-4ef1-8a83-68c7b2ce7a7c-catalog-content\") pod \"redhat-marketplace-lvw5w\" (UID: \"d03c6574-9476-4ef1-8a83-68c7b2ce7a7c\") " pod="openshift-marketplace/redhat-marketplace-lvw5w" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.582651 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nr7h\" (UniqueName: \"kubernetes.io/projected/d03c6574-9476-4ef1-8a83-68c7b2ce7a7c-kube-api-access-6nr7h\") pod \"redhat-marketplace-lvw5w\" (UID: \"d03c6574-9476-4ef1-8a83-68c7b2ce7a7c\") " pod="openshift-marketplace/redhat-marketplace-lvw5w" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.582874 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0944aab1-01d8-43cc-91d8-cb913d4cad3a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0944aab1-01d8-43cc-91d8-cb913d4cad3a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.583032 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d03c6574-9476-4ef1-8a83-68c7b2ce7a7c-utilities\") pod \"redhat-marketplace-lvw5w\" (UID: \"d03c6574-9476-4ef1-8a83-68c7b2ce7a7c\") " pod="openshift-marketplace/redhat-marketplace-lvw5w" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.583081 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0944aab1-01d8-43cc-91d8-cb913d4cad3a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0944aab1-01d8-43cc-91d8-cb913d4cad3a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.683962 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d03c6574-9476-4ef1-8a83-68c7b2ce7a7c-catalog-content\") pod \"redhat-marketplace-lvw5w\" (UID: \"d03c6574-9476-4ef1-8a83-68c7b2ce7a7c\") " pod="openshift-marketplace/redhat-marketplace-lvw5w" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.684036 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nr7h\" (UniqueName: \"kubernetes.io/projected/d03c6574-9476-4ef1-8a83-68c7b2ce7a7c-kube-api-access-6nr7h\") pod \"redhat-marketplace-lvw5w\" (UID: \"d03c6574-9476-4ef1-8a83-68c7b2ce7a7c\") " pod="openshift-marketplace/redhat-marketplace-lvw5w" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.684126 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0944aab1-01d8-43cc-91d8-cb913d4cad3a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0944aab1-01d8-43cc-91d8-cb913d4cad3a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.684200 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d03c6574-9476-4ef1-8a83-68c7b2ce7a7c-utilities\") pod \"redhat-marketplace-lvw5w\" (UID: \"d03c6574-9476-4ef1-8a83-68c7b2ce7a7c\") " pod="openshift-marketplace/redhat-marketplace-lvw5w" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.684267 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0944aab1-01d8-43cc-91d8-cb913d4cad3a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0944aab1-01d8-43cc-91d8-cb913d4cad3a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.684221 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0944aab1-01d8-43cc-91d8-cb913d4cad3a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0944aab1-01d8-43cc-91d8-cb913d4cad3a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.684664 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d03c6574-9476-4ef1-8a83-68c7b2ce7a7c-catalog-content\") pod \"redhat-marketplace-lvw5w\" (UID: \"d03c6574-9476-4ef1-8a83-68c7b2ce7a7c\") " pod="openshift-marketplace/redhat-marketplace-lvw5w" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.684755 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d03c6574-9476-4ef1-8a83-68c7b2ce7a7c-utilities\") pod \"redhat-marketplace-lvw5w\" (UID: \"d03c6574-9476-4ef1-8a83-68c7b2ce7a7c\") " pod="openshift-marketplace/redhat-marketplace-lvw5w" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.700300 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0944aab1-01d8-43cc-91d8-cb913d4cad3a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0944aab1-01d8-43cc-91d8-cb913d4cad3a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.714275 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nr7h\" (UniqueName: \"kubernetes.io/projected/d03c6574-9476-4ef1-8a83-68c7b2ce7a7c-kube-api-access-6nr7h\") pod \"redhat-marketplace-lvw5w\" (UID: \"d03c6574-9476-4ef1-8a83-68c7b2ce7a7c\") " pod="openshift-marketplace/redhat-marketplace-lvw5w" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.800992 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.802740 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401035-dht2z" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.821924 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-424gh"] Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.822526 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lvw5w" Nov 25 09:26:56 crc kubenswrapper[4776]: W1125 09:26:56.832743 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda7dabec0_1f3d_4ab3_818f_de4c9fb4e9a4.slice/crio-29cbd3bbb0b0c03ff1f16785fea923b23d55106e2b74a01dc4a9c330f735b571 WatchSource:0}: Error finding container 29cbd3bbb0b0c03ff1f16785fea923b23d55106e2b74a01dc4a9c330f735b571: Status 404 returned error can't find the container with id 29cbd3bbb0b0c03ff1f16785fea923b23d55106e2b74a01dc4a9c330f735b571 Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.858686 4776 patch_prober.go:28] interesting pod/downloads-7954f5f757-r979m container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.858740 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-r979m" podUID="529b4e1d-e99c-4342-8cff-68c1e31a776f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.858692 4776 patch_prober.go:28] interesting pod/downloads-7954f5f757-r979m container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.858996 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r979m" podUID="529b4e1d-e99c-4342-8cff-68c1e31a776f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.874004 4776 patch_prober.go:28] interesting pod/router-default-5444994796-8nrlj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:26:56 crc kubenswrapper[4776]: [-]has-synced failed: reason withheld Nov 25 09:26:56 crc kubenswrapper[4776]: [+]process-running ok Nov 25 09:26:56 crc kubenswrapper[4776]: healthz check failed Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.874088 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nrlj" podUID="02eb8668-3740-4714-98ae-e993184549aa" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.887495 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fdwr\" (UniqueName: \"kubernetes.io/projected/4b51fc9b-c461-4e80-9a56-4f0e43e02399-kube-api-access-8fdwr\") pod \"4b51fc9b-c461-4e80-9a56-4f0e43e02399\" (UID: \"4b51fc9b-c461-4e80-9a56-4f0e43e02399\") " Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.887581 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b51fc9b-c461-4e80-9a56-4f0e43e02399-config-volume\") pod \"4b51fc9b-c461-4e80-9a56-4f0e43e02399\" (UID: \"4b51fc9b-c461-4e80-9a56-4f0e43e02399\") " Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.887617 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b51fc9b-c461-4e80-9a56-4f0e43e02399-secret-volume\") pod \"4b51fc9b-c461-4e80-9a56-4f0e43e02399\" (UID: \"4b51fc9b-c461-4e80-9a56-4f0e43e02399\") " Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.888740 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b51fc9b-c461-4e80-9a56-4f0e43e02399-config-volume" (OuterVolumeSpecName: "config-volume") pod "4b51fc9b-c461-4e80-9a56-4f0e43e02399" (UID: "4b51fc9b-c461-4e80-9a56-4f0e43e02399"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.895617 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b51fc9b-c461-4e80-9a56-4f0e43e02399-kube-api-access-8fdwr" (OuterVolumeSpecName: "kube-api-access-8fdwr") pod "4b51fc9b-c461-4e80-9a56-4f0e43e02399" (UID: "4b51fc9b-c461-4e80-9a56-4f0e43e02399"). InnerVolumeSpecName "kube-api-access-8fdwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.900886 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b51fc9b-c461-4e80-9a56-4f0e43e02399-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4b51fc9b-c461-4e80-9a56-4f0e43e02399" (UID: "4b51fc9b-c461-4e80-9a56-4f0e43e02399"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.962231 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.970203 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-srdjj" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.991378 4776 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4b51fc9b-c461-4e80-9a56-4f0e43e02399-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.991725 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fdwr\" (UniqueName: \"kubernetes.io/projected/4b51fc9b-c461-4e80-9a56-4f0e43e02399-kube-api-access-8fdwr\") on node \"crc\" DevicePath \"\"" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.991739 4776 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b51fc9b-c461-4e80-9a56-4f0e43e02399-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 09:26:56 crc kubenswrapper[4776]: I1125 09:26:56.996308 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.039774 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.112718 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vtzcb"] Nov 25 09:26:57 crc kubenswrapper[4776]: E1125 09:26:57.112976 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b51fc9b-c461-4e80-9a56-4f0e43e02399" containerName="collect-profiles" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.112999 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b51fc9b-c461-4e80-9a56-4f0e43e02399" containerName="collect-profiles" Nov 25 09:26:57 crc kubenswrapper[4776]: W1125 09:26:57.120716 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod0944aab1_01d8_43cc_91d8_cb913d4cad3a.slice/crio-a2c18f46bc3c2186320aea86cba4bfc0fbd221a4177de4394da07673655d87ba WatchSource:0}: Error finding container a2c18f46bc3c2186320aea86cba4bfc0fbd221a4177de4394da07673655d87ba: Status 404 returned error can't find the container with id a2c18f46bc3c2186320aea86cba4bfc0fbd221a4177de4394da07673655d87ba Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.127126 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b51fc9b-c461-4e80-9a56-4f0e43e02399" containerName="collect-profiles" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.128256 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vtzcb"] Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.128356 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vtzcb" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.129930 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.176482 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.176514 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.188836 4776 patch_prober.go:28] interesting pod/console-f9d7485db-bhgv2 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.188879 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-bhgv2" podUID="04fe29b5-d8c6-46c5-9d1f-76005e225fd6" containerName="console" probeResult="failure" output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.195769 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f04466da-02ff-40ea-bbf4-5761ab1d93b6-catalog-content\") pod \"redhat-operators-vtzcb\" (UID: \"f04466da-02ff-40ea-bbf4-5761ab1d93b6\") " pod="openshift-marketplace/redhat-operators-vtzcb" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.195863 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f04466da-02ff-40ea-bbf4-5761ab1d93b6-utilities\") pod \"redhat-operators-vtzcb\" (UID: \"f04466da-02ff-40ea-bbf4-5761ab1d93b6\") " pod="openshift-marketplace/redhat-operators-vtzcb" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.195881 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmb8m\" (UniqueName: \"kubernetes.io/projected/f04466da-02ff-40ea-bbf4-5761ab1d93b6-kube-api-access-kmb8m\") pod \"redhat-operators-vtzcb\" (UID: \"f04466da-02ff-40ea-bbf4-5761ab1d93b6\") " pod="openshift-marketplace/redhat-operators-vtzcb" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.297008 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f04466da-02ff-40ea-bbf4-5761ab1d93b6-utilities\") pod \"redhat-operators-vtzcb\" (UID: \"f04466da-02ff-40ea-bbf4-5761ab1d93b6\") " pod="openshift-marketplace/redhat-operators-vtzcb" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.297084 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmb8m\" (UniqueName: \"kubernetes.io/projected/f04466da-02ff-40ea-bbf4-5761ab1d93b6-kube-api-access-kmb8m\") pod \"redhat-operators-vtzcb\" (UID: \"f04466da-02ff-40ea-bbf4-5761ab1d93b6\") " pod="openshift-marketplace/redhat-operators-vtzcb" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.297194 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f04466da-02ff-40ea-bbf4-5761ab1d93b6-catalog-content\") pod \"redhat-operators-vtzcb\" (UID: \"f04466da-02ff-40ea-bbf4-5761ab1d93b6\") " pod="openshift-marketplace/redhat-operators-vtzcb" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.299272 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f04466da-02ff-40ea-bbf4-5761ab1d93b6-utilities\") pod \"redhat-operators-vtzcb\" (UID: \"f04466da-02ff-40ea-bbf4-5761ab1d93b6\") " pod="openshift-marketplace/redhat-operators-vtzcb" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.299742 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f04466da-02ff-40ea-bbf4-5761ab1d93b6-catalog-content\") pod \"redhat-operators-vtzcb\" (UID: \"f04466da-02ff-40ea-bbf4-5761ab1d93b6\") " pod="openshift-marketplace/redhat-operators-vtzcb" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.316199 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lvw5w"] Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.328046 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmb8m\" (UniqueName: \"kubernetes.io/projected/f04466da-02ff-40ea-bbf4-5761ab1d93b6-kube-api-access-kmb8m\") pod \"redhat-operators-vtzcb\" (UID: \"f04466da-02ff-40ea-bbf4-5761ab1d93b6\") " pod="openshift-marketplace/redhat-operators-vtzcb" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.466385 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bhq2s"] Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.467833 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bhq2s" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.471630 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bhq2s"] Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.499603 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6rll\" (UniqueName: \"kubernetes.io/projected/906f7079-2d0f-4205-8c68-4e66f0f25ab7-kube-api-access-j6rll\") pod \"redhat-operators-bhq2s\" (UID: \"906f7079-2d0f-4205-8c68-4e66f0f25ab7\") " pod="openshift-marketplace/redhat-operators-bhq2s" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.499652 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/906f7079-2d0f-4205-8c68-4e66f0f25ab7-catalog-content\") pod \"redhat-operators-bhq2s\" (UID: \"906f7079-2d0f-4205-8c68-4e66f0f25ab7\") " pod="openshift-marketplace/redhat-operators-bhq2s" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.499678 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/906f7079-2d0f-4205-8c68-4e66f0f25ab7-utilities\") pod \"redhat-operators-bhq2s\" (UID: \"906f7079-2d0f-4205-8c68-4e66f0f25ab7\") " pod="openshift-marketplace/redhat-operators-bhq2s" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.546888 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vtzcb" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.554878 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"0944aab1-01d8-43cc-91d8-cb913d4cad3a","Type":"ContainerStarted","Data":"a2c18f46bc3c2186320aea86cba4bfc0fbd221a4177de4394da07673655d87ba"} Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.558912 4776 generic.go:334] "Generic (PLEG): container finished" podID="a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4" containerID="da412c58869886ce3eb63d409b39f983a7659b26bb5805a0dc254bd7646f6ae0" exitCode=0 Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.559431 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-424gh" event={"ID":"a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4","Type":"ContainerDied","Data":"da412c58869886ce3eb63d409b39f983a7659b26bb5805a0dc254bd7646f6ae0"} Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.559483 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-424gh" event={"ID":"a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4","Type":"ContainerStarted","Data":"29cbd3bbb0b0c03ff1f16785fea923b23d55106e2b74a01dc4a9c330f735b571"} Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.571611 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lvw5w" event={"ID":"d03c6574-9476-4ef1-8a83-68c7b2ce7a7c","Type":"ContainerStarted","Data":"cc57b99a40a352b676061f940813b6ab071c32ec4c696dbfce8c0cf5a64db228"} Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.581381 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401035-dht2z" event={"ID":"4b51fc9b-c461-4e80-9a56-4f0e43e02399","Type":"ContainerDied","Data":"8f34a2a36c30d5862e517fa9c193c6f488947582dd60db9420d739b1aff92ede"} Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.581468 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f34a2a36c30d5862e517fa9c193c6f488947582dd60db9420d739b1aff92ede" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.581581 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401035-dht2z" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.601816 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6rll\" (UniqueName: \"kubernetes.io/projected/906f7079-2d0f-4205-8c68-4e66f0f25ab7-kube-api-access-j6rll\") pod \"redhat-operators-bhq2s\" (UID: \"906f7079-2d0f-4205-8c68-4e66f0f25ab7\") " pod="openshift-marketplace/redhat-operators-bhq2s" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.603214 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/906f7079-2d0f-4205-8c68-4e66f0f25ab7-catalog-content\") pod \"redhat-operators-bhq2s\" (UID: \"906f7079-2d0f-4205-8c68-4e66f0f25ab7\") " pod="openshift-marketplace/redhat-operators-bhq2s" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.603260 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/906f7079-2d0f-4205-8c68-4e66f0f25ab7-utilities\") pod \"redhat-operators-bhq2s\" (UID: \"906f7079-2d0f-4205-8c68-4e66f0f25ab7\") " pod="openshift-marketplace/redhat-operators-bhq2s" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.603958 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/906f7079-2d0f-4205-8c68-4e66f0f25ab7-catalog-content\") pod \"redhat-operators-bhq2s\" (UID: \"906f7079-2d0f-4205-8c68-4e66f0f25ab7\") " pod="openshift-marketplace/redhat-operators-bhq2s" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.604222 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/906f7079-2d0f-4205-8c68-4e66f0f25ab7-utilities\") pod \"redhat-operators-bhq2s\" (UID: \"906f7079-2d0f-4205-8c68-4e66f0f25ab7\") " pod="openshift-marketplace/redhat-operators-bhq2s" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.628567 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6rll\" (UniqueName: \"kubernetes.io/projected/906f7079-2d0f-4205-8c68-4e66f0f25ab7-kube-api-access-j6rll\") pod \"redhat-operators-bhq2s\" (UID: \"906f7079-2d0f-4205-8c68-4e66f0f25ab7\") " pod="openshift-marketplace/redhat-operators-bhq2s" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.801809 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bhq2s" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.865457 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vtzcb"] Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.872325 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-8nrlj" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.875541 4776 patch_prober.go:28] interesting pod/router-default-5444994796-8nrlj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:26:57 crc kubenswrapper[4776]: [-]has-synced failed: reason withheld Nov 25 09:26:57 crc kubenswrapper[4776]: [+]process-running ok Nov 25 09:26:57 crc kubenswrapper[4776]: healthz check failed Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.875594 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nrlj" podUID="02eb8668-3740-4714-98ae-e993184549aa" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:26:57 crc kubenswrapper[4776]: I1125 09:26:57.923689 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-5n2sp" Nov 25 09:26:58 crc kubenswrapper[4776]: I1125 09:26:58.026208 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bhq2s"] Nov 25 09:26:58 crc kubenswrapper[4776]: W1125 09:26:58.061341 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod906f7079_2d0f_4205_8c68_4e66f0f25ab7.slice/crio-ca6f8f4aa81800ad5d4211f2cb2bf9f027a07fcfffa99d357f83178db6094fb7 WatchSource:0}: Error finding container ca6f8f4aa81800ad5d4211f2cb2bf9f027a07fcfffa99d357f83178db6094fb7: Status 404 returned error can't find the container with id ca6f8f4aa81800ad5d4211f2cb2bf9f027a07fcfffa99d357f83178db6094fb7 Nov 25 09:26:58 crc kubenswrapper[4776]: I1125 09:26:58.364949 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 09:26:58 crc kubenswrapper[4776]: I1125 09:26:58.366101 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 09:26:58 crc kubenswrapper[4776]: I1125 09:26:58.370985 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 09:26:58 crc kubenswrapper[4776]: I1125 09:26:58.404661 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 09:26:58 crc kubenswrapper[4776]: I1125 09:26:58.404675 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 09:26:58 crc kubenswrapper[4776]: I1125 09:26:58.416274 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/01bc6d7e-5810-42c8-94e9-77fc02765d38-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"01bc6d7e-5810-42c8-94e9-77fc02765d38\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 09:26:58 crc kubenswrapper[4776]: I1125 09:26:58.416349 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/01bc6d7e-5810-42c8-94e9-77fc02765d38-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"01bc6d7e-5810-42c8-94e9-77fc02765d38\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 09:26:58 crc kubenswrapper[4776]: I1125 09:26:58.517981 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/01bc6d7e-5810-42c8-94e9-77fc02765d38-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"01bc6d7e-5810-42c8-94e9-77fc02765d38\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 09:26:58 crc kubenswrapper[4776]: I1125 09:26:58.518035 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/01bc6d7e-5810-42c8-94e9-77fc02765d38-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"01bc6d7e-5810-42c8-94e9-77fc02765d38\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 09:26:58 crc kubenswrapper[4776]: I1125 09:26:58.518181 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/01bc6d7e-5810-42c8-94e9-77fc02765d38-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"01bc6d7e-5810-42c8-94e9-77fc02765d38\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 09:26:58 crc kubenswrapper[4776]: I1125 09:26:58.539449 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/01bc6d7e-5810-42c8-94e9-77fc02765d38-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"01bc6d7e-5810-42c8-94e9-77fc02765d38\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 09:26:58 crc kubenswrapper[4776]: I1125 09:26:58.587264 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bhq2s" event={"ID":"906f7079-2d0f-4205-8c68-4e66f0f25ab7","Type":"ContainerStarted","Data":"ca6f8f4aa81800ad5d4211f2cb2bf9f027a07fcfffa99d357f83178db6094fb7"} Nov 25 09:26:58 crc kubenswrapper[4776]: I1125 09:26:58.588689 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"0944aab1-01d8-43cc-91d8-cb913d4cad3a","Type":"ContainerStarted","Data":"ea35a3396c53c39770bc80293500a97d8915fe41b5f1597fa92fb2df461771b7"} Nov 25 09:26:58 crc kubenswrapper[4776]: I1125 09:26:58.601969 4776 generic.go:334] "Generic (PLEG): container finished" podID="f04466da-02ff-40ea-bbf4-5761ab1d93b6" containerID="b2636cec13e2f0b6bb7f7c8dc2b36d065469205fabc5ba15607d8b20d9edb287" exitCode=0 Nov 25 09:26:58 crc kubenswrapper[4776]: I1125 09:26:58.602032 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vtzcb" event={"ID":"f04466da-02ff-40ea-bbf4-5761ab1d93b6","Type":"ContainerDied","Data":"b2636cec13e2f0b6bb7f7c8dc2b36d065469205fabc5ba15607d8b20d9edb287"} Nov 25 09:26:58 crc kubenswrapper[4776]: I1125 09:26:58.602058 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vtzcb" event={"ID":"f04466da-02ff-40ea-bbf4-5761ab1d93b6","Type":"ContainerStarted","Data":"2819732fdc9b853b0d8e99b18f825f87998b08daa12eb26045bd78f2ca6db2aa"} Nov 25 09:26:58 crc kubenswrapper[4776]: I1125 09:26:58.608518 4776 generic.go:334] "Generic (PLEG): container finished" podID="d03c6574-9476-4ef1-8a83-68c7b2ce7a7c" containerID="8682a296ed548c21d62d6f831c43389df48184c667a1917ffb3ea8bc60ee9da8" exitCode=0 Nov 25 09:26:58 crc kubenswrapper[4776]: I1125 09:26:58.608566 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lvw5w" event={"ID":"d03c6574-9476-4ef1-8a83-68c7b2ce7a7c","Type":"ContainerDied","Data":"8682a296ed548c21d62d6f831c43389df48184c667a1917ffb3ea8bc60ee9da8"} Nov 25 09:26:58 crc kubenswrapper[4776]: I1125 09:26:58.610518 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.610505844 podStartE2EDuration="2.610505844s" podCreationTimestamp="2025-11-25 09:26:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:26:58.607612373 +0000 UTC m=+163.648671916" watchObservedRunningTime="2025-11-25 09:26:58.610505844 +0000 UTC m=+163.651565397" Nov 25 09:26:58 crc kubenswrapper[4776]: I1125 09:26:58.727042 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 09:26:58 crc kubenswrapper[4776]: I1125 09:26:58.874509 4776 patch_prober.go:28] interesting pod/router-default-5444994796-8nrlj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:26:58 crc kubenswrapper[4776]: [-]has-synced failed: reason withheld Nov 25 09:26:58 crc kubenswrapper[4776]: [+]process-running ok Nov 25 09:26:58 crc kubenswrapper[4776]: healthz check failed Nov 25 09:26:58 crc kubenswrapper[4776]: I1125 09:26:58.874789 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nrlj" podUID="02eb8668-3740-4714-98ae-e993184549aa" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:26:59 crc kubenswrapper[4776]: I1125 09:26:59.194762 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 09:26:59 crc kubenswrapper[4776]: W1125 09:26:59.210363 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod01bc6d7e_5810_42c8_94e9_77fc02765d38.slice/crio-c6ec4d52acb5642392db55956207e68f1ddd690057eaa83027bd4023dd2f396e WatchSource:0}: Error finding container c6ec4d52acb5642392db55956207e68f1ddd690057eaa83027bd4023dd2f396e: Status 404 returned error can't find the container with id c6ec4d52acb5642392db55956207e68f1ddd690057eaa83027bd4023dd2f396e Nov 25 09:26:59 crc kubenswrapper[4776]: I1125 09:26:59.614435 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"01bc6d7e-5810-42c8-94e9-77fc02765d38","Type":"ContainerStarted","Data":"c6ec4d52acb5642392db55956207e68f1ddd690057eaa83027bd4023dd2f396e"} Nov 25 09:26:59 crc kubenswrapper[4776]: I1125 09:26:59.617901 4776 generic.go:334] "Generic (PLEG): container finished" podID="906f7079-2d0f-4205-8c68-4e66f0f25ab7" containerID="e469a9bc244eab25c648d83980d20bd80cc80fdf843f3d91433cad0f1394c02e" exitCode=0 Nov 25 09:26:59 crc kubenswrapper[4776]: I1125 09:26:59.617996 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bhq2s" event={"ID":"906f7079-2d0f-4205-8c68-4e66f0f25ab7","Type":"ContainerDied","Data":"e469a9bc244eab25c648d83980d20bd80cc80fdf843f3d91433cad0f1394c02e"} Nov 25 09:26:59 crc kubenswrapper[4776]: I1125 09:26:59.632642 4776 generic.go:334] "Generic (PLEG): container finished" podID="0944aab1-01d8-43cc-91d8-cb913d4cad3a" containerID="ea35a3396c53c39770bc80293500a97d8915fe41b5f1597fa92fb2df461771b7" exitCode=0 Nov 25 09:26:59 crc kubenswrapper[4776]: I1125 09:26:59.632677 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"0944aab1-01d8-43cc-91d8-cb913d4cad3a","Type":"ContainerDied","Data":"ea35a3396c53c39770bc80293500a97d8915fe41b5f1597fa92fb2df461771b7"} Nov 25 09:26:59 crc kubenswrapper[4776]: I1125 09:26:59.874475 4776 patch_prober.go:28] interesting pod/router-default-5444994796-8nrlj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:26:59 crc kubenswrapper[4776]: [-]has-synced failed: reason withheld Nov 25 09:26:59 crc kubenswrapper[4776]: [+]process-running ok Nov 25 09:26:59 crc kubenswrapper[4776]: healthz check failed Nov 25 09:26:59 crc kubenswrapper[4776]: I1125 09:26:59.874530 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nrlj" podUID="02eb8668-3740-4714-98ae-e993184549aa" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:27:00 crc kubenswrapper[4776]: I1125 09:27:00.520036 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:27:00 crc kubenswrapper[4776]: I1125 09:27:00.656316 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"01bc6d7e-5810-42c8-94e9-77fc02765d38","Type":"ContainerStarted","Data":"cd3e76c60a588bc87211dd8479cca72b870914c3434d01f01ace0474d150cccb"} Nov 25 09:27:00 crc kubenswrapper[4776]: I1125 09:27:00.875437 4776 patch_prober.go:28] interesting pod/router-default-5444994796-8nrlj container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:27:00 crc kubenswrapper[4776]: [+]has-synced ok Nov 25 09:27:00 crc kubenswrapper[4776]: [+]process-running ok Nov 25 09:27:00 crc kubenswrapper[4776]: healthz check failed Nov 25 09:27:00 crc kubenswrapper[4776]: I1125 09:27:00.875512 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8nrlj" podUID="02eb8668-3740-4714-98ae-e993184549aa" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:27:00 crc kubenswrapper[4776]: I1125 09:27:00.990392 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 09:27:01 crc kubenswrapper[4776]: I1125 09:27:01.061255 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0944aab1-01d8-43cc-91d8-cb913d4cad3a-kube-api-access\") pod \"0944aab1-01d8-43cc-91d8-cb913d4cad3a\" (UID: \"0944aab1-01d8-43cc-91d8-cb913d4cad3a\") " Nov 25 09:27:01 crc kubenswrapper[4776]: I1125 09:27:01.061450 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0944aab1-01d8-43cc-91d8-cb913d4cad3a-kubelet-dir\") pod \"0944aab1-01d8-43cc-91d8-cb913d4cad3a\" (UID: \"0944aab1-01d8-43cc-91d8-cb913d4cad3a\") " Nov 25 09:27:01 crc kubenswrapper[4776]: I1125 09:27:01.061576 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0944aab1-01d8-43cc-91d8-cb913d4cad3a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "0944aab1-01d8-43cc-91d8-cb913d4cad3a" (UID: "0944aab1-01d8-43cc-91d8-cb913d4cad3a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:27:01 crc kubenswrapper[4776]: I1125 09:27:01.065354 4776 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0944aab1-01d8-43cc-91d8-cb913d4cad3a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 09:27:01 crc kubenswrapper[4776]: I1125 09:27:01.081158 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0944aab1-01d8-43cc-91d8-cb913d4cad3a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0944aab1-01d8-43cc-91d8-cb913d4cad3a" (UID: "0944aab1-01d8-43cc-91d8-cb913d4cad3a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:27:01 crc kubenswrapper[4776]: I1125 09:27:01.166500 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0944aab1-01d8-43cc-91d8-cb913d4cad3a-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 09:27:01 crc kubenswrapper[4776]: I1125 09:27:01.673676 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 09:27:01 crc kubenswrapper[4776]: I1125 09:27:01.678004 4776 generic.go:334] "Generic (PLEG): container finished" podID="01bc6d7e-5810-42c8-94e9-77fc02765d38" containerID="cd3e76c60a588bc87211dd8479cca72b870914c3434d01f01ace0474d150cccb" exitCode=0 Nov 25 09:27:01 crc kubenswrapper[4776]: I1125 09:27:01.689837 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"0944aab1-01d8-43cc-91d8-cb913d4cad3a","Type":"ContainerDied","Data":"a2c18f46bc3c2186320aea86cba4bfc0fbd221a4177de4394da07673655d87ba"} Nov 25 09:27:01 crc kubenswrapper[4776]: I1125 09:27:01.689887 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2c18f46bc3c2186320aea86cba4bfc0fbd221a4177de4394da07673655d87ba" Nov 25 09:27:01 crc kubenswrapper[4776]: I1125 09:27:01.689899 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"01bc6d7e-5810-42c8-94e9-77fc02765d38","Type":"ContainerDied","Data":"cd3e76c60a588bc87211dd8479cca72b870914c3434d01f01ace0474d150cccb"} Nov 25 09:27:01 crc kubenswrapper[4776]: I1125 09:27:01.878730 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-8nrlj" Nov 25 09:27:01 crc kubenswrapper[4776]: I1125 09:27:01.884390 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-8nrlj" Nov 25 09:27:02 crc kubenswrapper[4776]: I1125 09:27:02.953917 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-7vqhc" Nov 25 09:27:04 crc kubenswrapper[4776]: I1125 09:27:04.002813 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-metrics-certs\") pod \"network-metrics-daemon-nvc68\" (UID: \"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\") " pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:27:04 crc kubenswrapper[4776]: I1125 09:27:04.009022 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087-metrics-certs\") pod \"network-metrics-daemon-nvc68\" (UID: \"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087\") " pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:27:04 crc kubenswrapper[4776]: I1125 09:27:04.103527 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-nvc68" Nov 25 09:27:06 crc kubenswrapper[4776]: I1125 09:27:06.677048 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 09:27:06 crc kubenswrapper[4776]: I1125 09:27:06.731287 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"01bc6d7e-5810-42c8-94e9-77fc02765d38","Type":"ContainerDied","Data":"c6ec4d52acb5642392db55956207e68f1ddd690057eaa83027bd4023dd2f396e"} Nov 25 09:27:06 crc kubenswrapper[4776]: I1125 09:27:06.731345 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6ec4d52acb5642392db55956207e68f1ddd690057eaa83027bd4023dd2f396e" Nov 25 09:27:06 crc kubenswrapper[4776]: I1125 09:27:06.731393 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 09:27:06 crc kubenswrapper[4776]: I1125 09:27:06.743774 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/01bc6d7e-5810-42c8-94e9-77fc02765d38-kubelet-dir\") pod \"01bc6d7e-5810-42c8-94e9-77fc02765d38\" (UID: \"01bc6d7e-5810-42c8-94e9-77fc02765d38\") " Nov 25 09:27:06 crc kubenswrapper[4776]: I1125 09:27:06.743837 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/01bc6d7e-5810-42c8-94e9-77fc02765d38-kube-api-access\") pod \"01bc6d7e-5810-42c8-94e9-77fc02765d38\" (UID: \"01bc6d7e-5810-42c8-94e9-77fc02765d38\") " Nov 25 09:27:06 crc kubenswrapper[4776]: I1125 09:27:06.744002 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/01bc6d7e-5810-42c8-94e9-77fc02765d38-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "01bc6d7e-5810-42c8-94e9-77fc02765d38" (UID: "01bc6d7e-5810-42c8-94e9-77fc02765d38"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:27:06 crc kubenswrapper[4776]: I1125 09:27:06.744338 4776 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/01bc6d7e-5810-42c8-94e9-77fc02765d38-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 09:27:06 crc kubenswrapper[4776]: I1125 09:27:06.753748 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01bc6d7e-5810-42c8-94e9-77fc02765d38-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "01bc6d7e-5810-42c8-94e9-77fc02765d38" (UID: "01bc6d7e-5810-42c8-94e9-77fc02765d38"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:27:06 crc kubenswrapper[4776]: I1125 09:27:06.845901 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/01bc6d7e-5810-42c8-94e9-77fc02765d38-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 09:27:06 crc kubenswrapper[4776]: I1125 09:27:06.859503 4776 patch_prober.go:28] interesting pod/downloads-7954f5f757-r979m container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Nov 25 09:27:06 crc kubenswrapper[4776]: I1125 09:27:06.859579 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-r979m" podUID="529b4e1d-e99c-4342-8cff-68c1e31a776f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Nov 25 09:27:06 crc kubenswrapper[4776]: I1125 09:27:06.859641 4776 patch_prober.go:28] interesting pod/downloads-7954f5f757-r979m container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Nov 25 09:27:06 crc kubenswrapper[4776]: I1125 09:27:06.859772 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r979m" podUID="529b4e1d-e99c-4342-8cff-68c1e31a776f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.25:8080/\": dial tcp 10.217.0.25:8080: connect: connection refused" Nov 25 09:27:07 crc kubenswrapper[4776]: I1125 09:27:07.179400 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:27:07 crc kubenswrapper[4776]: I1125 09:27:07.183211 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:27:13 crc kubenswrapper[4776]: I1125 09:27:13.701649 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:27:14 crc kubenswrapper[4776]: I1125 09:27:14.093813 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:27:16 crc kubenswrapper[4776]: I1125 09:27:16.868517 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-r979m" Nov 25 09:27:17 crc kubenswrapper[4776]: I1125 09:27:17.818904 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:27:17 crc kubenswrapper[4776]: I1125 09:27:17.818992 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:27:25 crc kubenswrapper[4776]: E1125 09:27:25.815940 4776 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 25 09:27:25 crc kubenswrapper[4776]: E1125 09:27:25.816622 4776 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-txb6p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-zd2l7_openshift-marketplace(2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 09:27:25 crc kubenswrapper[4776]: E1125 09:27:25.817789 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-zd2l7" podUID="2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6" Nov 25 09:27:26 crc kubenswrapper[4776]: E1125 09:27:26.599711 4776 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 25 09:27:26 crc kubenswrapper[4776]: E1125 09:27:26.599943 4776 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8nqbw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-4hbp8_openshift-marketplace(b9a42349-e392-40c2-a6e8-32917240a4ad): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 09:27:26 crc kubenswrapper[4776]: E1125 09:27:26.601239 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-4hbp8" podUID="b9a42349-e392-40c2-a6e8-32917240a4ad" Nov 25 09:27:27 crc kubenswrapper[4776]: E1125 09:27:27.183269 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-zd2l7" podUID="2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6" Nov 25 09:27:27 crc kubenswrapper[4776]: E1125 09:27:27.183341 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-4hbp8" podUID="b9a42349-e392-40c2-a6e8-32917240a4ad" Nov 25 09:27:27 crc kubenswrapper[4776]: I1125 09:27:27.887121 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwxqt" Nov 25 09:27:28 crc kubenswrapper[4776]: E1125 09:27:28.854361 4776 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 25 09:27:28 crc kubenswrapper[4776]: E1125 09:27:28.854518 4776 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rfs2q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-rpmxl_openshift-marketplace(fa898cba-424e-4a13-93de-0e869ccc6780): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 09:27:28 crc kubenswrapper[4776]: E1125 09:27:28.855687 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-rpmxl" podUID="fa898cba-424e-4a13-93de-0e869ccc6780" Nov 25 09:27:29 crc kubenswrapper[4776]: E1125 09:27:29.003991 4776 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 25 09:27:29 crc kubenswrapper[4776]: E1125 09:27:29.004193 4776 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-66kjv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-xz58l_openshift-marketplace(e057eef9-0e1f-4272-8525-3a4350250d34): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 09:27:29 crc kubenswrapper[4776]: E1125 09:27:29.005413 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-xz58l" podUID="e057eef9-0e1f-4272-8525-3a4350250d34" Nov 25 09:27:32 crc kubenswrapper[4776]: E1125 09:27:32.408722 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-rpmxl" podUID="fa898cba-424e-4a13-93de-0e869ccc6780" Nov 25 09:27:32 crc kubenswrapper[4776]: E1125 09:27:32.408841 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xz58l" podUID="e057eef9-0e1f-4272-8525-3a4350250d34" Nov 25 09:27:32 crc kubenswrapper[4776]: E1125 09:27:32.431149 4776 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 25 09:27:32 crc kubenswrapper[4776]: E1125 09:27:32.431284 4776 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j6rll,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-bhq2s_openshift-marketplace(906f7079-2d0f-4205-8c68-4e66f0f25ab7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 09:27:32 crc kubenswrapper[4776]: E1125 09:27:32.432363 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-bhq2s" podUID="906f7079-2d0f-4205-8c68-4e66f0f25ab7" Nov 25 09:27:32 crc kubenswrapper[4776]: E1125 09:27:32.432407 4776 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 25 09:27:32 crc kubenswrapper[4776]: E1125 09:27:32.432463 4776 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kmb8m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-vtzcb_openshift-marketplace(f04466da-02ff-40ea-bbf4-5761ab1d93b6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 09:27:32 crc kubenswrapper[4776]: E1125 09:27:32.433916 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-vtzcb" podUID="f04466da-02ff-40ea-bbf4-5761ab1d93b6" Nov 25 09:27:32 crc kubenswrapper[4776]: I1125 09:27:32.628191 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-nvc68"] Nov 25 09:27:32 crc kubenswrapper[4776]: I1125 09:27:32.899992 4776 generic.go:334] "Generic (PLEG): container finished" podID="a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4" containerID="7ba5a9ee88320e50a4c8851248d2c590c2254ea1f44a2e1e8a165ddcbe7bf9d9" exitCode=0 Nov 25 09:27:32 crc kubenswrapper[4776]: I1125 09:27:32.900039 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-424gh" event={"ID":"a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4","Type":"ContainerDied","Data":"7ba5a9ee88320e50a4c8851248d2c590c2254ea1f44a2e1e8a165ddcbe7bf9d9"} Nov 25 09:27:32 crc kubenswrapper[4776]: I1125 09:27:32.901972 4776 generic.go:334] "Generic (PLEG): container finished" podID="d03c6574-9476-4ef1-8a83-68c7b2ce7a7c" containerID="550ac2e6526bd78ca43ddc52db2ec48621ed4c99aeb5da893b2ce811b31d6a0a" exitCode=0 Nov 25 09:27:32 crc kubenswrapper[4776]: I1125 09:27:32.902035 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lvw5w" event={"ID":"d03c6574-9476-4ef1-8a83-68c7b2ce7a7c","Type":"ContainerDied","Data":"550ac2e6526bd78ca43ddc52db2ec48621ed4c99aeb5da893b2ce811b31d6a0a"} Nov 25 09:27:32 crc kubenswrapper[4776]: I1125 09:27:32.904579 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-nvc68" event={"ID":"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087","Type":"ContainerStarted","Data":"77128408be4f4f23513410a2a1f191a6f2ee45b8f2b1e20fbefe6bc2cfba8d74"} Nov 25 09:27:32 crc kubenswrapper[4776]: I1125 09:27:32.904607 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-nvc68" event={"ID":"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087","Type":"ContainerStarted","Data":"8c1d748502ef5b7badfaf34072fafa5fc828b09858480376dfb1518d12cde2cc"} Nov 25 09:27:32 crc kubenswrapper[4776]: E1125 09:27:32.906381 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-bhq2s" podUID="906f7079-2d0f-4205-8c68-4e66f0f25ab7" Nov 25 09:27:32 crc kubenswrapper[4776]: E1125 09:27:32.906446 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-vtzcb" podUID="f04466da-02ff-40ea-bbf4-5761ab1d93b6" Nov 25 09:27:33 crc kubenswrapper[4776]: I1125 09:27:33.912874 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-nvc68" event={"ID":"4ce4494d-8f1f-4c9c-bbd3-a7946ebe3087","Type":"ContainerStarted","Data":"21cec81d70b69b3cb3ed2325743c628e556c6a0e725a23479494e83875623270"} Nov 25 09:27:33 crc kubenswrapper[4776]: I1125 09:27:33.915436 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-424gh" event={"ID":"a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4","Type":"ContainerStarted","Data":"f35834a96b6f4a6905e8dfd87eed8775cb1d8e2f2e8090f1436ffd51750dfc10"} Nov 25 09:27:33 crc kubenswrapper[4776]: I1125 09:27:33.917816 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lvw5w" event={"ID":"d03c6574-9476-4ef1-8a83-68c7b2ce7a7c","Type":"ContainerStarted","Data":"d2bfa7e181c5e7b0ef5cf046e89bde5d6648a980d989f69f6bf142def31cd4ee"} Nov 25 09:27:33 crc kubenswrapper[4776]: I1125 09:27:33.956827 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-424gh" podStartSLOduration=2.222152697 podStartE2EDuration="37.956807455s" podCreationTimestamp="2025-11-25 09:26:56 +0000 UTC" firstStartedPulling="2025-11-25 09:26:57.602859026 +0000 UTC m=+162.643918579" lastFinishedPulling="2025-11-25 09:27:33.337513784 +0000 UTC m=+198.378573337" observedRunningTime="2025-11-25 09:27:33.95343761 +0000 UTC m=+198.994497163" watchObservedRunningTime="2025-11-25 09:27:33.956807455 +0000 UTC m=+198.997867018" Nov 25 09:27:33 crc kubenswrapper[4776]: I1125 09:27:33.957080 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-nvc68" podStartSLOduration=172.957055551 podStartE2EDuration="2m52.957055551s" podCreationTimestamp="2025-11-25 09:24:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:27:33.938349961 +0000 UTC m=+198.979409534" watchObservedRunningTime="2025-11-25 09:27:33.957055551 +0000 UTC m=+198.998115114" Nov 25 09:27:33 crc kubenswrapper[4776]: I1125 09:27:33.975729 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lvw5w" podStartSLOduration=3.06247905 podStartE2EDuration="37.97571133s" podCreationTimestamp="2025-11-25 09:26:56 +0000 UTC" firstStartedPulling="2025-11-25 09:26:58.611885878 +0000 UTC m=+163.652945431" lastFinishedPulling="2025-11-25 09:27:33.525118158 +0000 UTC m=+198.566177711" observedRunningTime="2025-11-25 09:27:33.972009097 +0000 UTC m=+199.013068680" watchObservedRunningTime="2025-11-25 09:27:33.97571133 +0000 UTC m=+199.016770883" Nov 25 09:27:36 crc kubenswrapper[4776]: I1125 09:27:36.415944 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-424gh" Nov 25 09:27:36 crc kubenswrapper[4776]: I1125 09:27:36.416480 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-424gh" Nov 25 09:27:36 crc kubenswrapper[4776]: I1125 09:27:36.589585 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-424gh" Nov 25 09:27:36 crc kubenswrapper[4776]: I1125 09:27:36.823040 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lvw5w" Nov 25 09:27:36 crc kubenswrapper[4776]: I1125 09:27:36.823120 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lvw5w" Nov 25 09:27:36 crc kubenswrapper[4776]: I1125 09:27:36.889808 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lvw5w" Nov 25 09:27:39 crc kubenswrapper[4776]: I1125 09:27:39.953374 4776 generic.go:334] "Generic (PLEG): container finished" podID="b9a42349-e392-40c2-a6e8-32917240a4ad" containerID="d449736365749162bf20d43848a01157e34a14559ad0bdeb42a86596048ce284" exitCode=0 Nov 25 09:27:39 crc kubenswrapper[4776]: I1125 09:27:39.953505 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4hbp8" event={"ID":"b9a42349-e392-40c2-a6e8-32917240a4ad","Type":"ContainerDied","Data":"d449736365749162bf20d43848a01157e34a14559ad0bdeb42a86596048ce284"} Nov 25 09:27:40 crc kubenswrapper[4776]: I1125 09:27:40.961310 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4hbp8" event={"ID":"b9a42349-e392-40c2-a6e8-32917240a4ad","Type":"ContainerStarted","Data":"433260df8679fd7846bf1b2320365b2c4bb4e3519fcc542ec8827f0138650f26"} Nov 25 09:27:40 crc kubenswrapper[4776]: I1125 09:27:40.984835 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4hbp8" podStartSLOduration=2.104814007 podStartE2EDuration="46.984817652s" podCreationTimestamp="2025-11-25 09:26:54 +0000 UTC" firstStartedPulling="2025-11-25 09:26:55.521778047 +0000 UTC m=+160.562837600" lastFinishedPulling="2025-11-25 09:27:40.401781692 +0000 UTC m=+205.442841245" observedRunningTime="2025-11-25 09:27:40.982674818 +0000 UTC m=+206.023734381" watchObservedRunningTime="2025-11-25 09:27:40.984817652 +0000 UTC m=+206.025877215" Nov 25 09:27:41 crc kubenswrapper[4776]: I1125 09:27:41.967631 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zd2l7" event={"ID":"2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6","Type":"ContainerStarted","Data":"67807704b3a78d02a654d934ffc5c9d67cddb72a357b2d264879685d28d66525"} Nov 25 09:27:42 crc kubenswrapper[4776]: I1125 09:27:42.974905 4776 generic.go:334] "Generic (PLEG): container finished" podID="2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6" containerID="67807704b3a78d02a654d934ffc5c9d67cddb72a357b2d264879685d28d66525" exitCode=0 Nov 25 09:27:42 crc kubenswrapper[4776]: I1125 09:27:42.974954 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zd2l7" event={"ID":"2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6","Type":"ContainerDied","Data":"67807704b3a78d02a654d934ffc5c9d67cddb72a357b2d264879685d28d66525"} Nov 25 09:27:43 crc kubenswrapper[4776]: I1125 09:27:43.982031 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zd2l7" event={"ID":"2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6","Type":"ContainerStarted","Data":"6c98d48cbf363ef9564b4eb50524f27e3b2df256b9de3d10178317762cd09a92"} Nov 25 09:27:43 crc kubenswrapper[4776]: I1125 09:27:43.999838 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zd2l7" podStartSLOduration=3.070062668 podStartE2EDuration="50.999821972s" podCreationTimestamp="2025-11-25 09:26:53 +0000 UTC" firstStartedPulling="2025-11-25 09:26:55.536321426 +0000 UTC m=+160.577380979" lastFinishedPulling="2025-11-25 09:27:43.46608073 +0000 UTC m=+208.507140283" observedRunningTime="2025-11-25 09:27:43.998774115 +0000 UTC m=+209.039833668" watchObservedRunningTime="2025-11-25 09:27:43.999821972 +0000 UTC m=+209.040881525" Nov 25 09:27:44 crc kubenswrapper[4776]: I1125 09:27:44.181956 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zd2l7" Nov 25 09:27:44 crc kubenswrapper[4776]: I1125 09:27:44.182028 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zd2l7" Nov 25 09:27:44 crc kubenswrapper[4776]: I1125 09:27:44.610889 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4hbp8" Nov 25 09:27:44 crc kubenswrapper[4776]: I1125 09:27:44.611192 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4hbp8" Nov 25 09:27:44 crc kubenswrapper[4776]: I1125 09:27:44.650980 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4hbp8" Nov 25 09:27:45 crc kubenswrapper[4776]: I1125 09:27:45.219335 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-zd2l7" podUID="2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6" containerName="registry-server" probeResult="failure" output=< Nov 25 09:27:45 crc kubenswrapper[4776]: timeout: failed to connect service ":50051" within 1s Nov 25 09:27:45 crc kubenswrapper[4776]: > Nov 25 09:27:45 crc kubenswrapper[4776]: I1125 09:27:45.905284 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8rhfr"] Nov 25 09:27:45 crc kubenswrapper[4776]: I1125 09:27:45.991676 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vtzcb" event={"ID":"f04466da-02ff-40ea-bbf4-5761ab1d93b6","Type":"ContainerStarted","Data":"5bd3d984c7e13c499dee93d8ca51ce2f2e500ac426765a6852d8a8d35962b914"} Nov 25 09:27:46 crc kubenswrapper[4776]: I1125 09:27:46.460495 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-424gh" Nov 25 09:27:46 crc kubenswrapper[4776]: I1125 09:27:46.865708 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lvw5w" Nov 25 09:27:46 crc kubenswrapper[4776]: I1125 09:27:46.998117 4776 generic.go:334] "Generic (PLEG): container finished" podID="e057eef9-0e1f-4272-8525-3a4350250d34" containerID="99ddbbab9d1679f0278d7ea97c61135a720d506e7ea38657567c3796cfe8b7e5" exitCode=0 Nov 25 09:27:46 crc kubenswrapper[4776]: I1125 09:27:46.998179 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xz58l" event={"ID":"e057eef9-0e1f-4272-8525-3a4350250d34","Type":"ContainerDied","Data":"99ddbbab9d1679f0278d7ea97c61135a720d506e7ea38657567c3796cfe8b7e5"} Nov 25 09:27:47 crc kubenswrapper[4776]: I1125 09:27:47.002294 4776 generic.go:334] "Generic (PLEG): container finished" podID="f04466da-02ff-40ea-bbf4-5761ab1d93b6" containerID="5bd3d984c7e13c499dee93d8ca51ce2f2e500ac426765a6852d8a8d35962b914" exitCode=0 Nov 25 09:27:47 crc kubenswrapper[4776]: I1125 09:27:47.002333 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vtzcb" event={"ID":"f04466da-02ff-40ea-bbf4-5761ab1d93b6","Type":"ContainerDied","Data":"5bd3d984c7e13c499dee93d8ca51ce2f2e500ac426765a6852d8a8d35962b914"} Nov 25 09:27:47 crc kubenswrapper[4776]: I1125 09:27:47.818246 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:27:47 crc kubenswrapper[4776]: I1125 09:27:47.818583 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:27:47 crc kubenswrapper[4776]: I1125 09:27:47.818635 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 09:27:47 crc kubenswrapper[4776]: I1125 09:27:47.819213 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 09:27:47 crc kubenswrapper[4776]: I1125 09:27:47.819308 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb" gracePeriod=600 Nov 25 09:27:48 crc kubenswrapper[4776]: I1125 09:27:48.018103 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpmxl" event={"ID":"fa898cba-424e-4a13-93de-0e869ccc6780","Type":"ContainerStarted","Data":"e280e4312dc56ebbc0fe57df6397be0c895fe3fb45aad3d30e73dddaabd4f357"} Nov 25 09:27:48 crc kubenswrapper[4776]: I1125 09:27:48.022828 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb" exitCode=0 Nov 25 09:27:48 crc kubenswrapper[4776]: I1125 09:27:48.022902 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb"} Nov 25 09:27:48 crc kubenswrapper[4776]: I1125 09:27:48.028180 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bhq2s" event={"ID":"906f7079-2d0f-4205-8c68-4e66f0f25ab7","Type":"ContainerStarted","Data":"a424b18e6bfd5d5e15fd2e35b8d9d449a5a12102d82bdee34952ba7300ddfe54"} Nov 25 09:27:49 crc kubenswrapper[4776]: I1125 09:27:49.035856 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa898cba-424e-4a13-93de-0e869ccc6780" containerID="e280e4312dc56ebbc0fe57df6397be0c895fe3fb45aad3d30e73dddaabd4f357" exitCode=0 Nov 25 09:27:49 crc kubenswrapper[4776]: I1125 09:27:49.036058 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpmxl" event={"ID":"fa898cba-424e-4a13-93de-0e869ccc6780","Type":"ContainerDied","Data":"e280e4312dc56ebbc0fe57df6397be0c895fe3fb45aad3d30e73dddaabd4f357"} Nov 25 09:27:49 crc kubenswrapper[4776]: I1125 09:27:49.040469 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"2898978600d851d8d30b75f90d2f6443dc65ff39be5ce74362b19bd50031aee8"} Nov 25 09:27:49 crc kubenswrapper[4776]: I1125 09:27:49.042883 4776 generic.go:334] "Generic (PLEG): container finished" podID="906f7079-2d0f-4205-8c68-4e66f0f25ab7" containerID="a424b18e6bfd5d5e15fd2e35b8d9d449a5a12102d82bdee34952ba7300ddfe54" exitCode=0 Nov 25 09:27:49 crc kubenswrapper[4776]: I1125 09:27:49.042957 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bhq2s" event={"ID":"906f7079-2d0f-4205-8c68-4e66f0f25ab7","Type":"ContainerDied","Data":"a424b18e6bfd5d5e15fd2e35b8d9d449a5a12102d82bdee34952ba7300ddfe54"} Nov 25 09:27:49 crc kubenswrapper[4776]: I1125 09:27:49.044943 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vtzcb" event={"ID":"f04466da-02ff-40ea-bbf4-5761ab1d93b6","Type":"ContainerStarted","Data":"fcb02491623a58ad77246475d2fb42c5b52591265c3a51bb480c9e4acb1a3b1a"} Nov 25 09:27:49 crc kubenswrapper[4776]: I1125 09:27:49.047519 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xz58l" event={"ID":"e057eef9-0e1f-4272-8525-3a4350250d34","Type":"ContainerStarted","Data":"f4c230a58885c09d82df1d41502ae961d28381fb3d5a7e0c27cebfeed965d2c3"} Nov 25 09:27:49 crc kubenswrapper[4776]: I1125 09:27:49.092792 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vtzcb" podStartSLOduration=2.598089873 podStartE2EDuration="52.092774796s" podCreationTimestamp="2025-11-25 09:26:57 +0000 UTC" firstStartedPulling="2025-11-25 09:26:58.604176538 +0000 UTC m=+163.645236091" lastFinishedPulling="2025-11-25 09:27:48.098861461 +0000 UTC m=+213.139921014" observedRunningTime="2025-11-25 09:27:49.090385496 +0000 UTC m=+214.131445059" watchObservedRunningTime="2025-11-25 09:27:49.092774796 +0000 UTC m=+214.133834349" Nov 25 09:27:49 crc kubenswrapper[4776]: I1125 09:27:49.121743 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xz58l" podStartSLOduration=2.735459552 podStartE2EDuration="55.121723223s" podCreationTimestamp="2025-11-25 09:26:54 +0000 UTC" firstStartedPulling="2025-11-25 09:26:55.52715712 +0000 UTC m=+160.568216673" lastFinishedPulling="2025-11-25 09:27:47.913420791 +0000 UTC m=+212.954480344" observedRunningTime="2025-11-25 09:27:49.120965254 +0000 UTC m=+214.162024807" watchObservedRunningTime="2025-11-25 09:27:49.121723223 +0000 UTC m=+214.162782776" Nov 25 09:27:49 crc kubenswrapper[4776]: I1125 09:27:49.528635 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lvw5w"] Nov 25 09:27:49 crc kubenswrapper[4776]: I1125 09:27:49.529090 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lvw5w" podUID="d03c6574-9476-4ef1-8a83-68c7b2ce7a7c" containerName="registry-server" containerID="cri-o://d2bfa7e181c5e7b0ef5cf046e89bde5d6648a980d989f69f6bf142def31cd4ee" gracePeriod=2 Nov 25 09:27:49 crc kubenswrapper[4776]: I1125 09:27:49.923006 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lvw5w" Nov 25 09:27:50 crc kubenswrapper[4776]: I1125 09:27:50.042589 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nr7h\" (UniqueName: \"kubernetes.io/projected/d03c6574-9476-4ef1-8a83-68c7b2ce7a7c-kube-api-access-6nr7h\") pod \"d03c6574-9476-4ef1-8a83-68c7b2ce7a7c\" (UID: \"d03c6574-9476-4ef1-8a83-68c7b2ce7a7c\") " Nov 25 09:27:50 crc kubenswrapper[4776]: I1125 09:27:50.042657 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d03c6574-9476-4ef1-8a83-68c7b2ce7a7c-catalog-content\") pod \"d03c6574-9476-4ef1-8a83-68c7b2ce7a7c\" (UID: \"d03c6574-9476-4ef1-8a83-68c7b2ce7a7c\") " Nov 25 09:27:50 crc kubenswrapper[4776]: I1125 09:27:50.042719 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d03c6574-9476-4ef1-8a83-68c7b2ce7a7c-utilities\") pod \"d03c6574-9476-4ef1-8a83-68c7b2ce7a7c\" (UID: \"d03c6574-9476-4ef1-8a83-68c7b2ce7a7c\") " Nov 25 09:27:50 crc kubenswrapper[4776]: I1125 09:27:50.043646 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d03c6574-9476-4ef1-8a83-68c7b2ce7a7c-utilities" (OuterVolumeSpecName: "utilities") pod "d03c6574-9476-4ef1-8a83-68c7b2ce7a7c" (UID: "d03c6574-9476-4ef1-8a83-68c7b2ce7a7c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:27:50 crc kubenswrapper[4776]: I1125 09:27:50.054327 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d03c6574-9476-4ef1-8a83-68c7b2ce7a7c-kube-api-access-6nr7h" (OuterVolumeSpecName: "kube-api-access-6nr7h") pod "d03c6574-9476-4ef1-8a83-68c7b2ce7a7c" (UID: "d03c6574-9476-4ef1-8a83-68c7b2ce7a7c"). InnerVolumeSpecName "kube-api-access-6nr7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:27:50 crc kubenswrapper[4776]: I1125 09:27:50.059529 4776 generic.go:334] "Generic (PLEG): container finished" podID="d03c6574-9476-4ef1-8a83-68c7b2ce7a7c" containerID="d2bfa7e181c5e7b0ef5cf046e89bde5d6648a980d989f69f6bf142def31cd4ee" exitCode=0 Nov 25 09:27:50 crc kubenswrapper[4776]: I1125 09:27:50.059598 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lvw5w" event={"ID":"d03c6574-9476-4ef1-8a83-68c7b2ce7a7c","Type":"ContainerDied","Data":"d2bfa7e181c5e7b0ef5cf046e89bde5d6648a980d989f69f6bf142def31cd4ee"} Nov 25 09:27:50 crc kubenswrapper[4776]: I1125 09:27:50.059629 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lvw5w" event={"ID":"d03c6574-9476-4ef1-8a83-68c7b2ce7a7c","Type":"ContainerDied","Data":"cc57b99a40a352b676061f940813b6ab071c32ec4c696dbfce8c0cf5a64db228"} Nov 25 09:27:50 crc kubenswrapper[4776]: I1125 09:27:50.059651 4776 scope.go:117] "RemoveContainer" containerID="d2bfa7e181c5e7b0ef5cf046e89bde5d6648a980d989f69f6bf142def31cd4ee" Nov 25 09:27:50 crc kubenswrapper[4776]: I1125 09:27:50.059792 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lvw5w" Nov 25 09:27:50 crc kubenswrapper[4776]: I1125 09:27:50.065949 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpmxl" event={"ID":"fa898cba-424e-4a13-93de-0e869ccc6780","Type":"ContainerStarted","Data":"94e87c7fb1fddfd2ba93e6b4a99d1442e85b19798a5aee37c32dc0a9628a3c96"} Nov 25 09:27:50 crc kubenswrapper[4776]: I1125 09:27:50.073327 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bhq2s" event={"ID":"906f7079-2d0f-4205-8c68-4e66f0f25ab7","Type":"ContainerStarted","Data":"208eec9295acdad309b5995cddb470d33db8ae415b6759010f1b92ec115fe1eb"} Nov 25 09:27:50 crc kubenswrapper[4776]: I1125 09:27:50.075364 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d03c6574-9476-4ef1-8a83-68c7b2ce7a7c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d03c6574-9476-4ef1-8a83-68c7b2ce7a7c" (UID: "d03c6574-9476-4ef1-8a83-68c7b2ce7a7c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:27:50 crc kubenswrapper[4776]: I1125 09:27:50.080613 4776 scope.go:117] "RemoveContainer" containerID="550ac2e6526bd78ca43ddc52db2ec48621ed4c99aeb5da893b2ce811b31d6a0a" Nov 25 09:27:50 crc kubenswrapper[4776]: I1125 09:27:50.089942 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rpmxl" podStartSLOduration=2.198082957 podStartE2EDuration="56.089924272s" podCreationTimestamp="2025-11-25 09:26:54 +0000 UTC" firstStartedPulling="2025-11-25 09:26:55.531431925 +0000 UTC m=+160.572491518" lastFinishedPulling="2025-11-25 09:27:49.42327328 +0000 UTC m=+214.464332833" observedRunningTime="2025-11-25 09:27:50.088810884 +0000 UTC m=+215.129870447" watchObservedRunningTime="2025-11-25 09:27:50.089924272 +0000 UTC m=+215.130983825" Nov 25 09:27:50 crc kubenswrapper[4776]: I1125 09:27:50.097024 4776 scope.go:117] "RemoveContainer" containerID="8682a296ed548c21d62d6f831c43389df48184c667a1917ffb3ea8bc60ee9da8" Nov 25 09:27:50 crc kubenswrapper[4776]: I1125 09:27:50.114239 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bhq2s" podStartSLOduration=3.200716888 podStartE2EDuration="53.114215652s" podCreationTimestamp="2025-11-25 09:26:57 +0000 UTC" firstStartedPulling="2025-11-25 09:26:59.623285219 +0000 UTC m=+164.664344772" lastFinishedPulling="2025-11-25 09:27:49.536783983 +0000 UTC m=+214.577843536" observedRunningTime="2025-11-25 09:27:50.113531855 +0000 UTC m=+215.154591408" watchObservedRunningTime="2025-11-25 09:27:50.114215652 +0000 UTC m=+215.155275205" Nov 25 09:27:50 crc kubenswrapper[4776]: I1125 09:27:50.131643 4776 scope.go:117] "RemoveContainer" containerID="d2bfa7e181c5e7b0ef5cf046e89bde5d6648a980d989f69f6bf142def31cd4ee" Nov 25 09:27:50 crc kubenswrapper[4776]: E1125 09:27:50.132376 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2bfa7e181c5e7b0ef5cf046e89bde5d6648a980d989f69f6bf142def31cd4ee\": container with ID starting with d2bfa7e181c5e7b0ef5cf046e89bde5d6648a980d989f69f6bf142def31cd4ee not found: ID does not exist" containerID="d2bfa7e181c5e7b0ef5cf046e89bde5d6648a980d989f69f6bf142def31cd4ee" Nov 25 09:27:50 crc kubenswrapper[4776]: I1125 09:27:50.132609 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2bfa7e181c5e7b0ef5cf046e89bde5d6648a980d989f69f6bf142def31cd4ee"} err="failed to get container status \"d2bfa7e181c5e7b0ef5cf046e89bde5d6648a980d989f69f6bf142def31cd4ee\": rpc error: code = NotFound desc = could not find container \"d2bfa7e181c5e7b0ef5cf046e89bde5d6648a980d989f69f6bf142def31cd4ee\": container with ID starting with d2bfa7e181c5e7b0ef5cf046e89bde5d6648a980d989f69f6bf142def31cd4ee not found: ID does not exist" Nov 25 09:27:50 crc kubenswrapper[4776]: I1125 09:27:50.132734 4776 scope.go:117] "RemoveContainer" containerID="550ac2e6526bd78ca43ddc52db2ec48621ed4c99aeb5da893b2ce811b31d6a0a" Nov 25 09:27:50 crc kubenswrapper[4776]: E1125 09:27:50.133089 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"550ac2e6526bd78ca43ddc52db2ec48621ed4c99aeb5da893b2ce811b31d6a0a\": container with ID starting with 550ac2e6526bd78ca43ddc52db2ec48621ed4c99aeb5da893b2ce811b31d6a0a not found: ID does not exist" containerID="550ac2e6526bd78ca43ddc52db2ec48621ed4c99aeb5da893b2ce811b31d6a0a" Nov 25 09:27:50 crc kubenswrapper[4776]: I1125 09:27:50.133199 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"550ac2e6526bd78ca43ddc52db2ec48621ed4c99aeb5da893b2ce811b31d6a0a"} err="failed to get container status \"550ac2e6526bd78ca43ddc52db2ec48621ed4c99aeb5da893b2ce811b31d6a0a\": rpc error: code = NotFound desc = could not find container \"550ac2e6526bd78ca43ddc52db2ec48621ed4c99aeb5da893b2ce811b31d6a0a\": container with ID starting with 550ac2e6526bd78ca43ddc52db2ec48621ed4c99aeb5da893b2ce811b31d6a0a not found: ID does not exist" Nov 25 09:27:50 crc kubenswrapper[4776]: I1125 09:27:50.133304 4776 scope.go:117] "RemoveContainer" containerID="8682a296ed548c21d62d6f831c43389df48184c667a1917ffb3ea8bc60ee9da8" Nov 25 09:27:50 crc kubenswrapper[4776]: E1125 09:27:50.133838 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8682a296ed548c21d62d6f831c43389df48184c667a1917ffb3ea8bc60ee9da8\": container with ID starting with 8682a296ed548c21d62d6f831c43389df48184c667a1917ffb3ea8bc60ee9da8 not found: ID does not exist" containerID="8682a296ed548c21d62d6f831c43389df48184c667a1917ffb3ea8bc60ee9da8" Nov 25 09:27:50 crc kubenswrapper[4776]: I1125 09:27:50.133882 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8682a296ed548c21d62d6f831c43389df48184c667a1917ffb3ea8bc60ee9da8"} err="failed to get container status \"8682a296ed548c21d62d6f831c43389df48184c667a1917ffb3ea8bc60ee9da8\": rpc error: code = NotFound desc = could not find container \"8682a296ed548c21d62d6f831c43389df48184c667a1917ffb3ea8bc60ee9da8\": container with ID starting with 8682a296ed548c21d62d6f831c43389df48184c667a1917ffb3ea8bc60ee9da8 not found: ID does not exist" Nov 25 09:27:50 crc kubenswrapper[4776]: I1125 09:27:50.144791 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d03c6574-9476-4ef1-8a83-68c7b2ce7a7c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:27:50 crc kubenswrapper[4776]: I1125 09:27:50.145595 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d03c6574-9476-4ef1-8a83-68c7b2ce7a7c-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:27:50 crc kubenswrapper[4776]: I1125 09:27:50.145718 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nr7h\" (UniqueName: \"kubernetes.io/projected/d03c6574-9476-4ef1-8a83-68c7b2ce7a7c-kube-api-access-6nr7h\") on node \"crc\" DevicePath \"\"" Nov 25 09:27:50 crc kubenswrapper[4776]: I1125 09:27:50.393915 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lvw5w"] Nov 25 09:27:50 crc kubenswrapper[4776]: I1125 09:27:50.397173 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lvw5w"] Nov 25 09:27:51 crc kubenswrapper[4776]: I1125 09:27:51.668806 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d03c6574-9476-4ef1-8a83-68c7b2ce7a7c" path="/var/lib/kubelet/pods/d03c6574-9476-4ef1-8a83-68c7b2ce7a7c/volumes" Nov 25 09:27:54 crc kubenswrapper[4776]: I1125 09:27:54.238273 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zd2l7" Nov 25 09:27:54 crc kubenswrapper[4776]: I1125 09:27:54.284685 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zd2l7" Nov 25 09:27:54 crc kubenswrapper[4776]: I1125 09:27:54.403585 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xz58l" Nov 25 09:27:54 crc kubenswrapper[4776]: I1125 09:27:54.403634 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xz58l" Nov 25 09:27:54 crc kubenswrapper[4776]: I1125 09:27:54.439948 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xz58l" Nov 25 09:27:54 crc kubenswrapper[4776]: I1125 09:27:54.668320 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4hbp8" Nov 25 09:27:54 crc kubenswrapper[4776]: I1125 09:27:54.800286 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rpmxl" Nov 25 09:27:54 crc kubenswrapper[4776]: I1125 09:27:54.800350 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rpmxl" Nov 25 09:27:54 crc kubenswrapper[4776]: I1125 09:27:54.858871 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rpmxl" Nov 25 09:27:55 crc kubenswrapper[4776]: I1125 09:27:55.146944 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rpmxl" Nov 25 09:27:55 crc kubenswrapper[4776]: I1125 09:27:55.154058 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xz58l" Nov 25 09:27:56 crc kubenswrapper[4776]: I1125 09:27:56.131458 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4hbp8"] Nov 25 09:27:56 crc kubenswrapper[4776]: I1125 09:27:56.131730 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4hbp8" podUID="b9a42349-e392-40c2-a6e8-32917240a4ad" containerName="registry-server" containerID="cri-o://433260df8679fd7846bf1b2320365b2c4bb4e3519fcc542ec8827f0138650f26" gracePeriod=2 Nov 25 09:27:57 crc kubenswrapper[4776]: I1125 09:27:57.253845 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rpmxl"] Nov 25 09:27:57 crc kubenswrapper[4776]: I1125 09:27:57.267409 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rpmxl" podUID="fa898cba-424e-4a13-93de-0e869ccc6780" containerName="registry-server" containerID="cri-o://94e87c7fb1fddfd2ba93e6b4a99d1442e85b19798a5aee37c32dc0a9628a3c96" gracePeriod=2 Nov 25 09:27:57 crc kubenswrapper[4776]: I1125 09:27:57.548247 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vtzcb" Nov 25 09:27:57 crc kubenswrapper[4776]: I1125 09:27:57.548603 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vtzcb" Nov 25 09:27:57 crc kubenswrapper[4776]: I1125 09:27:57.581371 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vtzcb" Nov 25 09:27:57 crc kubenswrapper[4776]: I1125 09:27:57.802719 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bhq2s" Nov 25 09:27:57 crc kubenswrapper[4776]: I1125 09:27:57.803182 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bhq2s" Nov 25 09:27:57 crc kubenswrapper[4776]: I1125 09:27:57.871747 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bhq2s" Nov 25 09:27:58 crc kubenswrapper[4776]: I1125 09:27:58.189184 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bhq2s" Nov 25 09:27:58 crc kubenswrapper[4776]: I1125 09:27:58.206206 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vtzcb" Nov 25 09:27:59 crc kubenswrapper[4776]: I1125 09:27:59.149965 4776 generic.go:334] "Generic (PLEG): container finished" podID="b9a42349-e392-40c2-a6e8-32917240a4ad" containerID="433260df8679fd7846bf1b2320365b2c4bb4e3519fcc542ec8827f0138650f26" exitCode=0 Nov 25 09:27:59 crc kubenswrapper[4776]: I1125 09:27:59.150039 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4hbp8" event={"ID":"b9a42349-e392-40c2-a6e8-32917240a4ad","Type":"ContainerDied","Data":"433260df8679fd7846bf1b2320365b2c4bb4e3519fcc542ec8827f0138650f26"} Nov 25 09:27:59 crc kubenswrapper[4776]: I1125 09:27:59.155026 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa898cba-424e-4a13-93de-0e869ccc6780" containerID="94e87c7fb1fddfd2ba93e6b4a99d1442e85b19798a5aee37c32dc0a9628a3c96" exitCode=0 Nov 25 09:27:59 crc kubenswrapper[4776]: I1125 09:27:59.155115 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpmxl" event={"ID":"fa898cba-424e-4a13-93de-0e869ccc6780","Type":"ContainerDied","Data":"94e87c7fb1fddfd2ba93e6b4a99d1442e85b19798a5aee37c32dc0a9628a3c96"} Nov 25 09:27:59 crc kubenswrapper[4776]: I1125 09:27:59.902525 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4hbp8" Nov 25 09:27:59 crc kubenswrapper[4776]: I1125 09:27:59.996097 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nqbw\" (UniqueName: \"kubernetes.io/projected/b9a42349-e392-40c2-a6e8-32917240a4ad-kube-api-access-8nqbw\") pod \"b9a42349-e392-40c2-a6e8-32917240a4ad\" (UID: \"b9a42349-e392-40c2-a6e8-32917240a4ad\") " Nov 25 09:27:59 crc kubenswrapper[4776]: I1125 09:27:59.996163 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9a42349-e392-40c2-a6e8-32917240a4ad-utilities\") pod \"b9a42349-e392-40c2-a6e8-32917240a4ad\" (UID: \"b9a42349-e392-40c2-a6e8-32917240a4ad\") " Nov 25 09:27:59 crc kubenswrapper[4776]: I1125 09:27:59.996246 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9a42349-e392-40c2-a6e8-32917240a4ad-catalog-content\") pod \"b9a42349-e392-40c2-a6e8-32917240a4ad\" (UID: \"b9a42349-e392-40c2-a6e8-32917240a4ad\") " Nov 25 09:27:59 crc kubenswrapper[4776]: I1125 09:27:59.997681 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9a42349-e392-40c2-a6e8-32917240a4ad-utilities" (OuterVolumeSpecName: "utilities") pod "b9a42349-e392-40c2-a6e8-32917240a4ad" (UID: "b9a42349-e392-40c2-a6e8-32917240a4ad"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:28:00 crc kubenswrapper[4776]: I1125 09:28:00.003924 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9a42349-e392-40c2-a6e8-32917240a4ad-kube-api-access-8nqbw" (OuterVolumeSpecName: "kube-api-access-8nqbw") pod "b9a42349-e392-40c2-a6e8-32917240a4ad" (UID: "b9a42349-e392-40c2-a6e8-32917240a4ad"). InnerVolumeSpecName "kube-api-access-8nqbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:28:00 crc kubenswrapper[4776]: I1125 09:28:00.036982 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9a42349-e392-40c2-a6e8-32917240a4ad-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b9a42349-e392-40c2-a6e8-32917240a4ad" (UID: "b9a42349-e392-40c2-a6e8-32917240a4ad"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:28:00 crc kubenswrapper[4776]: I1125 09:28:00.097807 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nqbw\" (UniqueName: \"kubernetes.io/projected/b9a42349-e392-40c2-a6e8-32917240a4ad-kube-api-access-8nqbw\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:00 crc kubenswrapper[4776]: I1125 09:28:00.097868 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9a42349-e392-40c2-a6e8-32917240a4ad-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:00 crc kubenswrapper[4776]: I1125 09:28:00.097889 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9a42349-e392-40c2-a6e8-32917240a4ad-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:00 crc kubenswrapper[4776]: I1125 09:28:00.166016 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4hbp8" Nov 25 09:28:00 crc kubenswrapper[4776]: I1125 09:28:00.166003 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4hbp8" event={"ID":"b9a42349-e392-40c2-a6e8-32917240a4ad","Type":"ContainerDied","Data":"82f4bfa4b3677700320ef1f2ff5b571fe3d8fc39e6f88f04be2c994030940ed5"} Nov 25 09:28:00 crc kubenswrapper[4776]: I1125 09:28:00.166120 4776 scope.go:117] "RemoveContainer" containerID="433260df8679fd7846bf1b2320365b2c4bb4e3519fcc542ec8827f0138650f26" Nov 25 09:28:00 crc kubenswrapper[4776]: I1125 09:28:00.190446 4776 scope.go:117] "RemoveContainer" containerID="d449736365749162bf20d43848a01157e34a14559ad0bdeb42a86596048ce284" Nov 25 09:28:00 crc kubenswrapper[4776]: I1125 09:28:00.220402 4776 scope.go:117] "RemoveContainer" containerID="f8c42243e16782a0ddbc885783ad5d47d190bd89b4e9f78a77fa89969136cf8a" Nov 25 09:28:00 crc kubenswrapper[4776]: I1125 09:28:00.220710 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4hbp8"] Nov 25 09:28:00 crc kubenswrapper[4776]: I1125 09:28:00.224213 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4hbp8"] Nov 25 09:28:00 crc kubenswrapper[4776]: I1125 09:28:00.432951 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rpmxl" Nov 25 09:28:00 crc kubenswrapper[4776]: I1125 09:28:00.605315 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa898cba-424e-4a13-93de-0e869ccc6780-catalog-content\") pod \"fa898cba-424e-4a13-93de-0e869ccc6780\" (UID: \"fa898cba-424e-4a13-93de-0e869ccc6780\") " Nov 25 09:28:00 crc kubenswrapper[4776]: I1125 09:28:00.605376 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa898cba-424e-4a13-93de-0e869ccc6780-utilities\") pod \"fa898cba-424e-4a13-93de-0e869ccc6780\" (UID: \"fa898cba-424e-4a13-93de-0e869ccc6780\") " Nov 25 09:28:00 crc kubenswrapper[4776]: I1125 09:28:00.605520 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfs2q\" (UniqueName: \"kubernetes.io/projected/fa898cba-424e-4a13-93de-0e869ccc6780-kube-api-access-rfs2q\") pod \"fa898cba-424e-4a13-93de-0e869ccc6780\" (UID: \"fa898cba-424e-4a13-93de-0e869ccc6780\") " Nov 25 09:28:00 crc kubenswrapper[4776]: I1125 09:28:00.606158 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa898cba-424e-4a13-93de-0e869ccc6780-utilities" (OuterVolumeSpecName: "utilities") pod "fa898cba-424e-4a13-93de-0e869ccc6780" (UID: "fa898cba-424e-4a13-93de-0e869ccc6780"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:28:00 crc kubenswrapper[4776]: I1125 09:28:00.610430 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa898cba-424e-4a13-93de-0e869ccc6780-kube-api-access-rfs2q" (OuterVolumeSpecName: "kube-api-access-rfs2q") pod "fa898cba-424e-4a13-93de-0e869ccc6780" (UID: "fa898cba-424e-4a13-93de-0e869ccc6780"). InnerVolumeSpecName "kube-api-access-rfs2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:28:00 crc kubenswrapper[4776]: I1125 09:28:00.707128 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfs2q\" (UniqueName: \"kubernetes.io/projected/fa898cba-424e-4a13-93de-0e869ccc6780-kube-api-access-rfs2q\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:00 crc kubenswrapper[4776]: I1125 09:28:00.707170 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa898cba-424e-4a13-93de-0e869ccc6780-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:01 crc kubenswrapper[4776]: I1125 09:28:01.126847 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bhq2s"] Nov 25 09:28:01 crc kubenswrapper[4776]: I1125 09:28:01.175179 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rpmxl" event={"ID":"fa898cba-424e-4a13-93de-0e869ccc6780","Type":"ContainerDied","Data":"a558cbb45ddf4d73617df1e89c9d52a41fd0c17bf70b9bfd8408284a23af4162"} Nov 25 09:28:01 crc kubenswrapper[4776]: I1125 09:28:01.175239 4776 scope.go:117] "RemoveContainer" containerID="94e87c7fb1fddfd2ba93e6b4a99d1442e85b19798a5aee37c32dc0a9628a3c96" Nov 25 09:28:01 crc kubenswrapper[4776]: I1125 09:28:01.175192 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rpmxl" Nov 25 09:28:01 crc kubenswrapper[4776]: I1125 09:28:01.175419 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bhq2s" podUID="906f7079-2d0f-4205-8c68-4e66f0f25ab7" containerName="registry-server" containerID="cri-o://208eec9295acdad309b5995cddb470d33db8ae415b6759010f1b92ec115fe1eb" gracePeriod=2 Nov 25 09:28:01 crc kubenswrapper[4776]: I1125 09:28:01.190969 4776 scope.go:117] "RemoveContainer" containerID="e280e4312dc56ebbc0fe57df6397be0c895fe3fb45aad3d30e73dddaabd4f357" Nov 25 09:28:01 crc kubenswrapper[4776]: I1125 09:28:01.207975 4776 scope.go:117] "RemoveContainer" containerID="65ffd35e733c330afec288d913f6b319055816eb0c048d59c0bfa50eb662e9cd" Nov 25 09:28:01 crc kubenswrapper[4776]: I1125 09:28:01.676784 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9a42349-e392-40c2-a6e8-32917240a4ad" path="/var/lib/kubelet/pods/b9a42349-e392-40c2-a6e8-32917240a4ad/volumes" Nov 25 09:28:01 crc kubenswrapper[4776]: I1125 09:28:01.691573 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa898cba-424e-4a13-93de-0e869ccc6780-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fa898cba-424e-4a13-93de-0e869ccc6780" (UID: "fa898cba-424e-4a13-93de-0e869ccc6780"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:28:01 crc kubenswrapper[4776]: I1125 09:28:01.720665 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa898cba-424e-4a13-93de-0e869ccc6780-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:01 crc kubenswrapper[4776]: I1125 09:28:01.825852 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rpmxl"] Nov 25 09:28:01 crc kubenswrapper[4776]: I1125 09:28:01.828332 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rpmxl"] Nov 25 09:28:03 crc kubenswrapper[4776]: I1125 09:28:03.193508 4776 generic.go:334] "Generic (PLEG): container finished" podID="906f7079-2d0f-4205-8c68-4e66f0f25ab7" containerID="208eec9295acdad309b5995cddb470d33db8ae415b6759010f1b92ec115fe1eb" exitCode=0 Nov 25 09:28:03 crc kubenswrapper[4776]: I1125 09:28:03.193561 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bhq2s" event={"ID":"906f7079-2d0f-4205-8c68-4e66f0f25ab7","Type":"ContainerDied","Data":"208eec9295acdad309b5995cddb470d33db8ae415b6759010f1b92ec115fe1eb"} Nov 25 09:28:03 crc kubenswrapper[4776]: I1125 09:28:03.614676 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bhq2s" Nov 25 09:28:03 crc kubenswrapper[4776]: I1125 09:28:03.669087 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa898cba-424e-4a13-93de-0e869ccc6780" path="/var/lib/kubelet/pods/fa898cba-424e-4a13-93de-0e869ccc6780/volumes" Nov 25 09:28:03 crc kubenswrapper[4776]: I1125 09:28:03.750507 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/906f7079-2d0f-4205-8c68-4e66f0f25ab7-catalog-content\") pod \"906f7079-2d0f-4205-8c68-4e66f0f25ab7\" (UID: \"906f7079-2d0f-4205-8c68-4e66f0f25ab7\") " Nov 25 09:28:03 crc kubenswrapper[4776]: I1125 09:28:03.750569 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6rll\" (UniqueName: \"kubernetes.io/projected/906f7079-2d0f-4205-8c68-4e66f0f25ab7-kube-api-access-j6rll\") pod \"906f7079-2d0f-4205-8c68-4e66f0f25ab7\" (UID: \"906f7079-2d0f-4205-8c68-4e66f0f25ab7\") " Nov 25 09:28:03 crc kubenswrapper[4776]: I1125 09:28:03.750658 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/906f7079-2d0f-4205-8c68-4e66f0f25ab7-utilities\") pod \"906f7079-2d0f-4205-8c68-4e66f0f25ab7\" (UID: \"906f7079-2d0f-4205-8c68-4e66f0f25ab7\") " Nov 25 09:28:03 crc kubenswrapper[4776]: I1125 09:28:03.751475 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/906f7079-2d0f-4205-8c68-4e66f0f25ab7-utilities" (OuterVolumeSpecName: "utilities") pod "906f7079-2d0f-4205-8c68-4e66f0f25ab7" (UID: "906f7079-2d0f-4205-8c68-4e66f0f25ab7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:28:03 crc kubenswrapper[4776]: I1125 09:28:03.758365 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/906f7079-2d0f-4205-8c68-4e66f0f25ab7-kube-api-access-j6rll" (OuterVolumeSpecName: "kube-api-access-j6rll") pod "906f7079-2d0f-4205-8c68-4e66f0f25ab7" (UID: "906f7079-2d0f-4205-8c68-4e66f0f25ab7"). InnerVolumeSpecName "kube-api-access-j6rll". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:28:03 crc kubenswrapper[4776]: I1125 09:28:03.851957 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6rll\" (UniqueName: \"kubernetes.io/projected/906f7079-2d0f-4205-8c68-4e66f0f25ab7-kube-api-access-j6rll\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:03 crc kubenswrapper[4776]: I1125 09:28:03.851983 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/906f7079-2d0f-4205-8c68-4e66f0f25ab7-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:04 crc kubenswrapper[4776]: I1125 09:28:04.203428 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bhq2s" event={"ID":"906f7079-2d0f-4205-8c68-4e66f0f25ab7","Type":"ContainerDied","Data":"ca6f8f4aa81800ad5d4211f2cb2bf9f027a07fcfffa99d357f83178db6094fb7"} Nov 25 09:28:04 crc kubenswrapper[4776]: I1125 09:28:04.203492 4776 scope.go:117] "RemoveContainer" containerID="208eec9295acdad309b5995cddb470d33db8ae415b6759010f1b92ec115fe1eb" Nov 25 09:28:04 crc kubenswrapper[4776]: I1125 09:28:04.203527 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bhq2s" Nov 25 09:28:04 crc kubenswrapper[4776]: I1125 09:28:04.226421 4776 scope.go:117] "RemoveContainer" containerID="a424b18e6bfd5d5e15fd2e35b8d9d449a5a12102d82bdee34952ba7300ddfe54" Nov 25 09:28:04 crc kubenswrapper[4776]: I1125 09:28:04.248327 4776 scope.go:117] "RemoveContainer" containerID="e469a9bc244eab25c648d83980d20bd80cc80fdf843f3d91433cad0f1394c02e" Nov 25 09:28:04 crc kubenswrapper[4776]: I1125 09:28:04.550545 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/906f7079-2d0f-4205-8c68-4e66f0f25ab7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "906f7079-2d0f-4205-8c68-4e66f0f25ab7" (UID: "906f7079-2d0f-4205-8c68-4e66f0f25ab7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:28:04 crc kubenswrapper[4776]: I1125 09:28:04.565754 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/906f7079-2d0f-4205-8c68-4e66f0f25ab7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:04 crc kubenswrapper[4776]: I1125 09:28:04.828394 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bhq2s"] Nov 25 09:28:04 crc kubenswrapper[4776]: I1125 09:28:04.831996 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bhq2s"] Nov 25 09:28:05 crc kubenswrapper[4776]: I1125 09:28:05.675278 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="906f7079-2d0f-4205-8c68-4e66f0f25ab7" path="/var/lib/kubelet/pods/906f7079-2d0f-4205-8c68-4e66f0f25ab7/volumes" Nov 25 09:28:10 crc kubenswrapper[4776]: I1125 09:28:10.944674 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" podUID="c5ae11bc-db7c-4665-88e4-ef3b0eb83581" containerName="oauth-openshift" containerID="cri-o://d884031de43de28107048c44a1903141869dd6318f0e8ec733b6b2eb1ce4ac13" gracePeriod=15 Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.247257 4776 generic.go:334] "Generic (PLEG): container finished" podID="c5ae11bc-db7c-4665-88e4-ef3b0eb83581" containerID="d884031de43de28107048c44a1903141869dd6318f0e8ec733b6b2eb1ce4ac13" exitCode=0 Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.247331 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" event={"ID":"c5ae11bc-db7c-4665-88e4-ef3b0eb83581","Type":"ContainerDied","Data":"d884031de43de28107048c44a1903141869dd6318f0e8ec733b6b2eb1ce4ac13"} Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.324326 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.452757 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-serving-cert\") pod \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.452847 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-cliconfig\") pod \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.452873 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-trusted-ca-bundle\") pod \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.452907 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-router-certs\") pod \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.452931 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dj8fq\" (UniqueName: \"kubernetes.io/projected/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-kube-api-access-dj8fq\") pod \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.452963 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-user-template-provider-selection\") pod \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.452992 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-user-template-error\") pod \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.453025 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-service-ca\") pod \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.453056 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-session\") pod \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.453343 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-audit-dir\") pod \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.453378 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-user-idp-0-file-data\") pod \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.453409 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-user-template-login\") pod \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.453466 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-audit-policies\") pod \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.453499 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-ocp-branding-template\") pod \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\" (UID: \"c5ae11bc-db7c-4665-88e4-ef3b0eb83581\") " Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.454036 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "c5ae11bc-db7c-4665-88e4-ef3b0eb83581" (UID: "c5ae11bc-db7c-4665-88e4-ef3b0eb83581"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.454403 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "c5ae11bc-db7c-4665-88e4-ef3b0eb83581" (UID: "c5ae11bc-db7c-4665-88e4-ef3b0eb83581"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.454760 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "c5ae11bc-db7c-4665-88e4-ef3b0eb83581" (UID: "c5ae11bc-db7c-4665-88e4-ef3b0eb83581"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.455287 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "c5ae11bc-db7c-4665-88e4-ef3b0eb83581" (UID: "c5ae11bc-db7c-4665-88e4-ef3b0eb83581"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.455362 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "c5ae11bc-db7c-4665-88e4-ef3b0eb83581" (UID: "c5ae11bc-db7c-4665-88e4-ef3b0eb83581"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.459777 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "c5ae11bc-db7c-4665-88e4-ef3b0eb83581" (UID: "c5ae11bc-db7c-4665-88e4-ef3b0eb83581"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.460144 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "c5ae11bc-db7c-4665-88e4-ef3b0eb83581" (UID: "c5ae11bc-db7c-4665-88e4-ef3b0eb83581"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.460415 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "c5ae11bc-db7c-4665-88e4-ef3b0eb83581" (UID: "c5ae11bc-db7c-4665-88e4-ef3b0eb83581"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.460700 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-kube-api-access-dj8fq" (OuterVolumeSpecName: "kube-api-access-dj8fq") pod "c5ae11bc-db7c-4665-88e4-ef3b0eb83581" (UID: "c5ae11bc-db7c-4665-88e4-ef3b0eb83581"). InnerVolumeSpecName "kube-api-access-dj8fq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.460933 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "c5ae11bc-db7c-4665-88e4-ef3b0eb83581" (UID: "c5ae11bc-db7c-4665-88e4-ef3b0eb83581"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.460933 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "c5ae11bc-db7c-4665-88e4-ef3b0eb83581" (UID: "c5ae11bc-db7c-4665-88e4-ef3b0eb83581"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.461286 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "c5ae11bc-db7c-4665-88e4-ef3b0eb83581" (UID: "c5ae11bc-db7c-4665-88e4-ef3b0eb83581"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.461522 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "c5ae11bc-db7c-4665-88e4-ef3b0eb83581" (UID: "c5ae11bc-db7c-4665-88e4-ef3b0eb83581"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.461758 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "c5ae11bc-db7c-4665-88e4-ef3b0eb83581" (UID: "c5ae11bc-db7c-4665-88e4-ef3b0eb83581"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.554490 4776 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.554542 4776 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.554557 4776 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.554572 4776 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.554585 4776 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.554600 4776 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.554613 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dj8fq\" (UniqueName: \"kubernetes.io/projected/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-kube-api-access-dj8fq\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.554625 4776 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.554639 4776 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.554651 4776 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.554662 4776 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.554673 4776 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.554684 4776 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:11 crc kubenswrapper[4776]: I1125 09:28:11.554696 4776 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c5ae11bc-db7c-4665-88e4-ef3b0eb83581-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.242998 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5d7d85f8c7-jc455"] Nov 25 09:28:12 crc kubenswrapper[4776]: E1125 09:28:12.243507 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d03c6574-9476-4ef1-8a83-68c7b2ce7a7c" containerName="extract-utilities" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.243519 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="d03c6574-9476-4ef1-8a83-68c7b2ce7a7c" containerName="extract-utilities" Nov 25 09:28:12 crc kubenswrapper[4776]: E1125 09:28:12.243529 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a42349-e392-40c2-a6e8-32917240a4ad" containerName="extract-content" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.243535 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a42349-e392-40c2-a6e8-32917240a4ad" containerName="extract-content" Nov 25 09:28:12 crc kubenswrapper[4776]: E1125 09:28:12.243543 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="906f7079-2d0f-4205-8c68-4e66f0f25ab7" containerName="registry-server" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.243549 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="906f7079-2d0f-4205-8c68-4e66f0f25ab7" containerName="registry-server" Nov 25 09:28:12 crc kubenswrapper[4776]: E1125 09:28:12.243558 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa898cba-424e-4a13-93de-0e869ccc6780" containerName="registry-server" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.243564 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa898cba-424e-4a13-93de-0e869ccc6780" containerName="registry-server" Nov 25 09:28:12 crc kubenswrapper[4776]: E1125 09:28:12.243572 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="906f7079-2d0f-4205-8c68-4e66f0f25ab7" containerName="extract-utilities" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.243577 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="906f7079-2d0f-4205-8c68-4e66f0f25ab7" containerName="extract-utilities" Nov 25 09:28:12 crc kubenswrapper[4776]: E1125 09:28:12.243583 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa898cba-424e-4a13-93de-0e869ccc6780" containerName="extract-content" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.243589 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa898cba-424e-4a13-93de-0e869ccc6780" containerName="extract-content" Nov 25 09:28:12 crc kubenswrapper[4776]: E1125 09:28:12.243596 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01bc6d7e-5810-42c8-94e9-77fc02765d38" containerName="pruner" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.243602 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="01bc6d7e-5810-42c8-94e9-77fc02765d38" containerName="pruner" Nov 25 09:28:12 crc kubenswrapper[4776]: E1125 09:28:12.243616 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="906f7079-2d0f-4205-8c68-4e66f0f25ab7" containerName="extract-content" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.243621 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="906f7079-2d0f-4205-8c68-4e66f0f25ab7" containerName="extract-content" Nov 25 09:28:12 crc kubenswrapper[4776]: E1125 09:28:12.243628 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d03c6574-9476-4ef1-8a83-68c7b2ce7a7c" containerName="registry-server" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.243633 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="d03c6574-9476-4ef1-8a83-68c7b2ce7a7c" containerName="registry-server" Nov 25 09:28:12 crc kubenswrapper[4776]: E1125 09:28:12.243644 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0944aab1-01d8-43cc-91d8-cb913d4cad3a" containerName="pruner" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.243650 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="0944aab1-01d8-43cc-91d8-cb913d4cad3a" containerName="pruner" Nov 25 09:28:12 crc kubenswrapper[4776]: E1125 09:28:12.243660 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d03c6574-9476-4ef1-8a83-68c7b2ce7a7c" containerName="extract-content" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.243666 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="d03c6574-9476-4ef1-8a83-68c7b2ce7a7c" containerName="extract-content" Nov 25 09:28:12 crc kubenswrapper[4776]: E1125 09:28:12.243673 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5ae11bc-db7c-4665-88e4-ef3b0eb83581" containerName="oauth-openshift" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.243678 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5ae11bc-db7c-4665-88e4-ef3b0eb83581" containerName="oauth-openshift" Nov 25 09:28:12 crc kubenswrapper[4776]: E1125 09:28:12.243686 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a42349-e392-40c2-a6e8-32917240a4ad" containerName="extract-utilities" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.243692 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a42349-e392-40c2-a6e8-32917240a4ad" containerName="extract-utilities" Nov 25 09:28:12 crc kubenswrapper[4776]: E1125 09:28:12.243702 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a42349-e392-40c2-a6e8-32917240a4ad" containerName="registry-server" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.243707 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a42349-e392-40c2-a6e8-32917240a4ad" containerName="registry-server" Nov 25 09:28:12 crc kubenswrapper[4776]: E1125 09:28:12.243716 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa898cba-424e-4a13-93de-0e869ccc6780" containerName="extract-utilities" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.243722 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa898cba-424e-4a13-93de-0e869ccc6780" containerName="extract-utilities" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.243810 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa898cba-424e-4a13-93de-0e869ccc6780" containerName="registry-server" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.243822 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="d03c6574-9476-4ef1-8a83-68c7b2ce7a7c" containerName="registry-server" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.243832 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="0944aab1-01d8-43cc-91d8-cb913d4cad3a" containerName="pruner" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.243843 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="906f7079-2d0f-4205-8c68-4e66f0f25ab7" containerName="registry-server" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.243852 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5ae11bc-db7c-4665-88e4-ef3b0eb83581" containerName="oauth-openshift" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.243860 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="01bc6d7e-5810-42c8-94e9-77fc02765d38" containerName="pruner" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.243869 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9a42349-e392-40c2-a6e8-32917240a4ad" containerName="registry-server" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.244282 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.253948 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" event={"ID":"c5ae11bc-db7c-4665-88e4-ef3b0eb83581","Type":"ContainerDied","Data":"c7be2d6b7d9f8dacd74ac864092f2282a5ad82be9eda5a52b5c5b3c23e02f46a"} Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.254009 4776 scope.go:117] "RemoveContainer" containerID="d884031de43de28107048c44a1903141869dd6318f0e8ec733b6b2eb1ce4ac13" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.254005 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8rhfr" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.259035 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5d7d85f8c7-jc455"] Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.301571 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8rhfr"] Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.307203 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8rhfr"] Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.363402 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.363470 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-system-router-certs\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.363489 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.363507 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.363532 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.363580 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9z4t\" (UniqueName: \"kubernetes.io/projected/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-kube-api-access-p9z4t\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.363602 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.363704 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-user-template-login\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.363865 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-system-service-ca\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.363913 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-audit-policies\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.363982 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-system-session\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.364061 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-audit-dir\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.364115 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-user-template-error\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.364152 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.465487 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9z4t\" (UniqueName: \"kubernetes.io/projected/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-kube-api-access-p9z4t\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.465536 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.465556 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-user-template-login\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.465603 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-system-service-ca\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.465622 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-audit-policies\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.465641 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-system-session\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.465661 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-audit-dir\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.465686 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-user-template-error\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.465704 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.465725 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.465750 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.465767 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-system-router-certs\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.465787 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.465809 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.466675 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-system-service-ca\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.466723 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-audit-policies\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.466788 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-audit-dir\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.467152 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.467302 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.470643 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.470677 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-system-session\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.470689 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.470729 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-system-router-certs\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.473028 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.473302 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-user-template-login\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.480947 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.481487 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-v4-0-config-user-template-error\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.486493 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9z4t\" (UniqueName: \"kubernetes.io/projected/2fb32bf9-0b4e-4424-bdd2-e13a01f291d9-kube-api-access-p9z4t\") pod \"oauth-openshift-5d7d85f8c7-jc455\" (UID: \"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9\") " pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.563993 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:12 crc kubenswrapper[4776]: I1125 09:28:12.962173 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5d7d85f8c7-jc455"] Nov 25 09:28:13 crc kubenswrapper[4776]: I1125 09:28:13.262190 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" event={"ID":"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9","Type":"ContainerStarted","Data":"870c91557b1be4ec1c2c2e8ffdd38c2e262e24ed62d6dce90754fe3d68e9ad76"} Nov 25 09:28:13 crc kubenswrapper[4776]: I1125 09:28:13.669814 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5ae11bc-db7c-4665-88e4-ef3b0eb83581" path="/var/lib/kubelet/pods/c5ae11bc-db7c-4665-88e4-ef3b0eb83581/volumes" Nov 25 09:28:14 crc kubenswrapper[4776]: I1125 09:28:14.268948 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" event={"ID":"2fb32bf9-0b4e-4424-bdd2-e13a01f291d9","Type":"ContainerStarted","Data":"16542a3dc33e4995efc99f0b111bc484906ad87e187cf452aca64528c75ac8e0"} Nov 25 09:28:14 crc kubenswrapper[4776]: I1125 09:28:14.269200 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:14 crc kubenswrapper[4776]: I1125 09:28:14.273905 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" Nov 25 09:28:14 crc kubenswrapper[4776]: I1125 09:28:14.292503 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5d7d85f8c7-jc455" podStartSLOduration=29.292481293 podStartE2EDuration="29.292481293s" podCreationTimestamp="2025-11-25 09:27:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:28:14.291763675 +0000 UTC m=+239.332823248" watchObservedRunningTime="2025-11-25 09:28:14.292481293 +0000 UTC m=+239.333540856" Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.391866 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zd2l7"] Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.392899 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zd2l7" podUID="2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6" containerName="registry-server" containerID="cri-o://6c98d48cbf363ef9564b4eb50524f27e3b2df256b9de3d10178317762cd09a92" gracePeriod=30 Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.395437 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xz58l"] Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.395693 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xz58l" podUID="e057eef9-0e1f-4272-8525-3a4350250d34" containerName="registry-server" containerID="cri-o://f4c230a58885c09d82df1d41502ae961d28381fb3d5a7e0c27cebfeed965d2c3" gracePeriod=30 Nov 25 09:28:44 crc kubenswrapper[4776]: E1125 09:28:44.414045 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f4c230a58885c09d82df1d41502ae961d28381fb3d5a7e0c27cebfeed965d2c3" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.416595 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5n2sp"] Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.416830 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-5n2sp" podUID="e8765bc6-c8e3-40d7-a061-211b1cad5df2" containerName="marketplace-operator" containerID="cri-o://aa1237b8018ef898ff3df28567501bf0a4db0a98d91104b753dc8e81f603467a" gracePeriod=30 Nov 25 09:28:44 crc kubenswrapper[4776]: E1125 09:28:44.430878 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f4c230a58885c09d82df1d41502ae961d28381fb3d5a7e0c27cebfeed965d2c3" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 09:28:44 crc kubenswrapper[4776]: E1125 09:28:44.437990 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4c230a58885c09d82df1d41502ae961d28381fb3d5a7e0c27cebfeed965d2c3 is running failed: container process not found" containerID="f4c230a58885c09d82df1d41502ae961d28381fb3d5a7e0c27cebfeed965d2c3" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 09:28:44 crc kubenswrapper[4776]: E1125 09:28:44.438147 4776 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4c230a58885c09d82df1d41502ae961d28381fb3d5a7e0c27cebfeed965d2c3 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-xz58l" podUID="e057eef9-0e1f-4272-8525-3a4350250d34" containerName="registry-server" Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.450439 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-424gh"] Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.450786 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-424gh" podUID="a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4" containerName="registry-server" containerID="cri-o://f35834a96b6f4a6905e8dfd87eed8775cb1d8e2f2e8090f1436ffd51750dfc10" gracePeriod=30 Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.464809 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vtzcb"] Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.465003 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vtzcb" podUID="f04466da-02ff-40ea-bbf4-5761ab1d93b6" containerName="registry-server" containerID="cri-o://fcb02491623a58ad77246475d2fb42c5b52591265c3a51bb480c9e4acb1a3b1a" gracePeriod=30 Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.468307 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xcd5g"] Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.469217 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xcd5g" Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.471231 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xcd5g"] Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.589656 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c7a039e1-a9b6-4fc9-b297-6cb3eb6618f2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xcd5g\" (UID: \"c7a039e1-a9b6-4fc9-b297-6cb3eb6618f2\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcd5g" Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.589725 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c7a039e1-a9b6-4fc9-b297-6cb3eb6618f2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xcd5g\" (UID: \"c7a039e1-a9b6-4fc9-b297-6cb3eb6618f2\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcd5g" Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.589747 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh42v\" (UniqueName: \"kubernetes.io/projected/c7a039e1-a9b6-4fc9-b297-6cb3eb6618f2-kube-api-access-mh42v\") pod \"marketplace-operator-79b997595-xcd5g\" (UID: \"c7a039e1-a9b6-4fc9-b297-6cb3eb6618f2\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcd5g" Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.690960 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c7a039e1-a9b6-4fc9-b297-6cb3eb6618f2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xcd5g\" (UID: \"c7a039e1-a9b6-4fc9-b297-6cb3eb6618f2\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcd5g" Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.691005 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mh42v\" (UniqueName: \"kubernetes.io/projected/c7a039e1-a9b6-4fc9-b297-6cb3eb6618f2-kube-api-access-mh42v\") pod \"marketplace-operator-79b997595-xcd5g\" (UID: \"c7a039e1-a9b6-4fc9-b297-6cb3eb6618f2\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcd5g" Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.691096 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c7a039e1-a9b6-4fc9-b297-6cb3eb6618f2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xcd5g\" (UID: \"c7a039e1-a9b6-4fc9-b297-6cb3eb6618f2\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcd5g" Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.692763 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c7a039e1-a9b6-4fc9-b297-6cb3eb6618f2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xcd5g\" (UID: \"c7a039e1-a9b6-4fc9-b297-6cb3eb6618f2\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcd5g" Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.699712 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c7a039e1-a9b6-4fc9-b297-6cb3eb6618f2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xcd5g\" (UID: \"c7a039e1-a9b6-4fc9-b297-6cb3eb6618f2\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcd5g" Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.709474 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh42v\" (UniqueName: \"kubernetes.io/projected/c7a039e1-a9b6-4fc9-b297-6cb3eb6618f2-kube-api-access-mh42v\") pod \"marketplace-operator-79b997595-xcd5g\" (UID: \"c7a039e1-a9b6-4fc9-b297-6cb3eb6618f2\") " pod="openshift-marketplace/marketplace-operator-79b997595-xcd5g" Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.851380 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xcd5g" Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.852592 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zd2l7" Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.858252 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5n2sp" Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.871918 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xz58l" Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.883326 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vtzcb" Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.888339 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-424gh" Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.996550 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txb6p\" (UniqueName: \"kubernetes.io/projected/2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6-kube-api-access-txb6p\") pod \"2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6\" (UID: \"2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6\") " Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.996585 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e8765bc6-c8e3-40d7-a061-211b1cad5df2-marketplace-operator-metrics\") pod \"e8765bc6-c8e3-40d7-a061-211b1cad5df2\" (UID: \"e8765bc6-c8e3-40d7-a061-211b1cad5df2\") " Nov 25 09:28:44 crc kubenswrapper[4776]: I1125 09:28:44.996619 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f04466da-02ff-40ea-bbf4-5761ab1d93b6-catalog-content\") pod \"f04466da-02ff-40ea-bbf4-5761ab1d93b6\" (UID: \"f04466da-02ff-40ea-bbf4-5761ab1d93b6\") " Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.002995 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6-kube-api-access-txb6p" (OuterVolumeSpecName: "kube-api-access-txb6p") pod "2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6" (UID: "2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6"). InnerVolumeSpecName "kube-api-access-txb6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.008421 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8765bc6-c8e3-40d7-a061-211b1cad5df2-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "e8765bc6-c8e3-40d7-a061-211b1cad5df2" (UID: "e8765bc6-c8e3-40d7-a061-211b1cad5df2"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.008569 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f04466da-02ff-40ea-bbf4-5761ab1d93b6-utilities\") pod \"f04466da-02ff-40ea-bbf4-5761ab1d93b6\" (UID: \"f04466da-02ff-40ea-bbf4-5761ab1d93b6\") " Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.008623 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6-utilities\") pod \"2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6\" (UID: \"2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6\") " Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.008650 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4-utilities\") pod \"a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4\" (UID: \"a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4\") " Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.009230 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f04466da-02ff-40ea-bbf4-5761ab1d93b6-utilities" (OuterVolumeSpecName: "utilities") pod "f04466da-02ff-40ea-bbf4-5761ab1d93b6" (UID: "f04466da-02ff-40ea-bbf4-5761ab1d93b6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.009332 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6-utilities" (OuterVolumeSpecName: "utilities") pod "2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6" (UID: "2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.009389 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4-catalog-content\") pod \"a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4\" (UID: \"a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4\") " Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.009422 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6-catalog-content\") pod \"2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6\" (UID: \"2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6\") " Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.009439 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmb8m\" (UniqueName: \"kubernetes.io/projected/f04466da-02ff-40ea-bbf4-5761ab1d93b6-kube-api-access-kmb8m\") pod \"f04466da-02ff-40ea-bbf4-5761ab1d93b6\" (UID: \"f04466da-02ff-40ea-bbf4-5761ab1d93b6\") " Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.009465 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e057eef9-0e1f-4272-8525-3a4350250d34-utilities\") pod \"e057eef9-0e1f-4272-8525-3a4350250d34\" (UID: \"e057eef9-0e1f-4272-8525-3a4350250d34\") " Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.009486 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qm8bb\" (UniqueName: \"kubernetes.io/projected/e8765bc6-c8e3-40d7-a061-211b1cad5df2-kube-api-access-qm8bb\") pod \"e8765bc6-c8e3-40d7-a061-211b1cad5df2\" (UID: \"e8765bc6-c8e3-40d7-a061-211b1cad5df2\") " Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.009507 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdjdp\" (UniqueName: \"kubernetes.io/projected/a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4-kube-api-access-bdjdp\") pod \"a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4\" (UID: \"a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4\") " Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.009557 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e057eef9-0e1f-4272-8525-3a4350250d34-catalog-content\") pod \"e057eef9-0e1f-4272-8525-3a4350250d34\" (UID: \"e057eef9-0e1f-4272-8525-3a4350250d34\") " Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.009582 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e8765bc6-c8e3-40d7-a061-211b1cad5df2-marketplace-trusted-ca\") pod \"e8765bc6-c8e3-40d7-a061-211b1cad5df2\" (UID: \"e8765bc6-c8e3-40d7-a061-211b1cad5df2\") " Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.009606 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66kjv\" (UniqueName: \"kubernetes.io/projected/e057eef9-0e1f-4272-8525-3a4350250d34-kube-api-access-66kjv\") pod \"e057eef9-0e1f-4272-8525-3a4350250d34\" (UID: \"e057eef9-0e1f-4272-8525-3a4350250d34\") " Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.009882 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4-utilities" (OuterVolumeSpecName: "utilities") pod "a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4" (UID: "a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.010082 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txb6p\" (UniqueName: \"kubernetes.io/projected/2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6-kube-api-access-txb6p\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.010103 4776 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e8765bc6-c8e3-40d7-a061-211b1cad5df2-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.010114 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f04466da-02ff-40ea-bbf4-5761ab1d93b6-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.010125 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.010133 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.010969 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e057eef9-0e1f-4272-8525-3a4350250d34-utilities" (OuterVolumeSpecName: "utilities") pod "e057eef9-0e1f-4272-8525-3a4350250d34" (UID: "e057eef9-0e1f-4272-8525-3a4350250d34"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.010979 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8765bc6-c8e3-40d7-a061-211b1cad5df2-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "e8765bc6-c8e3-40d7-a061-211b1cad5df2" (UID: "e8765bc6-c8e3-40d7-a061-211b1cad5df2"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.013658 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e057eef9-0e1f-4272-8525-3a4350250d34-kube-api-access-66kjv" (OuterVolumeSpecName: "kube-api-access-66kjv") pod "e057eef9-0e1f-4272-8525-3a4350250d34" (UID: "e057eef9-0e1f-4272-8525-3a4350250d34"). InnerVolumeSpecName "kube-api-access-66kjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.016051 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f04466da-02ff-40ea-bbf4-5761ab1d93b6-kube-api-access-kmb8m" (OuterVolumeSpecName: "kube-api-access-kmb8m") pod "f04466da-02ff-40ea-bbf4-5761ab1d93b6" (UID: "f04466da-02ff-40ea-bbf4-5761ab1d93b6"). InnerVolumeSpecName "kube-api-access-kmb8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.016751 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8765bc6-c8e3-40d7-a061-211b1cad5df2-kube-api-access-qm8bb" (OuterVolumeSpecName: "kube-api-access-qm8bb") pod "e8765bc6-c8e3-40d7-a061-211b1cad5df2" (UID: "e8765bc6-c8e3-40d7-a061-211b1cad5df2"). InnerVolumeSpecName "kube-api-access-qm8bb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.040143 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4-kube-api-access-bdjdp" (OuterVolumeSpecName: "kube-api-access-bdjdp") pod "a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4" (UID: "a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4"). InnerVolumeSpecName "kube-api-access-bdjdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.043380 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4" (UID: "a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.058229 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xcd5g"] Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.082492 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e057eef9-0e1f-4272-8525-3a4350250d34-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e057eef9-0e1f-4272-8525-3a4350250d34" (UID: "e057eef9-0e1f-4272-8525-3a4350250d34"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.096503 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6" (UID: "2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.111448 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.111774 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.111785 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmb8m\" (UniqueName: \"kubernetes.io/projected/f04466da-02ff-40ea-bbf4-5761ab1d93b6-kube-api-access-kmb8m\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.111798 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e057eef9-0e1f-4272-8525-3a4350250d34-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.111806 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qm8bb\" (UniqueName: \"kubernetes.io/projected/e8765bc6-c8e3-40d7-a061-211b1cad5df2-kube-api-access-qm8bb\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.111814 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdjdp\" (UniqueName: \"kubernetes.io/projected/a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4-kube-api-access-bdjdp\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.111822 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e057eef9-0e1f-4272-8525-3a4350250d34-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.111832 4776 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e8765bc6-c8e3-40d7-a061-211b1cad5df2-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.111840 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66kjv\" (UniqueName: \"kubernetes.io/projected/e057eef9-0e1f-4272-8525-3a4350250d34-kube-api-access-66kjv\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.122385 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f04466da-02ff-40ea-bbf4-5761ab1d93b6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f04466da-02ff-40ea-bbf4-5761ab1d93b6" (UID: "f04466da-02ff-40ea-bbf4-5761ab1d93b6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.213572 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f04466da-02ff-40ea-bbf4-5761ab1d93b6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.459362 4776 generic.go:334] "Generic (PLEG): container finished" podID="a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4" containerID="f35834a96b6f4a6905e8dfd87eed8775cb1d8e2f2e8090f1436ffd51750dfc10" exitCode=0 Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.459394 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-424gh" event={"ID":"a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4","Type":"ContainerDied","Data":"f35834a96b6f4a6905e8dfd87eed8775cb1d8e2f2e8090f1436ffd51750dfc10"} Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.459428 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-424gh" event={"ID":"a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4","Type":"ContainerDied","Data":"29cbd3bbb0b0c03ff1f16785fea923b23d55106e2b74a01dc4a9c330f735b571"} Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.459453 4776 scope.go:117] "RemoveContainer" containerID="f35834a96b6f4a6905e8dfd87eed8775cb1d8e2f2e8090f1436ffd51750dfc10" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.459491 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-424gh" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.463979 4776 generic.go:334] "Generic (PLEG): container finished" podID="e8765bc6-c8e3-40d7-a061-211b1cad5df2" containerID="aa1237b8018ef898ff3df28567501bf0a4db0a98d91104b753dc8e81f603467a" exitCode=0 Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.464051 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5n2sp" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.464029 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5n2sp" event={"ID":"e8765bc6-c8e3-40d7-a061-211b1cad5df2","Type":"ContainerDied","Data":"aa1237b8018ef898ff3df28567501bf0a4db0a98d91104b753dc8e81f603467a"} Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.464185 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5n2sp" event={"ID":"e8765bc6-c8e3-40d7-a061-211b1cad5df2","Type":"ContainerDied","Data":"c03b771376224415a70b8bf2e6d4b61d05e29b7ba7909208777791fbe3b54f5c"} Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.466633 4776 generic.go:334] "Generic (PLEG): container finished" podID="e057eef9-0e1f-4272-8525-3a4350250d34" containerID="f4c230a58885c09d82df1d41502ae961d28381fb3d5a7e0c27cebfeed965d2c3" exitCode=0 Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.466665 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xz58l" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.466713 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xz58l" event={"ID":"e057eef9-0e1f-4272-8525-3a4350250d34","Type":"ContainerDied","Data":"f4c230a58885c09d82df1d41502ae961d28381fb3d5a7e0c27cebfeed965d2c3"} Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.466750 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xz58l" event={"ID":"e057eef9-0e1f-4272-8525-3a4350250d34","Type":"ContainerDied","Data":"f1b8f0f731be5211e5ca88944ed7d31c2f317b2b939e3f627c0162fb7ac9201a"} Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.472498 4776 generic.go:334] "Generic (PLEG): container finished" podID="f04466da-02ff-40ea-bbf4-5761ab1d93b6" containerID="fcb02491623a58ad77246475d2fb42c5b52591265c3a51bb480c9e4acb1a3b1a" exitCode=0 Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.472554 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vtzcb" event={"ID":"f04466da-02ff-40ea-bbf4-5761ab1d93b6","Type":"ContainerDied","Data":"fcb02491623a58ad77246475d2fb42c5b52591265c3a51bb480c9e4acb1a3b1a"} Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.472574 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vtzcb" event={"ID":"f04466da-02ff-40ea-bbf4-5761ab1d93b6","Type":"ContainerDied","Data":"2819732fdc9b853b0d8e99b18f825f87998b08daa12eb26045bd78f2ca6db2aa"} Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.472632 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vtzcb" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.477591 4776 scope.go:117] "RemoveContainer" containerID="7ba5a9ee88320e50a4c8851248d2c590c2254ea1f44a2e1e8a165ddcbe7bf9d9" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.480303 4776 generic.go:334] "Generic (PLEG): container finished" podID="2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6" containerID="6c98d48cbf363ef9564b4eb50524f27e3b2df256b9de3d10178317762cd09a92" exitCode=0 Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.480366 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zd2l7" event={"ID":"2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6","Type":"ContainerDied","Data":"6c98d48cbf363ef9564b4eb50524f27e3b2df256b9de3d10178317762cd09a92"} Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.480376 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zd2l7" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.480393 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zd2l7" event={"ID":"2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6","Type":"ContainerDied","Data":"1000b0616d2704456e7d987010b9bd00f6990de2c2a44dd67d4fa61f3ff8536a"} Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.482171 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xcd5g" event={"ID":"c7a039e1-a9b6-4fc9-b297-6cb3eb6618f2","Type":"ContainerStarted","Data":"5b7c7c6e60e2e33481de624bab47e6cd7248e64818857b8e53b075ba725ab305"} Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.482237 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xcd5g" event={"ID":"c7a039e1-a9b6-4fc9-b297-6cb3eb6618f2","Type":"ContainerStarted","Data":"6f0a5ff2a212572df96973b640e6e187401dfc2f4629d227e395833582202ca1"} Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.482722 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-xcd5g" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.485705 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-xcd5g" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.494429 4776 scope.go:117] "RemoveContainer" containerID="da412c58869886ce3eb63d409b39f983a7659b26bb5805a0dc254bd7646f6ae0" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.497360 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-xcd5g" podStartSLOduration=1.497342706 podStartE2EDuration="1.497342706s" podCreationTimestamp="2025-11-25 09:28:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:28:45.496580907 +0000 UTC m=+270.537640460" watchObservedRunningTime="2025-11-25 09:28:45.497342706 +0000 UTC m=+270.538402259" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.524139 4776 scope.go:117] "RemoveContainer" containerID="f35834a96b6f4a6905e8dfd87eed8775cb1d8e2f2e8090f1436ffd51750dfc10" Nov 25 09:28:45 crc kubenswrapper[4776]: E1125 09:28:45.524630 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f35834a96b6f4a6905e8dfd87eed8775cb1d8e2f2e8090f1436ffd51750dfc10\": container with ID starting with f35834a96b6f4a6905e8dfd87eed8775cb1d8e2f2e8090f1436ffd51750dfc10 not found: ID does not exist" containerID="f35834a96b6f4a6905e8dfd87eed8775cb1d8e2f2e8090f1436ffd51750dfc10" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.524768 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f35834a96b6f4a6905e8dfd87eed8775cb1d8e2f2e8090f1436ffd51750dfc10"} err="failed to get container status \"f35834a96b6f4a6905e8dfd87eed8775cb1d8e2f2e8090f1436ffd51750dfc10\": rpc error: code = NotFound desc = could not find container \"f35834a96b6f4a6905e8dfd87eed8775cb1d8e2f2e8090f1436ffd51750dfc10\": container with ID starting with f35834a96b6f4a6905e8dfd87eed8775cb1d8e2f2e8090f1436ffd51750dfc10 not found: ID does not exist" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.524867 4776 scope.go:117] "RemoveContainer" containerID="7ba5a9ee88320e50a4c8851248d2c590c2254ea1f44a2e1e8a165ddcbe7bf9d9" Nov 25 09:28:45 crc kubenswrapper[4776]: E1125 09:28:45.539529 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ba5a9ee88320e50a4c8851248d2c590c2254ea1f44a2e1e8a165ddcbe7bf9d9\": container with ID starting with 7ba5a9ee88320e50a4c8851248d2c590c2254ea1f44a2e1e8a165ddcbe7bf9d9 not found: ID does not exist" containerID="7ba5a9ee88320e50a4c8851248d2c590c2254ea1f44a2e1e8a165ddcbe7bf9d9" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.539754 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ba5a9ee88320e50a4c8851248d2c590c2254ea1f44a2e1e8a165ddcbe7bf9d9"} err="failed to get container status \"7ba5a9ee88320e50a4c8851248d2c590c2254ea1f44a2e1e8a165ddcbe7bf9d9\": rpc error: code = NotFound desc = could not find container \"7ba5a9ee88320e50a4c8851248d2c590c2254ea1f44a2e1e8a165ddcbe7bf9d9\": container with ID starting with 7ba5a9ee88320e50a4c8851248d2c590c2254ea1f44a2e1e8a165ddcbe7bf9d9 not found: ID does not exist" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.539919 4776 scope.go:117] "RemoveContainer" containerID="da412c58869886ce3eb63d409b39f983a7659b26bb5805a0dc254bd7646f6ae0" Nov 25 09:28:45 crc kubenswrapper[4776]: E1125 09:28:45.543424 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da412c58869886ce3eb63d409b39f983a7659b26bb5805a0dc254bd7646f6ae0\": container with ID starting with da412c58869886ce3eb63d409b39f983a7659b26bb5805a0dc254bd7646f6ae0 not found: ID does not exist" containerID="da412c58869886ce3eb63d409b39f983a7659b26bb5805a0dc254bd7646f6ae0" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.543465 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da412c58869886ce3eb63d409b39f983a7659b26bb5805a0dc254bd7646f6ae0"} err="failed to get container status \"da412c58869886ce3eb63d409b39f983a7659b26bb5805a0dc254bd7646f6ae0\": rpc error: code = NotFound desc = could not find container \"da412c58869886ce3eb63d409b39f983a7659b26bb5805a0dc254bd7646f6ae0\": container with ID starting with da412c58869886ce3eb63d409b39f983a7659b26bb5805a0dc254bd7646f6ae0 not found: ID does not exist" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.543487 4776 scope.go:117] "RemoveContainer" containerID="aa1237b8018ef898ff3df28567501bf0a4db0a98d91104b753dc8e81f603467a" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.546029 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xz58l"] Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.555248 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xz58l"] Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.559849 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-424gh"] Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.564926 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-424gh"] Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.568115 4776 scope.go:117] "RemoveContainer" containerID="aa1237b8018ef898ff3df28567501bf0a4db0a98d91104b753dc8e81f603467a" Nov 25 09:28:45 crc kubenswrapper[4776]: E1125 09:28:45.568461 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa1237b8018ef898ff3df28567501bf0a4db0a98d91104b753dc8e81f603467a\": container with ID starting with aa1237b8018ef898ff3df28567501bf0a4db0a98d91104b753dc8e81f603467a not found: ID does not exist" containerID="aa1237b8018ef898ff3df28567501bf0a4db0a98d91104b753dc8e81f603467a" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.568496 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa1237b8018ef898ff3df28567501bf0a4db0a98d91104b753dc8e81f603467a"} err="failed to get container status \"aa1237b8018ef898ff3df28567501bf0a4db0a98d91104b753dc8e81f603467a\": rpc error: code = NotFound desc = could not find container \"aa1237b8018ef898ff3df28567501bf0a4db0a98d91104b753dc8e81f603467a\": container with ID starting with aa1237b8018ef898ff3df28567501bf0a4db0a98d91104b753dc8e81f603467a not found: ID does not exist" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.568520 4776 scope.go:117] "RemoveContainer" containerID="f4c230a58885c09d82df1d41502ae961d28381fb3d5a7e0c27cebfeed965d2c3" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.570257 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5n2sp"] Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.574723 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5n2sp"] Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.584247 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vtzcb"] Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.585087 4776 scope.go:117] "RemoveContainer" containerID="99ddbbab9d1679f0278d7ea97c61135a720d506e7ea38657567c3796cfe8b7e5" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.587677 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vtzcb"] Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.592027 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zd2l7"] Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.595751 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zd2l7"] Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.605465 4776 scope.go:117] "RemoveContainer" containerID="7caabe31431087b47cbe5e03f5d129421ae23d1e4a00404793e64b1b43c78160" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.620228 4776 scope.go:117] "RemoveContainer" containerID="f4c230a58885c09d82df1d41502ae961d28381fb3d5a7e0c27cebfeed965d2c3" Nov 25 09:28:45 crc kubenswrapper[4776]: E1125 09:28:45.621243 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4c230a58885c09d82df1d41502ae961d28381fb3d5a7e0c27cebfeed965d2c3\": container with ID starting with f4c230a58885c09d82df1d41502ae961d28381fb3d5a7e0c27cebfeed965d2c3 not found: ID does not exist" containerID="f4c230a58885c09d82df1d41502ae961d28381fb3d5a7e0c27cebfeed965d2c3" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.621271 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4c230a58885c09d82df1d41502ae961d28381fb3d5a7e0c27cebfeed965d2c3"} err="failed to get container status \"f4c230a58885c09d82df1d41502ae961d28381fb3d5a7e0c27cebfeed965d2c3\": rpc error: code = NotFound desc = could not find container \"f4c230a58885c09d82df1d41502ae961d28381fb3d5a7e0c27cebfeed965d2c3\": container with ID starting with f4c230a58885c09d82df1d41502ae961d28381fb3d5a7e0c27cebfeed965d2c3 not found: ID does not exist" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.621294 4776 scope.go:117] "RemoveContainer" containerID="99ddbbab9d1679f0278d7ea97c61135a720d506e7ea38657567c3796cfe8b7e5" Nov 25 09:28:45 crc kubenswrapper[4776]: E1125 09:28:45.621743 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99ddbbab9d1679f0278d7ea97c61135a720d506e7ea38657567c3796cfe8b7e5\": container with ID starting with 99ddbbab9d1679f0278d7ea97c61135a720d506e7ea38657567c3796cfe8b7e5 not found: ID does not exist" containerID="99ddbbab9d1679f0278d7ea97c61135a720d506e7ea38657567c3796cfe8b7e5" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.621788 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99ddbbab9d1679f0278d7ea97c61135a720d506e7ea38657567c3796cfe8b7e5"} err="failed to get container status \"99ddbbab9d1679f0278d7ea97c61135a720d506e7ea38657567c3796cfe8b7e5\": rpc error: code = NotFound desc = could not find container \"99ddbbab9d1679f0278d7ea97c61135a720d506e7ea38657567c3796cfe8b7e5\": container with ID starting with 99ddbbab9d1679f0278d7ea97c61135a720d506e7ea38657567c3796cfe8b7e5 not found: ID does not exist" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.621817 4776 scope.go:117] "RemoveContainer" containerID="7caabe31431087b47cbe5e03f5d129421ae23d1e4a00404793e64b1b43c78160" Nov 25 09:28:45 crc kubenswrapper[4776]: E1125 09:28:45.622244 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7caabe31431087b47cbe5e03f5d129421ae23d1e4a00404793e64b1b43c78160\": container with ID starting with 7caabe31431087b47cbe5e03f5d129421ae23d1e4a00404793e64b1b43c78160 not found: ID does not exist" containerID="7caabe31431087b47cbe5e03f5d129421ae23d1e4a00404793e64b1b43c78160" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.622276 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7caabe31431087b47cbe5e03f5d129421ae23d1e4a00404793e64b1b43c78160"} err="failed to get container status \"7caabe31431087b47cbe5e03f5d129421ae23d1e4a00404793e64b1b43c78160\": rpc error: code = NotFound desc = could not find container \"7caabe31431087b47cbe5e03f5d129421ae23d1e4a00404793e64b1b43c78160\": container with ID starting with 7caabe31431087b47cbe5e03f5d129421ae23d1e4a00404793e64b1b43c78160 not found: ID does not exist" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.622292 4776 scope.go:117] "RemoveContainer" containerID="fcb02491623a58ad77246475d2fb42c5b52591265c3a51bb480c9e4acb1a3b1a" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.635820 4776 scope.go:117] "RemoveContainer" containerID="5bd3d984c7e13c499dee93d8ca51ce2f2e500ac426765a6852d8a8d35962b914" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.648457 4776 scope.go:117] "RemoveContainer" containerID="b2636cec13e2f0b6bb7f7c8dc2b36d065469205fabc5ba15607d8b20d9edb287" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.663505 4776 scope.go:117] "RemoveContainer" containerID="fcb02491623a58ad77246475d2fb42c5b52591265c3a51bb480c9e4acb1a3b1a" Nov 25 09:28:45 crc kubenswrapper[4776]: E1125 09:28:45.663930 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcb02491623a58ad77246475d2fb42c5b52591265c3a51bb480c9e4acb1a3b1a\": container with ID starting with fcb02491623a58ad77246475d2fb42c5b52591265c3a51bb480c9e4acb1a3b1a not found: ID does not exist" containerID="fcb02491623a58ad77246475d2fb42c5b52591265c3a51bb480c9e4acb1a3b1a" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.663974 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcb02491623a58ad77246475d2fb42c5b52591265c3a51bb480c9e4acb1a3b1a"} err="failed to get container status \"fcb02491623a58ad77246475d2fb42c5b52591265c3a51bb480c9e4acb1a3b1a\": rpc error: code = NotFound desc = could not find container \"fcb02491623a58ad77246475d2fb42c5b52591265c3a51bb480c9e4acb1a3b1a\": container with ID starting with fcb02491623a58ad77246475d2fb42c5b52591265c3a51bb480c9e4acb1a3b1a not found: ID does not exist" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.663999 4776 scope.go:117] "RemoveContainer" containerID="5bd3d984c7e13c499dee93d8ca51ce2f2e500ac426765a6852d8a8d35962b914" Nov 25 09:28:45 crc kubenswrapper[4776]: E1125 09:28:45.664296 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bd3d984c7e13c499dee93d8ca51ce2f2e500ac426765a6852d8a8d35962b914\": container with ID starting with 5bd3d984c7e13c499dee93d8ca51ce2f2e500ac426765a6852d8a8d35962b914 not found: ID does not exist" containerID="5bd3d984c7e13c499dee93d8ca51ce2f2e500ac426765a6852d8a8d35962b914" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.664319 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bd3d984c7e13c499dee93d8ca51ce2f2e500ac426765a6852d8a8d35962b914"} err="failed to get container status \"5bd3d984c7e13c499dee93d8ca51ce2f2e500ac426765a6852d8a8d35962b914\": rpc error: code = NotFound desc = could not find container \"5bd3d984c7e13c499dee93d8ca51ce2f2e500ac426765a6852d8a8d35962b914\": container with ID starting with 5bd3d984c7e13c499dee93d8ca51ce2f2e500ac426765a6852d8a8d35962b914 not found: ID does not exist" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.664335 4776 scope.go:117] "RemoveContainer" containerID="b2636cec13e2f0b6bb7f7c8dc2b36d065469205fabc5ba15607d8b20d9edb287" Nov 25 09:28:45 crc kubenswrapper[4776]: E1125 09:28:45.664572 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2636cec13e2f0b6bb7f7c8dc2b36d065469205fabc5ba15607d8b20d9edb287\": container with ID starting with b2636cec13e2f0b6bb7f7c8dc2b36d065469205fabc5ba15607d8b20d9edb287 not found: ID does not exist" containerID="b2636cec13e2f0b6bb7f7c8dc2b36d065469205fabc5ba15607d8b20d9edb287" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.664612 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2636cec13e2f0b6bb7f7c8dc2b36d065469205fabc5ba15607d8b20d9edb287"} err="failed to get container status \"b2636cec13e2f0b6bb7f7c8dc2b36d065469205fabc5ba15607d8b20d9edb287\": rpc error: code = NotFound desc = could not find container \"b2636cec13e2f0b6bb7f7c8dc2b36d065469205fabc5ba15607d8b20d9edb287\": container with ID starting with b2636cec13e2f0b6bb7f7c8dc2b36d065469205fabc5ba15607d8b20d9edb287 not found: ID does not exist" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.664644 4776 scope.go:117] "RemoveContainer" containerID="6c98d48cbf363ef9564b4eb50524f27e3b2df256b9de3d10178317762cd09a92" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.669698 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6" path="/var/lib/kubelet/pods/2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6/volumes" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.670425 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4" path="/var/lib/kubelet/pods/a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4/volumes" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.671128 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e057eef9-0e1f-4272-8525-3a4350250d34" path="/var/lib/kubelet/pods/e057eef9-0e1f-4272-8525-3a4350250d34/volumes" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.672389 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8765bc6-c8e3-40d7-a061-211b1cad5df2" path="/var/lib/kubelet/pods/e8765bc6-c8e3-40d7-a061-211b1cad5df2/volumes" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.672911 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f04466da-02ff-40ea-bbf4-5761ab1d93b6" path="/var/lib/kubelet/pods/f04466da-02ff-40ea-bbf4-5761ab1d93b6/volumes" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.678912 4776 scope.go:117] "RemoveContainer" containerID="67807704b3a78d02a654d934ffc5c9d67cddb72a357b2d264879685d28d66525" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.692670 4776 scope.go:117] "RemoveContainer" containerID="69d6eade4eba1db1723bfebcd0379bb921a12d7d5d045d56597e365d50ea6184" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.710001 4776 scope.go:117] "RemoveContainer" containerID="6c98d48cbf363ef9564b4eb50524f27e3b2df256b9de3d10178317762cd09a92" Nov 25 09:28:45 crc kubenswrapper[4776]: E1125 09:28:45.710659 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c98d48cbf363ef9564b4eb50524f27e3b2df256b9de3d10178317762cd09a92\": container with ID starting with 6c98d48cbf363ef9564b4eb50524f27e3b2df256b9de3d10178317762cd09a92 not found: ID does not exist" containerID="6c98d48cbf363ef9564b4eb50524f27e3b2df256b9de3d10178317762cd09a92" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.710721 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c98d48cbf363ef9564b4eb50524f27e3b2df256b9de3d10178317762cd09a92"} err="failed to get container status \"6c98d48cbf363ef9564b4eb50524f27e3b2df256b9de3d10178317762cd09a92\": rpc error: code = NotFound desc = could not find container \"6c98d48cbf363ef9564b4eb50524f27e3b2df256b9de3d10178317762cd09a92\": container with ID starting with 6c98d48cbf363ef9564b4eb50524f27e3b2df256b9de3d10178317762cd09a92 not found: ID does not exist" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.710751 4776 scope.go:117] "RemoveContainer" containerID="67807704b3a78d02a654d934ffc5c9d67cddb72a357b2d264879685d28d66525" Nov 25 09:28:45 crc kubenswrapper[4776]: E1125 09:28:45.711164 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67807704b3a78d02a654d934ffc5c9d67cddb72a357b2d264879685d28d66525\": container with ID starting with 67807704b3a78d02a654d934ffc5c9d67cddb72a357b2d264879685d28d66525 not found: ID does not exist" containerID="67807704b3a78d02a654d934ffc5c9d67cddb72a357b2d264879685d28d66525" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.711210 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67807704b3a78d02a654d934ffc5c9d67cddb72a357b2d264879685d28d66525"} err="failed to get container status \"67807704b3a78d02a654d934ffc5c9d67cddb72a357b2d264879685d28d66525\": rpc error: code = NotFound desc = could not find container \"67807704b3a78d02a654d934ffc5c9d67cddb72a357b2d264879685d28d66525\": container with ID starting with 67807704b3a78d02a654d934ffc5c9d67cddb72a357b2d264879685d28d66525 not found: ID does not exist" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.711238 4776 scope.go:117] "RemoveContainer" containerID="69d6eade4eba1db1723bfebcd0379bb921a12d7d5d045d56597e365d50ea6184" Nov 25 09:28:45 crc kubenswrapper[4776]: E1125 09:28:45.711550 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69d6eade4eba1db1723bfebcd0379bb921a12d7d5d045d56597e365d50ea6184\": container with ID starting with 69d6eade4eba1db1723bfebcd0379bb921a12d7d5d045d56597e365d50ea6184 not found: ID does not exist" containerID="69d6eade4eba1db1723bfebcd0379bb921a12d7d5d045d56597e365d50ea6184" Nov 25 09:28:45 crc kubenswrapper[4776]: I1125 09:28:45.711578 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69d6eade4eba1db1723bfebcd0379bb921a12d7d5d045d56597e365d50ea6184"} err="failed to get container status \"69d6eade4eba1db1723bfebcd0379bb921a12d7d5d045d56597e365d50ea6184\": rpc error: code = NotFound desc = could not find container \"69d6eade4eba1db1723bfebcd0379bb921a12d7d5d045d56597e365d50ea6184\": container with ID starting with 69d6eade4eba1db1723bfebcd0379bb921a12d7d5d045d56597e365d50ea6184 not found: ID does not exist" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.607309 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sz9dx"] Nov 25 09:28:46 crc kubenswrapper[4776]: E1125 09:28:46.608043 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f04466da-02ff-40ea-bbf4-5761ab1d93b6" containerName="extract-utilities" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.608085 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="f04466da-02ff-40ea-bbf4-5761ab1d93b6" containerName="extract-utilities" Nov 25 09:28:46 crc kubenswrapper[4776]: E1125 09:28:46.608103 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4" containerName="extract-content" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.608114 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4" containerName="extract-content" Nov 25 09:28:46 crc kubenswrapper[4776]: E1125 09:28:46.608133 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e057eef9-0e1f-4272-8525-3a4350250d34" containerName="extract-utilities" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.608146 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="e057eef9-0e1f-4272-8525-3a4350250d34" containerName="extract-utilities" Nov 25 09:28:46 crc kubenswrapper[4776]: E1125 09:28:46.608163 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6" containerName="extract-utilities" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.608173 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6" containerName="extract-utilities" Nov 25 09:28:46 crc kubenswrapper[4776]: E1125 09:28:46.608195 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f04466da-02ff-40ea-bbf4-5761ab1d93b6" containerName="registry-server" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.608206 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="f04466da-02ff-40ea-bbf4-5761ab1d93b6" containerName="registry-server" Nov 25 09:28:46 crc kubenswrapper[4776]: E1125 09:28:46.608220 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6" containerName="extract-content" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.608230 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6" containerName="extract-content" Nov 25 09:28:46 crc kubenswrapper[4776]: E1125 09:28:46.608247 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4" containerName="extract-utilities" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.608256 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4" containerName="extract-utilities" Nov 25 09:28:46 crc kubenswrapper[4776]: E1125 09:28:46.608271 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4" containerName="registry-server" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.608281 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4" containerName="registry-server" Nov 25 09:28:46 crc kubenswrapper[4776]: E1125 09:28:46.608295 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6" containerName="registry-server" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.608304 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6" containerName="registry-server" Nov 25 09:28:46 crc kubenswrapper[4776]: E1125 09:28:46.608318 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e057eef9-0e1f-4272-8525-3a4350250d34" containerName="registry-server" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.608333 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="e057eef9-0e1f-4272-8525-3a4350250d34" containerName="registry-server" Nov 25 09:28:46 crc kubenswrapper[4776]: E1125 09:28:46.608348 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f04466da-02ff-40ea-bbf4-5761ab1d93b6" containerName="extract-content" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.608359 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="f04466da-02ff-40ea-bbf4-5761ab1d93b6" containerName="extract-content" Nov 25 09:28:46 crc kubenswrapper[4776]: E1125 09:28:46.608372 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e057eef9-0e1f-4272-8525-3a4350250d34" containerName="extract-content" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.608384 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="e057eef9-0e1f-4272-8525-3a4350250d34" containerName="extract-content" Nov 25 09:28:46 crc kubenswrapper[4776]: E1125 09:28:46.608402 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8765bc6-c8e3-40d7-a061-211b1cad5df2" containerName="marketplace-operator" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.608413 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8765bc6-c8e3-40d7-a061-211b1cad5df2" containerName="marketplace-operator" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.608564 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7dabec0-1f3d-4ab3-818f-de4c9fb4e9a4" containerName="registry-server" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.608582 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8765bc6-c8e3-40d7-a061-211b1cad5df2" containerName="marketplace-operator" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.608595 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="f04466da-02ff-40ea-bbf4-5761ab1d93b6" containerName="registry-server" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.608615 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e8d2c90-9ed7-496e-8ae2-49b6b84bf6f6" containerName="registry-server" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.608632 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="e057eef9-0e1f-4272-8525-3a4350250d34" containerName="registry-server" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.609756 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sz9dx" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.612040 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.664781 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sz9dx"] Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.730875 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n5md\" (UniqueName: \"kubernetes.io/projected/cb60184b-c3b3-4ddb-90da-01c62fb183d7-kube-api-access-9n5md\") pod \"redhat-marketplace-sz9dx\" (UID: \"cb60184b-c3b3-4ddb-90da-01c62fb183d7\") " pod="openshift-marketplace/redhat-marketplace-sz9dx" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.730948 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb60184b-c3b3-4ddb-90da-01c62fb183d7-utilities\") pod \"redhat-marketplace-sz9dx\" (UID: \"cb60184b-c3b3-4ddb-90da-01c62fb183d7\") " pod="openshift-marketplace/redhat-marketplace-sz9dx" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.731098 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb60184b-c3b3-4ddb-90da-01c62fb183d7-catalog-content\") pod \"redhat-marketplace-sz9dx\" (UID: \"cb60184b-c3b3-4ddb-90da-01c62fb183d7\") " pod="openshift-marketplace/redhat-marketplace-sz9dx" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.806362 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5pmz9"] Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.807290 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5pmz9" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.809201 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.819458 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5pmz9"] Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.832004 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n5md\" (UniqueName: \"kubernetes.io/projected/cb60184b-c3b3-4ddb-90da-01c62fb183d7-kube-api-access-9n5md\") pod \"redhat-marketplace-sz9dx\" (UID: \"cb60184b-c3b3-4ddb-90da-01c62fb183d7\") " pod="openshift-marketplace/redhat-marketplace-sz9dx" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.832092 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb60184b-c3b3-4ddb-90da-01c62fb183d7-utilities\") pod \"redhat-marketplace-sz9dx\" (UID: \"cb60184b-c3b3-4ddb-90da-01c62fb183d7\") " pod="openshift-marketplace/redhat-marketplace-sz9dx" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.832165 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb60184b-c3b3-4ddb-90da-01c62fb183d7-catalog-content\") pod \"redhat-marketplace-sz9dx\" (UID: \"cb60184b-c3b3-4ddb-90da-01c62fb183d7\") " pod="openshift-marketplace/redhat-marketplace-sz9dx" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.832913 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb60184b-c3b3-4ddb-90da-01c62fb183d7-catalog-content\") pod \"redhat-marketplace-sz9dx\" (UID: \"cb60184b-c3b3-4ddb-90da-01c62fb183d7\") " pod="openshift-marketplace/redhat-marketplace-sz9dx" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.833313 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb60184b-c3b3-4ddb-90da-01c62fb183d7-utilities\") pod \"redhat-marketplace-sz9dx\" (UID: \"cb60184b-c3b3-4ddb-90da-01c62fb183d7\") " pod="openshift-marketplace/redhat-marketplace-sz9dx" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.852052 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n5md\" (UniqueName: \"kubernetes.io/projected/cb60184b-c3b3-4ddb-90da-01c62fb183d7-kube-api-access-9n5md\") pod \"redhat-marketplace-sz9dx\" (UID: \"cb60184b-c3b3-4ddb-90da-01c62fb183d7\") " pod="openshift-marketplace/redhat-marketplace-sz9dx" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.926664 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sz9dx" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.933180 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e5ec892-fe87-4b96-8cf1-cd81510223c2-utilities\") pod \"redhat-operators-5pmz9\" (UID: \"6e5ec892-fe87-4b96-8cf1-cd81510223c2\") " pod="openshift-marketplace/redhat-operators-5pmz9" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.933279 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e5ec892-fe87-4b96-8cf1-cd81510223c2-catalog-content\") pod \"redhat-operators-5pmz9\" (UID: \"6e5ec892-fe87-4b96-8cf1-cd81510223c2\") " pod="openshift-marketplace/redhat-operators-5pmz9" Nov 25 09:28:46 crc kubenswrapper[4776]: I1125 09:28:46.933306 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cfxn\" (UniqueName: \"kubernetes.io/projected/6e5ec892-fe87-4b96-8cf1-cd81510223c2-kube-api-access-7cfxn\") pod \"redhat-operators-5pmz9\" (UID: \"6e5ec892-fe87-4b96-8cf1-cd81510223c2\") " pod="openshift-marketplace/redhat-operators-5pmz9" Nov 25 09:28:47 crc kubenswrapper[4776]: I1125 09:28:47.036743 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e5ec892-fe87-4b96-8cf1-cd81510223c2-utilities\") pod \"redhat-operators-5pmz9\" (UID: \"6e5ec892-fe87-4b96-8cf1-cd81510223c2\") " pod="openshift-marketplace/redhat-operators-5pmz9" Nov 25 09:28:47 crc kubenswrapper[4776]: I1125 09:28:47.037950 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e5ec892-fe87-4b96-8cf1-cd81510223c2-catalog-content\") pod \"redhat-operators-5pmz9\" (UID: \"6e5ec892-fe87-4b96-8cf1-cd81510223c2\") " pod="openshift-marketplace/redhat-operators-5pmz9" Nov 25 09:28:47 crc kubenswrapper[4776]: I1125 09:28:47.037975 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cfxn\" (UniqueName: \"kubernetes.io/projected/6e5ec892-fe87-4b96-8cf1-cd81510223c2-kube-api-access-7cfxn\") pod \"redhat-operators-5pmz9\" (UID: \"6e5ec892-fe87-4b96-8cf1-cd81510223c2\") " pod="openshift-marketplace/redhat-operators-5pmz9" Nov 25 09:28:47 crc kubenswrapper[4776]: I1125 09:28:47.038007 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e5ec892-fe87-4b96-8cf1-cd81510223c2-utilities\") pod \"redhat-operators-5pmz9\" (UID: \"6e5ec892-fe87-4b96-8cf1-cd81510223c2\") " pod="openshift-marketplace/redhat-operators-5pmz9" Nov 25 09:28:47 crc kubenswrapper[4776]: I1125 09:28:47.038249 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e5ec892-fe87-4b96-8cf1-cd81510223c2-catalog-content\") pod \"redhat-operators-5pmz9\" (UID: \"6e5ec892-fe87-4b96-8cf1-cd81510223c2\") " pod="openshift-marketplace/redhat-operators-5pmz9" Nov 25 09:28:47 crc kubenswrapper[4776]: I1125 09:28:47.055817 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cfxn\" (UniqueName: \"kubernetes.io/projected/6e5ec892-fe87-4b96-8cf1-cd81510223c2-kube-api-access-7cfxn\") pod \"redhat-operators-5pmz9\" (UID: \"6e5ec892-fe87-4b96-8cf1-cd81510223c2\") " pod="openshift-marketplace/redhat-operators-5pmz9" Nov 25 09:28:47 crc kubenswrapper[4776]: I1125 09:28:47.095189 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sz9dx"] Nov 25 09:28:47 crc kubenswrapper[4776]: W1125 09:28:47.101432 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb60184b_c3b3_4ddb_90da_01c62fb183d7.slice/crio-38b27280dbeed3e1d3cedfbc8c46c1225237738cb83036381392dfa94bddfe64 WatchSource:0}: Error finding container 38b27280dbeed3e1d3cedfbc8c46c1225237738cb83036381392dfa94bddfe64: Status 404 returned error can't find the container with id 38b27280dbeed3e1d3cedfbc8c46c1225237738cb83036381392dfa94bddfe64 Nov 25 09:28:47 crc kubenswrapper[4776]: I1125 09:28:47.125534 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5pmz9" Nov 25 09:28:47 crc kubenswrapper[4776]: I1125 09:28:47.305945 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5pmz9"] Nov 25 09:28:47 crc kubenswrapper[4776]: I1125 09:28:47.502943 4776 generic.go:334] "Generic (PLEG): container finished" podID="6e5ec892-fe87-4b96-8cf1-cd81510223c2" containerID="c0b73e794a052f0a8f8878e647049945c4acef32db1870f61712db52b3d31824" exitCode=0 Nov 25 09:28:47 crc kubenswrapper[4776]: I1125 09:28:47.502980 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5pmz9" event={"ID":"6e5ec892-fe87-4b96-8cf1-cd81510223c2","Type":"ContainerDied","Data":"c0b73e794a052f0a8f8878e647049945c4acef32db1870f61712db52b3d31824"} Nov 25 09:28:47 crc kubenswrapper[4776]: I1125 09:28:47.503011 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5pmz9" event={"ID":"6e5ec892-fe87-4b96-8cf1-cd81510223c2","Type":"ContainerStarted","Data":"5fd6015dfcf80b26481da08305ce53e411ae6669f8a5a9fb6ed8498ec0eeb7c9"} Nov 25 09:28:47 crc kubenswrapper[4776]: I1125 09:28:47.506951 4776 generic.go:334] "Generic (PLEG): container finished" podID="cb60184b-c3b3-4ddb-90da-01c62fb183d7" containerID="9aedebca27685cb1bb2b4ce5685bd87f0ab7932bf5bdfae6fe4a20388ff39206" exitCode=0 Nov 25 09:28:47 crc kubenswrapper[4776]: I1125 09:28:47.507040 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sz9dx" event={"ID":"cb60184b-c3b3-4ddb-90da-01c62fb183d7","Type":"ContainerDied","Data":"9aedebca27685cb1bb2b4ce5685bd87f0ab7932bf5bdfae6fe4a20388ff39206"} Nov 25 09:28:47 crc kubenswrapper[4776]: I1125 09:28:47.507090 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sz9dx" event={"ID":"cb60184b-c3b3-4ddb-90da-01c62fb183d7","Type":"ContainerStarted","Data":"38b27280dbeed3e1d3cedfbc8c46c1225237738cb83036381392dfa94bddfe64"} Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.005375 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4mztr"] Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.007044 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4mztr" Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.008569 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.017466 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4mztr"] Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.161973 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bd6a235-d6fa-4e0a-90fe-61c36fd645d4-utilities\") pod \"certified-operators-4mztr\" (UID: \"3bd6a235-d6fa-4e0a-90fe-61c36fd645d4\") " pod="openshift-marketplace/certified-operators-4mztr" Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.162033 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mpz8\" (UniqueName: \"kubernetes.io/projected/3bd6a235-d6fa-4e0a-90fe-61c36fd645d4-kube-api-access-4mpz8\") pod \"certified-operators-4mztr\" (UID: \"3bd6a235-d6fa-4e0a-90fe-61c36fd645d4\") " pod="openshift-marketplace/certified-operators-4mztr" Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.162091 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bd6a235-d6fa-4e0a-90fe-61c36fd645d4-catalog-content\") pod \"certified-operators-4mztr\" (UID: \"3bd6a235-d6fa-4e0a-90fe-61c36fd645d4\") " pod="openshift-marketplace/certified-operators-4mztr" Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.205964 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qtfp5"] Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.206950 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qtfp5" Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.208714 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.217440 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qtfp5"] Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.263730 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mpz8\" (UniqueName: \"kubernetes.io/projected/3bd6a235-d6fa-4e0a-90fe-61c36fd645d4-kube-api-access-4mpz8\") pod \"certified-operators-4mztr\" (UID: \"3bd6a235-d6fa-4e0a-90fe-61c36fd645d4\") " pod="openshift-marketplace/certified-operators-4mztr" Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.263805 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bd6a235-d6fa-4e0a-90fe-61c36fd645d4-catalog-content\") pod \"certified-operators-4mztr\" (UID: \"3bd6a235-d6fa-4e0a-90fe-61c36fd645d4\") " pod="openshift-marketplace/certified-operators-4mztr" Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.263915 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bd6a235-d6fa-4e0a-90fe-61c36fd645d4-utilities\") pod \"certified-operators-4mztr\" (UID: \"3bd6a235-d6fa-4e0a-90fe-61c36fd645d4\") " pod="openshift-marketplace/certified-operators-4mztr" Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.264388 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bd6a235-d6fa-4e0a-90fe-61c36fd645d4-catalog-content\") pod \"certified-operators-4mztr\" (UID: \"3bd6a235-d6fa-4e0a-90fe-61c36fd645d4\") " pod="openshift-marketplace/certified-operators-4mztr" Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.264451 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bd6a235-d6fa-4e0a-90fe-61c36fd645d4-utilities\") pod \"certified-operators-4mztr\" (UID: \"3bd6a235-d6fa-4e0a-90fe-61c36fd645d4\") " pod="openshift-marketplace/certified-operators-4mztr" Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.282970 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mpz8\" (UniqueName: \"kubernetes.io/projected/3bd6a235-d6fa-4e0a-90fe-61c36fd645d4-kube-api-access-4mpz8\") pod \"certified-operators-4mztr\" (UID: \"3bd6a235-d6fa-4e0a-90fe-61c36fd645d4\") " pod="openshift-marketplace/certified-operators-4mztr" Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.320683 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4mztr" Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.365184 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/803c71a8-0f91-4fa4-949f-5995ae8af48c-catalog-content\") pod \"community-operators-qtfp5\" (UID: \"803c71a8-0f91-4fa4-949f-5995ae8af48c\") " pod="openshift-marketplace/community-operators-qtfp5" Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.365548 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/803c71a8-0f91-4fa4-949f-5995ae8af48c-utilities\") pod \"community-operators-qtfp5\" (UID: \"803c71a8-0f91-4fa4-949f-5995ae8af48c\") " pod="openshift-marketplace/community-operators-qtfp5" Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.365616 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgd96\" (UniqueName: \"kubernetes.io/projected/803c71a8-0f91-4fa4-949f-5995ae8af48c-kube-api-access-sgd96\") pod \"community-operators-qtfp5\" (UID: \"803c71a8-0f91-4fa4-949f-5995ae8af48c\") " pod="openshift-marketplace/community-operators-qtfp5" Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.467123 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/803c71a8-0f91-4fa4-949f-5995ae8af48c-utilities\") pod \"community-operators-qtfp5\" (UID: \"803c71a8-0f91-4fa4-949f-5995ae8af48c\") " pod="openshift-marketplace/community-operators-qtfp5" Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.467198 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgd96\" (UniqueName: \"kubernetes.io/projected/803c71a8-0f91-4fa4-949f-5995ae8af48c-kube-api-access-sgd96\") pod \"community-operators-qtfp5\" (UID: \"803c71a8-0f91-4fa4-949f-5995ae8af48c\") " pod="openshift-marketplace/community-operators-qtfp5" Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.467244 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/803c71a8-0f91-4fa4-949f-5995ae8af48c-catalog-content\") pod \"community-operators-qtfp5\" (UID: \"803c71a8-0f91-4fa4-949f-5995ae8af48c\") " pod="openshift-marketplace/community-operators-qtfp5" Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.467668 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/803c71a8-0f91-4fa4-949f-5995ae8af48c-utilities\") pod \"community-operators-qtfp5\" (UID: \"803c71a8-0f91-4fa4-949f-5995ae8af48c\") " pod="openshift-marketplace/community-operators-qtfp5" Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.467691 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/803c71a8-0f91-4fa4-949f-5995ae8af48c-catalog-content\") pod \"community-operators-qtfp5\" (UID: \"803c71a8-0f91-4fa4-949f-5995ae8af48c\") " pod="openshift-marketplace/community-operators-qtfp5" Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.487142 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgd96\" (UniqueName: \"kubernetes.io/projected/803c71a8-0f91-4fa4-949f-5995ae8af48c-kube-api-access-sgd96\") pod \"community-operators-qtfp5\" (UID: \"803c71a8-0f91-4fa4-949f-5995ae8af48c\") " pod="openshift-marketplace/community-operators-qtfp5" Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.498387 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4mztr"] Nov 25 09:28:49 crc kubenswrapper[4776]: W1125 09:28:49.501288 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bd6a235_d6fa_4e0a_90fe_61c36fd645d4.slice/crio-b563edd385fd045a89430f80cb581ced5bf54527abd1f29aec7af00dba156812 WatchSource:0}: Error finding container b563edd385fd045a89430f80cb581ced5bf54527abd1f29aec7af00dba156812: Status 404 returned error can't find the container with id b563edd385fd045a89430f80cb581ced5bf54527abd1f29aec7af00dba156812 Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.517312 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4mztr" event={"ID":"3bd6a235-d6fa-4e0a-90fe-61c36fd645d4","Type":"ContainerStarted","Data":"b563edd385fd045a89430f80cb581ced5bf54527abd1f29aec7af00dba156812"} Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.519565 4776 generic.go:334] "Generic (PLEG): container finished" podID="6e5ec892-fe87-4b96-8cf1-cd81510223c2" containerID="f8e5993f36165eff18d03e7547b4e2f10e329241eebfa3ac4c016b22aa9e558d" exitCode=0 Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.519655 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5pmz9" event={"ID":"6e5ec892-fe87-4b96-8cf1-cd81510223c2","Type":"ContainerDied","Data":"f8e5993f36165eff18d03e7547b4e2f10e329241eebfa3ac4c016b22aa9e558d"} Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.522049 4776 generic.go:334] "Generic (PLEG): container finished" podID="cb60184b-c3b3-4ddb-90da-01c62fb183d7" containerID="41109a7841a7e66d98b2c01c9f0616769f8a732fca80f5b69bd2bbcb82038b24" exitCode=0 Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.522182 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sz9dx" event={"ID":"cb60184b-c3b3-4ddb-90da-01c62fb183d7","Type":"ContainerDied","Data":"41109a7841a7e66d98b2c01c9f0616769f8a732fca80f5b69bd2bbcb82038b24"} Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.535958 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qtfp5" Nov 25 09:28:49 crc kubenswrapper[4776]: I1125 09:28:49.782082 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qtfp5"] Nov 25 09:28:49 crc kubenswrapper[4776]: W1125 09:28:49.787630 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod803c71a8_0f91_4fa4_949f_5995ae8af48c.slice/crio-5f58603cdec81d729b4f9b8301e53651ddaa2cf75eaff06f2f129e7bb09407c6 WatchSource:0}: Error finding container 5f58603cdec81d729b4f9b8301e53651ddaa2cf75eaff06f2f129e7bb09407c6: Status 404 returned error can't find the container with id 5f58603cdec81d729b4f9b8301e53651ddaa2cf75eaff06f2f129e7bb09407c6 Nov 25 09:28:50 crc kubenswrapper[4776]: I1125 09:28:50.529930 4776 generic.go:334] "Generic (PLEG): container finished" podID="803c71a8-0f91-4fa4-949f-5995ae8af48c" containerID="fec2192b4efe1e64c830a28f73961187fe306ad1bf78b5f07cc939292a8078fb" exitCode=0 Nov 25 09:28:50 crc kubenswrapper[4776]: I1125 09:28:50.530024 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qtfp5" event={"ID":"803c71a8-0f91-4fa4-949f-5995ae8af48c","Type":"ContainerDied","Data":"fec2192b4efe1e64c830a28f73961187fe306ad1bf78b5f07cc939292a8078fb"} Nov 25 09:28:50 crc kubenswrapper[4776]: I1125 09:28:50.530354 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qtfp5" event={"ID":"803c71a8-0f91-4fa4-949f-5995ae8af48c","Type":"ContainerStarted","Data":"5f58603cdec81d729b4f9b8301e53651ddaa2cf75eaff06f2f129e7bb09407c6"} Nov 25 09:28:50 crc kubenswrapper[4776]: I1125 09:28:50.535619 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5pmz9" event={"ID":"6e5ec892-fe87-4b96-8cf1-cd81510223c2","Type":"ContainerStarted","Data":"bafa88355c4379e175b5e32901bb1406224a3638a16138f21dfda8d0184d235f"} Nov 25 09:28:50 crc kubenswrapper[4776]: I1125 09:28:50.538999 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sz9dx" event={"ID":"cb60184b-c3b3-4ddb-90da-01c62fb183d7","Type":"ContainerStarted","Data":"5a54c4e47dd74b39a826f000d59f8eb077423e6d045825dc5209aad85660ae0f"} Nov 25 09:28:50 crc kubenswrapper[4776]: I1125 09:28:50.540727 4776 generic.go:334] "Generic (PLEG): container finished" podID="3bd6a235-d6fa-4e0a-90fe-61c36fd645d4" containerID="5686eecf274b670fdc58379788adcae83c5f637e042dc38f99d4d8f3efeafb91" exitCode=0 Nov 25 09:28:50 crc kubenswrapper[4776]: I1125 09:28:50.540758 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4mztr" event={"ID":"3bd6a235-d6fa-4e0a-90fe-61c36fd645d4","Type":"ContainerDied","Data":"5686eecf274b670fdc58379788adcae83c5f637e042dc38f99d4d8f3efeafb91"} Nov 25 09:28:50 crc kubenswrapper[4776]: I1125 09:28:50.601056 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sz9dx" podStartSLOduration=2.178626622 podStartE2EDuration="4.601041015s" podCreationTimestamp="2025-11-25 09:28:46 +0000 UTC" firstStartedPulling="2025-11-25 09:28:47.514055545 +0000 UTC m=+272.555115098" lastFinishedPulling="2025-11-25 09:28:49.936469938 +0000 UTC m=+274.977529491" observedRunningTime="2025-11-25 09:28:50.597198419 +0000 UTC m=+275.638257982" watchObservedRunningTime="2025-11-25 09:28:50.601041015 +0000 UTC m=+275.642100568" Nov 25 09:28:50 crc kubenswrapper[4776]: I1125 09:28:50.615932 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5pmz9" podStartSLOduration=1.9850561770000001 podStartE2EDuration="4.615917108s" podCreationTimestamp="2025-11-25 09:28:46 +0000 UTC" firstStartedPulling="2025-11-25 09:28:47.504314611 +0000 UTC m=+272.545374164" lastFinishedPulling="2025-11-25 09:28:50.135175552 +0000 UTC m=+275.176235095" observedRunningTime="2025-11-25 09:28:50.613716633 +0000 UTC m=+275.654776186" watchObservedRunningTime="2025-11-25 09:28:50.615917108 +0000 UTC m=+275.656976661" Nov 25 09:28:51 crc kubenswrapper[4776]: I1125 09:28:51.546866 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4mztr" event={"ID":"3bd6a235-d6fa-4e0a-90fe-61c36fd645d4","Type":"ContainerStarted","Data":"76c02d64794adfd88244525abe367e225d920ff5adb6f7417c9a1713066a15e9"} Nov 25 09:28:51 crc kubenswrapper[4776]: I1125 09:28:51.548558 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qtfp5" event={"ID":"803c71a8-0f91-4fa4-949f-5995ae8af48c","Type":"ContainerStarted","Data":"d2c0e9024beda44de31e92248c6399dc5d8b0df27bc1e2cd4df7f0e99d38e835"} Nov 25 09:28:52 crc kubenswrapper[4776]: I1125 09:28:52.555600 4776 generic.go:334] "Generic (PLEG): container finished" podID="3bd6a235-d6fa-4e0a-90fe-61c36fd645d4" containerID="76c02d64794adfd88244525abe367e225d920ff5adb6f7417c9a1713066a15e9" exitCode=0 Nov 25 09:28:52 crc kubenswrapper[4776]: I1125 09:28:52.555685 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4mztr" event={"ID":"3bd6a235-d6fa-4e0a-90fe-61c36fd645d4","Type":"ContainerDied","Data":"76c02d64794adfd88244525abe367e225d920ff5adb6f7417c9a1713066a15e9"} Nov 25 09:28:52 crc kubenswrapper[4776]: I1125 09:28:52.558403 4776 generic.go:334] "Generic (PLEG): container finished" podID="803c71a8-0f91-4fa4-949f-5995ae8af48c" containerID="d2c0e9024beda44de31e92248c6399dc5d8b0df27bc1e2cd4df7f0e99d38e835" exitCode=0 Nov 25 09:28:52 crc kubenswrapper[4776]: I1125 09:28:52.558458 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qtfp5" event={"ID":"803c71a8-0f91-4fa4-949f-5995ae8af48c","Type":"ContainerDied","Data":"d2c0e9024beda44de31e92248c6399dc5d8b0df27bc1e2cd4df7f0e99d38e835"} Nov 25 09:28:54 crc kubenswrapper[4776]: I1125 09:28:54.570872 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4mztr" event={"ID":"3bd6a235-d6fa-4e0a-90fe-61c36fd645d4","Type":"ContainerStarted","Data":"9a54ea40141c6492c16647254890d7480212f504211cdd7182de33956d03007f"} Nov 25 09:28:54 crc kubenswrapper[4776]: I1125 09:28:54.573896 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qtfp5" event={"ID":"803c71a8-0f91-4fa4-949f-5995ae8af48c","Type":"ContainerStarted","Data":"f0a9e85f5f20ee1bfc4671b51d58a387cec9925e959e9adab277d5e5d1ad976f"} Nov 25 09:28:54 crc kubenswrapper[4776]: I1125 09:28:54.596007 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4mztr" podStartSLOduration=4.159713847 podStartE2EDuration="6.595986247s" podCreationTimestamp="2025-11-25 09:28:48 +0000 UTC" firstStartedPulling="2025-11-25 09:28:50.542122098 +0000 UTC m=+275.583181651" lastFinishedPulling="2025-11-25 09:28:52.978394488 +0000 UTC m=+278.019454051" observedRunningTime="2025-11-25 09:28:54.591494834 +0000 UTC m=+279.632554407" watchObservedRunningTime="2025-11-25 09:28:54.595986247 +0000 UTC m=+279.637045800" Nov 25 09:28:54 crc kubenswrapper[4776]: I1125 09:28:54.615142 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qtfp5" podStartSLOduration=3.114244746 podStartE2EDuration="5.615118957s" podCreationTimestamp="2025-11-25 09:28:49 +0000 UTC" firstStartedPulling="2025-11-25 09:28:50.533322247 +0000 UTC m=+275.574381810" lastFinishedPulling="2025-11-25 09:28:53.034196468 +0000 UTC m=+278.075256021" observedRunningTime="2025-11-25 09:28:54.61285588 +0000 UTC m=+279.653915433" watchObservedRunningTime="2025-11-25 09:28:54.615118957 +0000 UTC m=+279.656178510" Nov 25 09:28:56 crc kubenswrapper[4776]: I1125 09:28:56.926934 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sz9dx" Nov 25 09:28:56 crc kubenswrapper[4776]: I1125 09:28:56.927327 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sz9dx" Nov 25 09:28:56 crc kubenswrapper[4776]: I1125 09:28:56.979840 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sz9dx" Nov 25 09:28:57 crc kubenswrapper[4776]: I1125 09:28:57.126575 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5pmz9" Nov 25 09:28:57 crc kubenswrapper[4776]: I1125 09:28:57.126904 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5pmz9" Nov 25 09:28:57 crc kubenswrapper[4776]: I1125 09:28:57.169096 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5pmz9" Nov 25 09:28:57 crc kubenswrapper[4776]: I1125 09:28:57.625748 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sz9dx" Nov 25 09:28:57 crc kubenswrapper[4776]: I1125 09:28:57.627041 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5pmz9" Nov 25 09:28:59 crc kubenswrapper[4776]: I1125 09:28:59.321033 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4mztr" Nov 25 09:28:59 crc kubenswrapper[4776]: I1125 09:28:59.321370 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4mztr" Nov 25 09:28:59 crc kubenswrapper[4776]: I1125 09:28:59.368722 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4mztr" Nov 25 09:28:59 crc kubenswrapper[4776]: I1125 09:28:59.537056 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qtfp5" Nov 25 09:28:59 crc kubenswrapper[4776]: I1125 09:28:59.537250 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qtfp5" Nov 25 09:28:59 crc kubenswrapper[4776]: I1125 09:28:59.571345 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qtfp5" Nov 25 09:28:59 crc kubenswrapper[4776]: I1125 09:28:59.632669 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qtfp5" Nov 25 09:28:59 crc kubenswrapper[4776]: I1125 09:28:59.632776 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4mztr" Nov 25 09:30:00 crc kubenswrapper[4776]: I1125 09:30:00.132105 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401050-2dn8b"] Nov 25 09:30:00 crc kubenswrapper[4776]: I1125 09:30:00.133499 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401050-2dn8b" Nov 25 09:30:00 crc kubenswrapper[4776]: I1125 09:30:00.135580 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 09:30:00 crc kubenswrapper[4776]: I1125 09:30:00.135714 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 09:30:00 crc kubenswrapper[4776]: I1125 09:30:00.138911 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401050-2dn8b"] Nov 25 09:30:00 crc kubenswrapper[4776]: I1125 09:30:00.241179 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgshl\" (UniqueName: \"kubernetes.io/projected/e37acd8f-6b60-44d4-9d6f-21fd887ed739-kube-api-access-vgshl\") pod \"collect-profiles-29401050-2dn8b\" (UID: \"e37acd8f-6b60-44d4-9d6f-21fd887ed739\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401050-2dn8b" Nov 25 09:30:00 crc kubenswrapper[4776]: I1125 09:30:00.241261 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e37acd8f-6b60-44d4-9d6f-21fd887ed739-config-volume\") pod \"collect-profiles-29401050-2dn8b\" (UID: \"e37acd8f-6b60-44d4-9d6f-21fd887ed739\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401050-2dn8b" Nov 25 09:30:00 crc kubenswrapper[4776]: I1125 09:30:00.241484 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e37acd8f-6b60-44d4-9d6f-21fd887ed739-secret-volume\") pod \"collect-profiles-29401050-2dn8b\" (UID: \"e37acd8f-6b60-44d4-9d6f-21fd887ed739\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401050-2dn8b" Nov 25 09:30:00 crc kubenswrapper[4776]: I1125 09:30:00.342310 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e37acd8f-6b60-44d4-9d6f-21fd887ed739-secret-volume\") pod \"collect-profiles-29401050-2dn8b\" (UID: \"e37acd8f-6b60-44d4-9d6f-21fd887ed739\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401050-2dn8b" Nov 25 09:30:00 crc kubenswrapper[4776]: I1125 09:30:00.342369 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgshl\" (UniqueName: \"kubernetes.io/projected/e37acd8f-6b60-44d4-9d6f-21fd887ed739-kube-api-access-vgshl\") pod \"collect-profiles-29401050-2dn8b\" (UID: \"e37acd8f-6b60-44d4-9d6f-21fd887ed739\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401050-2dn8b" Nov 25 09:30:00 crc kubenswrapper[4776]: I1125 09:30:00.342457 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e37acd8f-6b60-44d4-9d6f-21fd887ed739-config-volume\") pod \"collect-profiles-29401050-2dn8b\" (UID: \"e37acd8f-6b60-44d4-9d6f-21fd887ed739\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401050-2dn8b" Nov 25 09:30:00 crc kubenswrapper[4776]: I1125 09:30:00.343431 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e37acd8f-6b60-44d4-9d6f-21fd887ed739-config-volume\") pod \"collect-profiles-29401050-2dn8b\" (UID: \"e37acd8f-6b60-44d4-9d6f-21fd887ed739\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401050-2dn8b" Nov 25 09:30:00 crc kubenswrapper[4776]: I1125 09:30:00.354188 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e37acd8f-6b60-44d4-9d6f-21fd887ed739-secret-volume\") pod \"collect-profiles-29401050-2dn8b\" (UID: \"e37acd8f-6b60-44d4-9d6f-21fd887ed739\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401050-2dn8b" Nov 25 09:30:00 crc kubenswrapper[4776]: I1125 09:30:00.357892 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgshl\" (UniqueName: \"kubernetes.io/projected/e37acd8f-6b60-44d4-9d6f-21fd887ed739-kube-api-access-vgshl\") pod \"collect-profiles-29401050-2dn8b\" (UID: \"e37acd8f-6b60-44d4-9d6f-21fd887ed739\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401050-2dn8b" Nov 25 09:30:00 crc kubenswrapper[4776]: I1125 09:30:00.448539 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401050-2dn8b" Nov 25 09:30:00 crc kubenswrapper[4776]: I1125 09:30:00.623057 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401050-2dn8b"] Nov 25 09:30:00 crc kubenswrapper[4776]: I1125 09:30:00.954642 4776 generic.go:334] "Generic (PLEG): container finished" podID="e37acd8f-6b60-44d4-9d6f-21fd887ed739" containerID="252bc9b9dbc3c6b6f3ef21f4a2d7fa9d52b302ae1da040029a791e6056f1b8e5" exitCode=0 Nov 25 09:30:00 crc kubenswrapper[4776]: I1125 09:30:00.954694 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401050-2dn8b" event={"ID":"e37acd8f-6b60-44d4-9d6f-21fd887ed739","Type":"ContainerDied","Data":"252bc9b9dbc3c6b6f3ef21f4a2d7fa9d52b302ae1da040029a791e6056f1b8e5"} Nov 25 09:30:00 crc kubenswrapper[4776]: I1125 09:30:00.954945 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401050-2dn8b" event={"ID":"e37acd8f-6b60-44d4-9d6f-21fd887ed739","Type":"ContainerStarted","Data":"f267eded3fe9a5a1fd0c6954e5e26fea2a46d17db22963707341aac428d1f906"} Nov 25 09:30:02 crc kubenswrapper[4776]: I1125 09:30:02.193093 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401050-2dn8b" Nov 25 09:30:02 crc kubenswrapper[4776]: I1125 09:30:02.367439 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e37acd8f-6b60-44d4-9d6f-21fd887ed739-secret-volume\") pod \"e37acd8f-6b60-44d4-9d6f-21fd887ed739\" (UID: \"e37acd8f-6b60-44d4-9d6f-21fd887ed739\") " Nov 25 09:30:02 crc kubenswrapper[4776]: I1125 09:30:02.367905 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e37acd8f-6b60-44d4-9d6f-21fd887ed739-config-volume\") pod \"e37acd8f-6b60-44d4-9d6f-21fd887ed739\" (UID: \"e37acd8f-6b60-44d4-9d6f-21fd887ed739\") " Nov 25 09:30:02 crc kubenswrapper[4776]: I1125 09:30:02.368130 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgshl\" (UniqueName: \"kubernetes.io/projected/e37acd8f-6b60-44d4-9d6f-21fd887ed739-kube-api-access-vgshl\") pod \"e37acd8f-6b60-44d4-9d6f-21fd887ed739\" (UID: \"e37acd8f-6b60-44d4-9d6f-21fd887ed739\") " Nov 25 09:30:02 crc kubenswrapper[4776]: I1125 09:30:02.368694 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e37acd8f-6b60-44d4-9d6f-21fd887ed739-config-volume" (OuterVolumeSpecName: "config-volume") pod "e37acd8f-6b60-44d4-9d6f-21fd887ed739" (UID: "e37acd8f-6b60-44d4-9d6f-21fd887ed739"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:30:02 crc kubenswrapper[4776]: I1125 09:30:02.369348 4776 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e37acd8f-6b60-44d4-9d6f-21fd887ed739-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 09:30:02 crc kubenswrapper[4776]: I1125 09:30:02.376356 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e37acd8f-6b60-44d4-9d6f-21fd887ed739-kube-api-access-vgshl" (OuterVolumeSpecName: "kube-api-access-vgshl") pod "e37acd8f-6b60-44d4-9d6f-21fd887ed739" (UID: "e37acd8f-6b60-44d4-9d6f-21fd887ed739"). InnerVolumeSpecName "kube-api-access-vgshl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:30:02 crc kubenswrapper[4776]: I1125 09:30:02.376554 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e37acd8f-6b60-44d4-9d6f-21fd887ed739-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e37acd8f-6b60-44d4-9d6f-21fd887ed739" (UID: "e37acd8f-6b60-44d4-9d6f-21fd887ed739"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:30:02 crc kubenswrapper[4776]: I1125 09:30:02.470786 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgshl\" (UniqueName: \"kubernetes.io/projected/e37acd8f-6b60-44d4-9d6f-21fd887ed739-kube-api-access-vgshl\") on node \"crc\" DevicePath \"\"" Nov 25 09:30:02 crc kubenswrapper[4776]: I1125 09:30:02.470861 4776 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e37acd8f-6b60-44d4-9d6f-21fd887ed739-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 09:30:02 crc kubenswrapper[4776]: I1125 09:30:02.966626 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401050-2dn8b" event={"ID":"e37acd8f-6b60-44d4-9d6f-21fd887ed739","Type":"ContainerDied","Data":"f267eded3fe9a5a1fd0c6954e5e26fea2a46d17db22963707341aac428d1f906"} Nov 25 09:30:02 crc kubenswrapper[4776]: I1125 09:30:02.966668 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f267eded3fe9a5a1fd0c6954e5e26fea2a46d17db22963707341aac428d1f906" Nov 25 09:30:02 crc kubenswrapper[4776]: I1125 09:30:02.966679 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401050-2dn8b" Nov 25 09:30:17 crc kubenswrapper[4776]: I1125 09:30:17.818852 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:30:17 crc kubenswrapper[4776]: I1125 09:30:17.819398 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:30:47 crc kubenswrapper[4776]: I1125 09:30:47.818198 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:30:47 crc kubenswrapper[4776]: I1125 09:30:47.818953 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:31:17 crc kubenswrapper[4776]: I1125 09:31:17.818983 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:31:17 crc kubenswrapper[4776]: I1125 09:31:17.819566 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:31:17 crc kubenswrapper[4776]: I1125 09:31:17.819620 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 09:31:17 crc kubenswrapper[4776]: I1125 09:31:17.820175 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2898978600d851d8d30b75f90d2f6443dc65ff39be5ce74362b19bd50031aee8"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 09:31:17 crc kubenswrapper[4776]: I1125 09:31:17.820233 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://2898978600d851d8d30b75f90d2f6443dc65ff39be5ce74362b19bd50031aee8" gracePeriod=600 Nov 25 09:31:18 crc kubenswrapper[4776]: I1125 09:31:18.381124 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="2898978600d851d8d30b75f90d2f6443dc65ff39be5ce74362b19bd50031aee8" exitCode=0 Nov 25 09:31:18 crc kubenswrapper[4776]: I1125 09:31:18.381228 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"2898978600d851d8d30b75f90d2f6443dc65ff39be5ce74362b19bd50031aee8"} Nov 25 09:31:18 crc kubenswrapper[4776]: I1125 09:31:18.381427 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"5ae1ab323d058608828b15daefde2387d82c4579f51d09882253c9347334d6f7"} Nov 25 09:31:18 crc kubenswrapper[4776]: I1125 09:31:18.381447 4776 scope.go:117] "RemoveContainer" containerID="c02788c9cbbba68da9f9a9ab605115dd2a73755e351a703509d8d705a05811bb" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.234034 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-7n6zc"] Nov 25 09:31:20 crc kubenswrapper[4776]: E1125 09:31:20.234965 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e37acd8f-6b60-44d4-9d6f-21fd887ed739" containerName="collect-profiles" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.234987 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="e37acd8f-6b60-44d4-9d6f-21fd887ed739" containerName="collect-profiles" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.235173 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="e37acd8f-6b60-44d4-9d6f-21fd887ed739" containerName="collect-profiles" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.235657 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.249794 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-7n6zc"] Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.386142 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-7n6zc\" (UID: \"a52eb557-6695-4e7f-adb3-cd35a9b407d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.386203 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a52eb557-6695-4e7f-adb3-cd35a9b407d7-bound-sa-token\") pod \"image-registry-66df7c8f76-7n6zc\" (UID: \"a52eb557-6695-4e7f-adb3-cd35a9b407d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.386249 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a52eb557-6695-4e7f-adb3-cd35a9b407d7-ca-trust-extracted\") pod \"image-registry-66df7c8f76-7n6zc\" (UID: \"a52eb557-6695-4e7f-adb3-cd35a9b407d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.386268 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fq8bk\" (UniqueName: \"kubernetes.io/projected/a52eb557-6695-4e7f-adb3-cd35a9b407d7-kube-api-access-fq8bk\") pod \"image-registry-66df7c8f76-7n6zc\" (UID: \"a52eb557-6695-4e7f-adb3-cd35a9b407d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.386285 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a52eb557-6695-4e7f-adb3-cd35a9b407d7-registry-tls\") pod \"image-registry-66df7c8f76-7n6zc\" (UID: \"a52eb557-6695-4e7f-adb3-cd35a9b407d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.386301 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a52eb557-6695-4e7f-adb3-cd35a9b407d7-installation-pull-secrets\") pod \"image-registry-66df7c8f76-7n6zc\" (UID: \"a52eb557-6695-4e7f-adb3-cd35a9b407d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.386330 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a52eb557-6695-4e7f-adb3-cd35a9b407d7-registry-certificates\") pod \"image-registry-66df7c8f76-7n6zc\" (UID: \"a52eb557-6695-4e7f-adb3-cd35a9b407d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.386349 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a52eb557-6695-4e7f-adb3-cd35a9b407d7-trusted-ca\") pod \"image-registry-66df7c8f76-7n6zc\" (UID: \"a52eb557-6695-4e7f-adb3-cd35a9b407d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.405157 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-7n6zc\" (UID: \"a52eb557-6695-4e7f-adb3-cd35a9b407d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.487370 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a52eb557-6695-4e7f-adb3-cd35a9b407d7-trusted-ca\") pod \"image-registry-66df7c8f76-7n6zc\" (UID: \"a52eb557-6695-4e7f-adb3-cd35a9b407d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.487444 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a52eb557-6695-4e7f-adb3-cd35a9b407d7-bound-sa-token\") pod \"image-registry-66df7c8f76-7n6zc\" (UID: \"a52eb557-6695-4e7f-adb3-cd35a9b407d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.487489 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a52eb557-6695-4e7f-adb3-cd35a9b407d7-ca-trust-extracted\") pod \"image-registry-66df7c8f76-7n6zc\" (UID: \"a52eb557-6695-4e7f-adb3-cd35a9b407d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.487513 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fq8bk\" (UniqueName: \"kubernetes.io/projected/a52eb557-6695-4e7f-adb3-cd35a9b407d7-kube-api-access-fq8bk\") pod \"image-registry-66df7c8f76-7n6zc\" (UID: \"a52eb557-6695-4e7f-adb3-cd35a9b407d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.487531 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a52eb557-6695-4e7f-adb3-cd35a9b407d7-registry-tls\") pod \"image-registry-66df7c8f76-7n6zc\" (UID: \"a52eb557-6695-4e7f-adb3-cd35a9b407d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.487550 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a52eb557-6695-4e7f-adb3-cd35a9b407d7-installation-pull-secrets\") pod \"image-registry-66df7c8f76-7n6zc\" (UID: \"a52eb557-6695-4e7f-adb3-cd35a9b407d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.487580 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a52eb557-6695-4e7f-adb3-cd35a9b407d7-registry-certificates\") pod \"image-registry-66df7c8f76-7n6zc\" (UID: \"a52eb557-6695-4e7f-adb3-cd35a9b407d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.488176 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a52eb557-6695-4e7f-adb3-cd35a9b407d7-ca-trust-extracted\") pod \"image-registry-66df7c8f76-7n6zc\" (UID: \"a52eb557-6695-4e7f-adb3-cd35a9b407d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.488781 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a52eb557-6695-4e7f-adb3-cd35a9b407d7-registry-certificates\") pod \"image-registry-66df7c8f76-7n6zc\" (UID: \"a52eb557-6695-4e7f-adb3-cd35a9b407d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.488917 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a52eb557-6695-4e7f-adb3-cd35a9b407d7-trusted-ca\") pod \"image-registry-66df7c8f76-7n6zc\" (UID: \"a52eb557-6695-4e7f-adb3-cd35a9b407d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.493895 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a52eb557-6695-4e7f-adb3-cd35a9b407d7-installation-pull-secrets\") pod \"image-registry-66df7c8f76-7n6zc\" (UID: \"a52eb557-6695-4e7f-adb3-cd35a9b407d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.496132 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a52eb557-6695-4e7f-adb3-cd35a9b407d7-registry-tls\") pod \"image-registry-66df7c8f76-7n6zc\" (UID: \"a52eb557-6695-4e7f-adb3-cd35a9b407d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.505402 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a52eb557-6695-4e7f-adb3-cd35a9b407d7-bound-sa-token\") pod \"image-registry-66df7c8f76-7n6zc\" (UID: \"a52eb557-6695-4e7f-adb3-cd35a9b407d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.505742 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fq8bk\" (UniqueName: \"kubernetes.io/projected/a52eb557-6695-4e7f-adb3-cd35a9b407d7-kube-api-access-fq8bk\") pod \"image-registry-66df7c8f76-7n6zc\" (UID: \"a52eb557-6695-4e7f-adb3-cd35a9b407d7\") " pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.553653 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" Nov 25 09:31:20 crc kubenswrapper[4776]: I1125 09:31:20.736592 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-7n6zc"] Nov 25 09:31:20 crc kubenswrapper[4776]: W1125 09:31:20.745331 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda52eb557_6695_4e7f_adb3_cd35a9b407d7.slice/crio-1db2a8e3cb48fa457e8a98542011c09ab32acd2fc47845307d2e6a67f95e3261 WatchSource:0}: Error finding container 1db2a8e3cb48fa457e8a98542011c09ab32acd2fc47845307d2e6a67f95e3261: Status 404 returned error can't find the container with id 1db2a8e3cb48fa457e8a98542011c09ab32acd2fc47845307d2e6a67f95e3261 Nov 25 09:31:21 crc kubenswrapper[4776]: I1125 09:31:21.397429 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" event={"ID":"a52eb557-6695-4e7f-adb3-cd35a9b407d7","Type":"ContainerStarted","Data":"0fa5544c90b48e43ea91ccab99b52cd1c5e92852b9d0a2db5d052ada6f6a0cc3"} Nov 25 09:31:21 crc kubenswrapper[4776]: I1125 09:31:21.397817 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" event={"ID":"a52eb557-6695-4e7f-adb3-cd35a9b407d7","Type":"ContainerStarted","Data":"1db2a8e3cb48fa457e8a98542011c09ab32acd2fc47845307d2e6a67f95e3261"} Nov 25 09:31:21 crc kubenswrapper[4776]: I1125 09:31:21.397858 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" Nov 25 09:31:21 crc kubenswrapper[4776]: I1125 09:31:21.436585 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" podStartSLOduration=1.436558806 podStartE2EDuration="1.436558806s" podCreationTimestamp="2025-11-25 09:31:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:31:21.429047687 +0000 UTC m=+426.470107230" watchObservedRunningTime="2025-11-25 09:31:21.436558806 +0000 UTC m=+426.477618369" Nov 25 09:31:40 crc kubenswrapper[4776]: I1125 09:31:40.558073 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-7n6zc" Nov 25 09:31:40 crc kubenswrapper[4776]: I1125 09:31:40.607920 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-tjf22"] Nov 25 09:32:05 crc kubenswrapper[4776]: I1125 09:32:05.645813 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" podUID="2441a0f3-b420-4a64-803b-81082cc15874" containerName="registry" containerID="cri-o://1426561e5140b314f69ecb69456d3fdf76ef4d0201b681d26b5d01bc154eaedd" gracePeriod=30 Nov 25 09:32:05 crc kubenswrapper[4776]: I1125 09:32:05.964532 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.014817 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2441a0f3-b420-4a64-803b-81082cc15874-bound-sa-token\") pod \"2441a0f3-b420-4a64-803b-81082cc15874\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.014867 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2441a0f3-b420-4a64-803b-81082cc15874-trusted-ca\") pod \"2441a0f3-b420-4a64-803b-81082cc15874\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.014918 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2441a0f3-b420-4a64-803b-81082cc15874-registry-certificates\") pod \"2441a0f3-b420-4a64-803b-81082cc15874\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.014996 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwv8h\" (UniqueName: \"kubernetes.io/projected/2441a0f3-b420-4a64-803b-81082cc15874-kube-api-access-mwv8h\") pod \"2441a0f3-b420-4a64-803b-81082cc15874\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.015021 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2441a0f3-b420-4a64-803b-81082cc15874-ca-trust-extracted\") pod \"2441a0f3-b420-4a64-803b-81082cc15874\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.015054 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2441a0f3-b420-4a64-803b-81082cc15874-installation-pull-secrets\") pod \"2441a0f3-b420-4a64-803b-81082cc15874\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.015087 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2441a0f3-b420-4a64-803b-81082cc15874-registry-tls\") pod \"2441a0f3-b420-4a64-803b-81082cc15874\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.015188 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"2441a0f3-b420-4a64-803b-81082cc15874\" (UID: \"2441a0f3-b420-4a64-803b-81082cc15874\") " Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.015975 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2441a0f3-b420-4a64-803b-81082cc15874-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "2441a0f3-b420-4a64-803b-81082cc15874" (UID: "2441a0f3-b420-4a64-803b-81082cc15874"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.016110 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2441a0f3-b420-4a64-803b-81082cc15874-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "2441a0f3-b420-4a64-803b-81082cc15874" (UID: "2441a0f3-b420-4a64-803b-81082cc15874"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.020847 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2441a0f3-b420-4a64-803b-81082cc15874-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "2441a0f3-b420-4a64-803b-81082cc15874" (UID: "2441a0f3-b420-4a64-803b-81082cc15874"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.021521 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2441a0f3-b420-4a64-803b-81082cc15874-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "2441a0f3-b420-4a64-803b-81082cc15874" (UID: "2441a0f3-b420-4a64-803b-81082cc15874"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.021709 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2441a0f3-b420-4a64-803b-81082cc15874-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "2441a0f3-b420-4a64-803b-81082cc15874" (UID: "2441a0f3-b420-4a64-803b-81082cc15874"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.022006 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2441a0f3-b420-4a64-803b-81082cc15874-kube-api-access-mwv8h" (OuterVolumeSpecName: "kube-api-access-mwv8h") pod "2441a0f3-b420-4a64-803b-81082cc15874" (UID: "2441a0f3-b420-4a64-803b-81082cc15874"). InnerVolumeSpecName "kube-api-access-mwv8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.028679 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "2441a0f3-b420-4a64-803b-81082cc15874" (UID: "2441a0f3-b420-4a64-803b-81082cc15874"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.034612 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2441a0f3-b420-4a64-803b-81082cc15874-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "2441a0f3-b420-4a64-803b-81082cc15874" (UID: "2441a0f3-b420-4a64-803b-81082cc15874"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.115999 4776 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2441a0f3-b420-4a64-803b-81082cc15874-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.116029 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwv8h\" (UniqueName: \"kubernetes.io/projected/2441a0f3-b420-4a64-803b-81082cc15874-kube-api-access-mwv8h\") on node \"crc\" DevicePath \"\"" Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.116038 4776 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2441a0f3-b420-4a64-803b-81082cc15874-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.116053 4776 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2441a0f3-b420-4a64-803b-81082cc15874-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.116062 4776 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2441a0f3-b420-4a64-803b-81082cc15874-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.116100 4776 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2441a0f3-b420-4a64-803b-81082cc15874-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.116109 4776 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2441a0f3-b420-4a64-803b-81082cc15874-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.677902 4776 generic.go:334] "Generic (PLEG): container finished" podID="2441a0f3-b420-4a64-803b-81082cc15874" containerID="1426561e5140b314f69ecb69456d3fdf76ef4d0201b681d26b5d01bc154eaedd" exitCode=0 Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.677974 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" event={"ID":"2441a0f3-b420-4a64-803b-81082cc15874","Type":"ContainerDied","Data":"1426561e5140b314f69ecb69456d3fdf76ef4d0201b681d26b5d01bc154eaedd"} Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.678014 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" event={"ID":"2441a0f3-b420-4a64-803b-81082cc15874","Type":"ContainerDied","Data":"c760fdf52209f8d7e7374c21df5d5a74e223b27cc07cb83155583ea1c8e24fcb"} Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.678041 4776 scope.go:117] "RemoveContainer" containerID="1426561e5140b314f69ecb69456d3fdf76ef4d0201b681d26b5d01bc154eaedd" Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.678271 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-tjf22" Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.704739 4776 scope.go:117] "RemoveContainer" containerID="1426561e5140b314f69ecb69456d3fdf76ef4d0201b681d26b5d01bc154eaedd" Nov 25 09:32:06 crc kubenswrapper[4776]: E1125 09:32:06.705180 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1426561e5140b314f69ecb69456d3fdf76ef4d0201b681d26b5d01bc154eaedd\": container with ID starting with 1426561e5140b314f69ecb69456d3fdf76ef4d0201b681d26b5d01bc154eaedd not found: ID does not exist" containerID="1426561e5140b314f69ecb69456d3fdf76ef4d0201b681d26b5d01bc154eaedd" Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.705246 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1426561e5140b314f69ecb69456d3fdf76ef4d0201b681d26b5d01bc154eaedd"} err="failed to get container status \"1426561e5140b314f69ecb69456d3fdf76ef4d0201b681d26b5d01bc154eaedd\": rpc error: code = NotFound desc = could not find container \"1426561e5140b314f69ecb69456d3fdf76ef4d0201b681d26b5d01bc154eaedd\": container with ID starting with 1426561e5140b314f69ecb69456d3fdf76ef4d0201b681d26b5d01bc154eaedd not found: ID does not exist" Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.722872 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-tjf22"] Nov 25 09:32:06 crc kubenswrapper[4776]: I1125 09:32:06.727101 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-tjf22"] Nov 25 09:32:07 crc kubenswrapper[4776]: I1125 09:32:07.671206 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2441a0f3-b420-4a64-803b-81082cc15874" path="/var/lib/kubelet/pods/2441a0f3-b420-4a64-803b-81082cc15874/volumes" Nov 25 09:33:47 crc kubenswrapper[4776]: I1125 09:33:47.818634 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:33:47 crc kubenswrapper[4776]: I1125 09:33:47.819388 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:34:17 crc kubenswrapper[4776]: I1125 09:34:17.818662 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:34:17 crc kubenswrapper[4776]: I1125 09:34:17.819755 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:34:47 crc kubenswrapper[4776]: I1125 09:34:47.818244 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:34:47 crc kubenswrapper[4776]: I1125 09:34:47.818696 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:34:47 crc kubenswrapper[4776]: I1125 09:34:47.818749 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 09:34:47 crc kubenswrapper[4776]: I1125 09:34:47.819507 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5ae1ab323d058608828b15daefde2387d82c4579f51d09882253c9347334d6f7"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 09:34:47 crc kubenswrapper[4776]: I1125 09:34:47.819571 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://5ae1ab323d058608828b15daefde2387d82c4579f51d09882253c9347334d6f7" gracePeriod=600 Nov 25 09:34:48 crc kubenswrapper[4776]: I1125 09:34:48.641385 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="5ae1ab323d058608828b15daefde2387d82c4579f51d09882253c9347334d6f7" exitCode=0 Nov 25 09:34:48 crc kubenswrapper[4776]: I1125 09:34:48.641416 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"5ae1ab323d058608828b15daefde2387d82c4579f51d09882253c9347334d6f7"} Nov 25 09:34:48 crc kubenswrapper[4776]: I1125 09:34:48.641725 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"dc4f980209957fcb2e8415b8fd0ea1aba1b60952959f82bbdfd9dd9f2991ea9c"} Nov 25 09:34:48 crc kubenswrapper[4776]: I1125 09:34:48.641750 4776 scope.go:117] "RemoveContainer" containerID="2898978600d851d8d30b75f90d2f6443dc65ff39be5ce74362b19bd50031aee8" Nov 25 09:36:33 crc kubenswrapper[4776]: I1125 09:36:33.631713 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-r7x57"] Nov 25 09:36:33 crc kubenswrapper[4776]: I1125 09:36:33.633686 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295" gracePeriod=30 Nov 25 09:36:33 crc kubenswrapper[4776]: I1125 09:36:33.633822 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="nbdb" containerID="cri-o://8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49" gracePeriod=30 Nov 25 09:36:33 crc kubenswrapper[4776]: I1125 09:36:33.633858 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="northd" containerID="cri-o://55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516" gracePeriod=30 Nov 25 09:36:33 crc kubenswrapper[4776]: I1125 09:36:33.633695 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="ovn-controller" containerID="cri-o://af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1" gracePeriod=30 Nov 25 09:36:33 crc kubenswrapper[4776]: I1125 09:36:33.633907 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="kube-rbac-proxy-node" containerID="cri-o://c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188" gracePeriod=30 Nov 25 09:36:33 crc kubenswrapper[4776]: I1125 09:36:33.633948 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="ovn-acl-logging" containerID="cri-o://954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760" gracePeriod=30 Nov 25 09:36:33 crc kubenswrapper[4776]: I1125 09:36:33.633989 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="sbdb" containerID="cri-o://957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5" gracePeriod=30 Nov 25 09:36:33 crc kubenswrapper[4776]: I1125 09:36:33.679981 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="ovnkube-controller" containerID="cri-o://fdec62f7fcca2fe281153267dffee23264c5d60c8d63fe9d451cd8699b962ae7" gracePeriod=30 Nov 25 09:36:33 crc kubenswrapper[4776]: E1125 09:36:33.754284 4776 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34502968_eb8d_47fa_83c2_7f6f6735ce2d.slice/crio-af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34502968_eb8d_47fa_83c2_7f6f6735ce2d.slice/crio-c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34502968_eb8d_47fa_83c2_7f6f6735ce2d.slice/crio-conmon-c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34502968_eb8d_47fa_83c2_7f6f6735ce2d.slice/crio-conmon-954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd4c37fb2_c7d2_4155_8dd9_13ed9c7d3f8c.slice/crio-ea0a18f856c9c659938df76dc703106bc870a8067586cfcda259e8dfe225e2ab.scope\": RecentStats: unable to find data in memory cache]" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.230754 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-r7x57_34502968-eb8d-47fa-83c2-7f6f6735ce2d/ovnkube-controller/3.log" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.233522 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-r7x57_34502968-eb8d-47fa-83c2-7f6f6735ce2d/ovn-acl-logging/0.log" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.234049 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-r7x57_34502968-eb8d-47fa-83c2-7f6f6735ce2d/ovn-controller/0.log" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.234471 4776 generic.go:334] "Generic (PLEG): container finished" podID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerID="fdec62f7fcca2fe281153267dffee23264c5d60c8d63fe9d451cd8699b962ae7" exitCode=0 Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.234509 4776 generic.go:334] "Generic (PLEG): container finished" podID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerID="957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5" exitCode=0 Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.234522 4776 generic.go:334] "Generic (PLEG): container finished" podID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerID="8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49" exitCode=0 Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.234536 4776 generic.go:334] "Generic (PLEG): container finished" podID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerID="55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516" exitCode=0 Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.234546 4776 generic.go:334] "Generic (PLEG): container finished" podID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerID="c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295" exitCode=0 Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.234558 4776 generic.go:334] "Generic (PLEG): container finished" podID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerID="c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188" exitCode=0 Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.234567 4776 generic.go:334] "Generic (PLEG): container finished" podID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerID="954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760" exitCode=143 Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.234576 4776 generic.go:334] "Generic (PLEG): container finished" podID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerID="af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1" exitCode=143 Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.234572 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" event={"ID":"34502968-eb8d-47fa-83c2-7f6f6735ce2d","Type":"ContainerDied","Data":"fdec62f7fcca2fe281153267dffee23264c5d60c8d63fe9d451cd8699b962ae7"} Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.234624 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" event={"ID":"34502968-eb8d-47fa-83c2-7f6f6735ce2d","Type":"ContainerDied","Data":"957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5"} Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.234646 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" event={"ID":"34502968-eb8d-47fa-83c2-7f6f6735ce2d","Type":"ContainerDied","Data":"8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49"} Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.234662 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" event={"ID":"34502968-eb8d-47fa-83c2-7f6f6735ce2d","Type":"ContainerDied","Data":"55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516"} Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.234679 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" event={"ID":"34502968-eb8d-47fa-83c2-7f6f6735ce2d","Type":"ContainerDied","Data":"c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295"} Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.234695 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" event={"ID":"34502968-eb8d-47fa-83c2-7f6f6735ce2d","Type":"ContainerDied","Data":"c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188"} Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.234711 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" event={"ID":"34502968-eb8d-47fa-83c2-7f6f6735ce2d","Type":"ContainerDied","Data":"954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760"} Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.234730 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" event={"ID":"34502968-eb8d-47fa-83c2-7f6f6735ce2d","Type":"ContainerDied","Data":"af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1"} Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.234733 4776 scope.go:117] "RemoveContainer" containerID="3c5b9f801210cc7df25d48a313b74f7a347bfd5cba0855ee054fc48baf53e3b0" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.237496 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5q5c6_d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c/kube-multus/2.log" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.238138 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5q5c6_d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c/kube-multus/1.log" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.238214 4776 generic.go:334] "Generic (PLEG): container finished" podID="d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c" containerID="ea0a18f856c9c659938df76dc703106bc870a8067586cfcda259e8dfe225e2ab" exitCode=2 Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.238255 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5q5c6" event={"ID":"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c","Type":"ContainerDied","Data":"ea0a18f856c9c659938df76dc703106bc870a8067586cfcda259e8dfe225e2ab"} Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.238893 4776 scope.go:117] "RemoveContainer" containerID="ea0a18f856c9c659938df76dc703106bc870a8067586cfcda259e8dfe225e2ab" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.342981 4776 scope.go:117] "RemoveContainer" containerID="d7b1ea649a64d7a158675b397137593f6b901e762b67261d2e2872eeedc0d895" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.415617 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-r7x57_34502968-eb8d-47fa-83c2-7f6f6735ce2d/ovn-acl-logging/0.log" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.416099 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-r7x57_34502968-eb8d-47fa-83c2-7f6f6735ce2d/ovn-controller/0.log" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.416452 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.461061 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4cx97"] Nov 25 09:36:34 crc kubenswrapper[4776]: E1125 09:36:34.461286 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="northd" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.461298 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="northd" Nov 25 09:36:34 crc kubenswrapper[4776]: E1125 09:36:34.461309 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="ovn-acl-logging" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.461317 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="ovn-acl-logging" Nov 25 09:36:34 crc kubenswrapper[4776]: E1125 09:36:34.461327 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="ovnkube-controller" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.461333 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="ovnkube-controller" Nov 25 09:36:34 crc kubenswrapper[4776]: E1125 09:36:34.461339 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="kubecfg-setup" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.461344 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="kubecfg-setup" Nov 25 09:36:34 crc kubenswrapper[4776]: E1125 09:36:34.461356 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="kube-rbac-proxy-node" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.461362 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="kube-rbac-proxy-node" Nov 25 09:36:34 crc kubenswrapper[4776]: E1125 09:36:34.461370 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="nbdb" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.461375 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="nbdb" Nov 25 09:36:34 crc kubenswrapper[4776]: E1125 09:36:34.461382 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.461387 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 09:36:34 crc kubenswrapper[4776]: E1125 09:36:34.461396 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="ovnkube-controller" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.461401 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="ovnkube-controller" Nov 25 09:36:34 crc kubenswrapper[4776]: E1125 09:36:34.461411 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="ovnkube-controller" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.461417 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="ovnkube-controller" Nov 25 09:36:34 crc kubenswrapper[4776]: E1125 09:36:34.461425 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="sbdb" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.461431 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="sbdb" Nov 25 09:36:34 crc kubenswrapper[4776]: E1125 09:36:34.461441 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2441a0f3-b420-4a64-803b-81082cc15874" containerName="registry" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.461447 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2441a0f3-b420-4a64-803b-81082cc15874" containerName="registry" Nov 25 09:36:34 crc kubenswrapper[4776]: E1125 09:36:34.461453 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="ovn-controller" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.461459 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="ovn-controller" Nov 25 09:36:34 crc kubenswrapper[4776]: E1125 09:36:34.461466 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="ovnkube-controller" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.461471 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="ovnkube-controller" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.461561 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="ovn-controller" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.461572 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="northd" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.461577 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="ovnkube-controller" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.461587 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="kube-rbac-proxy-node" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.461595 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="ovnkube-controller" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.461601 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="sbdb" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.461609 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="ovnkube-controller" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.461615 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="ovn-acl-logging" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.461623 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="2441a0f3-b420-4a64-803b-81082cc15874" containerName="registry" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.461629 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.461636 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="nbdb" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.461642 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="ovnkube-controller" Nov 25 09:36:34 crc kubenswrapper[4776]: E1125 09:36:34.461717 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="ovnkube-controller" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.461727 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="ovnkube-controller" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.461848 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" containerName="ovnkube-controller" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.463599 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.497528 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-run-netns\") pod \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.497650 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/34502968-eb8d-47fa-83c2-7f6f6735ce2d-env-overrides\") pod \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.497667 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "34502968-eb8d-47fa-83c2-7f6f6735ce2d" (UID: "34502968-eb8d-47fa-83c2-7f6f6735ce2d"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.497693 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-cni-netd\") pod \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.497760 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "34502968-eb8d-47fa-83c2-7f6f6735ce2d" (UID: "34502968-eb8d-47fa-83c2-7f6f6735ce2d"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.497773 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-slash\") pod \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.497825 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/34502968-eb8d-47fa-83c2-7f6f6735ce2d-ovn-node-metrics-cert\") pod \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.497831 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-slash" (OuterVolumeSpecName: "host-slash") pod "34502968-eb8d-47fa-83c2-7f6f6735ce2d" (UID: "34502968-eb8d-47fa-83c2-7f6f6735ce2d"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.497854 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-run-openvswitch\") pod \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.497880 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-kubelet\") pod \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.497902 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-cni-bin\") pod \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.497933 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-run-ovn-kubernetes\") pod \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.497955 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-node-log\") pod \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.497982 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/34502968-eb8d-47fa-83c2-7f6f6735ce2d-ovnkube-config\") pod \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.498002 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-etc-openvswitch\") pod \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.498036 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-run-ovn\") pod \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.498080 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34502968-eb8d-47fa-83c2-7f6f6735ce2d-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "34502968-eb8d-47fa-83c2-7f6f6735ce2d" (UID: "34502968-eb8d-47fa-83c2-7f6f6735ce2d"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.498097 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.498123 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "34502968-eb8d-47fa-83c2-7f6f6735ce2d" (UID: "34502968-eb8d-47fa-83c2-7f6f6735ce2d"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.498143 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/34502968-eb8d-47fa-83c2-7f6f6735ce2d-ovnkube-script-lib\") pod \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.498152 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "34502968-eb8d-47fa-83c2-7f6f6735ce2d" (UID: "34502968-eb8d-47fa-83c2-7f6f6735ce2d"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.498175 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "34502968-eb8d-47fa-83c2-7f6f6735ce2d" (UID: "34502968-eb8d-47fa-83c2-7f6f6735ce2d"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.498178 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmdxk\" (UniqueName: \"kubernetes.io/projected/34502968-eb8d-47fa-83c2-7f6f6735ce2d-kube-api-access-lmdxk\") pod \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.498207 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-systemd-units\") pod \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.498240 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-run-systemd\") pod \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.498271 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-log-socket\") pod \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.498306 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-var-lib-openvswitch\") pod \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\" (UID: \"34502968-eb8d-47fa-83c2-7f6f6735ce2d\") " Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.498552 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-node-log" (OuterVolumeSpecName: "node-log") pod "34502968-eb8d-47fa-83c2-7f6f6735ce2d" (UID: "34502968-eb8d-47fa-83c2-7f6f6735ce2d"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.498660 4776 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.498676 4776 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.498687 4776 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.498697 4776 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-node-log\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.498706 4776 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.498716 4776 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/34502968-eb8d-47fa-83c2-7f6f6735ce2d-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.498727 4776 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.498738 4776 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-slash\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.498767 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "34502968-eb8d-47fa-83c2-7f6f6735ce2d" (UID: "34502968-eb8d-47fa-83c2-7f6f6735ce2d"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.498791 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "34502968-eb8d-47fa-83c2-7f6f6735ce2d" (UID: "34502968-eb8d-47fa-83c2-7f6f6735ce2d"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.498953 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "34502968-eb8d-47fa-83c2-7f6f6735ce2d" (UID: "34502968-eb8d-47fa-83c2-7f6f6735ce2d"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.498996 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34502968-eb8d-47fa-83c2-7f6f6735ce2d-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "34502968-eb8d-47fa-83c2-7f6f6735ce2d" (UID: "34502968-eb8d-47fa-83c2-7f6f6735ce2d"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.499022 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "34502968-eb8d-47fa-83c2-7f6f6735ce2d" (UID: "34502968-eb8d-47fa-83c2-7f6f6735ce2d"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.499040 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "34502968-eb8d-47fa-83c2-7f6f6735ce2d" (UID: "34502968-eb8d-47fa-83c2-7f6f6735ce2d"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.499091 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-log-socket" (OuterVolumeSpecName: "log-socket") pod "34502968-eb8d-47fa-83c2-7f6f6735ce2d" (UID: "34502968-eb8d-47fa-83c2-7f6f6735ce2d"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.499143 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "34502968-eb8d-47fa-83c2-7f6f6735ce2d" (UID: "34502968-eb8d-47fa-83c2-7f6f6735ce2d"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.499486 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34502968-eb8d-47fa-83c2-7f6f6735ce2d-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "34502968-eb8d-47fa-83c2-7f6f6735ce2d" (UID: "34502968-eb8d-47fa-83c2-7f6f6735ce2d"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.504140 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34502968-eb8d-47fa-83c2-7f6f6735ce2d-kube-api-access-lmdxk" (OuterVolumeSpecName: "kube-api-access-lmdxk") pod "34502968-eb8d-47fa-83c2-7f6f6735ce2d" (UID: "34502968-eb8d-47fa-83c2-7f6f6735ce2d"). InnerVolumeSpecName "kube-api-access-lmdxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.505390 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34502968-eb8d-47fa-83c2-7f6f6735ce2d-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "34502968-eb8d-47fa-83c2-7f6f6735ce2d" (UID: "34502968-eb8d-47fa-83c2-7f6f6735ce2d"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.515906 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "34502968-eb8d-47fa-83c2-7f6f6735ce2d" (UID: "34502968-eb8d-47fa-83c2-7f6f6735ce2d"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.599788 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-host-run-netns\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.599854 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-host-cni-bin\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.599880 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-run-ovn\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.599921 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7rl6\" (UniqueName: \"kubernetes.io/projected/cd019a95-e207-4957-ae5f-a84104331f8f-kube-api-access-k7rl6\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.599989 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-host-kubelet\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.600023 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-node-log\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.600046 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-host-cni-netd\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.600091 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-log-socket\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.600177 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-host-run-ovn-kubernetes\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.600253 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cd019a95-e207-4957-ae5f-a84104331f8f-ovn-node-metrics-cert\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.600274 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-run-systemd\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.600291 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cd019a95-e207-4957-ae5f-a84104331f8f-ovnkube-config\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.600330 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/cd019a95-e207-4957-ae5f-a84104331f8f-ovnkube-script-lib\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.600355 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-etc-openvswitch\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.600372 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-run-openvswitch\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.600393 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cd019a95-e207-4957-ae5f-a84104331f8f-env-overrides\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.600409 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-var-lib-openvswitch\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.600456 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-host-slash\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.600475 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.600508 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-systemd-units\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.600562 4776 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.600574 4776 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/34502968-eb8d-47fa-83c2-7f6f6735ce2d-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.600586 4776 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.600594 4776 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/34502968-eb8d-47fa-83c2-7f6f6735ce2d-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.600604 4776 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.600612 4776 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.600622 4776 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.600631 4776 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/34502968-eb8d-47fa-83c2-7f6f6735ce2d-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.600640 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmdxk\" (UniqueName: \"kubernetes.io/projected/34502968-eb8d-47fa-83c2-7f6f6735ce2d-kube-api-access-lmdxk\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.600663 4776 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.600674 4776 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.600682 4776 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/34502968-eb8d-47fa-83c2-7f6f6735ce2d-log-socket\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.701294 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-host-kubelet\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.701335 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-node-log\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.701354 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-host-cni-netd\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.701375 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-log-socket\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.701392 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-host-run-ovn-kubernetes\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.701413 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cd019a95-e207-4957-ae5f-a84104331f8f-ovn-node-metrics-cert\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.701427 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-run-systemd\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.701443 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cd019a95-e207-4957-ae5f-a84104331f8f-ovnkube-config\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.701461 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/cd019a95-e207-4957-ae5f-a84104331f8f-ovnkube-script-lib\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.701481 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-etc-openvswitch\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.701495 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-run-openvswitch\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.701509 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cd019a95-e207-4957-ae5f-a84104331f8f-env-overrides\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.701525 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-var-lib-openvswitch\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.701550 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-host-slash\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.701565 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.701599 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-systemd-units\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.701615 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-host-run-netns\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.701633 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-host-cni-bin\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.701648 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-run-ovn\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.701665 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7rl6\" (UniqueName: \"kubernetes.io/projected/cd019a95-e207-4957-ae5f-a84104331f8f-kube-api-access-k7rl6\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.701993 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-host-kubelet\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.702026 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-node-log\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.702051 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-host-cni-netd\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.702094 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-log-socket\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.702124 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-host-run-ovn-kubernetes\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.702226 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-systemd-units\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.702256 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-var-lib-openvswitch\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.702302 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-host-run-netns\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.702279 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-host-slash\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.702340 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-run-openvswitch\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.702374 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-etc-openvswitch\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.702378 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.702425 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-run-ovn\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.702402 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-run-systemd\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.702488 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cd019a95-e207-4957-ae5f-a84104331f8f-host-cni-bin\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.703023 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/cd019a95-e207-4957-ae5f-a84104331f8f-ovnkube-script-lib\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.703278 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cd019a95-e207-4957-ae5f-a84104331f8f-env-overrides\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.703447 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cd019a95-e207-4957-ae5f-a84104331f8f-ovnkube-config\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.705967 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cd019a95-e207-4957-ae5f-a84104331f8f-ovn-node-metrics-cert\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.719489 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7rl6\" (UniqueName: \"kubernetes.io/projected/cd019a95-e207-4957-ae5f-a84104331f8f-kube-api-access-k7rl6\") pod \"ovnkube-node-4cx97\" (UID: \"cd019a95-e207-4957-ae5f-a84104331f8f\") " pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: I1125 09:36:34.776290 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:34 crc kubenswrapper[4776]: W1125 09:36:34.793998 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd019a95_e207_4957_ae5f_a84104331f8f.slice/crio-160d5e1e77710fcde52b5dc6c351dcff84d631fe8c0c09214226efea458735c6 WatchSource:0}: Error finding container 160d5e1e77710fcde52b5dc6c351dcff84d631fe8c0c09214226efea458735c6: Status 404 returned error can't find the container with id 160d5e1e77710fcde52b5dc6c351dcff84d631fe8c0c09214226efea458735c6 Nov 25 09:36:35 crc kubenswrapper[4776]: I1125 09:36:35.246057 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5q5c6_d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c/kube-multus/2.log" Nov 25 09:36:35 crc kubenswrapper[4776]: I1125 09:36:35.246497 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5q5c6" event={"ID":"d4c37fb2-c7d2-4155-8dd9-13ed9c7d3f8c","Type":"ContainerStarted","Data":"ed2a40a180331fd780ac6e1691c6fd3c7951b1fe73611f587511e8a5bb186b5f"} Nov 25 09:36:35 crc kubenswrapper[4776]: I1125 09:36:35.248016 4776 generic.go:334] "Generic (PLEG): container finished" podID="cd019a95-e207-4957-ae5f-a84104331f8f" containerID="66035217086e3fd5e0baa16e9a95c631300ada85b19bc0213e562cdb1d11ecaf" exitCode=0 Nov 25 09:36:35 crc kubenswrapper[4776]: I1125 09:36:35.248105 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" event={"ID":"cd019a95-e207-4957-ae5f-a84104331f8f","Type":"ContainerDied","Data":"66035217086e3fd5e0baa16e9a95c631300ada85b19bc0213e562cdb1d11ecaf"} Nov 25 09:36:35 crc kubenswrapper[4776]: I1125 09:36:35.248206 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" event={"ID":"cd019a95-e207-4957-ae5f-a84104331f8f","Type":"ContainerStarted","Data":"160d5e1e77710fcde52b5dc6c351dcff84d631fe8c0c09214226efea458735c6"} Nov 25 09:36:35 crc kubenswrapper[4776]: I1125 09:36:35.254329 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-r7x57_34502968-eb8d-47fa-83c2-7f6f6735ce2d/ovn-acl-logging/0.log" Nov 25 09:36:35 crc kubenswrapper[4776]: I1125 09:36:35.254717 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-r7x57_34502968-eb8d-47fa-83c2-7f6f6735ce2d/ovn-controller/0.log" Nov 25 09:36:35 crc kubenswrapper[4776]: I1125 09:36:35.254997 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" event={"ID":"34502968-eb8d-47fa-83c2-7f6f6735ce2d","Type":"ContainerDied","Data":"963473c38a206a0d639545c34a261dca74521650e71b6eedd3a3275052faeb01"} Nov 25 09:36:35 crc kubenswrapper[4776]: I1125 09:36:35.255039 4776 scope.go:117] "RemoveContainer" containerID="fdec62f7fcca2fe281153267dffee23264c5d60c8d63fe9d451cd8699b962ae7" Nov 25 09:36:35 crc kubenswrapper[4776]: I1125 09:36:35.255214 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-r7x57" Nov 25 09:36:35 crc kubenswrapper[4776]: I1125 09:36:35.281364 4776 scope.go:117] "RemoveContainer" containerID="957c4aaf01a1db4f3564d467bb7b57cbd3fbfe7395c029356db442cffd7a2ba5" Nov 25 09:36:35 crc kubenswrapper[4776]: I1125 09:36:35.321168 4776 scope.go:117] "RemoveContainer" containerID="8ddaf4e931cf870c3bb55f5a1f5b7d4b53d9b0768ccb69c45b06861653ed6a49" Nov 25 09:36:35 crc kubenswrapper[4776]: I1125 09:36:35.346647 4776 scope.go:117] "RemoveContainer" containerID="55f35423ffc46bdae27b5fc816df6322cbb7da9c10b6547eea336f5ab3a9e516" Nov 25 09:36:35 crc kubenswrapper[4776]: I1125 09:36:35.357429 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-r7x57"] Nov 25 09:36:35 crc kubenswrapper[4776]: I1125 09:36:35.360716 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-r7x57"] Nov 25 09:36:35 crc kubenswrapper[4776]: I1125 09:36:35.369844 4776 scope.go:117] "RemoveContainer" containerID="c906505931f62e682d64bc0de3f7016ba7c331956e5484185b3e4440519c3295" Nov 25 09:36:35 crc kubenswrapper[4776]: I1125 09:36:35.383851 4776 scope.go:117] "RemoveContainer" containerID="c309420c64bfa015ac6d75ce0833973730896b7c7ddbe445d13219630b055188" Nov 25 09:36:35 crc kubenswrapper[4776]: I1125 09:36:35.398968 4776 scope.go:117] "RemoveContainer" containerID="954bda4d6d5f7e02ee821f3ec13c9423a41a5ec2772b2a06d3a1aa1971438760" Nov 25 09:36:35 crc kubenswrapper[4776]: I1125 09:36:35.414244 4776 scope.go:117] "RemoveContainer" containerID="af3a265ca682a4cfdcdd6d7960ed41a86ce92cf1d1c8dcfb018c90c6d626fbe1" Nov 25 09:36:35 crc kubenswrapper[4776]: I1125 09:36:35.441449 4776 scope.go:117] "RemoveContainer" containerID="a20f288f67dfda909087263e6103bc767f18275a42f41b4b70e84e38e981085b" Nov 25 09:36:35 crc kubenswrapper[4776]: I1125 09:36:35.669445 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34502968-eb8d-47fa-83c2-7f6f6735ce2d" path="/var/lib/kubelet/pods/34502968-eb8d-47fa-83c2-7f6f6735ce2d/volumes" Nov 25 09:36:36 crc kubenswrapper[4776]: I1125 09:36:36.262983 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" event={"ID":"cd019a95-e207-4957-ae5f-a84104331f8f","Type":"ContainerStarted","Data":"901469832a093a40bbe1862f96453b5d87984e61272f88e4a7fcb894cd41adb5"} Nov 25 09:36:36 crc kubenswrapper[4776]: I1125 09:36:36.263285 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" event={"ID":"cd019a95-e207-4957-ae5f-a84104331f8f","Type":"ContainerStarted","Data":"5f88d014ef6186fa4665f06f781915ad1237e17e8b9cc6dd8b4fa11ad6ea93bf"} Nov 25 09:36:36 crc kubenswrapper[4776]: I1125 09:36:36.263300 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" event={"ID":"cd019a95-e207-4957-ae5f-a84104331f8f","Type":"ContainerStarted","Data":"66a174d5630e930b375d28b008dd1c26eb44a381142c15846135db6b76684438"} Nov 25 09:36:36 crc kubenswrapper[4776]: I1125 09:36:36.263312 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" event={"ID":"cd019a95-e207-4957-ae5f-a84104331f8f","Type":"ContainerStarted","Data":"acc4d83d41f02fe565cd999b87ec1e991038d2eb481719072a978ebd83bdb813"} Nov 25 09:36:36 crc kubenswrapper[4776]: I1125 09:36:36.263322 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" event={"ID":"cd019a95-e207-4957-ae5f-a84104331f8f","Type":"ContainerStarted","Data":"5a3220bfa70b9f8c839def10b9d740d063c9419c16dbda81e7017cb8924b0d43"} Nov 25 09:36:36 crc kubenswrapper[4776]: I1125 09:36:36.263331 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" event={"ID":"cd019a95-e207-4957-ae5f-a84104331f8f","Type":"ContainerStarted","Data":"0b34ff9cbed724c46f4c1dff947002e3f602b93da63678b19ab8ac6a030b30d7"} Nov 25 09:36:38 crc kubenswrapper[4776]: I1125 09:36:38.279691 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" event={"ID":"cd019a95-e207-4957-ae5f-a84104331f8f","Type":"ContainerStarted","Data":"a74594264f2d7d89df8e030b7abfbb5fdc4041b5a1aa6a31228b701b292f12f6"} Nov 25 09:36:38 crc kubenswrapper[4776]: I1125 09:36:38.984674 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-xhbgd"] Nov 25 09:36:38 crc kubenswrapper[4776]: I1125 09:36:38.985720 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-xhbgd" Nov 25 09:36:38 crc kubenswrapper[4776]: I1125 09:36:38.988199 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Nov 25 09:36:38 crc kubenswrapper[4776]: I1125 09:36:38.988472 4776 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-cnl7m" Nov 25 09:36:38 crc kubenswrapper[4776]: I1125 09:36:38.988930 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Nov 25 09:36:38 crc kubenswrapper[4776]: I1125 09:36:38.988978 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Nov 25 09:36:39 crc kubenswrapper[4776]: I1125 09:36:39.060355 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed-node-mnt\") pod \"crc-storage-crc-xhbgd\" (UID: \"f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed\") " pod="crc-storage/crc-storage-crc-xhbgd" Nov 25 09:36:39 crc kubenswrapper[4776]: I1125 09:36:39.060700 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed-crc-storage\") pod \"crc-storage-crc-xhbgd\" (UID: \"f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed\") " pod="crc-storage/crc-storage-crc-xhbgd" Nov 25 09:36:39 crc kubenswrapper[4776]: I1125 09:36:39.060780 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtbkb\" (UniqueName: \"kubernetes.io/projected/f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed-kube-api-access-qtbkb\") pod \"crc-storage-crc-xhbgd\" (UID: \"f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed\") " pod="crc-storage/crc-storage-crc-xhbgd" Nov 25 09:36:39 crc kubenswrapper[4776]: I1125 09:36:39.162231 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed-crc-storage\") pod \"crc-storage-crc-xhbgd\" (UID: \"f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed\") " pod="crc-storage/crc-storage-crc-xhbgd" Nov 25 09:36:39 crc kubenswrapper[4776]: I1125 09:36:39.162330 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtbkb\" (UniqueName: \"kubernetes.io/projected/f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed-kube-api-access-qtbkb\") pod \"crc-storage-crc-xhbgd\" (UID: \"f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed\") " pod="crc-storage/crc-storage-crc-xhbgd" Nov 25 09:36:39 crc kubenswrapper[4776]: I1125 09:36:39.162371 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed-node-mnt\") pod \"crc-storage-crc-xhbgd\" (UID: \"f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed\") " pod="crc-storage/crc-storage-crc-xhbgd" Nov 25 09:36:39 crc kubenswrapper[4776]: I1125 09:36:39.162805 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed-node-mnt\") pod \"crc-storage-crc-xhbgd\" (UID: \"f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed\") " pod="crc-storage/crc-storage-crc-xhbgd" Nov 25 09:36:39 crc kubenswrapper[4776]: I1125 09:36:39.163110 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed-crc-storage\") pod \"crc-storage-crc-xhbgd\" (UID: \"f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed\") " pod="crc-storage/crc-storage-crc-xhbgd" Nov 25 09:36:39 crc kubenswrapper[4776]: I1125 09:36:39.187461 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtbkb\" (UniqueName: \"kubernetes.io/projected/f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed-kube-api-access-qtbkb\") pod \"crc-storage-crc-xhbgd\" (UID: \"f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed\") " pod="crc-storage/crc-storage-crc-xhbgd" Nov 25 09:36:39 crc kubenswrapper[4776]: I1125 09:36:39.302220 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-xhbgd" Nov 25 09:36:39 crc kubenswrapper[4776]: E1125 09:36:39.379197 4776 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-xhbgd_crc-storage_f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed_0(3ed34a18bdce797c6bdd8461c8212ec92eb260acbc5f7f4740dc04b29a9b4c71): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 09:36:39 crc kubenswrapper[4776]: E1125 09:36:39.379291 4776 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-xhbgd_crc-storage_f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed_0(3ed34a18bdce797c6bdd8461c8212ec92eb260acbc5f7f4740dc04b29a9b4c71): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-xhbgd" Nov 25 09:36:39 crc kubenswrapper[4776]: E1125 09:36:39.379319 4776 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-xhbgd_crc-storage_f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed_0(3ed34a18bdce797c6bdd8461c8212ec92eb260acbc5f7f4740dc04b29a9b4c71): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-xhbgd" Nov 25 09:36:39 crc kubenswrapper[4776]: E1125 09:36:39.379378 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-xhbgd_crc-storage(f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-xhbgd_crc-storage(f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-xhbgd_crc-storage_f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed_0(3ed34a18bdce797c6bdd8461c8212ec92eb260acbc5f7f4740dc04b29a9b4c71): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-xhbgd" podUID="f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed" Nov 25 09:36:41 crc kubenswrapper[4776]: I1125 09:36:41.299934 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" event={"ID":"cd019a95-e207-4957-ae5f-a84104331f8f","Type":"ContainerStarted","Data":"40923c6b81da64121e583c40c1960e9921652d98168db67526d6c5b717c1934f"} Nov 25 09:36:41 crc kubenswrapper[4776]: I1125 09:36:41.300363 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:41 crc kubenswrapper[4776]: I1125 09:36:41.300383 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:41 crc kubenswrapper[4776]: I1125 09:36:41.349038 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" podStartSLOduration=7.349018255 podStartE2EDuration="7.349018255s" podCreationTimestamp="2025-11-25 09:36:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:36:41.348527502 +0000 UTC m=+746.389587095" watchObservedRunningTime="2025-11-25 09:36:41.349018255 +0000 UTC m=+746.390077808" Nov 25 09:36:41 crc kubenswrapper[4776]: I1125 09:36:41.367020 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:42 crc kubenswrapper[4776]: I1125 09:36:42.305329 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:42 crc kubenswrapper[4776]: I1125 09:36:42.328407 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-xhbgd"] Nov 25 09:36:42 crc kubenswrapper[4776]: I1125 09:36:42.328543 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-xhbgd" Nov 25 09:36:42 crc kubenswrapper[4776]: I1125 09:36:42.328969 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-xhbgd" Nov 25 09:36:42 crc kubenswrapper[4776]: I1125 09:36:42.348505 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:36:42 crc kubenswrapper[4776]: E1125 09:36:42.359709 4776 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-xhbgd_crc-storage_f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed_0(02591bf09afa5a5a2dd4f5a38dec8c3a79a025618b9d41e7daaffc8d692f986d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 09:36:42 crc kubenswrapper[4776]: E1125 09:36:42.359779 4776 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-xhbgd_crc-storage_f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed_0(02591bf09afa5a5a2dd4f5a38dec8c3a79a025618b9d41e7daaffc8d692f986d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-xhbgd" Nov 25 09:36:42 crc kubenswrapper[4776]: E1125 09:36:42.359804 4776 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-xhbgd_crc-storage_f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed_0(02591bf09afa5a5a2dd4f5a38dec8c3a79a025618b9d41e7daaffc8d692f986d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-xhbgd" Nov 25 09:36:42 crc kubenswrapper[4776]: E1125 09:36:42.359864 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-xhbgd_crc-storage(f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-xhbgd_crc-storage(f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-xhbgd_crc-storage_f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed_0(02591bf09afa5a5a2dd4f5a38dec8c3a79a025618b9d41e7daaffc8d692f986d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-xhbgd" podUID="f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed" Nov 25 09:36:51 crc kubenswrapper[4776]: I1125 09:36:51.401938 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fq5rl"] Nov 25 09:36:51 crc kubenswrapper[4776]: I1125 09:36:51.402725 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" podUID="490fba04-edd0-4edc-9f1e-4e08debf284f" containerName="controller-manager" containerID="cri-o://6ec4c7094e26a33c9942383e806dbf27649e9a5b3c66dc847760cd562f8587d7" gracePeriod=30 Nov 25 09:36:51 crc kubenswrapper[4776]: I1125 09:36:51.491233 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4"] Nov 25 09:36:51 crc kubenswrapper[4776]: I1125 09:36:51.491481 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4" podUID="c7ba5522-7755-4e0a-96e8-6d818e130fa2" containerName="route-controller-manager" containerID="cri-o://b3e28a8d029b15ccffe0ab864f1506542921325f7da5dfe877595c48db93c082" gracePeriod=30 Nov 25 09:36:51 crc kubenswrapper[4776]: I1125 09:36:51.865915 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" Nov 25 09:36:51 crc kubenswrapper[4776]: I1125 09:36:51.910926 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4" Nov 25 09:36:51 crc kubenswrapper[4776]: I1125 09:36:51.926258 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/490fba04-edd0-4edc-9f1e-4e08debf284f-proxy-ca-bundles\") pod \"490fba04-edd0-4edc-9f1e-4e08debf284f\" (UID: \"490fba04-edd0-4edc-9f1e-4e08debf284f\") " Nov 25 09:36:51 crc kubenswrapper[4776]: I1125 09:36:51.926348 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/490fba04-edd0-4edc-9f1e-4e08debf284f-serving-cert\") pod \"490fba04-edd0-4edc-9f1e-4e08debf284f\" (UID: \"490fba04-edd0-4edc-9f1e-4e08debf284f\") " Nov 25 09:36:51 crc kubenswrapper[4776]: I1125 09:36:51.926424 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/490fba04-edd0-4edc-9f1e-4e08debf284f-client-ca\") pod \"490fba04-edd0-4edc-9f1e-4e08debf284f\" (UID: \"490fba04-edd0-4edc-9f1e-4e08debf284f\") " Nov 25 09:36:51 crc kubenswrapper[4776]: I1125 09:36:51.926470 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/490fba04-edd0-4edc-9f1e-4e08debf284f-config\") pod \"490fba04-edd0-4edc-9f1e-4e08debf284f\" (UID: \"490fba04-edd0-4edc-9f1e-4e08debf284f\") " Nov 25 09:36:51 crc kubenswrapper[4776]: I1125 09:36:51.926519 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2jsz\" (UniqueName: \"kubernetes.io/projected/490fba04-edd0-4edc-9f1e-4e08debf284f-kube-api-access-x2jsz\") pod \"490fba04-edd0-4edc-9f1e-4e08debf284f\" (UID: \"490fba04-edd0-4edc-9f1e-4e08debf284f\") " Nov 25 09:36:51 crc kubenswrapper[4776]: I1125 09:36:51.927876 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/490fba04-edd0-4edc-9f1e-4e08debf284f-config" (OuterVolumeSpecName: "config") pod "490fba04-edd0-4edc-9f1e-4e08debf284f" (UID: "490fba04-edd0-4edc-9f1e-4e08debf284f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:36:51 crc kubenswrapper[4776]: I1125 09:36:51.927885 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/490fba04-edd0-4edc-9f1e-4e08debf284f-client-ca" (OuterVolumeSpecName: "client-ca") pod "490fba04-edd0-4edc-9f1e-4e08debf284f" (UID: "490fba04-edd0-4edc-9f1e-4e08debf284f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:36:51 crc kubenswrapper[4776]: I1125 09:36:51.928109 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/490fba04-edd0-4edc-9f1e-4e08debf284f-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "490fba04-edd0-4edc-9f1e-4e08debf284f" (UID: "490fba04-edd0-4edc-9f1e-4e08debf284f"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:36:51 crc kubenswrapper[4776]: I1125 09:36:51.977321 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/490fba04-edd0-4edc-9f1e-4e08debf284f-kube-api-access-x2jsz" (OuterVolumeSpecName: "kube-api-access-x2jsz") pod "490fba04-edd0-4edc-9f1e-4e08debf284f" (UID: "490fba04-edd0-4edc-9f1e-4e08debf284f"). InnerVolumeSpecName "kube-api-access-x2jsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:36:51 crc kubenswrapper[4776]: I1125 09:36:51.981153 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/490fba04-edd0-4edc-9f1e-4e08debf284f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "490fba04-edd0-4edc-9f1e-4e08debf284f" (UID: "490fba04-edd0-4edc-9f1e-4e08debf284f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.028581 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7ba5522-7755-4e0a-96e8-6d818e130fa2-config\") pod \"c7ba5522-7755-4e0a-96e8-6d818e130fa2\" (UID: \"c7ba5522-7755-4e0a-96e8-6d818e130fa2\") " Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.028753 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lm7sd\" (UniqueName: \"kubernetes.io/projected/c7ba5522-7755-4e0a-96e8-6d818e130fa2-kube-api-access-lm7sd\") pod \"c7ba5522-7755-4e0a-96e8-6d818e130fa2\" (UID: \"c7ba5522-7755-4e0a-96e8-6d818e130fa2\") " Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.028799 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7ba5522-7755-4e0a-96e8-6d818e130fa2-serving-cert\") pod \"c7ba5522-7755-4e0a-96e8-6d818e130fa2\" (UID: \"c7ba5522-7755-4e0a-96e8-6d818e130fa2\") " Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.028827 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c7ba5522-7755-4e0a-96e8-6d818e130fa2-client-ca\") pod \"c7ba5522-7755-4e0a-96e8-6d818e130fa2\" (UID: \"c7ba5522-7755-4e0a-96e8-6d818e130fa2\") " Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.029046 4776 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/490fba04-edd0-4edc-9f1e-4e08debf284f-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.029086 4776 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/490fba04-edd0-4edc-9f1e-4e08debf284f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.029099 4776 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/490fba04-edd0-4edc-9f1e-4e08debf284f-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.029110 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/490fba04-edd0-4edc-9f1e-4e08debf284f-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.029123 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2jsz\" (UniqueName: \"kubernetes.io/projected/490fba04-edd0-4edc-9f1e-4e08debf284f-kube-api-access-x2jsz\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.029457 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7ba5522-7755-4e0a-96e8-6d818e130fa2-config" (OuterVolumeSpecName: "config") pod "c7ba5522-7755-4e0a-96e8-6d818e130fa2" (UID: "c7ba5522-7755-4e0a-96e8-6d818e130fa2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.029544 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7ba5522-7755-4e0a-96e8-6d818e130fa2-client-ca" (OuterVolumeSpecName: "client-ca") pod "c7ba5522-7755-4e0a-96e8-6d818e130fa2" (UID: "c7ba5522-7755-4e0a-96e8-6d818e130fa2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.031739 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7ba5522-7755-4e0a-96e8-6d818e130fa2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c7ba5522-7755-4e0a-96e8-6d818e130fa2" (UID: "c7ba5522-7755-4e0a-96e8-6d818e130fa2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.032632 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7ba5522-7755-4e0a-96e8-6d818e130fa2-kube-api-access-lm7sd" (OuterVolumeSpecName: "kube-api-access-lm7sd") pod "c7ba5522-7755-4e0a-96e8-6d818e130fa2" (UID: "c7ba5522-7755-4e0a-96e8-6d818e130fa2"). InnerVolumeSpecName "kube-api-access-lm7sd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.130371 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lm7sd\" (UniqueName: \"kubernetes.io/projected/c7ba5522-7755-4e0a-96e8-6d818e130fa2-kube-api-access-lm7sd\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.130414 4776 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7ba5522-7755-4e0a-96e8-6d818e130fa2-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.130427 4776 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c7ba5522-7755-4e0a-96e8-6d818e130fa2-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.130439 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7ba5522-7755-4e0a-96e8-6d818e130fa2-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.367313 4776 generic.go:334] "Generic (PLEG): container finished" podID="c7ba5522-7755-4e0a-96e8-6d818e130fa2" containerID="b3e28a8d029b15ccffe0ab864f1506542921325f7da5dfe877595c48db93c082" exitCode=0 Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.367377 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.367379 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4" event={"ID":"c7ba5522-7755-4e0a-96e8-6d818e130fa2","Type":"ContainerDied","Data":"b3e28a8d029b15ccffe0ab864f1506542921325f7da5dfe877595c48db93c082"} Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.367553 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4" event={"ID":"c7ba5522-7755-4e0a-96e8-6d818e130fa2","Type":"ContainerDied","Data":"ac4a185653aa1a8f1db2726538a9079d4f34d1b3503b632a3403bee2e488996e"} Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.367630 4776 scope.go:117] "RemoveContainer" containerID="b3e28a8d029b15ccffe0ab864f1506542921325f7da5dfe877595c48db93c082" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.369549 4776 generic.go:334] "Generic (PLEG): container finished" podID="490fba04-edd0-4edc-9f1e-4e08debf284f" containerID="6ec4c7094e26a33c9942383e806dbf27649e9a5b3c66dc847760cd562f8587d7" exitCode=0 Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.369590 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" event={"ID":"490fba04-edd0-4edc-9f1e-4e08debf284f","Type":"ContainerDied","Data":"6ec4c7094e26a33c9942383e806dbf27649e9a5b3c66dc847760cd562f8587d7"} Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.369614 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" event={"ID":"490fba04-edd0-4edc-9f1e-4e08debf284f","Type":"ContainerDied","Data":"26614fee53750b1a6612a048fa02b3f83264c3b8996d41e6b69181ad31742c45"} Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.369595 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-fq5rl" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.384153 4776 scope.go:117] "RemoveContainer" containerID="b3e28a8d029b15ccffe0ab864f1506542921325f7da5dfe877595c48db93c082" Nov 25 09:36:52 crc kubenswrapper[4776]: E1125 09:36:52.384522 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3e28a8d029b15ccffe0ab864f1506542921325f7da5dfe877595c48db93c082\": container with ID starting with b3e28a8d029b15ccffe0ab864f1506542921325f7da5dfe877595c48db93c082 not found: ID does not exist" containerID="b3e28a8d029b15ccffe0ab864f1506542921325f7da5dfe877595c48db93c082" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.384557 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3e28a8d029b15ccffe0ab864f1506542921325f7da5dfe877595c48db93c082"} err="failed to get container status \"b3e28a8d029b15ccffe0ab864f1506542921325f7da5dfe877595c48db93c082\": rpc error: code = NotFound desc = could not find container \"b3e28a8d029b15ccffe0ab864f1506542921325f7da5dfe877595c48db93c082\": container with ID starting with b3e28a8d029b15ccffe0ab864f1506542921325f7da5dfe877595c48db93c082 not found: ID does not exist" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.384581 4776 scope.go:117] "RemoveContainer" containerID="6ec4c7094e26a33c9942383e806dbf27649e9a5b3c66dc847760cd562f8587d7" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.400326 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4"] Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.406884 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6sd4"] Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.413338 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fq5rl"] Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.414796 4776 scope.go:117] "RemoveContainer" containerID="6ec4c7094e26a33c9942383e806dbf27649e9a5b3c66dc847760cd562f8587d7" Nov 25 09:36:52 crc kubenswrapper[4776]: E1125 09:36:52.415404 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ec4c7094e26a33c9942383e806dbf27649e9a5b3c66dc847760cd562f8587d7\": container with ID starting with 6ec4c7094e26a33c9942383e806dbf27649e9a5b3c66dc847760cd562f8587d7 not found: ID does not exist" containerID="6ec4c7094e26a33c9942383e806dbf27649e9a5b3c66dc847760cd562f8587d7" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.415466 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ec4c7094e26a33c9942383e806dbf27649e9a5b3c66dc847760cd562f8587d7"} err="failed to get container status \"6ec4c7094e26a33c9942383e806dbf27649e9a5b3c66dc847760cd562f8587d7\": rpc error: code = NotFound desc = could not find container \"6ec4c7094e26a33c9942383e806dbf27649e9a5b3c66dc847760cd562f8587d7\": container with ID starting with 6ec4c7094e26a33c9942383e806dbf27649e9a5b3c66dc847760cd562f8587d7 not found: ID does not exist" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.416912 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-fq5rl"] Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.594374 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6d78d74874-8xpcp"] Nov 25 09:36:52 crc kubenswrapper[4776]: E1125 09:36:52.594665 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="490fba04-edd0-4edc-9f1e-4e08debf284f" containerName="controller-manager" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.594682 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="490fba04-edd0-4edc-9f1e-4e08debf284f" containerName="controller-manager" Nov 25 09:36:52 crc kubenswrapper[4776]: E1125 09:36:52.594708 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7ba5522-7755-4e0a-96e8-6d818e130fa2" containerName="route-controller-manager" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.594716 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7ba5522-7755-4e0a-96e8-6d818e130fa2" containerName="route-controller-manager" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.594839 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7ba5522-7755-4e0a-96e8-6d818e130fa2" containerName="route-controller-manager" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.594858 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="490fba04-edd0-4edc-9f1e-4e08debf284f" containerName="controller-manager" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.595330 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d78d74874-8xpcp" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.597102 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.597542 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.597546 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.597781 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.597895 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh"] Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.598042 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.598713 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.601309 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.602459 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.602591 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.602630 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.602592 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.602984 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.604701 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.606886 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.607992 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh"] Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.612080 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6d78d74874-8xpcp"] Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.636340 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b2d48bb3-265c-4043-97d4-465378eb9b43-proxy-ca-bundles\") pod \"controller-manager-6d78d74874-8xpcp\" (UID: \"b2d48bb3-265c-4043-97d4-465378eb9b43\") " pod="openshift-controller-manager/controller-manager-6d78d74874-8xpcp" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.636440 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2d48bb3-265c-4043-97d4-465378eb9b43-config\") pod \"controller-manager-6d78d74874-8xpcp\" (UID: \"b2d48bb3-265c-4043-97d4-465378eb9b43\") " pod="openshift-controller-manager/controller-manager-6d78d74874-8xpcp" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.636474 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b2d48bb3-265c-4043-97d4-465378eb9b43-client-ca\") pod \"controller-manager-6d78d74874-8xpcp\" (UID: \"b2d48bb3-265c-4043-97d4-465378eb9b43\") " pod="openshift-controller-manager/controller-manager-6d78d74874-8xpcp" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.636508 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csjd8\" (UniqueName: \"kubernetes.io/projected/b2d48bb3-265c-4043-97d4-465378eb9b43-kube-api-access-csjd8\") pod \"controller-manager-6d78d74874-8xpcp\" (UID: \"b2d48bb3-265c-4043-97d4-465378eb9b43\") " pod="openshift-controller-manager/controller-manager-6d78d74874-8xpcp" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.636532 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2d48bb3-265c-4043-97d4-465378eb9b43-serving-cert\") pod \"controller-manager-6d78d74874-8xpcp\" (UID: \"b2d48bb3-265c-4043-97d4-465378eb9b43\") " pod="openshift-controller-manager/controller-manager-6d78d74874-8xpcp" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.738170 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csjd8\" (UniqueName: \"kubernetes.io/projected/b2d48bb3-265c-4043-97d4-465378eb9b43-kube-api-access-csjd8\") pod \"controller-manager-6d78d74874-8xpcp\" (UID: \"b2d48bb3-265c-4043-97d4-465378eb9b43\") " pod="openshift-controller-manager/controller-manager-6d78d74874-8xpcp" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.738269 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2d48bb3-265c-4043-97d4-465378eb9b43-serving-cert\") pod \"controller-manager-6d78d74874-8xpcp\" (UID: \"b2d48bb3-265c-4043-97d4-465378eb9b43\") " pod="openshift-controller-manager/controller-manager-6d78d74874-8xpcp" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.738365 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4eebac0-e547-4f8e-944c-dff2a9c87ce8-config\") pod \"route-controller-manager-6f897c9b7-fjgvh\" (UID: \"b4eebac0-e547-4f8e-944c-dff2a9c87ce8\") " pod="openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.738430 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b2d48bb3-265c-4043-97d4-465378eb9b43-proxy-ca-bundles\") pod \"controller-manager-6d78d74874-8xpcp\" (UID: \"b2d48bb3-265c-4043-97d4-465378eb9b43\") " pod="openshift-controller-manager/controller-manager-6d78d74874-8xpcp" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.738723 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmv6k\" (UniqueName: \"kubernetes.io/projected/b4eebac0-e547-4f8e-944c-dff2a9c87ce8-kube-api-access-vmv6k\") pod \"route-controller-manager-6f897c9b7-fjgvh\" (UID: \"b4eebac0-e547-4f8e-944c-dff2a9c87ce8\") " pod="openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.738760 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2d48bb3-265c-4043-97d4-465378eb9b43-config\") pod \"controller-manager-6d78d74874-8xpcp\" (UID: \"b2d48bb3-265c-4043-97d4-465378eb9b43\") " pod="openshift-controller-manager/controller-manager-6d78d74874-8xpcp" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.738795 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4eebac0-e547-4f8e-944c-dff2a9c87ce8-serving-cert\") pod \"route-controller-manager-6f897c9b7-fjgvh\" (UID: \"b4eebac0-e547-4f8e-944c-dff2a9c87ce8\") " pod="openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.738837 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b2d48bb3-265c-4043-97d4-465378eb9b43-client-ca\") pod \"controller-manager-6d78d74874-8xpcp\" (UID: \"b2d48bb3-265c-4043-97d4-465378eb9b43\") " pod="openshift-controller-manager/controller-manager-6d78d74874-8xpcp" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.738884 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b4eebac0-e547-4f8e-944c-dff2a9c87ce8-client-ca\") pod \"route-controller-manager-6f897c9b7-fjgvh\" (UID: \"b4eebac0-e547-4f8e-944c-dff2a9c87ce8\") " pod="openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.741256 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b2d48bb3-265c-4043-97d4-465378eb9b43-client-ca\") pod \"controller-manager-6d78d74874-8xpcp\" (UID: \"b2d48bb3-265c-4043-97d4-465378eb9b43\") " pod="openshift-controller-manager/controller-manager-6d78d74874-8xpcp" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.741433 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2d48bb3-265c-4043-97d4-465378eb9b43-config\") pod \"controller-manager-6d78d74874-8xpcp\" (UID: \"b2d48bb3-265c-4043-97d4-465378eb9b43\") " pod="openshift-controller-manager/controller-manager-6d78d74874-8xpcp" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.742906 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b2d48bb3-265c-4043-97d4-465378eb9b43-proxy-ca-bundles\") pod \"controller-manager-6d78d74874-8xpcp\" (UID: \"b2d48bb3-265c-4043-97d4-465378eb9b43\") " pod="openshift-controller-manager/controller-manager-6d78d74874-8xpcp" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.743738 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2d48bb3-265c-4043-97d4-465378eb9b43-serving-cert\") pod \"controller-manager-6d78d74874-8xpcp\" (UID: \"b2d48bb3-265c-4043-97d4-465378eb9b43\") " pod="openshift-controller-manager/controller-manager-6d78d74874-8xpcp" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.771710 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csjd8\" (UniqueName: \"kubernetes.io/projected/b2d48bb3-265c-4043-97d4-465378eb9b43-kube-api-access-csjd8\") pod \"controller-manager-6d78d74874-8xpcp\" (UID: \"b2d48bb3-265c-4043-97d4-465378eb9b43\") " pod="openshift-controller-manager/controller-manager-6d78d74874-8xpcp" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.839955 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmv6k\" (UniqueName: \"kubernetes.io/projected/b4eebac0-e547-4f8e-944c-dff2a9c87ce8-kube-api-access-vmv6k\") pod \"route-controller-manager-6f897c9b7-fjgvh\" (UID: \"b4eebac0-e547-4f8e-944c-dff2a9c87ce8\") " pod="openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.839996 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4eebac0-e547-4f8e-944c-dff2a9c87ce8-serving-cert\") pod \"route-controller-manager-6f897c9b7-fjgvh\" (UID: \"b4eebac0-e547-4f8e-944c-dff2a9c87ce8\") " pod="openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.840022 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b4eebac0-e547-4f8e-944c-dff2a9c87ce8-client-ca\") pod \"route-controller-manager-6f897c9b7-fjgvh\" (UID: \"b4eebac0-e547-4f8e-944c-dff2a9c87ce8\") " pod="openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.840089 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4eebac0-e547-4f8e-944c-dff2a9c87ce8-config\") pod \"route-controller-manager-6f897c9b7-fjgvh\" (UID: \"b4eebac0-e547-4f8e-944c-dff2a9c87ce8\") " pod="openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.841237 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4eebac0-e547-4f8e-944c-dff2a9c87ce8-config\") pod \"route-controller-manager-6f897c9b7-fjgvh\" (UID: \"b4eebac0-e547-4f8e-944c-dff2a9c87ce8\") " pod="openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.841610 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b4eebac0-e547-4f8e-944c-dff2a9c87ce8-client-ca\") pod \"route-controller-manager-6f897c9b7-fjgvh\" (UID: \"b4eebac0-e547-4f8e-944c-dff2a9c87ce8\") " pod="openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.845282 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4eebac0-e547-4f8e-944c-dff2a9c87ce8-serving-cert\") pod \"route-controller-manager-6f897c9b7-fjgvh\" (UID: \"b4eebac0-e547-4f8e-944c-dff2a9c87ce8\") " pod="openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.877351 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmv6k\" (UniqueName: \"kubernetes.io/projected/b4eebac0-e547-4f8e-944c-dff2a9c87ce8-kube-api-access-vmv6k\") pod \"route-controller-manager-6f897c9b7-fjgvh\" (UID: \"b4eebac0-e547-4f8e-944c-dff2a9c87ce8\") " pod="openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.929485 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d78d74874-8xpcp" Nov 25 09:36:52 crc kubenswrapper[4776]: I1125 09:36:52.936654 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh" Nov 25 09:36:53 crc kubenswrapper[4776]: I1125 09:36:53.188245 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh"] Nov 25 09:36:53 crc kubenswrapper[4776]: I1125 09:36:53.195232 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6d78d74874-8xpcp"] Nov 25 09:36:53 crc kubenswrapper[4776]: W1125 09:36:53.198034 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2d48bb3_265c_4043_97d4_465378eb9b43.slice/crio-5a9717e1c4aed66cfe004d10b0906b6c9a17e4327351a2a887de2d79a847dc70 WatchSource:0}: Error finding container 5a9717e1c4aed66cfe004d10b0906b6c9a17e4327351a2a887de2d79a847dc70: Status 404 returned error can't find the container with id 5a9717e1c4aed66cfe004d10b0906b6c9a17e4327351a2a887de2d79a847dc70 Nov 25 09:36:53 crc kubenswrapper[4776]: I1125 09:36:53.378026 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d78d74874-8xpcp" event={"ID":"b2d48bb3-265c-4043-97d4-465378eb9b43","Type":"ContainerStarted","Data":"5a9717e1c4aed66cfe004d10b0906b6c9a17e4327351a2a887de2d79a847dc70"} Nov 25 09:36:53 crc kubenswrapper[4776]: I1125 09:36:53.445225 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh"] Nov 25 09:36:53 crc kubenswrapper[4776]: W1125 09:36:53.451456 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4eebac0_e547_4f8e_944c_dff2a9c87ce8.slice/crio-4fc14e6c2ee9ce816aa2f748f46cbd214c5c92f53212226394260d258eb766c2 WatchSource:0}: Error finding container 4fc14e6c2ee9ce816aa2f748f46cbd214c5c92f53212226394260d258eb766c2: Status 404 returned error can't find the container with id 4fc14e6c2ee9ce816aa2f748f46cbd214c5c92f53212226394260d258eb766c2 Nov 25 09:36:53 crc kubenswrapper[4776]: I1125 09:36:53.669200 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="490fba04-edd0-4edc-9f1e-4e08debf284f" path="/var/lib/kubelet/pods/490fba04-edd0-4edc-9f1e-4e08debf284f/volumes" Nov 25 09:36:53 crc kubenswrapper[4776]: I1125 09:36:53.670552 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7ba5522-7755-4e0a-96e8-6d818e130fa2" path="/var/lib/kubelet/pods/c7ba5522-7755-4e0a-96e8-6d818e130fa2/volumes" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.385041 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d78d74874-8xpcp" event={"ID":"b2d48bb3-265c-4043-97d4-465378eb9b43","Type":"ContainerStarted","Data":"7f1a7f0b358029d4afb8e37a98ffbf4fe8c3fba7640af05a508c1265716b16b0"} Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.385373 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6d78d74874-8xpcp" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.386999 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh" event={"ID":"b4eebac0-e547-4f8e-944c-dff2a9c87ce8","Type":"ContainerStarted","Data":"55be4cef4c54d6cf213ee19ebdc5361a021c9a9ac1bfed5413a0d0161e581e49"} Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.387054 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh" event={"ID":"b4eebac0-e547-4f8e-944c-dff2a9c87ce8","Type":"ContainerStarted","Data":"4fc14e6c2ee9ce816aa2f748f46cbd214c5c92f53212226394260d258eb766c2"} Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.387386 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh" podUID="b4eebac0-e547-4f8e-944c-dff2a9c87ce8" containerName="route-controller-manager" containerID="cri-o://55be4cef4c54d6cf213ee19ebdc5361a021c9a9ac1bfed5413a0d0161e581e49" gracePeriod=30 Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.387575 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.391807 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6d78d74874-8xpcp" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.396682 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.414484 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6d78d74874-8xpcp" podStartSLOduration=3.41445289 podStartE2EDuration="3.41445289s" podCreationTimestamp="2025-11-25 09:36:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:36:54.410204724 +0000 UTC m=+759.451264287" watchObservedRunningTime="2025-11-25 09:36:54.41445289 +0000 UTC m=+759.455512483" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.450596 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh" podStartSLOduration=3.450575043 podStartE2EDuration="3.450575043s" podCreationTimestamp="2025-11-25 09:36:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:36:54.448252805 +0000 UTC m=+759.489312368" watchObservedRunningTime="2025-11-25 09:36:54.450575043 +0000 UTC m=+759.491634596" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.758375 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.784528 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5cbcc5f946-bmrsg"] Nov 25 09:36:54 crc kubenswrapper[4776]: E1125 09:36:54.784739 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4eebac0-e547-4f8e-944c-dff2a9c87ce8" containerName="route-controller-manager" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.784751 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4eebac0-e547-4f8e-944c-dff2a9c87ce8" containerName="route-controller-manager" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.784835 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4eebac0-e547-4f8e-944c-dff2a9c87ce8" containerName="route-controller-manager" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.785204 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5cbcc5f946-bmrsg" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.815125 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5cbcc5f946-bmrsg"] Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.869503 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4eebac0-e547-4f8e-944c-dff2a9c87ce8-serving-cert\") pod \"b4eebac0-e547-4f8e-944c-dff2a9c87ce8\" (UID: \"b4eebac0-e547-4f8e-944c-dff2a9c87ce8\") " Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.869556 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmv6k\" (UniqueName: \"kubernetes.io/projected/b4eebac0-e547-4f8e-944c-dff2a9c87ce8-kube-api-access-vmv6k\") pod \"b4eebac0-e547-4f8e-944c-dff2a9c87ce8\" (UID: \"b4eebac0-e547-4f8e-944c-dff2a9c87ce8\") " Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.869596 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4eebac0-e547-4f8e-944c-dff2a9c87ce8-config\") pod \"b4eebac0-e547-4f8e-944c-dff2a9c87ce8\" (UID: \"b4eebac0-e547-4f8e-944c-dff2a9c87ce8\") " Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.869652 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b4eebac0-e547-4f8e-944c-dff2a9c87ce8-client-ca\") pod \"b4eebac0-e547-4f8e-944c-dff2a9c87ce8\" (UID: \"b4eebac0-e547-4f8e-944c-dff2a9c87ce8\") " Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.869783 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6aaa21be-7e3a-4adf-8d71-6fd41b5798a7-client-ca\") pod \"route-controller-manager-5cbcc5f946-bmrsg\" (UID: \"6aaa21be-7e3a-4adf-8d71-6fd41b5798a7\") " pod="openshift-route-controller-manager/route-controller-manager-5cbcc5f946-bmrsg" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.869835 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6aaa21be-7e3a-4adf-8d71-6fd41b5798a7-config\") pod \"route-controller-manager-5cbcc5f946-bmrsg\" (UID: \"6aaa21be-7e3a-4adf-8d71-6fd41b5798a7\") " pod="openshift-route-controller-manager/route-controller-manager-5cbcc5f946-bmrsg" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.869854 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6aaa21be-7e3a-4adf-8d71-6fd41b5798a7-serving-cert\") pod \"route-controller-manager-5cbcc5f946-bmrsg\" (UID: \"6aaa21be-7e3a-4adf-8d71-6fd41b5798a7\") " pod="openshift-route-controller-manager/route-controller-manager-5cbcc5f946-bmrsg" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.869892 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftqjp\" (UniqueName: \"kubernetes.io/projected/6aaa21be-7e3a-4adf-8d71-6fd41b5798a7-kube-api-access-ftqjp\") pod \"route-controller-manager-5cbcc5f946-bmrsg\" (UID: \"6aaa21be-7e3a-4adf-8d71-6fd41b5798a7\") " pod="openshift-route-controller-manager/route-controller-manager-5cbcc5f946-bmrsg" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.870332 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4eebac0-e547-4f8e-944c-dff2a9c87ce8-client-ca" (OuterVolumeSpecName: "client-ca") pod "b4eebac0-e547-4f8e-944c-dff2a9c87ce8" (UID: "b4eebac0-e547-4f8e-944c-dff2a9c87ce8"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.870343 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4eebac0-e547-4f8e-944c-dff2a9c87ce8-config" (OuterVolumeSpecName: "config") pod "b4eebac0-e547-4f8e-944c-dff2a9c87ce8" (UID: "b4eebac0-e547-4f8e-944c-dff2a9c87ce8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.875258 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4eebac0-e547-4f8e-944c-dff2a9c87ce8-kube-api-access-vmv6k" (OuterVolumeSpecName: "kube-api-access-vmv6k") pod "b4eebac0-e547-4f8e-944c-dff2a9c87ce8" (UID: "b4eebac0-e547-4f8e-944c-dff2a9c87ce8"). InnerVolumeSpecName "kube-api-access-vmv6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.876350 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4eebac0-e547-4f8e-944c-dff2a9c87ce8-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b4eebac0-e547-4f8e-944c-dff2a9c87ce8" (UID: "b4eebac0-e547-4f8e-944c-dff2a9c87ce8"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.971026 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6aaa21be-7e3a-4adf-8d71-6fd41b5798a7-client-ca\") pod \"route-controller-manager-5cbcc5f946-bmrsg\" (UID: \"6aaa21be-7e3a-4adf-8d71-6fd41b5798a7\") " pod="openshift-route-controller-manager/route-controller-manager-5cbcc5f946-bmrsg" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.972253 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6aaa21be-7e3a-4adf-8d71-6fd41b5798a7-config\") pod \"route-controller-manager-5cbcc5f946-bmrsg\" (UID: \"6aaa21be-7e3a-4adf-8d71-6fd41b5798a7\") " pod="openshift-route-controller-manager/route-controller-manager-5cbcc5f946-bmrsg" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.972291 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6aaa21be-7e3a-4adf-8d71-6fd41b5798a7-serving-cert\") pod \"route-controller-manager-5cbcc5f946-bmrsg\" (UID: \"6aaa21be-7e3a-4adf-8d71-6fd41b5798a7\") " pod="openshift-route-controller-manager/route-controller-manager-5cbcc5f946-bmrsg" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.972331 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftqjp\" (UniqueName: \"kubernetes.io/projected/6aaa21be-7e3a-4adf-8d71-6fd41b5798a7-kube-api-access-ftqjp\") pod \"route-controller-manager-5cbcc5f946-bmrsg\" (UID: \"6aaa21be-7e3a-4adf-8d71-6fd41b5798a7\") " pod="openshift-route-controller-manager/route-controller-manager-5cbcc5f946-bmrsg" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.972375 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6aaa21be-7e3a-4adf-8d71-6fd41b5798a7-client-ca\") pod \"route-controller-manager-5cbcc5f946-bmrsg\" (UID: \"6aaa21be-7e3a-4adf-8d71-6fd41b5798a7\") " pod="openshift-route-controller-manager/route-controller-manager-5cbcc5f946-bmrsg" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.972420 4776 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4eebac0-e547-4f8e-944c-dff2a9c87ce8-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.972434 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmv6k\" (UniqueName: \"kubernetes.io/projected/b4eebac0-e547-4f8e-944c-dff2a9c87ce8-kube-api-access-vmv6k\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.972450 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4eebac0-e547-4f8e-944c-dff2a9c87ce8-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.972461 4776 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b4eebac0-e547-4f8e-944c-dff2a9c87ce8-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.973945 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6aaa21be-7e3a-4adf-8d71-6fd41b5798a7-config\") pod \"route-controller-manager-5cbcc5f946-bmrsg\" (UID: \"6aaa21be-7e3a-4adf-8d71-6fd41b5798a7\") " pod="openshift-route-controller-manager/route-controller-manager-5cbcc5f946-bmrsg" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.983263 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6aaa21be-7e3a-4adf-8d71-6fd41b5798a7-serving-cert\") pod \"route-controller-manager-5cbcc5f946-bmrsg\" (UID: \"6aaa21be-7e3a-4adf-8d71-6fd41b5798a7\") " pod="openshift-route-controller-manager/route-controller-manager-5cbcc5f946-bmrsg" Nov 25 09:36:54 crc kubenswrapper[4776]: I1125 09:36:54.988697 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftqjp\" (UniqueName: \"kubernetes.io/projected/6aaa21be-7e3a-4adf-8d71-6fd41b5798a7-kube-api-access-ftqjp\") pod \"route-controller-manager-5cbcc5f946-bmrsg\" (UID: \"6aaa21be-7e3a-4adf-8d71-6fd41b5798a7\") " pod="openshift-route-controller-manager/route-controller-manager-5cbcc5f946-bmrsg" Nov 25 09:36:55 crc kubenswrapper[4776]: I1125 09:36:55.110211 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5cbcc5f946-bmrsg" Nov 25 09:36:55 crc kubenswrapper[4776]: I1125 09:36:55.330828 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5cbcc5f946-bmrsg"] Nov 25 09:36:55 crc kubenswrapper[4776]: I1125 09:36:55.391947 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5cbcc5f946-bmrsg" event={"ID":"6aaa21be-7e3a-4adf-8d71-6fd41b5798a7","Type":"ContainerStarted","Data":"7d8b4015259689ce5cf52d4eee1c50c4c48009a201e6ff4d6b2c92146eb79f67"} Nov 25 09:36:55 crc kubenswrapper[4776]: I1125 09:36:55.393557 4776 generic.go:334] "Generic (PLEG): container finished" podID="b4eebac0-e547-4f8e-944c-dff2a9c87ce8" containerID="55be4cef4c54d6cf213ee19ebdc5361a021c9a9ac1bfed5413a0d0161e581e49" exitCode=0 Nov 25 09:36:55 crc kubenswrapper[4776]: I1125 09:36:55.393603 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh" event={"ID":"b4eebac0-e547-4f8e-944c-dff2a9c87ce8","Type":"ContainerDied","Data":"55be4cef4c54d6cf213ee19ebdc5361a021c9a9ac1bfed5413a0d0161e581e49"} Nov 25 09:36:55 crc kubenswrapper[4776]: I1125 09:36:55.393622 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh" Nov 25 09:36:55 crc kubenswrapper[4776]: I1125 09:36:55.393635 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh" event={"ID":"b4eebac0-e547-4f8e-944c-dff2a9c87ce8","Type":"ContainerDied","Data":"4fc14e6c2ee9ce816aa2f748f46cbd214c5c92f53212226394260d258eb766c2"} Nov 25 09:36:55 crc kubenswrapper[4776]: I1125 09:36:55.393652 4776 scope.go:117] "RemoveContainer" containerID="55be4cef4c54d6cf213ee19ebdc5361a021c9a9ac1bfed5413a0d0161e581e49" Nov 25 09:36:55 crc kubenswrapper[4776]: I1125 09:36:55.416473 4776 scope.go:117] "RemoveContainer" containerID="55be4cef4c54d6cf213ee19ebdc5361a021c9a9ac1bfed5413a0d0161e581e49" Nov 25 09:36:55 crc kubenswrapper[4776]: E1125 09:36:55.417961 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55be4cef4c54d6cf213ee19ebdc5361a021c9a9ac1bfed5413a0d0161e581e49\": container with ID starting with 55be4cef4c54d6cf213ee19ebdc5361a021c9a9ac1bfed5413a0d0161e581e49 not found: ID does not exist" containerID="55be4cef4c54d6cf213ee19ebdc5361a021c9a9ac1bfed5413a0d0161e581e49" Nov 25 09:36:55 crc kubenswrapper[4776]: I1125 09:36:55.418306 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55be4cef4c54d6cf213ee19ebdc5361a021c9a9ac1bfed5413a0d0161e581e49"} err="failed to get container status \"55be4cef4c54d6cf213ee19ebdc5361a021c9a9ac1bfed5413a0d0161e581e49\": rpc error: code = NotFound desc = could not find container \"55be4cef4c54d6cf213ee19ebdc5361a021c9a9ac1bfed5413a0d0161e581e49\": container with ID starting with 55be4cef4c54d6cf213ee19ebdc5361a021c9a9ac1bfed5413a0d0161e581e49 not found: ID does not exist" Nov 25 09:36:55 crc kubenswrapper[4776]: I1125 09:36:55.419734 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh"] Nov 25 09:36:55 crc kubenswrapper[4776]: I1125 09:36:55.423635 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f897c9b7-fjgvh"] Nov 25 09:36:55 crc kubenswrapper[4776]: I1125 09:36:55.671900 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4eebac0-e547-4f8e-944c-dff2a9c87ce8" path="/var/lib/kubelet/pods/b4eebac0-e547-4f8e-944c-dff2a9c87ce8/volumes" Nov 25 09:36:56 crc kubenswrapper[4776]: I1125 09:36:56.403942 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5cbcc5f946-bmrsg" event={"ID":"6aaa21be-7e3a-4adf-8d71-6fd41b5798a7","Type":"ContainerStarted","Data":"771d2a70435e26c95c6f8005bc62c3a44a696d593b5a362857104de95eaf7230"} Nov 25 09:36:56 crc kubenswrapper[4776]: I1125 09:36:56.430505 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5cbcc5f946-bmrsg" podStartSLOduration=3.430486171 podStartE2EDuration="3.430486171s" podCreationTimestamp="2025-11-25 09:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:36:56.429316421 +0000 UTC m=+761.470376034" watchObservedRunningTime="2025-11-25 09:36:56.430486171 +0000 UTC m=+761.471545734" Nov 25 09:36:57 crc kubenswrapper[4776]: I1125 09:36:57.409905 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5cbcc5f946-bmrsg" Nov 25 09:36:57 crc kubenswrapper[4776]: I1125 09:36:57.421051 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5cbcc5f946-bmrsg" Nov 25 09:36:57 crc kubenswrapper[4776]: I1125 09:36:57.661522 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-xhbgd" Nov 25 09:36:57 crc kubenswrapper[4776]: I1125 09:36:57.661981 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-xhbgd" Nov 25 09:36:58 crc kubenswrapper[4776]: I1125 09:36:58.057168 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-xhbgd"] Nov 25 09:36:58 crc kubenswrapper[4776]: I1125 09:36:58.073321 4776 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 09:36:58 crc kubenswrapper[4776]: I1125 09:36:58.418322 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-xhbgd" event={"ID":"f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed","Type":"ContainerStarted","Data":"fb6a5bfec7e33c242ead54ce6e48237dd5ad79c0e30918d8013c68b2d3e57a15"} Nov 25 09:36:59 crc kubenswrapper[4776]: I1125 09:36:59.427533 4776 generic.go:334] "Generic (PLEG): container finished" podID="f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed" containerID="0bca6dcd4ef9e2845f88814b62bcb81d949aab1925dddb1c40352bbd094805bc" exitCode=0 Nov 25 09:36:59 crc kubenswrapper[4776]: I1125 09:36:59.427776 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-xhbgd" event={"ID":"f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed","Type":"ContainerDied","Data":"0bca6dcd4ef9e2845f88814b62bcb81d949aab1925dddb1c40352bbd094805bc"} Nov 25 09:37:00 crc kubenswrapper[4776]: I1125 09:37:00.040680 4776 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 09:37:00 crc kubenswrapper[4776]: I1125 09:37:00.816823 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-xhbgd" Nov 25 09:37:00 crc kubenswrapper[4776]: I1125 09:37:00.973377 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed-crc-storage\") pod \"f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed\" (UID: \"f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed\") " Nov 25 09:37:00 crc kubenswrapper[4776]: I1125 09:37:00.973508 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtbkb\" (UniqueName: \"kubernetes.io/projected/f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed-kube-api-access-qtbkb\") pod \"f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed\" (UID: \"f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed\") " Nov 25 09:37:00 crc kubenswrapper[4776]: I1125 09:37:00.973568 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed-node-mnt\") pod \"f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed\" (UID: \"f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed\") " Nov 25 09:37:00 crc kubenswrapper[4776]: I1125 09:37:00.973880 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed" (UID: "f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:37:00 crc kubenswrapper[4776]: I1125 09:37:00.974430 4776 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed-node-mnt\") on node \"crc\" DevicePath \"\"" Nov 25 09:37:00 crc kubenswrapper[4776]: I1125 09:37:00.979196 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed-kube-api-access-qtbkb" (OuterVolumeSpecName: "kube-api-access-qtbkb") pod "f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed" (UID: "f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed"). InnerVolumeSpecName "kube-api-access-qtbkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:37:00 crc kubenswrapper[4776]: I1125 09:37:00.988104 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed" (UID: "f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:37:01 crc kubenswrapper[4776]: I1125 09:37:01.075326 4776 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed-crc-storage\") on node \"crc\" DevicePath \"\"" Nov 25 09:37:01 crc kubenswrapper[4776]: I1125 09:37:01.075365 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtbkb\" (UniqueName: \"kubernetes.io/projected/f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed-kube-api-access-qtbkb\") on node \"crc\" DevicePath \"\"" Nov 25 09:37:01 crc kubenswrapper[4776]: I1125 09:37:01.439255 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-xhbgd" event={"ID":"f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed","Type":"ContainerDied","Data":"fb6a5bfec7e33c242ead54ce6e48237dd5ad79c0e30918d8013c68b2d3e57a15"} Nov 25 09:37:01 crc kubenswrapper[4776]: I1125 09:37:01.439293 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb6a5bfec7e33c242ead54ce6e48237dd5ad79c0e30918d8013c68b2d3e57a15" Nov 25 09:37:01 crc kubenswrapper[4776]: I1125 09:37:01.439300 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-xhbgd" Nov 25 09:37:04 crc kubenswrapper[4776]: I1125 09:37:04.801945 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4cx97" Nov 25 09:37:08 crc kubenswrapper[4776]: I1125 09:37:08.505749 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m"] Nov 25 09:37:08 crc kubenswrapper[4776]: E1125 09:37:08.506255 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed" containerName="storage" Nov 25 09:37:08 crc kubenswrapper[4776]: I1125 09:37:08.506269 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed" containerName="storage" Nov 25 09:37:08 crc kubenswrapper[4776]: I1125 09:37:08.506354 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed" containerName="storage" Nov 25 09:37:08 crc kubenswrapper[4776]: I1125 09:37:08.507040 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m" Nov 25 09:37:08 crc kubenswrapper[4776]: I1125 09:37:08.508683 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 25 09:37:08 crc kubenswrapper[4776]: I1125 09:37:08.519134 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m"] Nov 25 09:37:08 crc kubenswrapper[4776]: I1125 09:37:08.566393 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl98s\" (UniqueName: \"kubernetes.io/projected/b55a5ed7-b167-447e-b9ae-16ea2c904b39-kube-api-access-jl98s\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m\" (UID: \"b55a5ed7-b167-447e-b9ae-16ea2c904b39\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m" Nov 25 09:37:08 crc kubenswrapper[4776]: I1125 09:37:08.566468 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b55a5ed7-b167-447e-b9ae-16ea2c904b39-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m\" (UID: \"b55a5ed7-b167-447e-b9ae-16ea2c904b39\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m" Nov 25 09:37:08 crc kubenswrapper[4776]: I1125 09:37:08.566497 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b55a5ed7-b167-447e-b9ae-16ea2c904b39-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m\" (UID: \"b55a5ed7-b167-447e-b9ae-16ea2c904b39\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m" Nov 25 09:37:08 crc kubenswrapper[4776]: I1125 09:37:08.667263 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b55a5ed7-b167-447e-b9ae-16ea2c904b39-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m\" (UID: \"b55a5ed7-b167-447e-b9ae-16ea2c904b39\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m" Nov 25 09:37:08 crc kubenswrapper[4776]: I1125 09:37:08.667327 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b55a5ed7-b167-447e-b9ae-16ea2c904b39-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m\" (UID: \"b55a5ed7-b167-447e-b9ae-16ea2c904b39\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m" Nov 25 09:37:08 crc kubenswrapper[4776]: I1125 09:37:08.667422 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl98s\" (UniqueName: \"kubernetes.io/projected/b55a5ed7-b167-447e-b9ae-16ea2c904b39-kube-api-access-jl98s\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m\" (UID: \"b55a5ed7-b167-447e-b9ae-16ea2c904b39\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m" Nov 25 09:37:08 crc kubenswrapper[4776]: I1125 09:37:08.668146 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b55a5ed7-b167-447e-b9ae-16ea2c904b39-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m\" (UID: \"b55a5ed7-b167-447e-b9ae-16ea2c904b39\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m" Nov 25 09:37:08 crc kubenswrapper[4776]: I1125 09:37:08.668371 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b55a5ed7-b167-447e-b9ae-16ea2c904b39-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m\" (UID: \"b55a5ed7-b167-447e-b9ae-16ea2c904b39\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m" Nov 25 09:37:08 crc kubenswrapper[4776]: I1125 09:37:08.684912 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl98s\" (UniqueName: \"kubernetes.io/projected/b55a5ed7-b167-447e-b9ae-16ea2c904b39-kube-api-access-jl98s\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m\" (UID: \"b55a5ed7-b167-447e-b9ae-16ea2c904b39\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m" Nov 25 09:37:08 crc kubenswrapper[4776]: I1125 09:37:08.827870 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m" Nov 25 09:37:09 crc kubenswrapper[4776]: I1125 09:37:09.297800 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m"] Nov 25 09:37:09 crc kubenswrapper[4776]: W1125 09:37:09.304770 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb55a5ed7_b167_447e_b9ae_16ea2c904b39.slice/crio-2cb475320a08bed64238db7b6b6ec980c6da3cef6d72bd3edf1144ee179761ee WatchSource:0}: Error finding container 2cb475320a08bed64238db7b6b6ec980c6da3cef6d72bd3edf1144ee179761ee: Status 404 returned error can't find the container with id 2cb475320a08bed64238db7b6b6ec980c6da3cef6d72bd3edf1144ee179761ee Nov 25 09:37:09 crc kubenswrapper[4776]: I1125 09:37:09.484242 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m" event={"ID":"b55a5ed7-b167-447e-b9ae-16ea2c904b39","Type":"ContainerStarted","Data":"e9529303348c0080bd68c80a19cc7dd58af900cdf82d4dffb866a1959da95da2"} Nov 25 09:37:09 crc kubenswrapper[4776]: I1125 09:37:09.484578 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m" event={"ID":"b55a5ed7-b167-447e-b9ae-16ea2c904b39","Type":"ContainerStarted","Data":"2cb475320a08bed64238db7b6b6ec980c6da3cef6d72bd3edf1144ee179761ee"} Nov 25 09:37:10 crc kubenswrapper[4776]: I1125 09:37:10.490985 4776 generic.go:334] "Generic (PLEG): container finished" podID="b55a5ed7-b167-447e-b9ae-16ea2c904b39" containerID="e9529303348c0080bd68c80a19cc7dd58af900cdf82d4dffb866a1959da95da2" exitCode=0 Nov 25 09:37:10 crc kubenswrapper[4776]: I1125 09:37:10.491040 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m" event={"ID":"b55a5ed7-b167-447e-b9ae-16ea2c904b39","Type":"ContainerDied","Data":"e9529303348c0080bd68c80a19cc7dd58af900cdf82d4dffb866a1959da95da2"} Nov 25 09:37:10 crc kubenswrapper[4776]: I1125 09:37:10.861312 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x8w52"] Nov 25 09:37:10 crc kubenswrapper[4776]: I1125 09:37:10.862743 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x8w52" Nov 25 09:37:10 crc kubenswrapper[4776]: I1125 09:37:10.876825 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x8w52"] Nov 25 09:37:10 crc kubenswrapper[4776]: I1125 09:37:10.897893 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bht72\" (UniqueName: \"kubernetes.io/projected/6a8c7653-405c-477b-8fc1-032ddab19212-kube-api-access-bht72\") pod \"redhat-operators-x8w52\" (UID: \"6a8c7653-405c-477b-8fc1-032ddab19212\") " pod="openshift-marketplace/redhat-operators-x8w52" Nov 25 09:37:10 crc kubenswrapper[4776]: I1125 09:37:10.898218 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8c7653-405c-477b-8fc1-032ddab19212-utilities\") pod \"redhat-operators-x8w52\" (UID: \"6a8c7653-405c-477b-8fc1-032ddab19212\") " pod="openshift-marketplace/redhat-operators-x8w52" Nov 25 09:37:10 crc kubenswrapper[4776]: I1125 09:37:10.898261 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8c7653-405c-477b-8fc1-032ddab19212-catalog-content\") pod \"redhat-operators-x8w52\" (UID: \"6a8c7653-405c-477b-8fc1-032ddab19212\") " pod="openshift-marketplace/redhat-operators-x8w52" Nov 25 09:37:10 crc kubenswrapper[4776]: I1125 09:37:10.999180 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bht72\" (UniqueName: \"kubernetes.io/projected/6a8c7653-405c-477b-8fc1-032ddab19212-kube-api-access-bht72\") pod \"redhat-operators-x8w52\" (UID: \"6a8c7653-405c-477b-8fc1-032ddab19212\") " pod="openshift-marketplace/redhat-operators-x8w52" Nov 25 09:37:10 crc kubenswrapper[4776]: I1125 09:37:10.999232 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8c7653-405c-477b-8fc1-032ddab19212-utilities\") pod \"redhat-operators-x8w52\" (UID: \"6a8c7653-405c-477b-8fc1-032ddab19212\") " pod="openshift-marketplace/redhat-operators-x8w52" Nov 25 09:37:10 crc kubenswrapper[4776]: I1125 09:37:10.999276 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8c7653-405c-477b-8fc1-032ddab19212-catalog-content\") pod \"redhat-operators-x8w52\" (UID: \"6a8c7653-405c-477b-8fc1-032ddab19212\") " pod="openshift-marketplace/redhat-operators-x8w52" Nov 25 09:37:10 crc kubenswrapper[4776]: I1125 09:37:10.999679 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8c7653-405c-477b-8fc1-032ddab19212-catalog-content\") pod \"redhat-operators-x8w52\" (UID: \"6a8c7653-405c-477b-8fc1-032ddab19212\") " pod="openshift-marketplace/redhat-operators-x8w52" Nov 25 09:37:11 crc kubenswrapper[4776]: I1125 09:37:11.000028 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8c7653-405c-477b-8fc1-032ddab19212-utilities\") pod \"redhat-operators-x8w52\" (UID: \"6a8c7653-405c-477b-8fc1-032ddab19212\") " pod="openshift-marketplace/redhat-operators-x8w52" Nov 25 09:37:11 crc kubenswrapper[4776]: I1125 09:37:11.017014 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bht72\" (UniqueName: \"kubernetes.io/projected/6a8c7653-405c-477b-8fc1-032ddab19212-kube-api-access-bht72\") pod \"redhat-operators-x8w52\" (UID: \"6a8c7653-405c-477b-8fc1-032ddab19212\") " pod="openshift-marketplace/redhat-operators-x8w52" Nov 25 09:37:11 crc kubenswrapper[4776]: I1125 09:37:11.180949 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x8w52" Nov 25 09:37:11 crc kubenswrapper[4776]: I1125 09:37:11.575279 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x8w52"] Nov 25 09:37:12 crc kubenswrapper[4776]: I1125 09:37:12.503045 4776 generic.go:334] "Generic (PLEG): container finished" podID="b55a5ed7-b167-447e-b9ae-16ea2c904b39" containerID="9f18836aae0287ec6fcc37aeb564130426c815caa5ae321a9993e4790d9f781b" exitCode=0 Nov 25 09:37:12 crc kubenswrapper[4776]: I1125 09:37:12.503138 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m" event={"ID":"b55a5ed7-b167-447e-b9ae-16ea2c904b39","Type":"ContainerDied","Data":"9f18836aae0287ec6fcc37aeb564130426c815caa5ae321a9993e4790d9f781b"} Nov 25 09:37:12 crc kubenswrapper[4776]: I1125 09:37:12.505341 4776 generic.go:334] "Generic (PLEG): container finished" podID="6a8c7653-405c-477b-8fc1-032ddab19212" containerID="7e796145b2ccb43b7215f53072fb36d7b6267f182b9c06cc98e1cfbcae8465b0" exitCode=0 Nov 25 09:37:12 crc kubenswrapper[4776]: I1125 09:37:12.505403 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8w52" event={"ID":"6a8c7653-405c-477b-8fc1-032ddab19212","Type":"ContainerDied","Data":"7e796145b2ccb43b7215f53072fb36d7b6267f182b9c06cc98e1cfbcae8465b0"} Nov 25 09:37:12 crc kubenswrapper[4776]: I1125 09:37:12.505443 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8w52" event={"ID":"6a8c7653-405c-477b-8fc1-032ddab19212","Type":"ContainerStarted","Data":"0b544861fc85969b08d8576d710a826e23d463a1a13da66e06d018a9f8c00d60"} Nov 25 09:37:13 crc kubenswrapper[4776]: I1125 09:37:13.512754 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8w52" event={"ID":"6a8c7653-405c-477b-8fc1-032ddab19212","Type":"ContainerStarted","Data":"b66fce6a8e0ebdc7205e9f6ea2ed6745333ee9ed0c5c0d5a938ad2862320e144"} Nov 25 09:37:13 crc kubenswrapper[4776]: I1125 09:37:13.515418 4776 generic.go:334] "Generic (PLEG): container finished" podID="b55a5ed7-b167-447e-b9ae-16ea2c904b39" containerID="e0b16b2438a2a13fd6eb08b96c21989b1b9ae37c8d614429d724f838673ff5f9" exitCode=0 Nov 25 09:37:13 crc kubenswrapper[4776]: I1125 09:37:13.515453 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m" event={"ID":"b55a5ed7-b167-447e-b9ae-16ea2c904b39","Type":"ContainerDied","Data":"e0b16b2438a2a13fd6eb08b96c21989b1b9ae37c8d614429d724f838673ff5f9"} Nov 25 09:37:14 crc kubenswrapper[4776]: I1125 09:37:14.526507 4776 generic.go:334] "Generic (PLEG): container finished" podID="6a8c7653-405c-477b-8fc1-032ddab19212" containerID="b66fce6a8e0ebdc7205e9f6ea2ed6745333ee9ed0c5c0d5a938ad2862320e144" exitCode=0 Nov 25 09:37:14 crc kubenswrapper[4776]: I1125 09:37:14.526584 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8w52" event={"ID":"6a8c7653-405c-477b-8fc1-032ddab19212","Type":"ContainerDied","Data":"b66fce6a8e0ebdc7205e9f6ea2ed6745333ee9ed0c5c0d5a938ad2862320e144"} Nov 25 09:37:14 crc kubenswrapper[4776]: I1125 09:37:14.897391 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m" Nov 25 09:37:14 crc kubenswrapper[4776]: I1125 09:37:14.951544 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b55a5ed7-b167-447e-b9ae-16ea2c904b39-bundle\") pod \"b55a5ed7-b167-447e-b9ae-16ea2c904b39\" (UID: \"b55a5ed7-b167-447e-b9ae-16ea2c904b39\") " Nov 25 09:37:14 crc kubenswrapper[4776]: I1125 09:37:14.951665 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b55a5ed7-b167-447e-b9ae-16ea2c904b39-util\") pod \"b55a5ed7-b167-447e-b9ae-16ea2c904b39\" (UID: \"b55a5ed7-b167-447e-b9ae-16ea2c904b39\") " Nov 25 09:37:14 crc kubenswrapper[4776]: I1125 09:37:14.951702 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jl98s\" (UniqueName: \"kubernetes.io/projected/b55a5ed7-b167-447e-b9ae-16ea2c904b39-kube-api-access-jl98s\") pod \"b55a5ed7-b167-447e-b9ae-16ea2c904b39\" (UID: \"b55a5ed7-b167-447e-b9ae-16ea2c904b39\") " Nov 25 09:37:14 crc kubenswrapper[4776]: I1125 09:37:14.952700 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b55a5ed7-b167-447e-b9ae-16ea2c904b39-bundle" (OuterVolumeSpecName: "bundle") pod "b55a5ed7-b167-447e-b9ae-16ea2c904b39" (UID: "b55a5ed7-b167-447e-b9ae-16ea2c904b39"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:37:14 crc kubenswrapper[4776]: I1125 09:37:14.956589 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b55a5ed7-b167-447e-b9ae-16ea2c904b39-kube-api-access-jl98s" (OuterVolumeSpecName: "kube-api-access-jl98s") pod "b55a5ed7-b167-447e-b9ae-16ea2c904b39" (UID: "b55a5ed7-b167-447e-b9ae-16ea2c904b39"). InnerVolumeSpecName "kube-api-access-jl98s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:37:14 crc kubenswrapper[4776]: I1125 09:37:14.983428 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b55a5ed7-b167-447e-b9ae-16ea2c904b39-util" (OuterVolumeSpecName: "util") pod "b55a5ed7-b167-447e-b9ae-16ea2c904b39" (UID: "b55a5ed7-b167-447e-b9ae-16ea2c904b39"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:37:15 crc kubenswrapper[4776]: I1125 09:37:15.052904 4776 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b55a5ed7-b167-447e-b9ae-16ea2c904b39-util\") on node \"crc\" DevicePath \"\"" Nov 25 09:37:15 crc kubenswrapper[4776]: I1125 09:37:15.052954 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jl98s\" (UniqueName: \"kubernetes.io/projected/b55a5ed7-b167-447e-b9ae-16ea2c904b39-kube-api-access-jl98s\") on node \"crc\" DevicePath \"\"" Nov 25 09:37:15 crc kubenswrapper[4776]: I1125 09:37:15.052968 4776 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b55a5ed7-b167-447e-b9ae-16ea2c904b39-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:37:15 crc kubenswrapper[4776]: I1125 09:37:15.534181 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8w52" event={"ID":"6a8c7653-405c-477b-8fc1-032ddab19212","Type":"ContainerStarted","Data":"33c0c01b52e59cde188bd738eec4226c5f2f22c620682862403cd0f511405855"} Nov 25 09:37:15 crc kubenswrapper[4776]: I1125 09:37:15.536551 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m" event={"ID":"b55a5ed7-b167-447e-b9ae-16ea2c904b39","Type":"ContainerDied","Data":"2cb475320a08bed64238db7b6b6ec980c6da3cef6d72bd3edf1144ee179761ee"} Nov 25 09:37:15 crc kubenswrapper[4776]: I1125 09:37:15.536584 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2cb475320a08bed64238db7b6b6ec980c6da3cef6d72bd3edf1144ee179761ee" Nov 25 09:37:15 crc kubenswrapper[4776]: I1125 09:37:15.536608 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m" Nov 25 09:37:15 crc kubenswrapper[4776]: I1125 09:37:15.935855 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x8w52" podStartSLOduration=3.363916507 podStartE2EDuration="5.93583533s" podCreationTimestamp="2025-11-25 09:37:10 +0000 UTC" firstStartedPulling="2025-11-25 09:37:12.506309886 +0000 UTC m=+777.547369479" lastFinishedPulling="2025-11-25 09:37:15.078228749 +0000 UTC m=+780.119288302" observedRunningTime="2025-11-25 09:37:15.560680565 +0000 UTC m=+780.601740158" watchObservedRunningTime="2025-11-25 09:37:15.93583533 +0000 UTC m=+780.976894893" Nov 25 09:37:17 crc kubenswrapper[4776]: I1125 09:37:17.818137 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:37:17 crc kubenswrapper[4776]: I1125 09:37:17.818224 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:37:18 crc kubenswrapper[4776]: I1125 09:37:18.742357 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-klsvw"] Nov 25 09:37:18 crc kubenswrapper[4776]: E1125 09:37:18.742865 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b55a5ed7-b167-447e-b9ae-16ea2c904b39" containerName="extract" Nov 25 09:37:18 crc kubenswrapper[4776]: I1125 09:37:18.742885 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="b55a5ed7-b167-447e-b9ae-16ea2c904b39" containerName="extract" Nov 25 09:37:18 crc kubenswrapper[4776]: E1125 09:37:18.742902 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b55a5ed7-b167-447e-b9ae-16ea2c904b39" containerName="pull" Nov 25 09:37:18 crc kubenswrapper[4776]: I1125 09:37:18.742909 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="b55a5ed7-b167-447e-b9ae-16ea2c904b39" containerName="pull" Nov 25 09:37:18 crc kubenswrapper[4776]: E1125 09:37:18.742918 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b55a5ed7-b167-447e-b9ae-16ea2c904b39" containerName="util" Nov 25 09:37:18 crc kubenswrapper[4776]: I1125 09:37:18.742925 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="b55a5ed7-b167-447e-b9ae-16ea2c904b39" containerName="util" Nov 25 09:37:18 crc kubenswrapper[4776]: I1125 09:37:18.743034 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="b55a5ed7-b167-447e-b9ae-16ea2c904b39" containerName="extract" Nov 25 09:37:18 crc kubenswrapper[4776]: I1125 09:37:18.743423 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-klsvw" Nov 25 09:37:18 crc kubenswrapper[4776]: I1125 09:37:18.746086 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 25 09:37:18 crc kubenswrapper[4776]: I1125 09:37:18.746181 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 25 09:37:18 crc kubenswrapper[4776]: I1125 09:37:18.747930 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-7jlns" Nov 25 09:37:18 crc kubenswrapper[4776]: I1125 09:37:18.756846 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-klsvw"] Nov 25 09:37:18 crc kubenswrapper[4776]: I1125 09:37:18.896791 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtrvt\" (UniqueName: \"kubernetes.io/projected/87dd97b9-3af4-41c6-8c24-db941e082eb6-kube-api-access-jtrvt\") pod \"nmstate-operator-557fdffb88-klsvw\" (UID: \"87dd97b9-3af4-41c6-8c24-db941e082eb6\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-klsvw" Nov 25 09:37:18 crc kubenswrapper[4776]: I1125 09:37:18.998320 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtrvt\" (UniqueName: \"kubernetes.io/projected/87dd97b9-3af4-41c6-8c24-db941e082eb6-kube-api-access-jtrvt\") pod \"nmstate-operator-557fdffb88-klsvw\" (UID: \"87dd97b9-3af4-41c6-8c24-db941e082eb6\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-klsvw" Nov 25 09:37:19 crc kubenswrapper[4776]: I1125 09:37:19.018590 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtrvt\" (UniqueName: \"kubernetes.io/projected/87dd97b9-3af4-41c6-8c24-db941e082eb6-kube-api-access-jtrvt\") pod \"nmstate-operator-557fdffb88-klsvw\" (UID: \"87dd97b9-3af4-41c6-8c24-db941e082eb6\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-klsvw" Nov 25 09:37:19 crc kubenswrapper[4776]: I1125 09:37:19.063015 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-klsvw" Nov 25 09:37:19 crc kubenswrapper[4776]: I1125 09:37:19.497521 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-klsvw"] Nov 25 09:37:19 crc kubenswrapper[4776]: W1125 09:37:19.503217 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87dd97b9_3af4_41c6_8c24_db941e082eb6.slice/crio-e631593c5e3d493b9fa6233fbb652ec50b62dbeaecc99179e414b0cc0431dc8a WatchSource:0}: Error finding container e631593c5e3d493b9fa6233fbb652ec50b62dbeaecc99179e414b0cc0431dc8a: Status 404 returned error can't find the container with id e631593c5e3d493b9fa6233fbb652ec50b62dbeaecc99179e414b0cc0431dc8a Nov 25 09:37:19 crc kubenswrapper[4776]: I1125 09:37:19.559906 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-klsvw" event={"ID":"87dd97b9-3af4-41c6-8c24-db941e082eb6","Type":"ContainerStarted","Data":"e631593c5e3d493b9fa6233fbb652ec50b62dbeaecc99179e414b0cc0431dc8a"} Nov 25 09:37:21 crc kubenswrapper[4776]: I1125 09:37:21.181843 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x8w52" Nov 25 09:37:21 crc kubenswrapper[4776]: I1125 09:37:21.182207 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x8w52" Nov 25 09:37:21 crc kubenswrapper[4776]: I1125 09:37:21.240580 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x8w52" Nov 25 09:37:21 crc kubenswrapper[4776]: I1125 09:37:21.613553 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x8w52" Nov 25 09:37:22 crc kubenswrapper[4776]: I1125 09:37:22.575499 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-klsvw" event={"ID":"87dd97b9-3af4-41c6-8c24-db941e082eb6","Type":"ContainerStarted","Data":"25f5fe471279cbd1a47eb7eca0dd918d8723212e78c94c30e095d7137a2fad0b"} Nov 25 09:37:22 crc kubenswrapper[4776]: I1125 09:37:22.599838 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-557fdffb88-klsvw" podStartSLOduration=2.194220498 podStartE2EDuration="4.599811153s" podCreationTimestamp="2025-11-25 09:37:18 +0000 UTC" firstStartedPulling="2025-11-25 09:37:19.505681991 +0000 UTC m=+784.546741544" lastFinishedPulling="2025-11-25 09:37:21.911272646 +0000 UTC m=+786.952332199" observedRunningTime="2025-11-25 09:37:22.593375482 +0000 UTC m=+787.634435035" watchObservedRunningTime="2025-11-25 09:37:22.599811153 +0000 UTC m=+787.640870736" Nov 25 09:37:23 crc kubenswrapper[4776]: I1125 09:37:23.645472 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x8w52"] Nov 25 09:37:23 crc kubenswrapper[4776]: I1125 09:37:23.645824 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x8w52" podUID="6a8c7653-405c-477b-8fc1-032ddab19212" containerName="registry-server" containerID="cri-o://33c0c01b52e59cde188bd738eec4226c5f2f22c620682862403cd0f511405855" gracePeriod=2 Nov 25 09:37:24 crc kubenswrapper[4776]: I1125 09:37:24.081007 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x8w52" Nov 25 09:37:24 crc kubenswrapper[4776]: I1125 09:37:24.265687 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bht72\" (UniqueName: \"kubernetes.io/projected/6a8c7653-405c-477b-8fc1-032ddab19212-kube-api-access-bht72\") pod \"6a8c7653-405c-477b-8fc1-032ddab19212\" (UID: \"6a8c7653-405c-477b-8fc1-032ddab19212\") " Nov 25 09:37:24 crc kubenswrapper[4776]: I1125 09:37:24.265761 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8c7653-405c-477b-8fc1-032ddab19212-catalog-content\") pod \"6a8c7653-405c-477b-8fc1-032ddab19212\" (UID: \"6a8c7653-405c-477b-8fc1-032ddab19212\") " Nov 25 09:37:24 crc kubenswrapper[4776]: I1125 09:37:24.265811 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8c7653-405c-477b-8fc1-032ddab19212-utilities\") pod \"6a8c7653-405c-477b-8fc1-032ddab19212\" (UID: \"6a8c7653-405c-477b-8fc1-032ddab19212\") " Nov 25 09:37:24 crc kubenswrapper[4776]: I1125 09:37:24.266776 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a8c7653-405c-477b-8fc1-032ddab19212-utilities" (OuterVolumeSpecName: "utilities") pod "6a8c7653-405c-477b-8fc1-032ddab19212" (UID: "6a8c7653-405c-477b-8fc1-032ddab19212"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:37:24 crc kubenswrapper[4776]: I1125 09:37:24.272101 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a8c7653-405c-477b-8fc1-032ddab19212-kube-api-access-bht72" (OuterVolumeSpecName: "kube-api-access-bht72") pod "6a8c7653-405c-477b-8fc1-032ddab19212" (UID: "6a8c7653-405c-477b-8fc1-032ddab19212"). InnerVolumeSpecName "kube-api-access-bht72". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:37:24 crc kubenswrapper[4776]: I1125 09:37:24.367202 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a8c7653-405c-477b-8fc1-032ddab19212-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:37:24 crc kubenswrapper[4776]: I1125 09:37:24.367257 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bht72\" (UniqueName: \"kubernetes.io/projected/6a8c7653-405c-477b-8fc1-032ddab19212-kube-api-access-bht72\") on node \"crc\" DevicePath \"\"" Nov 25 09:37:24 crc kubenswrapper[4776]: I1125 09:37:24.561251 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a8c7653-405c-477b-8fc1-032ddab19212-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a8c7653-405c-477b-8fc1-032ddab19212" (UID: "6a8c7653-405c-477b-8fc1-032ddab19212"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:37:24 crc kubenswrapper[4776]: I1125 09:37:24.569406 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a8c7653-405c-477b-8fc1-032ddab19212-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:37:24 crc kubenswrapper[4776]: I1125 09:37:24.589036 4776 generic.go:334] "Generic (PLEG): container finished" podID="6a8c7653-405c-477b-8fc1-032ddab19212" containerID="33c0c01b52e59cde188bd738eec4226c5f2f22c620682862403cd0f511405855" exitCode=0 Nov 25 09:37:24 crc kubenswrapper[4776]: I1125 09:37:24.589105 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8w52" event={"ID":"6a8c7653-405c-477b-8fc1-032ddab19212","Type":"ContainerDied","Data":"33c0c01b52e59cde188bd738eec4226c5f2f22c620682862403cd0f511405855"} Nov 25 09:37:24 crc kubenswrapper[4776]: I1125 09:37:24.589145 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8w52" event={"ID":"6a8c7653-405c-477b-8fc1-032ddab19212","Type":"ContainerDied","Data":"0b544861fc85969b08d8576d710a826e23d463a1a13da66e06d018a9f8c00d60"} Nov 25 09:37:24 crc kubenswrapper[4776]: I1125 09:37:24.589153 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x8w52" Nov 25 09:37:24 crc kubenswrapper[4776]: I1125 09:37:24.589165 4776 scope.go:117] "RemoveContainer" containerID="33c0c01b52e59cde188bd738eec4226c5f2f22c620682862403cd0f511405855" Nov 25 09:37:24 crc kubenswrapper[4776]: I1125 09:37:24.608336 4776 scope.go:117] "RemoveContainer" containerID="b66fce6a8e0ebdc7205e9f6ea2ed6745333ee9ed0c5c0d5a938ad2862320e144" Nov 25 09:37:24 crc kubenswrapper[4776]: I1125 09:37:24.629151 4776 scope.go:117] "RemoveContainer" containerID="7e796145b2ccb43b7215f53072fb36d7b6267f182b9c06cc98e1cfbcae8465b0" Nov 25 09:37:24 crc kubenswrapper[4776]: I1125 09:37:24.631199 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x8w52"] Nov 25 09:37:24 crc kubenswrapper[4776]: I1125 09:37:24.633674 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x8w52"] Nov 25 09:37:24 crc kubenswrapper[4776]: I1125 09:37:24.656487 4776 scope.go:117] "RemoveContainer" containerID="33c0c01b52e59cde188bd738eec4226c5f2f22c620682862403cd0f511405855" Nov 25 09:37:24 crc kubenswrapper[4776]: E1125 09:37:24.657128 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33c0c01b52e59cde188bd738eec4226c5f2f22c620682862403cd0f511405855\": container with ID starting with 33c0c01b52e59cde188bd738eec4226c5f2f22c620682862403cd0f511405855 not found: ID does not exist" containerID="33c0c01b52e59cde188bd738eec4226c5f2f22c620682862403cd0f511405855" Nov 25 09:37:24 crc kubenswrapper[4776]: I1125 09:37:24.657182 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33c0c01b52e59cde188bd738eec4226c5f2f22c620682862403cd0f511405855"} err="failed to get container status \"33c0c01b52e59cde188bd738eec4226c5f2f22c620682862403cd0f511405855\": rpc error: code = NotFound desc = could not find container \"33c0c01b52e59cde188bd738eec4226c5f2f22c620682862403cd0f511405855\": container with ID starting with 33c0c01b52e59cde188bd738eec4226c5f2f22c620682862403cd0f511405855 not found: ID does not exist" Nov 25 09:37:24 crc kubenswrapper[4776]: I1125 09:37:24.657220 4776 scope.go:117] "RemoveContainer" containerID="b66fce6a8e0ebdc7205e9f6ea2ed6745333ee9ed0c5c0d5a938ad2862320e144" Nov 25 09:37:24 crc kubenswrapper[4776]: E1125 09:37:24.657741 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b66fce6a8e0ebdc7205e9f6ea2ed6745333ee9ed0c5c0d5a938ad2862320e144\": container with ID starting with b66fce6a8e0ebdc7205e9f6ea2ed6745333ee9ed0c5c0d5a938ad2862320e144 not found: ID does not exist" containerID="b66fce6a8e0ebdc7205e9f6ea2ed6745333ee9ed0c5c0d5a938ad2862320e144" Nov 25 09:37:24 crc kubenswrapper[4776]: I1125 09:37:24.657802 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b66fce6a8e0ebdc7205e9f6ea2ed6745333ee9ed0c5c0d5a938ad2862320e144"} err="failed to get container status \"b66fce6a8e0ebdc7205e9f6ea2ed6745333ee9ed0c5c0d5a938ad2862320e144\": rpc error: code = NotFound desc = could not find container \"b66fce6a8e0ebdc7205e9f6ea2ed6745333ee9ed0c5c0d5a938ad2862320e144\": container with ID starting with b66fce6a8e0ebdc7205e9f6ea2ed6745333ee9ed0c5c0d5a938ad2862320e144 not found: ID does not exist" Nov 25 09:37:24 crc kubenswrapper[4776]: I1125 09:37:24.657832 4776 scope.go:117] "RemoveContainer" containerID="7e796145b2ccb43b7215f53072fb36d7b6267f182b9c06cc98e1cfbcae8465b0" Nov 25 09:37:24 crc kubenswrapper[4776]: E1125 09:37:24.658442 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e796145b2ccb43b7215f53072fb36d7b6267f182b9c06cc98e1cfbcae8465b0\": container with ID starting with 7e796145b2ccb43b7215f53072fb36d7b6267f182b9c06cc98e1cfbcae8465b0 not found: ID does not exist" containerID="7e796145b2ccb43b7215f53072fb36d7b6267f182b9c06cc98e1cfbcae8465b0" Nov 25 09:37:24 crc kubenswrapper[4776]: I1125 09:37:24.658468 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e796145b2ccb43b7215f53072fb36d7b6267f182b9c06cc98e1cfbcae8465b0"} err="failed to get container status \"7e796145b2ccb43b7215f53072fb36d7b6267f182b9c06cc98e1cfbcae8465b0\": rpc error: code = NotFound desc = could not find container \"7e796145b2ccb43b7215f53072fb36d7b6267f182b9c06cc98e1cfbcae8465b0\": container with ID starting with 7e796145b2ccb43b7215f53072fb36d7b6267f182b9c06cc98e1cfbcae8465b0 not found: ID does not exist" Nov 25 09:37:25 crc kubenswrapper[4776]: I1125 09:37:25.667410 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a8c7653-405c-477b-8fc1-032ddab19212" path="/var/lib/kubelet/pods/6a8c7653-405c-477b-8fc1-032ddab19212/volumes" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.590489 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-8bncp"] Nov 25 09:37:28 crc kubenswrapper[4776]: E1125 09:37:28.590944 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a8c7653-405c-477b-8fc1-032ddab19212" containerName="extract-utilities" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.590956 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a8c7653-405c-477b-8fc1-032ddab19212" containerName="extract-utilities" Nov 25 09:37:28 crc kubenswrapper[4776]: E1125 09:37:28.590965 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a8c7653-405c-477b-8fc1-032ddab19212" containerName="extract-content" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.590971 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a8c7653-405c-477b-8fc1-032ddab19212" containerName="extract-content" Nov 25 09:37:28 crc kubenswrapper[4776]: E1125 09:37:28.590980 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a8c7653-405c-477b-8fc1-032ddab19212" containerName="registry-server" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.590986 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a8c7653-405c-477b-8fc1-032ddab19212" containerName="registry-server" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.591107 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a8c7653-405c-477b-8fc1-032ddab19212" containerName="registry-server" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.591647 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-8bncp" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.595154 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-r2hzs" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.604285 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-8bncp"] Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.614671 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-fv9bw"] Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.615331 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-fv9bw" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.621225 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/a4451cf3-5459-4208-bb86-49d3d60c180c-ovs-socket\") pod \"nmstate-handler-fv9bw\" (UID: \"a4451cf3-5459-4208-bb86-49d3d60c180c\") " pod="openshift-nmstate/nmstate-handler-fv9bw" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.621265 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/a4451cf3-5459-4208-bb86-49d3d60c180c-nmstate-lock\") pod \"nmstate-handler-fv9bw\" (UID: \"a4451cf3-5459-4208-bb86-49d3d60c180c\") " pod="openshift-nmstate/nmstate-handler-fv9bw" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.621315 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nm9gj\" (UniqueName: \"kubernetes.io/projected/a4451cf3-5459-4208-bb86-49d3d60c180c-kube-api-access-nm9gj\") pod \"nmstate-handler-fv9bw\" (UID: \"a4451cf3-5459-4208-bb86-49d3d60c180c\") " pod="openshift-nmstate/nmstate-handler-fv9bw" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.621345 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/a4451cf3-5459-4208-bb86-49d3d60c180c-dbus-socket\") pod \"nmstate-handler-fv9bw\" (UID: \"a4451cf3-5459-4208-bb86-49d3d60c180c\") " pod="openshift-nmstate/nmstate-handler-fv9bw" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.621369 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x78rm\" (UniqueName: \"kubernetes.io/projected/d070507a-81b9-4c71-b82e-5023abea5613-kube-api-access-x78rm\") pod \"nmstate-metrics-5dcf9c57c5-8bncp\" (UID: \"d070507a-81b9-4c71-b82e-5023abea5613\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-8bncp" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.633380 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-gdccx"] Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.633998 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-gdccx" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.635888 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.653713 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-gdccx"] Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.720693 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p5grn"] Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.721584 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p5grn" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.721728 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/a4451cf3-5459-4208-bb86-49d3d60c180c-dbus-socket\") pod \"nmstate-handler-fv9bw\" (UID: \"a4451cf3-5459-4208-bb86-49d3d60c180c\") " pod="openshift-nmstate/nmstate-handler-fv9bw" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.721790 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x78rm\" (UniqueName: \"kubernetes.io/projected/d070507a-81b9-4c71-b82e-5023abea5613-kube-api-access-x78rm\") pod \"nmstate-metrics-5dcf9c57c5-8bncp\" (UID: \"d070507a-81b9-4c71-b82e-5023abea5613\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-8bncp" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.721834 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh5jh\" (UniqueName: \"kubernetes.io/projected/cb651a64-745e-4587-b786-9e00604a5a77-kube-api-access-lh5jh\") pod \"nmstate-webhook-6b89b748d8-gdccx\" (UID: \"cb651a64-745e-4587-b786-9e00604a5a77\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-gdccx" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.721859 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/a4451cf3-5459-4208-bb86-49d3d60c180c-ovs-socket\") pod \"nmstate-handler-fv9bw\" (UID: \"a4451cf3-5459-4208-bb86-49d3d60c180c\") " pod="openshift-nmstate/nmstate-handler-fv9bw" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.721880 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/a4451cf3-5459-4208-bb86-49d3d60c180c-nmstate-lock\") pod \"nmstate-handler-fv9bw\" (UID: \"a4451cf3-5459-4208-bb86-49d3d60c180c\") " pod="openshift-nmstate/nmstate-handler-fv9bw" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.721896 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/cb651a64-745e-4587-b786-9e00604a5a77-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-gdccx\" (UID: \"cb651a64-745e-4587-b786-9e00604a5a77\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-gdccx" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.721923 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nm9gj\" (UniqueName: \"kubernetes.io/projected/a4451cf3-5459-4208-bb86-49d3d60c180c-kube-api-access-nm9gj\") pod \"nmstate-handler-fv9bw\" (UID: \"a4451cf3-5459-4208-bb86-49d3d60c180c\") " pod="openshift-nmstate/nmstate-handler-fv9bw" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.722585 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/a4451cf3-5459-4208-bb86-49d3d60c180c-ovs-socket\") pod \"nmstate-handler-fv9bw\" (UID: \"a4451cf3-5459-4208-bb86-49d3d60c180c\") " pod="openshift-nmstate/nmstate-handler-fv9bw" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.722755 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/a4451cf3-5459-4208-bb86-49d3d60c180c-nmstate-lock\") pod \"nmstate-handler-fv9bw\" (UID: \"a4451cf3-5459-4208-bb86-49d3d60c180c\") " pod="openshift-nmstate/nmstate-handler-fv9bw" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.724176 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/a4451cf3-5459-4208-bb86-49d3d60c180c-dbus-socket\") pod \"nmstate-handler-fv9bw\" (UID: \"a4451cf3-5459-4208-bb86-49d3d60c180c\") " pod="openshift-nmstate/nmstate-handler-fv9bw" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.725347 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.725370 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-zfxjr" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.725374 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.735005 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p5grn"] Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.748656 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nm9gj\" (UniqueName: \"kubernetes.io/projected/a4451cf3-5459-4208-bb86-49d3d60c180c-kube-api-access-nm9gj\") pod \"nmstate-handler-fv9bw\" (UID: \"a4451cf3-5459-4208-bb86-49d3d60c180c\") " pod="openshift-nmstate/nmstate-handler-fv9bw" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.748696 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x78rm\" (UniqueName: \"kubernetes.io/projected/d070507a-81b9-4c71-b82e-5023abea5613-kube-api-access-x78rm\") pod \"nmstate-metrics-5dcf9c57c5-8bncp\" (UID: \"d070507a-81b9-4c71-b82e-5023abea5613\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-8bncp" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.823271 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh5jh\" (UniqueName: \"kubernetes.io/projected/cb651a64-745e-4587-b786-9e00604a5a77-kube-api-access-lh5jh\") pod \"nmstate-webhook-6b89b748d8-gdccx\" (UID: \"cb651a64-745e-4587-b786-9e00604a5a77\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-gdccx" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.823474 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/cb651a64-745e-4587-b786-9e00604a5a77-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-gdccx\" (UID: \"cb651a64-745e-4587-b786-9e00604a5a77\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-gdccx" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.830890 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/cb651a64-745e-4587-b786-9e00604a5a77-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-gdccx\" (UID: \"cb651a64-745e-4587-b786-9e00604a5a77\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-gdccx" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.848019 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh5jh\" (UniqueName: \"kubernetes.io/projected/cb651a64-745e-4587-b786-9e00604a5a77-kube-api-access-lh5jh\") pod \"nmstate-webhook-6b89b748d8-gdccx\" (UID: \"cb651a64-745e-4587-b786-9e00604a5a77\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-gdccx" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.910416 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-8bncp" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.910433 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-647cc7864c-dg4zf"] Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.912318 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-647cc7864c-dg4zf" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.925217 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8fdef32b-fd6f-4280-9ca2-c912f40eb038-console-config\") pod \"console-647cc7864c-dg4zf\" (UID: \"8fdef32b-fd6f-4280-9ca2-c912f40eb038\") " pod="openshift-console/console-647cc7864c-dg4zf" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.925456 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8fdef32b-fd6f-4280-9ca2-c912f40eb038-trusted-ca-bundle\") pod \"console-647cc7864c-dg4zf\" (UID: \"8fdef32b-fd6f-4280-9ca2-c912f40eb038\") " pod="openshift-console/console-647cc7864c-dg4zf" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.925538 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8fdef32b-fd6f-4280-9ca2-c912f40eb038-console-oauth-config\") pod \"console-647cc7864c-dg4zf\" (UID: \"8fdef32b-fd6f-4280-9ca2-c912f40eb038\") " pod="openshift-console/console-647cc7864c-dg4zf" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.925598 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcqjj\" (UniqueName: \"kubernetes.io/projected/8fdef32b-fd6f-4280-9ca2-c912f40eb038-kube-api-access-hcqjj\") pod \"console-647cc7864c-dg4zf\" (UID: \"8fdef32b-fd6f-4280-9ca2-c912f40eb038\") " pod="openshift-console/console-647cc7864c-dg4zf" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.925670 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/a3994dc8-916f-4efb-b723-03650a70411d-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-p5grn\" (UID: \"a3994dc8-916f-4efb-b723-03650a70411d\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p5grn" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.925737 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8fdef32b-fd6f-4280-9ca2-c912f40eb038-console-serving-cert\") pod \"console-647cc7864c-dg4zf\" (UID: \"8fdef32b-fd6f-4280-9ca2-c912f40eb038\") " pod="openshift-console/console-647cc7864c-dg4zf" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.925798 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/a3994dc8-916f-4efb-b723-03650a70411d-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-p5grn\" (UID: \"a3994dc8-916f-4efb-b723-03650a70411d\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p5grn" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.925865 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z8m7\" (UniqueName: \"kubernetes.io/projected/a3994dc8-916f-4efb-b723-03650a70411d-kube-api-access-8z8m7\") pod \"nmstate-console-plugin-5874bd7bc5-p5grn\" (UID: \"a3994dc8-916f-4efb-b723-03650a70411d\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p5grn" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.925929 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8fdef32b-fd6f-4280-9ca2-c912f40eb038-service-ca\") pod \"console-647cc7864c-dg4zf\" (UID: \"8fdef32b-fd6f-4280-9ca2-c912f40eb038\") " pod="openshift-console/console-647cc7864c-dg4zf" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.925992 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8fdef32b-fd6f-4280-9ca2-c912f40eb038-oauth-serving-cert\") pod \"console-647cc7864c-dg4zf\" (UID: \"8fdef32b-fd6f-4280-9ca2-c912f40eb038\") " pod="openshift-console/console-647cc7864c-dg4zf" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.931719 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-fv9bw" Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.942644 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-647cc7864c-dg4zf"] Nov 25 09:37:28 crc kubenswrapper[4776]: I1125 09:37:28.951331 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-gdccx" Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.026709 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8fdef32b-fd6f-4280-9ca2-c912f40eb038-console-config\") pod \"console-647cc7864c-dg4zf\" (UID: \"8fdef32b-fd6f-4280-9ca2-c912f40eb038\") " pod="openshift-console/console-647cc7864c-dg4zf" Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.026759 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8fdef32b-fd6f-4280-9ca2-c912f40eb038-trusted-ca-bundle\") pod \"console-647cc7864c-dg4zf\" (UID: \"8fdef32b-fd6f-4280-9ca2-c912f40eb038\") " pod="openshift-console/console-647cc7864c-dg4zf" Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.026801 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8fdef32b-fd6f-4280-9ca2-c912f40eb038-console-oauth-config\") pod \"console-647cc7864c-dg4zf\" (UID: \"8fdef32b-fd6f-4280-9ca2-c912f40eb038\") " pod="openshift-console/console-647cc7864c-dg4zf" Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.026824 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcqjj\" (UniqueName: \"kubernetes.io/projected/8fdef32b-fd6f-4280-9ca2-c912f40eb038-kube-api-access-hcqjj\") pod \"console-647cc7864c-dg4zf\" (UID: \"8fdef32b-fd6f-4280-9ca2-c912f40eb038\") " pod="openshift-console/console-647cc7864c-dg4zf" Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.026895 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/a3994dc8-916f-4efb-b723-03650a70411d-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-p5grn\" (UID: \"a3994dc8-916f-4efb-b723-03650a70411d\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p5grn" Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.026916 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8fdef32b-fd6f-4280-9ca2-c912f40eb038-console-serving-cert\") pod \"console-647cc7864c-dg4zf\" (UID: \"8fdef32b-fd6f-4280-9ca2-c912f40eb038\") " pod="openshift-console/console-647cc7864c-dg4zf" Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.026937 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/a3994dc8-916f-4efb-b723-03650a70411d-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-p5grn\" (UID: \"a3994dc8-916f-4efb-b723-03650a70411d\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p5grn" Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.026971 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z8m7\" (UniqueName: \"kubernetes.io/projected/a3994dc8-916f-4efb-b723-03650a70411d-kube-api-access-8z8m7\") pod \"nmstate-console-plugin-5874bd7bc5-p5grn\" (UID: \"a3994dc8-916f-4efb-b723-03650a70411d\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p5grn" Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.026993 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8fdef32b-fd6f-4280-9ca2-c912f40eb038-service-ca\") pod \"console-647cc7864c-dg4zf\" (UID: \"8fdef32b-fd6f-4280-9ca2-c912f40eb038\") " pod="openshift-console/console-647cc7864c-dg4zf" Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.027020 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8fdef32b-fd6f-4280-9ca2-c912f40eb038-oauth-serving-cert\") pod \"console-647cc7864c-dg4zf\" (UID: \"8fdef32b-fd6f-4280-9ca2-c912f40eb038\") " pod="openshift-console/console-647cc7864c-dg4zf" Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.028233 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8fdef32b-fd6f-4280-9ca2-c912f40eb038-oauth-serving-cert\") pod \"console-647cc7864c-dg4zf\" (UID: \"8fdef32b-fd6f-4280-9ca2-c912f40eb038\") " pod="openshift-console/console-647cc7864c-dg4zf" Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.029011 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8fdef32b-fd6f-4280-9ca2-c912f40eb038-console-config\") pod \"console-647cc7864c-dg4zf\" (UID: \"8fdef32b-fd6f-4280-9ca2-c912f40eb038\") " pod="openshift-console/console-647cc7864c-dg4zf" Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.029935 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8fdef32b-fd6f-4280-9ca2-c912f40eb038-trusted-ca-bundle\") pod \"console-647cc7864c-dg4zf\" (UID: \"8fdef32b-fd6f-4280-9ca2-c912f40eb038\") " pod="openshift-console/console-647cc7864c-dg4zf" Nov 25 09:37:29 crc kubenswrapper[4776]: E1125 09:37:29.030005 4776 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Nov 25 09:37:29 crc kubenswrapper[4776]: E1125 09:37:29.030048 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a3994dc8-916f-4efb-b723-03650a70411d-plugin-serving-cert podName:a3994dc8-916f-4efb-b723-03650a70411d nodeName:}" failed. No retries permitted until 2025-11-25 09:37:29.530032723 +0000 UTC m=+794.571092276 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/a3994dc8-916f-4efb-b723-03650a70411d-plugin-serving-cert") pod "nmstate-console-plugin-5874bd7bc5-p5grn" (UID: "a3994dc8-916f-4efb-b723-03650a70411d") : secret "plugin-serving-cert" not found Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.031027 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8fdef32b-fd6f-4280-9ca2-c912f40eb038-service-ca\") pod \"console-647cc7864c-dg4zf\" (UID: \"8fdef32b-fd6f-4280-9ca2-c912f40eb038\") " pod="openshift-console/console-647cc7864c-dg4zf" Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.032168 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/a3994dc8-916f-4efb-b723-03650a70411d-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-p5grn\" (UID: \"a3994dc8-916f-4efb-b723-03650a70411d\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p5grn" Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.035469 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8fdef32b-fd6f-4280-9ca2-c912f40eb038-console-serving-cert\") pod \"console-647cc7864c-dg4zf\" (UID: \"8fdef32b-fd6f-4280-9ca2-c912f40eb038\") " pod="openshift-console/console-647cc7864c-dg4zf" Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.036415 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8fdef32b-fd6f-4280-9ca2-c912f40eb038-console-oauth-config\") pod \"console-647cc7864c-dg4zf\" (UID: \"8fdef32b-fd6f-4280-9ca2-c912f40eb038\") " pod="openshift-console/console-647cc7864c-dg4zf" Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.049738 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcqjj\" (UniqueName: \"kubernetes.io/projected/8fdef32b-fd6f-4280-9ca2-c912f40eb038-kube-api-access-hcqjj\") pod \"console-647cc7864c-dg4zf\" (UID: \"8fdef32b-fd6f-4280-9ca2-c912f40eb038\") " pod="openshift-console/console-647cc7864c-dg4zf" Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.054021 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z8m7\" (UniqueName: \"kubernetes.io/projected/a3994dc8-916f-4efb-b723-03650a70411d-kube-api-access-8z8m7\") pod \"nmstate-console-plugin-5874bd7bc5-p5grn\" (UID: \"a3994dc8-916f-4efb-b723-03650a70411d\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p5grn" Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.281162 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-647cc7864c-dg4zf" Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.317796 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-8bncp"] Nov 25 09:37:29 crc kubenswrapper[4776]: W1125 09:37:29.323869 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd070507a_81b9_4c71_b82e_5023abea5613.slice/crio-69a6c93376dff7849394f2645effef3ef955a0bcb1238b5833fbc5cf029b6163 WatchSource:0}: Error finding container 69a6c93376dff7849394f2645effef3ef955a0bcb1238b5833fbc5cf029b6163: Status 404 returned error can't find the container with id 69a6c93376dff7849394f2645effef3ef955a0bcb1238b5833fbc5cf029b6163 Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.404378 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-gdccx"] Nov 25 09:37:29 crc kubenswrapper[4776]: W1125 09:37:29.411549 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb651a64_745e_4587_b786_9e00604a5a77.slice/crio-c28722d1893b39abe884aa0502ee0dbaaa3ace471dc6743f0ed4e42d39b960ac WatchSource:0}: Error finding container c28722d1893b39abe884aa0502ee0dbaaa3ace471dc6743f0ed4e42d39b960ac: Status 404 returned error can't find the container with id c28722d1893b39abe884aa0502ee0dbaaa3ace471dc6743f0ed4e42d39b960ac Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.535867 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/a3994dc8-916f-4efb-b723-03650a70411d-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-p5grn\" (UID: \"a3994dc8-916f-4efb-b723-03650a70411d\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p5grn" Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.540094 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/a3994dc8-916f-4efb-b723-03650a70411d-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-p5grn\" (UID: \"a3994dc8-916f-4efb-b723-03650a70411d\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p5grn" Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.625956 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-gdccx" event={"ID":"cb651a64-745e-4587-b786-9e00604a5a77","Type":"ContainerStarted","Data":"c28722d1893b39abe884aa0502ee0dbaaa3ace471dc6743f0ed4e42d39b960ac"} Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.627346 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-fv9bw" event={"ID":"a4451cf3-5459-4208-bb86-49d3d60c180c","Type":"ContainerStarted","Data":"9fb26d7e8540456a861cf03790e2c8f8c00ad8be236942e5f402216cbf38ecc4"} Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.628636 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-8bncp" event={"ID":"d070507a-81b9-4c71-b82e-5023abea5613","Type":"ContainerStarted","Data":"69a6c93376dff7849394f2645effef3ef955a0bcb1238b5833fbc5cf029b6163"} Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.635977 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p5grn" Nov 25 09:37:29 crc kubenswrapper[4776]: I1125 09:37:29.679913 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-647cc7864c-dg4zf"] Nov 25 09:37:29 crc kubenswrapper[4776]: W1125 09:37:29.687090 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8fdef32b_fd6f_4280_9ca2_c912f40eb038.slice/crio-32de609bf182fa8ce0c35c0a66912d423b28d8ff88f9b02e38dbbb3be5d60b35 WatchSource:0}: Error finding container 32de609bf182fa8ce0c35c0a66912d423b28d8ff88f9b02e38dbbb3be5d60b35: Status 404 returned error can't find the container with id 32de609bf182fa8ce0c35c0a66912d423b28d8ff88f9b02e38dbbb3be5d60b35 Nov 25 09:37:30 crc kubenswrapper[4776]: I1125 09:37:30.068620 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p5grn"] Nov 25 09:37:30 crc kubenswrapper[4776]: W1125 09:37:30.072946 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda3994dc8_916f_4efb_b723_03650a70411d.slice/crio-a8f4f5ffdea442b7eef491ef09e93300ab25c7be2e0a38b2ad6c35d4c017f938 WatchSource:0}: Error finding container a8f4f5ffdea442b7eef491ef09e93300ab25c7be2e0a38b2ad6c35d4c017f938: Status 404 returned error can't find the container with id a8f4f5ffdea442b7eef491ef09e93300ab25c7be2e0a38b2ad6c35d4c017f938 Nov 25 09:37:30 crc kubenswrapper[4776]: I1125 09:37:30.638395 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-647cc7864c-dg4zf" event={"ID":"8fdef32b-fd6f-4280-9ca2-c912f40eb038","Type":"ContainerStarted","Data":"7b0ee336fc2949b1a9426a716cd1da3ebe0ea3f17dacd16a0c57c42ac8f35b1b"} Nov 25 09:37:30 crc kubenswrapper[4776]: I1125 09:37:30.639225 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-647cc7864c-dg4zf" event={"ID":"8fdef32b-fd6f-4280-9ca2-c912f40eb038","Type":"ContainerStarted","Data":"32de609bf182fa8ce0c35c0a66912d423b28d8ff88f9b02e38dbbb3be5d60b35"} Nov 25 09:37:30 crc kubenswrapper[4776]: I1125 09:37:30.641400 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p5grn" event={"ID":"a3994dc8-916f-4efb-b723-03650a70411d","Type":"ContainerStarted","Data":"a8f4f5ffdea442b7eef491ef09e93300ab25c7be2e0a38b2ad6c35d4c017f938"} Nov 25 09:37:30 crc kubenswrapper[4776]: I1125 09:37:30.658870 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-647cc7864c-dg4zf" podStartSLOduration=2.658852316 podStartE2EDuration="2.658852316s" podCreationTimestamp="2025-11-25 09:37:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:37:30.653593054 +0000 UTC m=+795.694652617" watchObservedRunningTime="2025-11-25 09:37:30.658852316 +0000 UTC m=+795.699911869" Nov 25 09:37:32 crc kubenswrapper[4776]: I1125 09:37:32.658595 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p5grn" event={"ID":"a3994dc8-916f-4efb-b723-03650a70411d","Type":"ContainerStarted","Data":"4092984ee21858026f5f5b333c5388629a5c1bbf4f19bf9f12a1a8ed8bdbad00"} Nov 25 09:37:32 crc kubenswrapper[4776]: I1125 09:37:32.660117 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-8bncp" event={"ID":"d070507a-81b9-4c71-b82e-5023abea5613","Type":"ContainerStarted","Data":"310c7d7068004b66b9383009e7f86b44715b6fea5a2dbf668d4199bcaecfa737"} Nov 25 09:37:32 crc kubenswrapper[4776]: I1125 09:37:32.661149 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-gdccx" event={"ID":"cb651a64-745e-4587-b786-9e00604a5a77","Type":"ContainerStarted","Data":"6e1aaabd6dd94ba62ecb8687c158dd0a4b6bbc956053a2650b28ddfb36c91809"} Nov 25 09:37:32 crc kubenswrapper[4776]: I1125 09:37:32.661684 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-gdccx" Nov 25 09:37:32 crc kubenswrapper[4776]: I1125 09:37:32.667998 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-fv9bw" event={"ID":"a4451cf3-5459-4208-bb86-49d3d60c180c","Type":"ContainerStarted","Data":"bedb8fe39e4022c6fff21edfa6c074d09703425f2e4af10a7e59808472f04dae"} Nov 25 09:37:32 crc kubenswrapper[4776]: I1125 09:37:32.668130 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-fv9bw" Nov 25 09:37:32 crc kubenswrapper[4776]: I1125 09:37:32.677477 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-p5grn" podStartSLOduration=2.35362721 podStartE2EDuration="4.677455942s" podCreationTimestamp="2025-11-25 09:37:28 +0000 UTC" firstStartedPulling="2025-11-25 09:37:30.076546315 +0000 UTC m=+795.117605888" lastFinishedPulling="2025-11-25 09:37:32.400375077 +0000 UTC m=+797.441434620" observedRunningTime="2025-11-25 09:37:32.675269617 +0000 UTC m=+797.716329180" watchObservedRunningTime="2025-11-25 09:37:32.677455942 +0000 UTC m=+797.718515515" Nov 25 09:37:32 crc kubenswrapper[4776]: I1125 09:37:32.715969 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-fv9bw" podStartSLOduration=2.1739759 podStartE2EDuration="4.715949804s" podCreationTimestamp="2025-11-25 09:37:28 +0000 UTC" firstStartedPulling="2025-11-25 09:37:28.979198022 +0000 UTC m=+794.020257575" lastFinishedPulling="2025-11-25 09:37:31.521171926 +0000 UTC m=+796.562231479" observedRunningTime="2025-11-25 09:37:32.712405975 +0000 UTC m=+797.753465528" watchObservedRunningTime="2025-11-25 09:37:32.715949804 +0000 UTC m=+797.757009377" Nov 25 09:37:32 crc kubenswrapper[4776]: I1125 09:37:32.737708 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-gdccx" podStartSLOduration=2.617633947 podStartE2EDuration="4.737670446s" podCreationTimestamp="2025-11-25 09:37:28 +0000 UTC" firstStartedPulling="2025-11-25 09:37:29.413754562 +0000 UTC m=+794.454814105" lastFinishedPulling="2025-11-25 09:37:31.533791051 +0000 UTC m=+796.574850604" observedRunningTime="2025-11-25 09:37:32.733264546 +0000 UTC m=+797.774324099" watchObservedRunningTime="2025-11-25 09:37:32.737670446 +0000 UTC m=+797.778730009" Nov 25 09:37:34 crc kubenswrapper[4776]: I1125 09:37:34.679916 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-8bncp" event={"ID":"d070507a-81b9-4c71-b82e-5023abea5613","Type":"ContainerStarted","Data":"2b7610a968434811729ec08102fa1975b8e16183f5c397fcde3b17e8e4876750"} Nov 25 09:37:34 crc kubenswrapper[4776]: I1125 09:37:34.698871 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-8bncp" podStartSLOduration=2.022567007 podStartE2EDuration="6.698847076s" podCreationTimestamp="2025-11-25 09:37:28 +0000 UTC" firstStartedPulling="2025-11-25 09:37:29.329287131 +0000 UTC m=+794.370346684" lastFinishedPulling="2025-11-25 09:37:34.0055672 +0000 UTC m=+799.046626753" observedRunningTime="2025-11-25 09:37:34.69703501 +0000 UTC m=+799.738094583" watchObservedRunningTime="2025-11-25 09:37:34.698847076 +0000 UTC m=+799.739906639" Nov 25 09:37:38 crc kubenswrapper[4776]: I1125 09:37:38.962767 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-fv9bw" Nov 25 09:37:39 crc kubenswrapper[4776]: I1125 09:37:39.282133 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-647cc7864c-dg4zf" Nov 25 09:37:39 crc kubenswrapper[4776]: I1125 09:37:39.282206 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-647cc7864c-dg4zf" Nov 25 09:37:39 crc kubenswrapper[4776]: I1125 09:37:39.286991 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-647cc7864c-dg4zf" Nov 25 09:37:39 crc kubenswrapper[4776]: I1125 09:37:39.714433 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-647cc7864c-dg4zf" Nov 25 09:37:39 crc kubenswrapper[4776]: I1125 09:37:39.775315 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-bhgv2"] Nov 25 09:37:44 crc kubenswrapper[4776]: I1125 09:37:44.424243 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7zjbg"] Nov 25 09:37:44 crc kubenswrapper[4776]: I1125 09:37:44.427346 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7zjbg" Nov 25 09:37:44 crc kubenswrapper[4776]: I1125 09:37:44.432373 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7zjbg"] Nov 25 09:37:44 crc kubenswrapper[4776]: I1125 09:37:44.531502 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ad7db17-c8be-4679-952c-14503eba40b6-catalog-content\") pod \"certified-operators-7zjbg\" (UID: \"5ad7db17-c8be-4679-952c-14503eba40b6\") " pod="openshift-marketplace/certified-operators-7zjbg" Nov 25 09:37:44 crc kubenswrapper[4776]: I1125 09:37:44.531585 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dt47p\" (UniqueName: \"kubernetes.io/projected/5ad7db17-c8be-4679-952c-14503eba40b6-kube-api-access-dt47p\") pod \"certified-operators-7zjbg\" (UID: \"5ad7db17-c8be-4679-952c-14503eba40b6\") " pod="openshift-marketplace/certified-operators-7zjbg" Nov 25 09:37:44 crc kubenswrapper[4776]: I1125 09:37:44.531672 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ad7db17-c8be-4679-952c-14503eba40b6-utilities\") pod \"certified-operators-7zjbg\" (UID: \"5ad7db17-c8be-4679-952c-14503eba40b6\") " pod="openshift-marketplace/certified-operators-7zjbg" Nov 25 09:37:44 crc kubenswrapper[4776]: I1125 09:37:44.632670 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ad7db17-c8be-4679-952c-14503eba40b6-catalog-content\") pod \"certified-operators-7zjbg\" (UID: \"5ad7db17-c8be-4679-952c-14503eba40b6\") " pod="openshift-marketplace/certified-operators-7zjbg" Nov 25 09:37:44 crc kubenswrapper[4776]: I1125 09:37:44.632822 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dt47p\" (UniqueName: \"kubernetes.io/projected/5ad7db17-c8be-4679-952c-14503eba40b6-kube-api-access-dt47p\") pod \"certified-operators-7zjbg\" (UID: \"5ad7db17-c8be-4679-952c-14503eba40b6\") " pod="openshift-marketplace/certified-operators-7zjbg" Nov 25 09:37:44 crc kubenswrapper[4776]: I1125 09:37:44.632892 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ad7db17-c8be-4679-952c-14503eba40b6-utilities\") pod \"certified-operators-7zjbg\" (UID: \"5ad7db17-c8be-4679-952c-14503eba40b6\") " pod="openshift-marketplace/certified-operators-7zjbg" Nov 25 09:37:44 crc kubenswrapper[4776]: I1125 09:37:44.633446 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ad7db17-c8be-4679-952c-14503eba40b6-utilities\") pod \"certified-operators-7zjbg\" (UID: \"5ad7db17-c8be-4679-952c-14503eba40b6\") " pod="openshift-marketplace/certified-operators-7zjbg" Nov 25 09:37:44 crc kubenswrapper[4776]: I1125 09:37:44.633631 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ad7db17-c8be-4679-952c-14503eba40b6-catalog-content\") pod \"certified-operators-7zjbg\" (UID: \"5ad7db17-c8be-4679-952c-14503eba40b6\") " pod="openshift-marketplace/certified-operators-7zjbg" Nov 25 09:37:44 crc kubenswrapper[4776]: I1125 09:37:44.653987 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dt47p\" (UniqueName: \"kubernetes.io/projected/5ad7db17-c8be-4679-952c-14503eba40b6-kube-api-access-dt47p\") pod \"certified-operators-7zjbg\" (UID: \"5ad7db17-c8be-4679-952c-14503eba40b6\") " pod="openshift-marketplace/certified-operators-7zjbg" Nov 25 09:37:44 crc kubenswrapper[4776]: I1125 09:37:44.744353 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7zjbg" Nov 25 09:37:44 crc kubenswrapper[4776]: I1125 09:37:44.981863 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7zjbg"] Nov 25 09:37:44 crc kubenswrapper[4776]: W1125 09:37:44.989303 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ad7db17_c8be_4679_952c_14503eba40b6.slice/crio-82116b97c41d60785f2aa1b002bed0fef4e29a0c3433cabd82ba4b641302801d WatchSource:0}: Error finding container 82116b97c41d60785f2aa1b002bed0fef4e29a0c3433cabd82ba4b641302801d: Status 404 returned error can't find the container with id 82116b97c41d60785f2aa1b002bed0fef4e29a0c3433cabd82ba4b641302801d Nov 25 09:37:45 crc kubenswrapper[4776]: I1125 09:37:45.745197 4776 generic.go:334] "Generic (PLEG): container finished" podID="5ad7db17-c8be-4679-952c-14503eba40b6" containerID="c306aa65939f55c56847a7a50dae2f80f7a148edc1e1d2a29c9705c9a1801a6b" exitCode=0 Nov 25 09:37:45 crc kubenswrapper[4776]: I1125 09:37:45.745319 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7zjbg" event={"ID":"5ad7db17-c8be-4679-952c-14503eba40b6","Type":"ContainerDied","Data":"c306aa65939f55c56847a7a50dae2f80f7a148edc1e1d2a29c9705c9a1801a6b"} Nov 25 09:37:45 crc kubenswrapper[4776]: I1125 09:37:45.745618 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7zjbg" event={"ID":"5ad7db17-c8be-4679-952c-14503eba40b6","Type":"ContainerStarted","Data":"82116b97c41d60785f2aa1b002bed0fef4e29a0c3433cabd82ba4b641302801d"} Nov 25 09:37:46 crc kubenswrapper[4776]: I1125 09:37:46.757526 4776 generic.go:334] "Generic (PLEG): container finished" podID="5ad7db17-c8be-4679-952c-14503eba40b6" containerID="383abbe443a8a6d6f5db562463adde88955bbf8ed71205948f00e006b7cb0907" exitCode=0 Nov 25 09:37:46 crc kubenswrapper[4776]: I1125 09:37:46.757580 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7zjbg" event={"ID":"5ad7db17-c8be-4679-952c-14503eba40b6","Type":"ContainerDied","Data":"383abbe443a8a6d6f5db562463adde88955bbf8ed71205948f00e006b7cb0907"} Nov 25 09:37:47 crc kubenswrapper[4776]: I1125 09:37:47.765911 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7zjbg" event={"ID":"5ad7db17-c8be-4679-952c-14503eba40b6","Type":"ContainerStarted","Data":"b28fd0add8a7a444972dd0aeb0209e26603c8bbcfa9a2365d5a0b165a3f963fd"} Nov 25 09:37:47 crc kubenswrapper[4776]: I1125 09:37:47.786668 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7zjbg" podStartSLOduration=2.312447334 podStartE2EDuration="3.786641244s" podCreationTimestamp="2025-11-25 09:37:44 +0000 UTC" firstStartedPulling="2025-11-25 09:37:45.746823639 +0000 UTC m=+810.787883192" lastFinishedPulling="2025-11-25 09:37:47.221017549 +0000 UTC m=+812.262077102" observedRunningTime="2025-11-25 09:37:47.780660335 +0000 UTC m=+812.821719908" watchObservedRunningTime="2025-11-25 09:37:47.786641244 +0000 UTC m=+812.827700817" Nov 25 09:37:47 crc kubenswrapper[4776]: I1125 09:37:47.818870 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:37:47 crc kubenswrapper[4776]: I1125 09:37:47.818933 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:37:48 crc kubenswrapper[4776]: I1125 09:37:48.959792 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-gdccx" Nov 25 09:37:50 crc kubenswrapper[4776]: I1125 09:37:50.800569 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fr2rf"] Nov 25 09:37:50 crc kubenswrapper[4776]: I1125 09:37:50.802187 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fr2rf" Nov 25 09:37:50 crc kubenswrapper[4776]: I1125 09:37:50.811457 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fr2rf"] Nov 25 09:37:50 crc kubenswrapper[4776]: I1125 09:37:50.908567 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqhpp\" (UniqueName: \"kubernetes.io/projected/c063c53a-8226-40c0-b803-93dc46813de0-kube-api-access-tqhpp\") pod \"redhat-marketplace-fr2rf\" (UID: \"c063c53a-8226-40c0-b803-93dc46813de0\") " pod="openshift-marketplace/redhat-marketplace-fr2rf" Nov 25 09:37:50 crc kubenswrapper[4776]: I1125 09:37:50.908798 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c063c53a-8226-40c0-b803-93dc46813de0-catalog-content\") pod \"redhat-marketplace-fr2rf\" (UID: \"c063c53a-8226-40c0-b803-93dc46813de0\") " pod="openshift-marketplace/redhat-marketplace-fr2rf" Nov 25 09:37:50 crc kubenswrapper[4776]: I1125 09:37:50.908925 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c063c53a-8226-40c0-b803-93dc46813de0-utilities\") pod \"redhat-marketplace-fr2rf\" (UID: \"c063c53a-8226-40c0-b803-93dc46813de0\") " pod="openshift-marketplace/redhat-marketplace-fr2rf" Nov 25 09:37:51 crc kubenswrapper[4776]: I1125 09:37:51.010403 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c063c53a-8226-40c0-b803-93dc46813de0-catalog-content\") pod \"redhat-marketplace-fr2rf\" (UID: \"c063c53a-8226-40c0-b803-93dc46813de0\") " pod="openshift-marketplace/redhat-marketplace-fr2rf" Nov 25 09:37:51 crc kubenswrapper[4776]: I1125 09:37:51.010477 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c063c53a-8226-40c0-b803-93dc46813de0-utilities\") pod \"redhat-marketplace-fr2rf\" (UID: \"c063c53a-8226-40c0-b803-93dc46813de0\") " pod="openshift-marketplace/redhat-marketplace-fr2rf" Nov 25 09:37:51 crc kubenswrapper[4776]: I1125 09:37:51.010706 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqhpp\" (UniqueName: \"kubernetes.io/projected/c063c53a-8226-40c0-b803-93dc46813de0-kube-api-access-tqhpp\") pod \"redhat-marketplace-fr2rf\" (UID: \"c063c53a-8226-40c0-b803-93dc46813de0\") " pod="openshift-marketplace/redhat-marketplace-fr2rf" Nov 25 09:37:51 crc kubenswrapper[4776]: I1125 09:37:51.011057 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c063c53a-8226-40c0-b803-93dc46813de0-catalog-content\") pod \"redhat-marketplace-fr2rf\" (UID: \"c063c53a-8226-40c0-b803-93dc46813de0\") " pod="openshift-marketplace/redhat-marketplace-fr2rf" Nov 25 09:37:51 crc kubenswrapper[4776]: I1125 09:37:51.011699 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c063c53a-8226-40c0-b803-93dc46813de0-utilities\") pod \"redhat-marketplace-fr2rf\" (UID: \"c063c53a-8226-40c0-b803-93dc46813de0\") " pod="openshift-marketplace/redhat-marketplace-fr2rf" Nov 25 09:37:51 crc kubenswrapper[4776]: I1125 09:37:51.033767 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqhpp\" (UniqueName: \"kubernetes.io/projected/c063c53a-8226-40c0-b803-93dc46813de0-kube-api-access-tqhpp\") pod \"redhat-marketplace-fr2rf\" (UID: \"c063c53a-8226-40c0-b803-93dc46813de0\") " pod="openshift-marketplace/redhat-marketplace-fr2rf" Nov 25 09:37:51 crc kubenswrapper[4776]: I1125 09:37:51.123430 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fr2rf" Nov 25 09:37:51 crc kubenswrapper[4776]: I1125 09:37:51.580359 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fr2rf"] Nov 25 09:37:51 crc kubenswrapper[4776]: W1125 09:37:51.592204 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc063c53a_8226_40c0_b803_93dc46813de0.slice/crio-e059683a702b428f701ef83254e77035cf655bdbcc3a060eb71e6fc108807833 WatchSource:0}: Error finding container e059683a702b428f701ef83254e77035cf655bdbcc3a060eb71e6fc108807833: Status 404 returned error can't find the container with id e059683a702b428f701ef83254e77035cf655bdbcc3a060eb71e6fc108807833 Nov 25 09:37:51 crc kubenswrapper[4776]: I1125 09:37:51.791474 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fr2rf" event={"ID":"c063c53a-8226-40c0-b803-93dc46813de0","Type":"ContainerStarted","Data":"e059683a702b428f701ef83254e77035cf655bdbcc3a060eb71e6fc108807833"} Nov 25 09:37:52 crc kubenswrapper[4776]: I1125 09:37:52.800884 4776 generic.go:334] "Generic (PLEG): container finished" podID="c063c53a-8226-40c0-b803-93dc46813de0" containerID="6af6bf2186be59fab74f999b94bc4919c28ae0212927f58cfdfa00d46325250a" exitCode=0 Nov 25 09:37:52 crc kubenswrapper[4776]: I1125 09:37:52.800927 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fr2rf" event={"ID":"c063c53a-8226-40c0-b803-93dc46813de0","Type":"ContainerDied","Data":"6af6bf2186be59fab74f999b94bc4919c28ae0212927f58cfdfa00d46325250a"} Nov 25 09:37:53 crc kubenswrapper[4776]: I1125 09:37:53.841135 4776 generic.go:334] "Generic (PLEG): container finished" podID="c063c53a-8226-40c0-b803-93dc46813de0" containerID="6c2dc7f617dc095d4ce901b20196daeca158dedd2272b216a5c373e4299b3829" exitCode=0 Nov 25 09:37:53 crc kubenswrapper[4776]: I1125 09:37:53.841380 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fr2rf" event={"ID":"c063c53a-8226-40c0-b803-93dc46813de0","Type":"ContainerDied","Data":"6c2dc7f617dc095d4ce901b20196daeca158dedd2272b216a5c373e4299b3829"} Nov 25 09:37:54 crc kubenswrapper[4776]: I1125 09:37:54.745204 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7zjbg" Nov 25 09:37:54 crc kubenswrapper[4776]: I1125 09:37:54.745632 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7zjbg" Nov 25 09:37:54 crc kubenswrapper[4776]: I1125 09:37:54.795394 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7zjbg" Nov 25 09:37:54 crc kubenswrapper[4776]: I1125 09:37:54.849176 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fr2rf" event={"ID":"c063c53a-8226-40c0-b803-93dc46813de0","Type":"ContainerStarted","Data":"650a5808155499b9b55b014659cef8831638ec359dff2ad758fa5706a58060c3"} Nov 25 09:37:54 crc kubenswrapper[4776]: I1125 09:37:54.874354 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fr2rf" podStartSLOduration=3.433306954 podStartE2EDuration="4.874335656s" podCreationTimestamp="2025-11-25 09:37:50 +0000 UTC" firstStartedPulling="2025-11-25 09:37:52.802582522 +0000 UTC m=+817.843642085" lastFinishedPulling="2025-11-25 09:37:54.243611224 +0000 UTC m=+819.284670787" observedRunningTime="2025-11-25 09:37:54.8717158 +0000 UTC m=+819.912775363" watchObservedRunningTime="2025-11-25 09:37:54.874335656 +0000 UTC m=+819.915395209" Nov 25 09:37:54 crc kubenswrapper[4776]: I1125 09:37:54.887137 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7zjbg" Nov 25 09:37:55 crc kubenswrapper[4776]: I1125 09:37:55.407591 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jnbfd"] Nov 25 09:37:55 crc kubenswrapper[4776]: I1125 09:37:55.408935 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jnbfd" Nov 25 09:37:55 crc kubenswrapper[4776]: I1125 09:37:55.417079 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jnbfd"] Nov 25 09:37:55 crc kubenswrapper[4776]: I1125 09:37:55.572493 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f473caf4-4336-45bc-97f0-b38e6109198e-utilities\") pod \"community-operators-jnbfd\" (UID: \"f473caf4-4336-45bc-97f0-b38e6109198e\") " pod="openshift-marketplace/community-operators-jnbfd" Nov 25 09:37:55 crc kubenswrapper[4776]: I1125 09:37:55.572547 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f473caf4-4336-45bc-97f0-b38e6109198e-catalog-content\") pod \"community-operators-jnbfd\" (UID: \"f473caf4-4336-45bc-97f0-b38e6109198e\") " pod="openshift-marketplace/community-operators-jnbfd" Nov 25 09:37:55 crc kubenswrapper[4776]: I1125 09:37:55.572575 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxjnk\" (UniqueName: \"kubernetes.io/projected/f473caf4-4336-45bc-97f0-b38e6109198e-kube-api-access-hxjnk\") pod \"community-operators-jnbfd\" (UID: \"f473caf4-4336-45bc-97f0-b38e6109198e\") " pod="openshift-marketplace/community-operators-jnbfd" Nov 25 09:37:55 crc kubenswrapper[4776]: I1125 09:37:55.674043 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f473caf4-4336-45bc-97f0-b38e6109198e-utilities\") pod \"community-operators-jnbfd\" (UID: \"f473caf4-4336-45bc-97f0-b38e6109198e\") " pod="openshift-marketplace/community-operators-jnbfd" Nov 25 09:37:55 crc kubenswrapper[4776]: I1125 09:37:55.674141 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f473caf4-4336-45bc-97f0-b38e6109198e-catalog-content\") pod \"community-operators-jnbfd\" (UID: \"f473caf4-4336-45bc-97f0-b38e6109198e\") " pod="openshift-marketplace/community-operators-jnbfd" Nov 25 09:37:55 crc kubenswrapper[4776]: I1125 09:37:55.674193 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxjnk\" (UniqueName: \"kubernetes.io/projected/f473caf4-4336-45bc-97f0-b38e6109198e-kube-api-access-hxjnk\") pod \"community-operators-jnbfd\" (UID: \"f473caf4-4336-45bc-97f0-b38e6109198e\") " pod="openshift-marketplace/community-operators-jnbfd" Nov 25 09:37:55 crc kubenswrapper[4776]: I1125 09:37:55.674692 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f473caf4-4336-45bc-97f0-b38e6109198e-utilities\") pod \"community-operators-jnbfd\" (UID: \"f473caf4-4336-45bc-97f0-b38e6109198e\") " pod="openshift-marketplace/community-operators-jnbfd" Nov 25 09:37:55 crc kubenswrapper[4776]: I1125 09:37:55.676408 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f473caf4-4336-45bc-97f0-b38e6109198e-catalog-content\") pod \"community-operators-jnbfd\" (UID: \"f473caf4-4336-45bc-97f0-b38e6109198e\") " pod="openshift-marketplace/community-operators-jnbfd" Nov 25 09:37:55 crc kubenswrapper[4776]: I1125 09:37:55.693678 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxjnk\" (UniqueName: \"kubernetes.io/projected/f473caf4-4336-45bc-97f0-b38e6109198e-kube-api-access-hxjnk\") pod \"community-operators-jnbfd\" (UID: \"f473caf4-4336-45bc-97f0-b38e6109198e\") " pod="openshift-marketplace/community-operators-jnbfd" Nov 25 09:37:55 crc kubenswrapper[4776]: I1125 09:37:55.727159 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jnbfd" Nov 25 09:37:56 crc kubenswrapper[4776]: I1125 09:37:56.209468 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jnbfd"] Nov 25 09:37:56 crc kubenswrapper[4776]: W1125 09:37:56.228172 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf473caf4_4336_45bc_97f0_b38e6109198e.slice/crio-7efedd7760396923e2970399f27c25f3b4d6aa1e25a426fd30f92d62906296cc WatchSource:0}: Error finding container 7efedd7760396923e2970399f27c25f3b4d6aa1e25a426fd30f92d62906296cc: Status 404 returned error can't find the container with id 7efedd7760396923e2970399f27c25f3b4d6aa1e25a426fd30f92d62906296cc Nov 25 09:37:56 crc kubenswrapper[4776]: I1125 09:37:56.878007 4776 generic.go:334] "Generic (PLEG): container finished" podID="f473caf4-4336-45bc-97f0-b38e6109198e" containerID="5848fcb8743291ac041828d178f8a474d57a052155ded6d8c35e96de7725a2df" exitCode=0 Nov 25 09:37:56 crc kubenswrapper[4776]: I1125 09:37:56.878107 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jnbfd" event={"ID":"f473caf4-4336-45bc-97f0-b38e6109198e","Type":"ContainerDied","Data":"5848fcb8743291ac041828d178f8a474d57a052155ded6d8c35e96de7725a2df"} Nov 25 09:37:56 crc kubenswrapper[4776]: I1125 09:37:56.879208 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jnbfd" event={"ID":"f473caf4-4336-45bc-97f0-b38e6109198e","Type":"ContainerStarted","Data":"7efedd7760396923e2970399f27c25f3b4d6aa1e25a426fd30f92d62906296cc"} Nov 25 09:37:57 crc kubenswrapper[4776]: I1125 09:37:57.792419 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7zjbg"] Nov 25 09:37:57 crc kubenswrapper[4776]: I1125 09:37:57.792919 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7zjbg" podUID="5ad7db17-c8be-4679-952c-14503eba40b6" containerName="registry-server" containerID="cri-o://b28fd0add8a7a444972dd0aeb0209e26603c8bbcfa9a2365d5a0b165a3f963fd" gracePeriod=2 Nov 25 09:37:58 crc kubenswrapper[4776]: I1125 09:37:58.157183 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7zjbg" Nov 25 09:37:58 crc kubenswrapper[4776]: I1125 09:37:58.207977 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ad7db17-c8be-4679-952c-14503eba40b6-utilities\") pod \"5ad7db17-c8be-4679-952c-14503eba40b6\" (UID: \"5ad7db17-c8be-4679-952c-14503eba40b6\") " Nov 25 09:37:58 crc kubenswrapper[4776]: I1125 09:37:58.208050 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dt47p\" (UniqueName: \"kubernetes.io/projected/5ad7db17-c8be-4679-952c-14503eba40b6-kube-api-access-dt47p\") pod \"5ad7db17-c8be-4679-952c-14503eba40b6\" (UID: \"5ad7db17-c8be-4679-952c-14503eba40b6\") " Nov 25 09:37:58 crc kubenswrapper[4776]: I1125 09:37:58.208118 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ad7db17-c8be-4679-952c-14503eba40b6-catalog-content\") pod \"5ad7db17-c8be-4679-952c-14503eba40b6\" (UID: \"5ad7db17-c8be-4679-952c-14503eba40b6\") " Nov 25 09:37:58 crc kubenswrapper[4776]: I1125 09:37:58.210608 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ad7db17-c8be-4679-952c-14503eba40b6-utilities" (OuterVolumeSpecName: "utilities") pod "5ad7db17-c8be-4679-952c-14503eba40b6" (UID: "5ad7db17-c8be-4679-952c-14503eba40b6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:37:58 crc kubenswrapper[4776]: I1125 09:37:58.214577 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ad7db17-c8be-4679-952c-14503eba40b6-kube-api-access-dt47p" (OuterVolumeSpecName: "kube-api-access-dt47p") pod "5ad7db17-c8be-4679-952c-14503eba40b6" (UID: "5ad7db17-c8be-4679-952c-14503eba40b6"). InnerVolumeSpecName "kube-api-access-dt47p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:37:58 crc kubenswrapper[4776]: I1125 09:37:58.273677 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ad7db17-c8be-4679-952c-14503eba40b6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ad7db17-c8be-4679-952c-14503eba40b6" (UID: "5ad7db17-c8be-4679-952c-14503eba40b6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:37:58 crc kubenswrapper[4776]: I1125 09:37:58.323019 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ad7db17-c8be-4679-952c-14503eba40b6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:37:58 crc kubenswrapper[4776]: I1125 09:37:58.323054 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ad7db17-c8be-4679-952c-14503eba40b6-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:37:58 crc kubenswrapper[4776]: I1125 09:37:58.323085 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dt47p\" (UniqueName: \"kubernetes.io/projected/5ad7db17-c8be-4679-952c-14503eba40b6-kube-api-access-dt47p\") on node \"crc\" DevicePath \"\"" Nov 25 09:37:58 crc kubenswrapper[4776]: I1125 09:37:58.895626 4776 generic.go:334] "Generic (PLEG): container finished" podID="5ad7db17-c8be-4679-952c-14503eba40b6" containerID="b28fd0add8a7a444972dd0aeb0209e26603c8bbcfa9a2365d5a0b165a3f963fd" exitCode=0 Nov 25 09:37:58 crc kubenswrapper[4776]: I1125 09:37:58.895693 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7zjbg" Nov 25 09:37:58 crc kubenswrapper[4776]: I1125 09:37:58.895711 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7zjbg" event={"ID":"5ad7db17-c8be-4679-952c-14503eba40b6","Type":"ContainerDied","Data":"b28fd0add8a7a444972dd0aeb0209e26603c8bbcfa9a2365d5a0b165a3f963fd"} Nov 25 09:37:58 crc kubenswrapper[4776]: I1125 09:37:58.896225 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7zjbg" event={"ID":"5ad7db17-c8be-4679-952c-14503eba40b6","Type":"ContainerDied","Data":"82116b97c41d60785f2aa1b002bed0fef4e29a0c3433cabd82ba4b641302801d"} Nov 25 09:37:58 crc kubenswrapper[4776]: I1125 09:37:58.896255 4776 scope.go:117] "RemoveContainer" containerID="b28fd0add8a7a444972dd0aeb0209e26603c8bbcfa9a2365d5a0b165a3f963fd" Nov 25 09:37:58 crc kubenswrapper[4776]: I1125 09:37:58.897923 4776 generic.go:334] "Generic (PLEG): container finished" podID="f473caf4-4336-45bc-97f0-b38e6109198e" containerID="3b614d8bb1878bd5316855217a6adbbfc0ef9e40bcc8ff8d8ac0849d190d68b2" exitCode=0 Nov 25 09:37:58 crc kubenswrapper[4776]: I1125 09:37:58.897967 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jnbfd" event={"ID":"f473caf4-4336-45bc-97f0-b38e6109198e","Type":"ContainerDied","Data":"3b614d8bb1878bd5316855217a6adbbfc0ef9e40bcc8ff8d8ac0849d190d68b2"} Nov 25 09:37:58 crc kubenswrapper[4776]: I1125 09:37:58.914372 4776 scope.go:117] "RemoveContainer" containerID="383abbe443a8a6d6f5db562463adde88955bbf8ed71205948f00e006b7cb0907" Nov 25 09:37:58 crc kubenswrapper[4776]: I1125 09:37:58.941289 4776 scope.go:117] "RemoveContainer" containerID="c306aa65939f55c56847a7a50dae2f80f7a148edc1e1d2a29c9705c9a1801a6b" Nov 25 09:37:58 crc kubenswrapper[4776]: I1125 09:37:58.953594 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7zjbg"] Nov 25 09:37:58 crc kubenswrapper[4776]: I1125 09:37:58.957423 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7zjbg"] Nov 25 09:37:58 crc kubenswrapper[4776]: I1125 09:37:58.976245 4776 scope.go:117] "RemoveContainer" containerID="b28fd0add8a7a444972dd0aeb0209e26603c8bbcfa9a2365d5a0b165a3f963fd" Nov 25 09:37:58 crc kubenswrapper[4776]: E1125 09:37:58.976719 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b28fd0add8a7a444972dd0aeb0209e26603c8bbcfa9a2365d5a0b165a3f963fd\": container with ID starting with b28fd0add8a7a444972dd0aeb0209e26603c8bbcfa9a2365d5a0b165a3f963fd not found: ID does not exist" containerID="b28fd0add8a7a444972dd0aeb0209e26603c8bbcfa9a2365d5a0b165a3f963fd" Nov 25 09:37:58 crc kubenswrapper[4776]: I1125 09:37:58.976761 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b28fd0add8a7a444972dd0aeb0209e26603c8bbcfa9a2365d5a0b165a3f963fd"} err="failed to get container status \"b28fd0add8a7a444972dd0aeb0209e26603c8bbcfa9a2365d5a0b165a3f963fd\": rpc error: code = NotFound desc = could not find container \"b28fd0add8a7a444972dd0aeb0209e26603c8bbcfa9a2365d5a0b165a3f963fd\": container with ID starting with b28fd0add8a7a444972dd0aeb0209e26603c8bbcfa9a2365d5a0b165a3f963fd not found: ID does not exist" Nov 25 09:37:58 crc kubenswrapper[4776]: I1125 09:37:58.976791 4776 scope.go:117] "RemoveContainer" containerID="383abbe443a8a6d6f5db562463adde88955bbf8ed71205948f00e006b7cb0907" Nov 25 09:37:58 crc kubenswrapper[4776]: E1125 09:37:58.977394 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"383abbe443a8a6d6f5db562463adde88955bbf8ed71205948f00e006b7cb0907\": container with ID starting with 383abbe443a8a6d6f5db562463adde88955bbf8ed71205948f00e006b7cb0907 not found: ID does not exist" containerID="383abbe443a8a6d6f5db562463adde88955bbf8ed71205948f00e006b7cb0907" Nov 25 09:37:58 crc kubenswrapper[4776]: I1125 09:37:58.977433 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"383abbe443a8a6d6f5db562463adde88955bbf8ed71205948f00e006b7cb0907"} err="failed to get container status \"383abbe443a8a6d6f5db562463adde88955bbf8ed71205948f00e006b7cb0907\": rpc error: code = NotFound desc = could not find container \"383abbe443a8a6d6f5db562463adde88955bbf8ed71205948f00e006b7cb0907\": container with ID starting with 383abbe443a8a6d6f5db562463adde88955bbf8ed71205948f00e006b7cb0907 not found: ID does not exist" Nov 25 09:37:58 crc kubenswrapper[4776]: I1125 09:37:58.977462 4776 scope.go:117] "RemoveContainer" containerID="c306aa65939f55c56847a7a50dae2f80f7a148edc1e1d2a29c9705c9a1801a6b" Nov 25 09:37:58 crc kubenswrapper[4776]: E1125 09:37:58.977829 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c306aa65939f55c56847a7a50dae2f80f7a148edc1e1d2a29c9705c9a1801a6b\": container with ID starting with c306aa65939f55c56847a7a50dae2f80f7a148edc1e1d2a29c9705c9a1801a6b not found: ID does not exist" containerID="c306aa65939f55c56847a7a50dae2f80f7a148edc1e1d2a29c9705c9a1801a6b" Nov 25 09:37:58 crc kubenswrapper[4776]: I1125 09:37:58.977941 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c306aa65939f55c56847a7a50dae2f80f7a148edc1e1d2a29c9705c9a1801a6b"} err="failed to get container status \"c306aa65939f55c56847a7a50dae2f80f7a148edc1e1d2a29c9705c9a1801a6b\": rpc error: code = NotFound desc = could not find container \"c306aa65939f55c56847a7a50dae2f80f7a148edc1e1d2a29c9705c9a1801a6b\": container with ID starting with c306aa65939f55c56847a7a50dae2f80f7a148edc1e1d2a29c9705c9a1801a6b not found: ID does not exist" Nov 25 09:37:59 crc kubenswrapper[4776]: I1125 09:37:59.669040 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ad7db17-c8be-4679-952c-14503eba40b6" path="/var/lib/kubelet/pods/5ad7db17-c8be-4679-952c-14503eba40b6/volumes" Nov 25 09:37:59 crc kubenswrapper[4776]: I1125 09:37:59.907657 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jnbfd" event={"ID":"f473caf4-4336-45bc-97f0-b38e6109198e","Type":"ContainerStarted","Data":"72b55da5dc7a663054b7784a8191794e545d6949a8be6d5e4c4354f1252d127f"} Nov 25 09:38:01 crc kubenswrapper[4776]: I1125 09:38:01.124303 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fr2rf" Nov 25 09:38:01 crc kubenswrapper[4776]: I1125 09:38:01.124376 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fr2rf" Nov 25 09:38:01 crc kubenswrapper[4776]: I1125 09:38:01.200517 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fr2rf" Nov 25 09:38:01 crc kubenswrapper[4776]: I1125 09:38:01.221252 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jnbfd" podStartSLOduration=3.796345726 podStartE2EDuration="6.221230114s" podCreationTimestamp="2025-11-25 09:37:55 +0000 UTC" firstStartedPulling="2025-11-25 09:37:56.881377221 +0000 UTC m=+821.922436784" lastFinishedPulling="2025-11-25 09:37:59.306261609 +0000 UTC m=+824.347321172" observedRunningTime="2025-11-25 09:37:59.92816141 +0000 UTC m=+824.969221003" watchObservedRunningTime="2025-11-25 09:38:01.221230114 +0000 UTC m=+826.262289667" Nov 25 09:38:01 crc kubenswrapper[4776]: I1125 09:38:01.433617 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq"] Nov 25 09:38:01 crc kubenswrapper[4776]: E1125 09:38:01.433831 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ad7db17-c8be-4679-952c-14503eba40b6" containerName="registry-server" Nov 25 09:38:01 crc kubenswrapper[4776]: I1125 09:38:01.433842 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ad7db17-c8be-4679-952c-14503eba40b6" containerName="registry-server" Nov 25 09:38:01 crc kubenswrapper[4776]: E1125 09:38:01.433854 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ad7db17-c8be-4679-952c-14503eba40b6" containerName="extract-content" Nov 25 09:38:01 crc kubenswrapper[4776]: I1125 09:38:01.433861 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ad7db17-c8be-4679-952c-14503eba40b6" containerName="extract-content" Nov 25 09:38:01 crc kubenswrapper[4776]: E1125 09:38:01.433869 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ad7db17-c8be-4679-952c-14503eba40b6" containerName="extract-utilities" Nov 25 09:38:01 crc kubenswrapper[4776]: I1125 09:38:01.433875 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ad7db17-c8be-4679-952c-14503eba40b6" containerName="extract-utilities" Nov 25 09:38:01 crc kubenswrapper[4776]: I1125 09:38:01.433965 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ad7db17-c8be-4679-952c-14503eba40b6" containerName="registry-server" Nov 25 09:38:01 crc kubenswrapper[4776]: I1125 09:38:01.434737 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq" Nov 25 09:38:01 crc kubenswrapper[4776]: I1125 09:38:01.437037 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 25 09:38:01 crc kubenswrapper[4776]: I1125 09:38:01.442090 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq"] Nov 25 09:38:01 crc kubenswrapper[4776]: I1125 09:38:01.561688 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c7cdbd4e-245e-4b7a-890d-c7695d445cb1-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq\" (UID: \"c7cdbd4e-245e-4b7a-890d-c7695d445cb1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq" Nov 25 09:38:01 crc kubenswrapper[4776]: I1125 09:38:01.561900 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw4xx\" (UniqueName: \"kubernetes.io/projected/c7cdbd4e-245e-4b7a-890d-c7695d445cb1-kube-api-access-kw4xx\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq\" (UID: \"c7cdbd4e-245e-4b7a-890d-c7695d445cb1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq" Nov 25 09:38:01 crc kubenswrapper[4776]: I1125 09:38:01.561965 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c7cdbd4e-245e-4b7a-890d-c7695d445cb1-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq\" (UID: \"c7cdbd4e-245e-4b7a-890d-c7695d445cb1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq" Nov 25 09:38:01 crc kubenswrapper[4776]: I1125 09:38:01.663346 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c7cdbd4e-245e-4b7a-890d-c7695d445cb1-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq\" (UID: \"c7cdbd4e-245e-4b7a-890d-c7695d445cb1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq" Nov 25 09:38:01 crc kubenswrapper[4776]: I1125 09:38:01.663425 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c7cdbd4e-245e-4b7a-890d-c7695d445cb1-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq\" (UID: \"c7cdbd4e-245e-4b7a-890d-c7695d445cb1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq" Nov 25 09:38:01 crc kubenswrapper[4776]: I1125 09:38:01.663476 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw4xx\" (UniqueName: \"kubernetes.io/projected/c7cdbd4e-245e-4b7a-890d-c7695d445cb1-kube-api-access-kw4xx\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq\" (UID: \"c7cdbd4e-245e-4b7a-890d-c7695d445cb1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq" Nov 25 09:38:01 crc kubenswrapper[4776]: I1125 09:38:01.663949 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c7cdbd4e-245e-4b7a-890d-c7695d445cb1-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq\" (UID: \"c7cdbd4e-245e-4b7a-890d-c7695d445cb1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq" Nov 25 09:38:01 crc kubenswrapper[4776]: I1125 09:38:01.664096 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c7cdbd4e-245e-4b7a-890d-c7695d445cb1-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq\" (UID: \"c7cdbd4e-245e-4b7a-890d-c7695d445cb1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq" Nov 25 09:38:01 crc kubenswrapper[4776]: I1125 09:38:01.691745 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw4xx\" (UniqueName: \"kubernetes.io/projected/c7cdbd4e-245e-4b7a-890d-c7695d445cb1-kube-api-access-kw4xx\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq\" (UID: \"c7cdbd4e-245e-4b7a-890d-c7695d445cb1\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq" Nov 25 09:38:01 crc kubenswrapper[4776]: I1125 09:38:01.756003 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq" Nov 25 09:38:01 crc kubenswrapper[4776]: I1125 09:38:01.958476 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fr2rf" Nov 25 09:38:02 crc kubenswrapper[4776]: I1125 09:38:02.148333 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq"] Nov 25 09:38:02 crc kubenswrapper[4776]: W1125 09:38:02.159107 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7cdbd4e_245e_4b7a_890d_c7695d445cb1.slice/crio-ca2113e7366ee92152eae53fad6b41cc1edd8ce8b2a241f31921d52a978860c9 WatchSource:0}: Error finding container ca2113e7366ee92152eae53fad6b41cc1edd8ce8b2a241f31921d52a978860c9: Status 404 returned error can't find the container with id ca2113e7366ee92152eae53fad6b41cc1edd8ce8b2a241f31921d52a978860c9 Nov 25 09:38:02 crc kubenswrapper[4776]: I1125 09:38:02.924250 4776 generic.go:334] "Generic (PLEG): container finished" podID="c7cdbd4e-245e-4b7a-890d-c7695d445cb1" containerID="f7877fe381ef415f30433ef59d98d183e0234baa52759ddd4c62dc4583fdcec0" exitCode=0 Nov 25 09:38:02 crc kubenswrapper[4776]: I1125 09:38:02.924332 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq" event={"ID":"c7cdbd4e-245e-4b7a-890d-c7695d445cb1","Type":"ContainerDied","Data":"f7877fe381ef415f30433ef59d98d183e0234baa52759ddd4c62dc4583fdcec0"} Nov 25 09:38:02 crc kubenswrapper[4776]: I1125 09:38:02.924368 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq" event={"ID":"c7cdbd4e-245e-4b7a-890d-c7695d445cb1","Type":"ContainerStarted","Data":"ca2113e7366ee92152eae53fad6b41cc1edd8ce8b2a241f31921d52a978860c9"} Nov 25 09:38:04 crc kubenswrapper[4776]: I1125 09:38:04.817381 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-bhgv2" podUID="04fe29b5-d8c6-46c5-9d1f-76005e225fd6" containerName="console" containerID="cri-o://cf7f59f46490d2b9142f335ae37337163611607d48c7ed1849cc2986f0939532" gracePeriod=15 Nov 25 09:38:04 crc kubenswrapper[4776]: I1125 09:38:04.939628 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-bhgv2_04fe29b5-d8c6-46c5-9d1f-76005e225fd6/console/0.log" Nov 25 09:38:04 crc kubenswrapper[4776]: I1125 09:38:04.940081 4776 generic.go:334] "Generic (PLEG): container finished" podID="04fe29b5-d8c6-46c5-9d1f-76005e225fd6" containerID="cf7f59f46490d2b9142f335ae37337163611607d48c7ed1849cc2986f0939532" exitCode=2 Nov 25 09:38:04 crc kubenswrapper[4776]: I1125 09:38:04.940121 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-bhgv2" event={"ID":"04fe29b5-d8c6-46c5-9d1f-76005e225fd6","Type":"ContainerDied","Data":"cf7f59f46490d2b9142f335ae37337163611607d48c7ed1849cc2986f0939532"} Nov 25 09:38:05 crc kubenswrapper[4776]: I1125 09:38:05.728664 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jnbfd" Nov 25 09:38:05 crc kubenswrapper[4776]: I1125 09:38:05.728725 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jnbfd" Nov 25 09:38:05 crc kubenswrapper[4776]: I1125 09:38:05.773187 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jnbfd" Nov 25 09:38:05 crc kubenswrapper[4776]: I1125 09:38:05.948600 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-bhgv2_04fe29b5-d8c6-46c5-9d1f-76005e225fd6/console/0.log" Nov 25 09:38:05 crc kubenswrapper[4776]: I1125 09:38:05.948688 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-bhgv2" event={"ID":"04fe29b5-d8c6-46c5-9d1f-76005e225fd6","Type":"ContainerDied","Data":"7822ef037f518dffb97dd42c557b0f98dbd1ff31fc53a6a167d11c58d44cdd27"} Nov 25 09:38:05 crc kubenswrapper[4776]: I1125 09:38:05.948729 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7822ef037f518dffb97dd42c557b0f98dbd1ff31fc53a6a167d11c58d44cdd27" Nov 25 09:38:05 crc kubenswrapper[4776]: I1125 09:38:05.992600 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jnbfd" Nov 25 09:38:05 crc kubenswrapper[4776]: I1125 09:38:05.997833 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-bhgv2_04fe29b5-d8c6-46c5-9d1f-76005e225fd6/console/0.log" Nov 25 09:38:05 crc kubenswrapper[4776]: I1125 09:38:05.997893 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.133870 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksp8p\" (UniqueName: \"kubernetes.io/projected/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-kube-api-access-ksp8p\") pod \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\" (UID: \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\") " Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.133947 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-trusted-ca-bundle\") pod \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\" (UID: \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\") " Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.133999 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-service-ca\") pod \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\" (UID: \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\") " Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.134059 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-oauth-serving-cert\") pod \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\" (UID: \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\") " Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.134128 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-console-serving-cert\") pod \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\" (UID: \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\") " Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.134170 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-console-oauth-config\") pod \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\" (UID: \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\") " Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.134203 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-console-config\") pod \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\" (UID: \"04fe29b5-d8c6-46c5-9d1f-76005e225fd6\") " Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.135299 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-service-ca" (OuterVolumeSpecName: "service-ca") pod "04fe29b5-d8c6-46c5-9d1f-76005e225fd6" (UID: "04fe29b5-d8c6-46c5-9d1f-76005e225fd6"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.135362 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "04fe29b5-d8c6-46c5-9d1f-76005e225fd6" (UID: "04fe29b5-d8c6-46c5-9d1f-76005e225fd6"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.135414 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "04fe29b5-d8c6-46c5-9d1f-76005e225fd6" (UID: "04fe29b5-d8c6-46c5-9d1f-76005e225fd6"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.135373 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-console-config" (OuterVolumeSpecName: "console-config") pod "04fe29b5-d8c6-46c5-9d1f-76005e225fd6" (UID: "04fe29b5-d8c6-46c5-9d1f-76005e225fd6"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.142286 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "04fe29b5-d8c6-46c5-9d1f-76005e225fd6" (UID: "04fe29b5-d8c6-46c5-9d1f-76005e225fd6"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.142319 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-kube-api-access-ksp8p" (OuterVolumeSpecName: "kube-api-access-ksp8p") pod "04fe29b5-d8c6-46c5-9d1f-76005e225fd6" (UID: "04fe29b5-d8c6-46c5-9d1f-76005e225fd6"). InnerVolumeSpecName "kube-api-access-ksp8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.143432 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "04fe29b5-d8c6-46c5-9d1f-76005e225fd6" (UID: "04fe29b5-d8c6-46c5-9d1f-76005e225fd6"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.194015 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fr2rf"] Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.194561 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fr2rf" podUID="c063c53a-8226-40c0-b803-93dc46813de0" containerName="registry-server" containerID="cri-o://650a5808155499b9b55b014659cef8831638ec359dff2ad758fa5706a58060c3" gracePeriod=2 Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.235844 4776 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.235895 4776 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.235908 4776 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.235919 4776 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.235932 4776 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.235944 4776 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-console-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.235954 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksp8p\" (UniqueName: \"kubernetes.io/projected/04fe29b5-d8c6-46c5-9d1f-76005e225fd6-kube-api-access-ksp8p\") on node \"crc\" DevicePath \"\"" Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.648803 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fr2rf" Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.741340 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c063c53a-8226-40c0-b803-93dc46813de0-catalog-content\") pod \"c063c53a-8226-40c0-b803-93dc46813de0\" (UID: \"c063c53a-8226-40c0-b803-93dc46813de0\") " Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.741458 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c063c53a-8226-40c0-b803-93dc46813de0-utilities\") pod \"c063c53a-8226-40c0-b803-93dc46813de0\" (UID: \"c063c53a-8226-40c0-b803-93dc46813de0\") " Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.741481 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqhpp\" (UniqueName: \"kubernetes.io/projected/c063c53a-8226-40c0-b803-93dc46813de0-kube-api-access-tqhpp\") pod \"c063c53a-8226-40c0-b803-93dc46813de0\" (UID: \"c063c53a-8226-40c0-b803-93dc46813de0\") " Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.742286 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c063c53a-8226-40c0-b803-93dc46813de0-utilities" (OuterVolumeSpecName: "utilities") pod "c063c53a-8226-40c0-b803-93dc46813de0" (UID: "c063c53a-8226-40c0-b803-93dc46813de0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.744750 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c063c53a-8226-40c0-b803-93dc46813de0-kube-api-access-tqhpp" (OuterVolumeSpecName: "kube-api-access-tqhpp") pod "c063c53a-8226-40c0-b803-93dc46813de0" (UID: "c063c53a-8226-40c0-b803-93dc46813de0"). InnerVolumeSpecName "kube-api-access-tqhpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.757173 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c063c53a-8226-40c0-b803-93dc46813de0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c063c53a-8226-40c0-b803-93dc46813de0" (UID: "c063c53a-8226-40c0-b803-93dc46813de0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.842514 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqhpp\" (UniqueName: \"kubernetes.io/projected/c063c53a-8226-40c0-b803-93dc46813de0-kube-api-access-tqhpp\") on node \"crc\" DevicePath \"\"" Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.842551 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c063c53a-8226-40c0-b803-93dc46813de0-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.842563 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c063c53a-8226-40c0-b803-93dc46813de0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.960995 4776 generic.go:334] "Generic (PLEG): container finished" podID="c063c53a-8226-40c0-b803-93dc46813de0" containerID="650a5808155499b9b55b014659cef8831638ec359dff2ad758fa5706a58060c3" exitCode=0 Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.961140 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fr2rf" event={"ID":"c063c53a-8226-40c0-b803-93dc46813de0","Type":"ContainerDied","Data":"650a5808155499b9b55b014659cef8831638ec359dff2ad758fa5706a58060c3"} Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.961171 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fr2rf" event={"ID":"c063c53a-8226-40c0-b803-93dc46813de0","Type":"ContainerDied","Data":"e059683a702b428f701ef83254e77035cf655bdbcc3a060eb71e6fc108807833"} Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.961241 4776 scope.go:117] "RemoveContainer" containerID="650a5808155499b9b55b014659cef8831638ec359dff2ad758fa5706a58060c3" Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.961767 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fr2rf" Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.967317 4776 generic.go:334] "Generic (PLEG): container finished" podID="c7cdbd4e-245e-4b7a-890d-c7695d445cb1" containerID="884e07ea7cedb5c9724f605b16c97af6fb269b6da2ca123fe8e41f3b75e1c489" exitCode=0 Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.968828 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq" event={"ID":"c7cdbd4e-245e-4b7a-890d-c7695d445cb1","Type":"ContainerDied","Data":"884e07ea7cedb5c9724f605b16c97af6fb269b6da2ca123fe8e41f3b75e1c489"} Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.968896 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-bhgv2" Nov 25 09:38:06 crc kubenswrapper[4776]: I1125 09:38:06.993297 4776 scope.go:117] "RemoveContainer" containerID="6c2dc7f617dc095d4ce901b20196daeca158dedd2272b216a5c373e4299b3829" Nov 25 09:38:07 crc kubenswrapper[4776]: I1125 09:38:07.011893 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-bhgv2"] Nov 25 09:38:07 crc kubenswrapper[4776]: I1125 09:38:07.012529 4776 scope.go:117] "RemoveContainer" containerID="6af6bf2186be59fab74f999b94bc4919c28ae0212927f58cfdfa00d46325250a" Nov 25 09:38:07 crc kubenswrapper[4776]: I1125 09:38:07.021152 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-bhgv2"] Nov 25 09:38:07 crc kubenswrapper[4776]: I1125 09:38:07.031914 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fr2rf"] Nov 25 09:38:07 crc kubenswrapper[4776]: I1125 09:38:07.036852 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fr2rf"] Nov 25 09:38:07 crc kubenswrapper[4776]: I1125 09:38:07.059130 4776 scope.go:117] "RemoveContainer" containerID="650a5808155499b9b55b014659cef8831638ec359dff2ad758fa5706a58060c3" Nov 25 09:38:07 crc kubenswrapper[4776]: E1125 09:38:07.059778 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"650a5808155499b9b55b014659cef8831638ec359dff2ad758fa5706a58060c3\": container with ID starting with 650a5808155499b9b55b014659cef8831638ec359dff2ad758fa5706a58060c3 not found: ID does not exist" containerID="650a5808155499b9b55b014659cef8831638ec359dff2ad758fa5706a58060c3" Nov 25 09:38:07 crc kubenswrapper[4776]: I1125 09:38:07.059846 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"650a5808155499b9b55b014659cef8831638ec359dff2ad758fa5706a58060c3"} err="failed to get container status \"650a5808155499b9b55b014659cef8831638ec359dff2ad758fa5706a58060c3\": rpc error: code = NotFound desc = could not find container \"650a5808155499b9b55b014659cef8831638ec359dff2ad758fa5706a58060c3\": container with ID starting with 650a5808155499b9b55b014659cef8831638ec359dff2ad758fa5706a58060c3 not found: ID does not exist" Nov 25 09:38:07 crc kubenswrapper[4776]: I1125 09:38:07.059874 4776 scope.go:117] "RemoveContainer" containerID="6c2dc7f617dc095d4ce901b20196daeca158dedd2272b216a5c373e4299b3829" Nov 25 09:38:07 crc kubenswrapper[4776]: E1125 09:38:07.060253 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c2dc7f617dc095d4ce901b20196daeca158dedd2272b216a5c373e4299b3829\": container with ID starting with 6c2dc7f617dc095d4ce901b20196daeca158dedd2272b216a5c373e4299b3829 not found: ID does not exist" containerID="6c2dc7f617dc095d4ce901b20196daeca158dedd2272b216a5c373e4299b3829" Nov 25 09:38:07 crc kubenswrapper[4776]: I1125 09:38:07.060290 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c2dc7f617dc095d4ce901b20196daeca158dedd2272b216a5c373e4299b3829"} err="failed to get container status \"6c2dc7f617dc095d4ce901b20196daeca158dedd2272b216a5c373e4299b3829\": rpc error: code = NotFound desc = could not find container \"6c2dc7f617dc095d4ce901b20196daeca158dedd2272b216a5c373e4299b3829\": container with ID starting with 6c2dc7f617dc095d4ce901b20196daeca158dedd2272b216a5c373e4299b3829 not found: ID does not exist" Nov 25 09:38:07 crc kubenswrapper[4776]: I1125 09:38:07.060303 4776 scope.go:117] "RemoveContainer" containerID="6af6bf2186be59fab74f999b94bc4919c28ae0212927f58cfdfa00d46325250a" Nov 25 09:38:07 crc kubenswrapper[4776]: E1125 09:38:07.060648 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6af6bf2186be59fab74f999b94bc4919c28ae0212927f58cfdfa00d46325250a\": container with ID starting with 6af6bf2186be59fab74f999b94bc4919c28ae0212927f58cfdfa00d46325250a not found: ID does not exist" containerID="6af6bf2186be59fab74f999b94bc4919c28ae0212927f58cfdfa00d46325250a" Nov 25 09:38:07 crc kubenswrapper[4776]: I1125 09:38:07.060680 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6af6bf2186be59fab74f999b94bc4919c28ae0212927f58cfdfa00d46325250a"} err="failed to get container status \"6af6bf2186be59fab74f999b94bc4919c28ae0212927f58cfdfa00d46325250a\": rpc error: code = NotFound desc = could not find container \"6af6bf2186be59fab74f999b94bc4919c28ae0212927f58cfdfa00d46325250a\": container with ID starting with 6af6bf2186be59fab74f999b94bc4919c28ae0212927f58cfdfa00d46325250a not found: ID does not exist" Nov 25 09:38:07 crc kubenswrapper[4776]: I1125 09:38:07.669965 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04fe29b5-d8c6-46c5-9d1f-76005e225fd6" path="/var/lib/kubelet/pods/04fe29b5-d8c6-46c5-9d1f-76005e225fd6/volumes" Nov 25 09:38:07 crc kubenswrapper[4776]: I1125 09:38:07.670681 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c063c53a-8226-40c0-b803-93dc46813de0" path="/var/lib/kubelet/pods/c063c53a-8226-40c0-b803-93dc46813de0/volumes" Nov 25 09:38:07 crc kubenswrapper[4776]: I1125 09:38:07.976229 4776 generic.go:334] "Generic (PLEG): container finished" podID="c7cdbd4e-245e-4b7a-890d-c7695d445cb1" containerID="c9bf218d15e53842436b3c8714e5f588ce4847adbbadc1c459d73e8e6717ae92" exitCode=0 Nov 25 09:38:07 crc kubenswrapper[4776]: I1125 09:38:07.976261 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq" event={"ID":"c7cdbd4e-245e-4b7a-890d-c7695d445cb1","Type":"ContainerDied","Data":"c9bf218d15e53842436b3c8714e5f588ce4847adbbadc1c459d73e8e6717ae92"} Nov 25 09:38:09 crc kubenswrapper[4776]: I1125 09:38:09.216280 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq" Nov 25 09:38:09 crc kubenswrapper[4776]: I1125 09:38:09.274041 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c7cdbd4e-245e-4b7a-890d-c7695d445cb1-util\") pod \"c7cdbd4e-245e-4b7a-890d-c7695d445cb1\" (UID: \"c7cdbd4e-245e-4b7a-890d-c7695d445cb1\") " Nov 25 09:38:09 crc kubenswrapper[4776]: I1125 09:38:09.274194 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kw4xx\" (UniqueName: \"kubernetes.io/projected/c7cdbd4e-245e-4b7a-890d-c7695d445cb1-kube-api-access-kw4xx\") pod \"c7cdbd4e-245e-4b7a-890d-c7695d445cb1\" (UID: \"c7cdbd4e-245e-4b7a-890d-c7695d445cb1\") " Nov 25 09:38:09 crc kubenswrapper[4776]: I1125 09:38:09.274237 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c7cdbd4e-245e-4b7a-890d-c7695d445cb1-bundle\") pod \"c7cdbd4e-245e-4b7a-890d-c7695d445cb1\" (UID: \"c7cdbd4e-245e-4b7a-890d-c7695d445cb1\") " Nov 25 09:38:09 crc kubenswrapper[4776]: I1125 09:38:09.275833 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7cdbd4e-245e-4b7a-890d-c7695d445cb1-bundle" (OuterVolumeSpecName: "bundle") pod "c7cdbd4e-245e-4b7a-890d-c7695d445cb1" (UID: "c7cdbd4e-245e-4b7a-890d-c7695d445cb1"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:38:09 crc kubenswrapper[4776]: I1125 09:38:09.278866 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7cdbd4e-245e-4b7a-890d-c7695d445cb1-kube-api-access-kw4xx" (OuterVolumeSpecName: "kube-api-access-kw4xx") pod "c7cdbd4e-245e-4b7a-890d-c7695d445cb1" (UID: "c7cdbd4e-245e-4b7a-890d-c7695d445cb1"). InnerVolumeSpecName "kube-api-access-kw4xx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:38:09 crc kubenswrapper[4776]: I1125 09:38:09.293618 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7cdbd4e-245e-4b7a-890d-c7695d445cb1-util" (OuterVolumeSpecName: "util") pod "c7cdbd4e-245e-4b7a-890d-c7695d445cb1" (UID: "c7cdbd4e-245e-4b7a-890d-c7695d445cb1"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:38:09 crc kubenswrapper[4776]: I1125 09:38:09.375791 4776 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c7cdbd4e-245e-4b7a-890d-c7695d445cb1-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:38:09 crc kubenswrapper[4776]: I1125 09:38:09.375824 4776 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c7cdbd4e-245e-4b7a-890d-c7695d445cb1-util\") on node \"crc\" DevicePath \"\"" Nov 25 09:38:09 crc kubenswrapper[4776]: I1125 09:38:09.375838 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kw4xx\" (UniqueName: \"kubernetes.io/projected/c7cdbd4e-245e-4b7a-890d-c7695d445cb1-kube-api-access-kw4xx\") on node \"crc\" DevicePath \"\"" Nov 25 09:38:09 crc kubenswrapper[4776]: I1125 09:38:09.797289 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jnbfd"] Nov 25 09:38:09 crc kubenswrapper[4776]: I1125 09:38:09.797690 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jnbfd" podUID="f473caf4-4336-45bc-97f0-b38e6109198e" containerName="registry-server" containerID="cri-o://72b55da5dc7a663054b7784a8191794e545d6949a8be6d5e4c4354f1252d127f" gracePeriod=2 Nov 25 09:38:09 crc kubenswrapper[4776]: I1125 09:38:09.989231 4776 generic.go:334] "Generic (PLEG): container finished" podID="f473caf4-4336-45bc-97f0-b38e6109198e" containerID="72b55da5dc7a663054b7784a8191794e545d6949a8be6d5e4c4354f1252d127f" exitCode=0 Nov 25 09:38:09 crc kubenswrapper[4776]: I1125 09:38:09.989312 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jnbfd" event={"ID":"f473caf4-4336-45bc-97f0-b38e6109198e","Type":"ContainerDied","Data":"72b55da5dc7a663054b7784a8191794e545d6949a8be6d5e4c4354f1252d127f"} Nov 25 09:38:09 crc kubenswrapper[4776]: I1125 09:38:09.991320 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq" event={"ID":"c7cdbd4e-245e-4b7a-890d-c7695d445cb1","Type":"ContainerDied","Data":"ca2113e7366ee92152eae53fad6b41cc1edd8ce8b2a241f31921d52a978860c9"} Nov 25 09:38:09 crc kubenswrapper[4776]: I1125 09:38:09.991357 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca2113e7366ee92152eae53fad6b41cc1edd8ce8b2a241f31921d52a978860c9" Nov 25 09:38:09 crc kubenswrapper[4776]: I1125 09:38:09.991413 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq" Nov 25 09:38:10 crc kubenswrapper[4776]: I1125 09:38:10.662968 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jnbfd" Nov 25 09:38:10 crc kubenswrapper[4776]: I1125 09:38:10.792102 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f473caf4-4336-45bc-97f0-b38e6109198e-utilities\") pod \"f473caf4-4336-45bc-97f0-b38e6109198e\" (UID: \"f473caf4-4336-45bc-97f0-b38e6109198e\") " Nov 25 09:38:10 crc kubenswrapper[4776]: I1125 09:38:10.792229 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxjnk\" (UniqueName: \"kubernetes.io/projected/f473caf4-4336-45bc-97f0-b38e6109198e-kube-api-access-hxjnk\") pod \"f473caf4-4336-45bc-97f0-b38e6109198e\" (UID: \"f473caf4-4336-45bc-97f0-b38e6109198e\") " Nov 25 09:38:10 crc kubenswrapper[4776]: I1125 09:38:10.792298 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f473caf4-4336-45bc-97f0-b38e6109198e-catalog-content\") pod \"f473caf4-4336-45bc-97f0-b38e6109198e\" (UID: \"f473caf4-4336-45bc-97f0-b38e6109198e\") " Nov 25 09:38:10 crc kubenswrapper[4776]: I1125 09:38:10.793708 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f473caf4-4336-45bc-97f0-b38e6109198e-utilities" (OuterVolumeSpecName: "utilities") pod "f473caf4-4336-45bc-97f0-b38e6109198e" (UID: "f473caf4-4336-45bc-97f0-b38e6109198e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:38:10 crc kubenswrapper[4776]: I1125 09:38:10.795249 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f473caf4-4336-45bc-97f0-b38e6109198e-kube-api-access-hxjnk" (OuterVolumeSpecName: "kube-api-access-hxjnk") pod "f473caf4-4336-45bc-97f0-b38e6109198e" (UID: "f473caf4-4336-45bc-97f0-b38e6109198e"). InnerVolumeSpecName "kube-api-access-hxjnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:38:10 crc kubenswrapper[4776]: I1125 09:38:10.838188 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f473caf4-4336-45bc-97f0-b38e6109198e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f473caf4-4336-45bc-97f0-b38e6109198e" (UID: "f473caf4-4336-45bc-97f0-b38e6109198e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:38:10 crc kubenswrapper[4776]: I1125 09:38:10.893763 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f473caf4-4336-45bc-97f0-b38e6109198e-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:38:10 crc kubenswrapper[4776]: I1125 09:38:10.893816 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxjnk\" (UniqueName: \"kubernetes.io/projected/f473caf4-4336-45bc-97f0-b38e6109198e-kube-api-access-hxjnk\") on node \"crc\" DevicePath \"\"" Nov 25 09:38:10 crc kubenswrapper[4776]: I1125 09:38:10.893827 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f473caf4-4336-45bc-97f0-b38e6109198e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:38:10 crc kubenswrapper[4776]: I1125 09:38:10.998088 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jnbfd" event={"ID":"f473caf4-4336-45bc-97f0-b38e6109198e","Type":"ContainerDied","Data":"7efedd7760396923e2970399f27c25f3b4d6aa1e25a426fd30f92d62906296cc"} Nov 25 09:38:10 crc kubenswrapper[4776]: I1125 09:38:10.998151 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jnbfd" Nov 25 09:38:10 crc kubenswrapper[4776]: I1125 09:38:10.998173 4776 scope.go:117] "RemoveContainer" containerID="72b55da5dc7a663054b7784a8191794e545d6949a8be6d5e4c4354f1252d127f" Nov 25 09:38:11 crc kubenswrapper[4776]: I1125 09:38:11.028294 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jnbfd"] Nov 25 09:38:11 crc kubenswrapper[4776]: I1125 09:38:11.032420 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jnbfd"] Nov 25 09:38:11 crc kubenswrapper[4776]: I1125 09:38:11.036556 4776 scope.go:117] "RemoveContainer" containerID="3b614d8bb1878bd5316855217a6adbbfc0ef9e40bcc8ff8d8ac0849d190d68b2" Nov 25 09:38:11 crc kubenswrapper[4776]: I1125 09:38:11.052770 4776 scope.go:117] "RemoveContainer" containerID="5848fcb8743291ac041828d178f8a474d57a052155ded6d8c35e96de7725a2df" Nov 25 09:38:11 crc kubenswrapper[4776]: I1125 09:38:11.671192 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f473caf4-4336-45bc-97f0-b38e6109198e" path="/var/lib/kubelet/pods/f473caf4-4336-45bc-97f0-b38e6109198e/volumes" Nov 25 09:38:15 crc kubenswrapper[4776]: I1125 09:38:15.901115 4776 scope.go:117] "RemoveContainer" containerID="cf7f59f46490d2b9142f335ae37337163611607d48c7ed1849cc2986f0939532" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.841289 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6bc7ff74-znqxl"] Nov 25 09:38:16 crc kubenswrapper[4776]: E1125 09:38:16.841692 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c063c53a-8226-40c0-b803-93dc46813de0" containerName="extract-utilities" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.841737 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="c063c53a-8226-40c0-b803-93dc46813de0" containerName="extract-utilities" Nov 25 09:38:16 crc kubenswrapper[4776]: E1125 09:38:16.841753 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04fe29b5-d8c6-46c5-9d1f-76005e225fd6" containerName="console" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.841764 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="04fe29b5-d8c6-46c5-9d1f-76005e225fd6" containerName="console" Nov 25 09:38:16 crc kubenswrapper[4776]: E1125 09:38:16.841776 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c063c53a-8226-40c0-b803-93dc46813de0" containerName="extract-content" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.841783 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="c063c53a-8226-40c0-b803-93dc46813de0" containerName="extract-content" Nov 25 09:38:16 crc kubenswrapper[4776]: E1125 09:38:16.841796 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f473caf4-4336-45bc-97f0-b38e6109198e" containerName="extract-content" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.841805 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="f473caf4-4336-45bc-97f0-b38e6109198e" containerName="extract-content" Nov 25 09:38:16 crc kubenswrapper[4776]: E1125 09:38:16.841817 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f473caf4-4336-45bc-97f0-b38e6109198e" containerName="registry-server" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.841825 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="f473caf4-4336-45bc-97f0-b38e6109198e" containerName="registry-server" Nov 25 09:38:16 crc kubenswrapper[4776]: E1125 09:38:16.841836 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7cdbd4e-245e-4b7a-890d-c7695d445cb1" containerName="util" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.841843 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7cdbd4e-245e-4b7a-890d-c7695d445cb1" containerName="util" Nov 25 09:38:16 crc kubenswrapper[4776]: E1125 09:38:16.841860 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f473caf4-4336-45bc-97f0-b38e6109198e" containerName="extract-utilities" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.841867 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="f473caf4-4336-45bc-97f0-b38e6109198e" containerName="extract-utilities" Nov 25 09:38:16 crc kubenswrapper[4776]: E1125 09:38:16.841877 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7cdbd4e-245e-4b7a-890d-c7695d445cb1" containerName="pull" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.841882 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7cdbd4e-245e-4b7a-890d-c7695d445cb1" containerName="pull" Nov 25 09:38:16 crc kubenswrapper[4776]: E1125 09:38:16.841891 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c063c53a-8226-40c0-b803-93dc46813de0" containerName="registry-server" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.841896 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="c063c53a-8226-40c0-b803-93dc46813de0" containerName="registry-server" Nov 25 09:38:16 crc kubenswrapper[4776]: E1125 09:38:16.841904 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7cdbd4e-245e-4b7a-890d-c7695d445cb1" containerName="extract" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.841909 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7cdbd4e-245e-4b7a-890d-c7695d445cb1" containerName="extract" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.842006 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="f473caf4-4336-45bc-97f0-b38e6109198e" containerName="registry-server" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.842023 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="04fe29b5-d8c6-46c5-9d1f-76005e225fd6" containerName="console" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.842031 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="c063c53a-8226-40c0-b803-93dc46813de0" containerName="registry-server" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.842038 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7cdbd4e-245e-4b7a-890d-c7695d445cb1" containerName="extract" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.842453 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6bc7ff74-znqxl" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.844388 4776 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.844656 4776 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-bstcg" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.844889 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.845032 4776 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.845213 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.863912 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6bc7ff74-znqxl"] Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.872120 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3700db74-6cd7-4370-8b44-fea7c84052af-apiservice-cert\") pod \"metallb-operator-controller-manager-6bc7ff74-znqxl\" (UID: \"3700db74-6cd7-4370-8b44-fea7c84052af\") " pod="metallb-system/metallb-operator-controller-manager-6bc7ff74-znqxl" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.872226 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3700db74-6cd7-4370-8b44-fea7c84052af-webhook-cert\") pod \"metallb-operator-controller-manager-6bc7ff74-znqxl\" (UID: \"3700db74-6cd7-4370-8b44-fea7c84052af\") " pod="metallb-system/metallb-operator-controller-manager-6bc7ff74-znqxl" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.872265 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtx49\" (UniqueName: \"kubernetes.io/projected/3700db74-6cd7-4370-8b44-fea7c84052af-kube-api-access-rtx49\") pod \"metallb-operator-controller-manager-6bc7ff74-znqxl\" (UID: \"3700db74-6cd7-4370-8b44-fea7c84052af\") " pod="metallb-system/metallb-operator-controller-manager-6bc7ff74-znqxl" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.974786 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3700db74-6cd7-4370-8b44-fea7c84052af-webhook-cert\") pod \"metallb-operator-controller-manager-6bc7ff74-znqxl\" (UID: \"3700db74-6cd7-4370-8b44-fea7c84052af\") " pod="metallb-system/metallb-operator-controller-manager-6bc7ff74-znqxl" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.974855 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtx49\" (UniqueName: \"kubernetes.io/projected/3700db74-6cd7-4370-8b44-fea7c84052af-kube-api-access-rtx49\") pod \"metallb-operator-controller-manager-6bc7ff74-znqxl\" (UID: \"3700db74-6cd7-4370-8b44-fea7c84052af\") " pod="metallb-system/metallb-operator-controller-manager-6bc7ff74-znqxl" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.974921 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3700db74-6cd7-4370-8b44-fea7c84052af-apiservice-cert\") pod \"metallb-operator-controller-manager-6bc7ff74-znqxl\" (UID: \"3700db74-6cd7-4370-8b44-fea7c84052af\") " pod="metallb-system/metallb-operator-controller-manager-6bc7ff74-znqxl" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.981202 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3700db74-6cd7-4370-8b44-fea7c84052af-apiservice-cert\") pod \"metallb-operator-controller-manager-6bc7ff74-znqxl\" (UID: \"3700db74-6cd7-4370-8b44-fea7c84052af\") " pod="metallb-system/metallb-operator-controller-manager-6bc7ff74-znqxl" Nov 25 09:38:16 crc kubenswrapper[4776]: I1125 09:38:16.989688 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3700db74-6cd7-4370-8b44-fea7c84052af-webhook-cert\") pod \"metallb-operator-controller-manager-6bc7ff74-znqxl\" (UID: \"3700db74-6cd7-4370-8b44-fea7c84052af\") " pod="metallb-system/metallb-operator-controller-manager-6bc7ff74-znqxl" Nov 25 09:38:17 crc kubenswrapper[4776]: I1125 09:38:17.002733 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtx49\" (UniqueName: \"kubernetes.io/projected/3700db74-6cd7-4370-8b44-fea7c84052af-kube-api-access-rtx49\") pod \"metallb-operator-controller-manager-6bc7ff74-znqxl\" (UID: \"3700db74-6cd7-4370-8b44-fea7c84052af\") " pod="metallb-system/metallb-operator-controller-manager-6bc7ff74-znqxl" Nov 25 09:38:17 crc kubenswrapper[4776]: I1125 09:38:17.083314 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5c67c9cb54-vgffr"] Nov 25 09:38:17 crc kubenswrapper[4776]: I1125 09:38:17.084106 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5c67c9cb54-vgffr" Nov 25 09:38:17 crc kubenswrapper[4776]: I1125 09:38:17.086008 4776 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 25 09:38:17 crc kubenswrapper[4776]: I1125 09:38:17.086661 4776 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 25 09:38:17 crc kubenswrapper[4776]: I1125 09:38:17.088459 4776 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-6xh8p" Nov 25 09:38:17 crc kubenswrapper[4776]: I1125 09:38:17.109048 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5c67c9cb54-vgffr"] Nov 25 09:38:17 crc kubenswrapper[4776]: I1125 09:38:17.156505 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6bc7ff74-znqxl" Nov 25 09:38:17 crc kubenswrapper[4776]: I1125 09:38:17.177508 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bggk\" (UniqueName: \"kubernetes.io/projected/12f6e05c-cc4d-4e0c-8a1a-bfe459e58d70-kube-api-access-5bggk\") pod \"metallb-operator-webhook-server-5c67c9cb54-vgffr\" (UID: \"12f6e05c-cc4d-4e0c-8a1a-bfe459e58d70\") " pod="metallb-system/metallb-operator-webhook-server-5c67c9cb54-vgffr" Nov 25 09:38:17 crc kubenswrapper[4776]: I1125 09:38:17.177596 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/12f6e05c-cc4d-4e0c-8a1a-bfe459e58d70-webhook-cert\") pod \"metallb-operator-webhook-server-5c67c9cb54-vgffr\" (UID: \"12f6e05c-cc4d-4e0c-8a1a-bfe459e58d70\") " pod="metallb-system/metallb-operator-webhook-server-5c67c9cb54-vgffr" Nov 25 09:38:17 crc kubenswrapper[4776]: I1125 09:38:17.177629 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/12f6e05c-cc4d-4e0c-8a1a-bfe459e58d70-apiservice-cert\") pod \"metallb-operator-webhook-server-5c67c9cb54-vgffr\" (UID: \"12f6e05c-cc4d-4e0c-8a1a-bfe459e58d70\") " pod="metallb-system/metallb-operator-webhook-server-5c67c9cb54-vgffr" Nov 25 09:38:17 crc kubenswrapper[4776]: I1125 09:38:17.292709 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bggk\" (UniqueName: \"kubernetes.io/projected/12f6e05c-cc4d-4e0c-8a1a-bfe459e58d70-kube-api-access-5bggk\") pod \"metallb-operator-webhook-server-5c67c9cb54-vgffr\" (UID: \"12f6e05c-cc4d-4e0c-8a1a-bfe459e58d70\") " pod="metallb-system/metallb-operator-webhook-server-5c67c9cb54-vgffr" Nov 25 09:38:17 crc kubenswrapper[4776]: I1125 09:38:17.293148 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/12f6e05c-cc4d-4e0c-8a1a-bfe459e58d70-webhook-cert\") pod \"metallb-operator-webhook-server-5c67c9cb54-vgffr\" (UID: \"12f6e05c-cc4d-4e0c-8a1a-bfe459e58d70\") " pod="metallb-system/metallb-operator-webhook-server-5c67c9cb54-vgffr" Nov 25 09:38:17 crc kubenswrapper[4776]: I1125 09:38:17.293184 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/12f6e05c-cc4d-4e0c-8a1a-bfe459e58d70-apiservice-cert\") pod \"metallb-operator-webhook-server-5c67c9cb54-vgffr\" (UID: \"12f6e05c-cc4d-4e0c-8a1a-bfe459e58d70\") " pod="metallb-system/metallb-operator-webhook-server-5c67c9cb54-vgffr" Nov 25 09:38:17 crc kubenswrapper[4776]: I1125 09:38:17.300579 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/12f6e05c-cc4d-4e0c-8a1a-bfe459e58d70-webhook-cert\") pod \"metallb-operator-webhook-server-5c67c9cb54-vgffr\" (UID: \"12f6e05c-cc4d-4e0c-8a1a-bfe459e58d70\") " pod="metallb-system/metallb-operator-webhook-server-5c67c9cb54-vgffr" Nov 25 09:38:17 crc kubenswrapper[4776]: I1125 09:38:17.312761 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/12f6e05c-cc4d-4e0c-8a1a-bfe459e58d70-apiservice-cert\") pod \"metallb-operator-webhook-server-5c67c9cb54-vgffr\" (UID: \"12f6e05c-cc4d-4e0c-8a1a-bfe459e58d70\") " pod="metallb-system/metallb-operator-webhook-server-5c67c9cb54-vgffr" Nov 25 09:38:17 crc kubenswrapper[4776]: I1125 09:38:17.318982 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bggk\" (UniqueName: \"kubernetes.io/projected/12f6e05c-cc4d-4e0c-8a1a-bfe459e58d70-kube-api-access-5bggk\") pod \"metallb-operator-webhook-server-5c67c9cb54-vgffr\" (UID: \"12f6e05c-cc4d-4e0c-8a1a-bfe459e58d70\") " pod="metallb-system/metallb-operator-webhook-server-5c67c9cb54-vgffr" Nov 25 09:38:17 crc kubenswrapper[4776]: I1125 09:38:17.400687 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5c67c9cb54-vgffr" Nov 25 09:38:17 crc kubenswrapper[4776]: I1125 09:38:17.631568 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6bc7ff74-znqxl"] Nov 25 09:38:17 crc kubenswrapper[4776]: W1125 09:38:17.638006 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3700db74_6cd7_4370_8b44_fea7c84052af.slice/crio-0567fb5dabe899832693dbcb3efdeee39cdcdce58f6698318e155cd7080b562b WatchSource:0}: Error finding container 0567fb5dabe899832693dbcb3efdeee39cdcdce58f6698318e155cd7080b562b: Status 404 returned error can't find the container with id 0567fb5dabe899832693dbcb3efdeee39cdcdce58f6698318e155cd7080b562b Nov 25 09:38:17 crc kubenswrapper[4776]: I1125 09:38:17.801617 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5c67c9cb54-vgffr"] Nov 25 09:38:17 crc kubenswrapper[4776]: W1125 09:38:17.811586 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12f6e05c_cc4d_4e0c_8a1a_bfe459e58d70.slice/crio-707e45af9560e60c4edc18236ff027fb8aee51f73b6793e924fab0de6eb106be WatchSource:0}: Error finding container 707e45af9560e60c4edc18236ff027fb8aee51f73b6793e924fab0de6eb106be: Status 404 returned error can't find the container with id 707e45af9560e60c4edc18236ff027fb8aee51f73b6793e924fab0de6eb106be Nov 25 09:38:17 crc kubenswrapper[4776]: I1125 09:38:17.818534 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:38:17 crc kubenswrapper[4776]: I1125 09:38:17.818580 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:38:17 crc kubenswrapper[4776]: I1125 09:38:17.818625 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 09:38:17 crc kubenswrapper[4776]: I1125 09:38:17.819152 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dc4f980209957fcb2e8415b8fd0ea1aba1b60952959f82bbdfd9dd9f2991ea9c"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 09:38:17 crc kubenswrapper[4776]: I1125 09:38:17.819231 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://dc4f980209957fcb2e8415b8fd0ea1aba1b60952959f82bbdfd9dd9f2991ea9c" gracePeriod=600 Nov 25 09:38:18 crc kubenswrapper[4776]: I1125 09:38:18.036705 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6bc7ff74-znqxl" event={"ID":"3700db74-6cd7-4370-8b44-fea7c84052af","Type":"ContainerStarted","Data":"0567fb5dabe899832693dbcb3efdeee39cdcdce58f6698318e155cd7080b562b"} Nov 25 09:38:18 crc kubenswrapper[4776]: I1125 09:38:18.038448 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="dc4f980209957fcb2e8415b8fd0ea1aba1b60952959f82bbdfd9dd9f2991ea9c" exitCode=0 Nov 25 09:38:18 crc kubenswrapper[4776]: I1125 09:38:18.038503 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"dc4f980209957fcb2e8415b8fd0ea1aba1b60952959f82bbdfd9dd9f2991ea9c"} Nov 25 09:38:18 crc kubenswrapper[4776]: I1125 09:38:18.038539 4776 scope.go:117] "RemoveContainer" containerID="5ae1ab323d058608828b15daefde2387d82c4579f51d09882253c9347334d6f7" Nov 25 09:38:18 crc kubenswrapper[4776]: I1125 09:38:18.039849 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5c67c9cb54-vgffr" event={"ID":"12f6e05c-cc4d-4e0c-8a1a-bfe459e58d70","Type":"ContainerStarted","Data":"707e45af9560e60c4edc18236ff027fb8aee51f73b6793e924fab0de6eb106be"} Nov 25 09:38:19 crc kubenswrapper[4776]: I1125 09:38:19.048274 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"6132d3cd67d80b93c28c890778e005f6af8e158a1aed030b0062e3edb3f984f8"} Nov 25 09:38:21 crc kubenswrapper[4776]: I1125 09:38:21.060677 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6bc7ff74-znqxl" event={"ID":"3700db74-6cd7-4370-8b44-fea7c84052af","Type":"ContainerStarted","Data":"31e13b1142c400270f5f0e6cf471a88e7326a956f014ca98b12b3d94cd17d8e1"} Nov 25 09:38:21 crc kubenswrapper[4776]: I1125 09:38:21.061295 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6bc7ff74-znqxl" Nov 25 09:38:21 crc kubenswrapper[4776]: I1125 09:38:21.080927 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6bc7ff74-znqxl" podStartSLOduration=2.320592324 podStartE2EDuration="5.080905695s" podCreationTimestamp="2025-11-25 09:38:16 +0000 UTC" firstStartedPulling="2025-11-25 09:38:17.640674894 +0000 UTC m=+842.681734447" lastFinishedPulling="2025-11-25 09:38:20.400988265 +0000 UTC m=+845.442047818" observedRunningTime="2025-11-25 09:38:21.079629313 +0000 UTC m=+846.120688886" watchObservedRunningTime="2025-11-25 09:38:21.080905695 +0000 UTC m=+846.121965248" Nov 25 09:38:25 crc kubenswrapper[4776]: I1125 09:38:25.082177 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5c67c9cb54-vgffr" event={"ID":"12f6e05c-cc4d-4e0c-8a1a-bfe459e58d70","Type":"ContainerStarted","Data":"97d7951aa2ae3e6d94f183fb4753cbf7505ff0da832ba5732e3ca84cf5aae28f"} Nov 25 09:38:25 crc kubenswrapper[4776]: I1125 09:38:25.082857 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5c67c9cb54-vgffr" Nov 25 09:38:25 crc kubenswrapper[4776]: I1125 09:38:25.102766 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5c67c9cb54-vgffr" podStartSLOduration=1.535601821 podStartE2EDuration="8.102750921s" podCreationTimestamp="2025-11-25 09:38:17 +0000 UTC" firstStartedPulling="2025-11-25 09:38:17.814502849 +0000 UTC m=+842.855562402" lastFinishedPulling="2025-11-25 09:38:24.381651949 +0000 UTC m=+849.422711502" observedRunningTime="2025-11-25 09:38:25.102253219 +0000 UTC m=+850.143312792" watchObservedRunningTime="2025-11-25 09:38:25.102750921 +0000 UTC m=+850.143810474" Nov 25 09:38:37 crc kubenswrapper[4776]: I1125 09:38:37.406275 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5c67c9cb54-vgffr" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.159746 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6bc7ff74-znqxl" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.779726 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-hj78r"] Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.782142 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-hj78r" Nov 25 09:38:57 crc kubenswrapper[4776]: W1125 09:38:57.784256 4776 reflector.go:561] object-"metallb-system"/"frr-startup": failed to list *v1.ConfigMap: configmaps "frr-startup" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Nov 25 09:38:57 crc kubenswrapper[4776]: E1125 09:38:57.784302 4776 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"frr-startup\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"frr-startup\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 09:38:57 crc kubenswrapper[4776]: W1125 09:38:57.784568 4776 reflector.go:561] object-"metallb-system"/"frr-k8s-daemon-dockercfg-rl6jz": failed to list *v1.Secret: secrets "frr-k8s-daemon-dockercfg-rl6jz" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Nov 25 09:38:57 crc kubenswrapper[4776]: E1125 09:38:57.784599 4776 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"frr-k8s-daemon-dockercfg-rl6jz\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"frr-k8s-daemon-dockercfg-rl6jz\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 09:38:57 crc kubenswrapper[4776]: W1125 09:38:57.784790 4776 reflector.go:561] object-"metallb-system"/"frr-k8s-certs-secret": failed to list *v1.Secret: secrets "frr-k8s-certs-secret" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Nov 25 09:38:57 crc kubenswrapper[4776]: E1125 09:38:57.784804 4776 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"frr-k8s-certs-secret\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"frr-k8s-certs-secret\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.785049 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-dfbc5"] Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.785835 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-dfbc5" Nov 25 09:38:57 crc kubenswrapper[4776]: W1125 09:38:57.786862 4776 reflector.go:561] object-"metallb-system"/"frr-k8s-webhook-server-cert": failed to list *v1.Secret: secrets "frr-k8s-webhook-server-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Nov 25 09:38:57 crc kubenswrapper[4776]: E1125 09:38:57.786980 4776 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"frr-k8s-webhook-server-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"frr-k8s-webhook-server-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.792719 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-dfbc5"] Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.893865 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6bd80dc6-3a88-459e-bfea-f2fc22aed477-metrics\") pod \"frr-k8s-hj78r\" (UID: \"6bd80dc6-3a88-459e-bfea-f2fc22aed477\") " pod="metallb-system/frr-k8s-hj78r" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.894167 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6bd80dc6-3a88-459e-bfea-f2fc22aed477-frr-startup\") pod \"frr-k8s-hj78r\" (UID: \"6bd80dc6-3a88-459e-bfea-f2fc22aed477\") " pod="metallb-system/frr-k8s-hj78r" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.894184 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/14a3e0bf-32be-4251-9d91-9e41b841d9e2-cert\") pod \"frr-k8s-webhook-server-6998585d5-dfbc5\" (UID: \"14a3e0bf-32be-4251-9d91-9e41b841d9e2\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-dfbc5" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.894220 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6bd80dc6-3a88-459e-bfea-f2fc22aed477-frr-conf\") pod \"frr-k8s-hj78r\" (UID: \"6bd80dc6-3a88-459e-bfea-f2fc22aed477\") " pod="metallb-system/frr-k8s-hj78r" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.894235 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6bd80dc6-3a88-459e-bfea-f2fc22aed477-metrics-certs\") pod \"frr-k8s-hj78r\" (UID: \"6bd80dc6-3a88-459e-bfea-f2fc22aed477\") " pod="metallb-system/frr-k8s-hj78r" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.894262 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ztf9\" (UniqueName: \"kubernetes.io/projected/14a3e0bf-32be-4251-9d91-9e41b841d9e2-kube-api-access-4ztf9\") pod \"frr-k8s-webhook-server-6998585d5-dfbc5\" (UID: \"14a3e0bf-32be-4251-9d91-9e41b841d9e2\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-dfbc5" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.894296 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fgtm\" (UniqueName: \"kubernetes.io/projected/6bd80dc6-3a88-459e-bfea-f2fc22aed477-kube-api-access-6fgtm\") pod \"frr-k8s-hj78r\" (UID: \"6bd80dc6-3a88-459e-bfea-f2fc22aed477\") " pod="metallb-system/frr-k8s-hj78r" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.894448 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6bd80dc6-3a88-459e-bfea-f2fc22aed477-frr-sockets\") pod \"frr-k8s-hj78r\" (UID: \"6bd80dc6-3a88-459e-bfea-f2fc22aed477\") " pod="metallb-system/frr-k8s-hj78r" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.894512 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6bd80dc6-3a88-459e-bfea-f2fc22aed477-reloader\") pod \"frr-k8s-hj78r\" (UID: \"6bd80dc6-3a88-459e-bfea-f2fc22aed477\") " pod="metallb-system/frr-k8s-hj78r" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.902943 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-p9fdc"] Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.904009 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-p9fdc" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.909536 4776 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.909732 4776 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.909769 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.909774 4776 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-q8rr5" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.930935 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-jvm2p"] Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.932007 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-jvm2p" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.934660 4776 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.953209 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-jvm2p"] Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.995582 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/294e21df-2814-43a8-975c-e0beb19b3e03-metrics-certs\") pod \"speaker-p9fdc\" (UID: \"294e21df-2814-43a8-975c-e0beb19b3e03\") " pod="metallb-system/speaker-p9fdc" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.995649 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6bd80dc6-3a88-459e-bfea-f2fc22aed477-frr-sockets\") pod \"frr-k8s-hj78r\" (UID: \"6bd80dc6-3a88-459e-bfea-f2fc22aed477\") " pod="metallb-system/frr-k8s-hj78r" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.995676 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6bd80dc6-3a88-459e-bfea-f2fc22aed477-reloader\") pod \"frr-k8s-hj78r\" (UID: \"6bd80dc6-3a88-459e-bfea-f2fc22aed477\") " pod="metallb-system/frr-k8s-hj78r" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.995736 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6bd80dc6-3a88-459e-bfea-f2fc22aed477-metrics\") pod \"frr-k8s-hj78r\" (UID: \"6bd80dc6-3a88-459e-bfea-f2fc22aed477\") " pod="metallb-system/frr-k8s-hj78r" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.995790 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6bd80dc6-3a88-459e-bfea-f2fc22aed477-frr-startup\") pod \"frr-k8s-hj78r\" (UID: \"6bd80dc6-3a88-459e-bfea-f2fc22aed477\") " pod="metallb-system/frr-k8s-hj78r" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.995818 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/14a3e0bf-32be-4251-9d91-9e41b841d9e2-cert\") pod \"frr-k8s-webhook-server-6998585d5-dfbc5\" (UID: \"14a3e0bf-32be-4251-9d91-9e41b841d9e2\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-dfbc5" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.995872 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6bd80dc6-3a88-459e-bfea-f2fc22aed477-frr-conf\") pod \"frr-k8s-hj78r\" (UID: \"6bd80dc6-3a88-459e-bfea-f2fc22aed477\") " pod="metallb-system/frr-k8s-hj78r" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.995897 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6bd80dc6-3a88-459e-bfea-f2fc22aed477-metrics-certs\") pod \"frr-k8s-hj78r\" (UID: \"6bd80dc6-3a88-459e-bfea-f2fc22aed477\") " pod="metallb-system/frr-k8s-hj78r" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.995926 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/294e21df-2814-43a8-975c-e0beb19b3e03-memberlist\") pod \"speaker-p9fdc\" (UID: \"294e21df-2814-43a8-975c-e0beb19b3e03\") " pod="metallb-system/speaker-p9fdc" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.995986 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ztf9\" (UniqueName: \"kubernetes.io/projected/14a3e0bf-32be-4251-9d91-9e41b841d9e2-kube-api-access-4ztf9\") pod \"frr-k8s-webhook-server-6998585d5-dfbc5\" (UID: \"14a3e0bf-32be-4251-9d91-9e41b841d9e2\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-dfbc5" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.996011 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fgtm\" (UniqueName: \"kubernetes.io/projected/6bd80dc6-3a88-459e-bfea-f2fc22aed477-kube-api-access-6fgtm\") pod \"frr-k8s-hj78r\" (UID: \"6bd80dc6-3a88-459e-bfea-f2fc22aed477\") " pod="metallb-system/frr-k8s-hj78r" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.996055 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/294e21df-2814-43a8-975c-e0beb19b3e03-metallb-excludel2\") pod \"speaker-p9fdc\" (UID: \"294e21df-2814-43a8-975c-e0beb19b3e03\") " pod="metallb-system/speaker-p9fdc" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.996619 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/6bd80dc6-3a88-459e-bfea-f2fc22aed477-frr-conf\") pod \"frr-k8s-hj78r\" (UID: \"6bd80dc6-3a88-459e-bfea-f2fc22aed477\") " pod="metallb-system/frr-k8s-hj78r" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.996928 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/6bd80dc6-3a88-459e-bfea-f2fc22aed477-frr-sockets\") pod \"frr-k8s-hj78r\" (UID: \"6bd80dc6-3a88-459e-bfea-f2fc22aed477\") " pod="metallb-system/frr-k8s-hj78r" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.997036 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/6bd80dc6-3a88-459e-bfea-f2fc22aed477-metrics\") pod \"frr-k8s-hj78r\" (UID: \"6bd80dc6-3a88-459e-bfea-f2fc22aed477\") " pod="metallb-system/frr-k8s-hj78r" Nov 25 09:38:57 crc kubenswrapper[4776]: I1125 09:38:57.997370 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/6bd80dc6-3a88-459e-bfea-f2fc22aed477-reloader\") pod \"frr-k8s-hj78r\" (UID: \"6bd80dc6-3a88-459e-bfea-f2fc22aed477\") " pod="metallb-system/frr-k8s-hj78r" Nov 25 09:38:58 crc kubenswrapper[4776]: I1125 09:38:58.021765 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fgtm\" (UniqueName: \"kubernetes.io/projected/6bd80dc6-3a88-459e-bfea-f2fc22aed477-kube-api-access-6fgtm\") pod \"frr-k8s-hj78r\" (UID: \"6bd80dc6-3a88-459e-bfea-f2fc22aed477\") " pod="metallb-system/frr-k8s-hj78r" Nov 25 09:38:58 crc kubenswrapper[4776]: I1125 09:38:58.021796 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ztf9\" (UniqueName: \"kubernetes.io/projected/14a3e0bf-32be-4251-9d91-9e41b841d9e2-kube-api-access-4ztf9\") pod \"frr-k8s-webhook-server-6998585d5-dfbc5\" (UID: \"14a3e0bf-32be-4251-9d91-9e41b841d9e2\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-dfbc5" Nov 25 09:38:58 crc kubenswrapper[4776]: I1125 09:38:58.097890 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kghph\" (UniqueName: \"kubernetes.io/projected/294e21df-2814-43a8-975c-e0beb19b3e03-kube-api-access-kghph\") pod \"speaker-p9fdc\" (UID: \"294e21df-2814-43a8-975c-e0beb19b3e03\") " pod="metallb-system/speaker-p9fdc" Nov 25 09:38:58 crc kubenswrapper[4776]: I1125 09:38:58.097993 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d74cede2-6a31-4173-a6af-87669c0804bf-metrics-certs\") pod \"controller-6c7b4b5f48-jvm2p\" (UID: \"d74cede2-6a31-4173-a6af-87669c0804bf\") " pod="metallb-system/controller-6c7b4b5f48-jvm2p" Nov 25 09:38:58 crc kubenswrapper[4776]: I1125 09:38:58.098028 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j8nn\" (UniqueName: \"kubernetes.io/projected/d74cede2-6a31-4173-a6af-87669c0804bf-kube-api-access-2j8nn\") pod \"controller-6c7b4b5f48-jvm2p\" (UID: \"d74cede2-6a31-4173-a6af-87669c0804bf\") " pod="metallb-system/controller-6c7b4b5f48-jvm2p" Nov 25 09:38:58 crc kubenswrapper[4776]: I1125 09:38:58.098209 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/294e21df-2814-43a8-975c-e0beb19b3e03-memberlist\") pod \"speaker-p9fdc\" (UID: \"294e21df-2814-43a8-975c-e0beb19b3e03\") " pod="metallb-system/speaker-p9fdc" Nov 25 09:38:58 crc kubenswrapper[4776]: I1125 09:38:58.098239 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d74cede2-6a31-4173-a6af-87669c0804bf-cert\") pod \"controller-6c7b4b5f48-jvm2p\" (UID: \"d74cede2-6a31-4173-a6af-87669c0804bf\") " pod="metallb-system/controller-6c7b4b5f48-jvm2p" Nov 25 09:38:58 crc kubenswrapper[4776]: I1125 09:38:58.098260 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/294e21df-2814-43a8-975c-e0beb19b3e03-metallb-excludel2\") pod \"speaker-p9fdc\" (UID: \"294e21df-2814-43a8-975c-e0beb19b3e03\") " pod="metallb-system/speaker-p9fdc" Nov 25 09:38:58 crc kubenswrapper[4776]: I1125 09:38:58.098304 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/294e21df-2814-43a8-975c-e0beb19b3e03-metrics-certs\") pod \"speaker-p9fdc\" (UID: \"294e21df-2814-43a8-975c-e0beb19b3e03\") " pod="metallb-system/speaker-p9fdc" Nov 25 09:38:58 crc kubenswrapper[4776]: E1125 09:38:58.098495 4776 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 25 09:38:58 crc kubenswrapper[4776]: E1125 09:38:58.098563 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/294e21df-2814-43a8-975c-e0beb19b3e03-memberlist podName:294e21df-2814-43a8-975c-e0beb19b3e03 nodeName:}" failed. No retries permitted until 2025-11-25 09:38:58.598542774 +0000 UTC m=+883.639602327 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/294e21df-2814-43a8-975c-e0beb19b3e03-memberlist") pod "speaker-p9fdc" (UID: "294e21df-2814-43a8-975c-e0beb19b3e03") : secret "metallb-memberlist" not found Nov 25 09:38:58 crc kubenswrapper[4776]: I1125 09:38:58.099458 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/294e21df-2814-43a8-975c-e0beb19b3e03-metallb-excludel2\") pod \"speaker-p9fdc\" (UID: \"294e21df-2814-43a8-975c-e0beb19b3e03\") " pod="metallb-system/speaker-p9fdc" Nov 25 09:38:58 crc kubenswrapper[4776]: I1125 09:38:58.100984 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/294e21df-2814-43a8-975c-e0beb19b3e03-metrics-certs\") pod \"speaker-p9fdc\" (UID: \"294e21df-2814-43a8-975c-e0beb19b3e03\") " pod="metallb-system/speaker-p9fdc" Nov 25 09:38:58 crc kubenswrapper[4776]: I1125 09:38:58.198920 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d74cede2-6a31-4173-a6af-87669c0804bf-cert\") pod \"controller-6c7b4b5f48-jvm2p\" (UID: \"d74cede2-6a31-4173-a6af-87669c0804bf\") " pod="metallb-system/controller-6c7b4b5f48-jvm2p" Nov 25 09:38:58 crc kubenswrapper[4776]: I1125 09:38:58.199014 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kghph\" (UniqueName: \"kubernetes.io/projected/294e21df-2814-43a8-975c-e0beb19b3e03-kube-api-access-kghph\") pod \"speaker-p9fdc\" (UID: \"294e21df-2814-43a8-975c-e0beb19b3e03\") " pod="metallb-system/speaker-p9fdc" Nov 25 09:38:58 crc kubenswrapper[4776]: I1125 09:38:58.199042 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d74cede2-6a31-4173-a6af-87669c0804bf-metrics-certs\") pod \"controller-6c7b4b5f48-jvm2p\" (UID: \"d74cede2-6a31-4173-a6af-87669c0804bf\") " pod="metallb-system/controller-6c7b4b5f48-jvm2p" Nov 25 09:38:58 crc kubenswrapper[4776]: I1125 09:38:58.199084 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j8nn\" (UniqueName: \"kubernetes.io/projected/d74cede2-6a31-4173-a6af-87669c0804bf-kube-api-access-2j8nn\") pod \"controller-6c7b4b5f48-jvm2p\" (UID: \"d74cede2-6a31-4173-a6af-87669c0804bf\") " pod="metallb-system/controller-6c7b4b5f48-jvm2p" Nov 25 09:38:58 crc kubenswrapper[4776]: I1125 09:38:58.202585 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d74cede2-6a31-4173-a6af-87669c0804bf-metrics-certs\") pod \"controller-6c7b4b5f48-jvm2p\" (UID: \"d74cede2-6a31-4173-a6af-87669c0804bf\") " pod="metallb-system/controller-6c7b4b5f48-jvm2p" Nov 25 09:38:58 crc kubenswrapper[4776]: I1125 09:38:58.209603 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d74cede2-6a31-4173-a6af-87669c0804bf-cert\") pod \"controller-6c7b4b5f48-jvm2p\" (UID: \"d74cede2-6a31-4173-a6af-87669c0804bf\") " pod="metallb-system/controller-6c7b4b5f48-jvm2p" Nov 25 09:38:58 crc kubenswrapper[4776]: I1125 09:38:58.217759 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kghph\" (UniqueName: \"kubernetes.io/projected/294e21df-2814-43a8-975c-e0beb19b3e03-kube-api-access-kghph\") pod \"speaker-p9fdc\" (UID: \"294e21df-2814-43a8-975c-e0beb19b3e03\") " pod="metallb-system/speaker-p9fdc" Nov 25 09:38:58 crc kubenswrapper[4776]: I1125 09:38:58.233562 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j8nn\" (UniqueName: \"kubernetes.io/projected/d74cede2-6a31-4173-a6af-87669c0804bf-kube-api-access-2j8nn\") pod \"controller-6c7b4b5f48-jvm2p\" (UID: \"d74cede2-6a31-4173-a6af-87669c0804bf\") " pod="metallb-system/controller-6c7b4b5f48-jvm2p" Nov 25 09:38:58 crc kubenswrapper[4776]: I1125 09:38:58.248844 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-jvm2p" Nov 25 09:38:58 crc kubenswrapper[4776]: I1125 09:38:58.605342 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/294e21df-2814-43a8-975c-e0beb19b3e03-memberlist\") pod \"speaker-p9fdc\" (UID: \"294e21df-2814-43a8-975c-e0beb19b3e03\") " pod="metallb-system/speaker-p9fdc" Nov 25 09:38:58 crc kubenswrapper[4776]: E1125 09:38:58.605500 4776 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 25 09:38:58 crc kubenswrapper[4776]: E1125 09:38:58.605840 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/294e21df-2814-43a8-975c-e0beb19b3e03-memberlist podName:294e21df-2814-43a8-975c-e0beb19b3e03 nodeName:}" failed. No retries permitted until 2025-11-25 09:38:59.60581189 +0000 UTC m=+884.646871483 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/294e21df-2814-43a8-975c-e0beb19b3e03-memberlist") pod "speaker-p9fdc" (UID: "294e21df-2814-43a8-975c-e0beb19b3e03") : secret "metallb-memberlist" not found Nov 25 09:38:58 crc kubenswrapper[4776]: I1125 09:38:58.721664 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-jvm2p"] Nov 25 09:38:58 crc kubenswrapper[4776]: I1125 09:38:58.958272 4776 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 25 09:38:58 crc kubenswrapper[4776]: I1125 09:38:58.965571 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6bd80dc6-3a88-459e-bfea-f2fc22aed477-metrics-certs\") pod \"frr-k8s-hj78r\" (UID: \"6bd80dc6-3a88-459e-bfea-f2fc22aed477\") " pod="metallb-system/frr-k8s-hj78r" Nov 25 09:38:58 crc kubenswrapper[4776]: E1125 09:38:58.996920 4776 configmap.go:193] Couldn't get configMap metallb-system/frr-startup: failed to sync configmap cache: timed out waiting for the condition Nov 25 09:38:58 crc kubenswrapper[4776]: E1125 09:38:58.996950 4776 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: failed to sync secret cache: timed out waiting for the condition Nov 25 09:38:58 crc kubenswrapper[4776]: E1125 09:38:58.997029 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6bd80dc6-3a88-459e-bfea-f2fc22aed477-frr-startup podName:6bd80dc6-3a88-459e-bfea-f2fc22aed477 nodeName:}" failed. No retries permitted until 2025-11-25 09:38:59.497005708 +0000 UTC m=+884.538065271 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "frr-startup" (UniqueName: "kubernetes.io/configmap/6bd80dc6-3a88-459e-bfea-f2fc22aed477-frr-startup") pod "frr-k8s-hj78r" (UID: "6bd80dc6-3a88-459e-bfea-f2fc22aed477") : failed to sync configmap cache: timed out waiting for the condition Nov 25 09:38:58 crc kubenswrapper[4776]: E1125 09:38:58.997117 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/14a3e0bf-32be-4251-9d91-9e41b841d9e2-cert podName:14a3e0bf-32be-4251-9d91-9e41b841d9e2 nodeName:}" failed. No retries permitted until 2025-11-25 09:38:59.497057289 +0000 UTC m=+884.538116842 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/14a3e0bf-32be-4251-9d91-9e41b841d9e2-cert") pod "frr-k8s-webhook-server-6998585d5-dfbc5" (UID: "14a3e0bf-32be-4251-9d91-9e41b841d9e2") : failed to sync secret cache: timed out waiting for the condition Nov 25 09:38:59 crc kubenswrapper[4776]: I1125 09:38:59.194767 4776 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-rl6jz" Nov 25 09:38:59 crc kubenswrapper[4776]: I1125 09:38:59.272408 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-jvm2p" event={"ID":"d74cede2-6a31-4173-a6af-87669c0804bf","Type":"ContainerStarted","Data":"01c80a7edfd853efc6f6cd082f7be99463f33f89493d213cdadcb2810c717ed0"} Nov 25 09:38:59 crc kubenswrapper[4776]: I1125 09:38:59.272456 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-jvm2p" event={"ID":"d74cede2-6a31-4173-a6af-87669c0804bf","Type":"ContainerStarted","Data":"2b878e42ef43ac1ca8fd417ff91467135f7856d09493deed4eb1ffaed868fe5c"} Nov 25 09:38:59 crc kubenswrapper[4776]: I1125 09:38:59.272469 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-jvm2p" event={"ID":"d74cede2-6a31-4173-a6af-87669c0804bf","Type":"ContainerStarted","Data":"7aeb58707d01c10c77fd704abd1ab8b3d80d07e47a465d0fd02dbe6fc5bee915"} Nov 25 09:38:59 crc kubenswrapper[4776]: I1125 09:38:59.272615 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-jvm2p" Nov 25 09:38:59 crc kubenswrapper[4776]: I1125 09:38:59.293848 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-jvm2p" podStartSLOduration=2.293822563 podStartE2EDuration="2.293822563s" podCreationTimestamp="2025-11-25 09:38:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:38:59.289433473 +0000 UTC m=+884.330493076" watchObservedRunningTime="2025-11-25 09:38:59.293822563 +0000 UTC m=+884.334882116" Nov 25 09:38:59 crc kubenswrapper[4776]: I1125 09:38:59.328583 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 25 09:38:59 crc kubenswrapper[4776]: I1125 09:38:59.371504 4776 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 25 09:38:59 crc kubenswrapper[4776]: I1125 09:38:59.515650 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6bd80dc6-3a88-459e-bfea-f2fc22aed477-frr-startup\") pod \"frr-k8s-hj78r\" (UID: \"6bd80dc6-3a88-459e-bfea-f2fc22aed477\") " pod="metallb-system/frr-k8s-hj78r" Nov 25 09:38:59 crc kubenswrapper[4776]: I1125 09:38:59.515700 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/14a3e0bf-32be-4251-9d91-9e41b841d9e2-cert\") pod \"frr-k8s-webhook-server-6998585d5-dfbc5\" (UID: \"14a3e0bf-32be-4251-9d91-9e41b841d9e2\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-dfbc5" Nov 25 09:38:59 crc kubenswrapper[4776]: I1125 09:38:59.516515 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/6bd80dc6-3a88-459e-bfea-f2fc22aed477-frr-startup\") pod \"frr-k8s-hj78r\" (UID: \"6bd80dc6-3a88-459e-bfea-f2fc22aed477\") " pod="metallb-system/frr-k8s-hj78r" Nov 25 09:38:59 crc kubenswrapper[4776]: I1125 09:38:59.529719 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/14a3e0bf-32be-4251-9d91-9e41b841d9e2-cert\") pod \"frr-k8s-webhook-server-6998585d5-dfbc5\" (UID: \"14a3e0bf-32be-4251-9d91-9e41b841d9e2\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-dfbc5" Nov 25 09:38:59 crc kubenswrapper[4776]: I1125 09:38:59.606051 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-hj78r" Nov 25 09:38:59 crc kubenswrapper[4776]: I1125 09:38:59.617484 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-dfbc5" Nov 25 09:38:59 crc kubenswrapper[4776]: I1125 09:38:59.619492 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/294e21df-2814-43a8-975c-e0beb19b3e03-memberlist\") pod \"speaker-p9fdc\" (UID: \"294e21df-2814-43a8-975c-e0beb19b3e03\") " pod="metallb-system/speaker-p9fdc" Nov 25 09:38:59 crc kubenswrapper[4776]: I1125 09:38:59.622921 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/294e21df-2814-43a8-975c-e0beb19b3e03-memberlist\") pod \"speaker-p9fdc\" (UID: \"294e21df-2814-43a8-975c-e0beb19b3e03\") " pod="metallb-system/speaker-p9fdc" Nov 25 09:38:59 crc kubenswrapper[4776]: I1125 09:38:59.729258 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-p9fdc" Nov 25 09:39:00 crc kubenswrapper[4776]: I1125 09:39:00.025443 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-dfbc5"] Nov 25 09:39:00 crc kubenswrapper[4776]: W1125 09:39:00.041641 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod14a3e0bf_32be_4251_9d91_9e41b841d9e2.slice/crio-b165f9aca4482659292f1d03f2737513cff36d467cb704c568e13eb73134b281 WatchSource:0}: Error finding container b165f9aca4482659292f1d03f2737513cff36d467cb704c568e13eb73134b281: Status 404 returned error can't find the container with id b165f9aca4482659292f1d03f2737513cff36d467cb704c568e13eb73134b281 Nov 25 09:39:00 crc kubenswrapper[4776]: I1125 09:39:00.279682 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-dfbc5" event={"ID":"14a3e0bf-32be-4251-9d91-9e41b841d9e2","Type":"ContainerStarted","Data":"b165f9aca4482659292f1d03f2737513cff36d467cb704c568e13eb73134b281"} Nov 25 09:39:00 crc kubenswrapper[4776]: I1125 09:39:00.280902 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hj78r" event={"ID":"6bd80dc6-3a88-459e-bfea-f2fc22aed477","Type":"ContainerStarted","Data":"3a81d044a3c8468fb1b9cc259df2b90c388c6780f489c448ff0b8d917c5f9f6f"} Nov 25 09:39:00 crc kubenswrapper[4776]: I1125 09:39:00.282817 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-p9fdc" event={"ID":"294e21df-2814-43a8-975c-e0beb19b3e03","Type":"ContainerStarted","Data":"c3dcc4e11aa6ca7fddbfe28c9b0285be685aea0215f52a6558cb1bb39f1f0279"} Nov 25 09:39:00 crc kubenswrapper[4776]: I1125 09:39:00.282866 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-p9fdc" event={"ID":"294e21df-2814-43a8-975c-e0beb19b3e03","Type":"ContainerStarted","Data":"6aba8dd3a5632e0f2ae84d26aa707417a00b47cd8d5281b27593bccf28a21728"} Nov 25 09:39:00 crc kubenswrapper[4776]: I1125 09:39:00.282879 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-p9fdc" event={"ID":"294e21df-2814-43a8-975c-e0beb19b3e03","Type":"ContainerStarted","Data":"b4d215a47b7952f19345ead6fb7c34ff9b31a2efa845ec71d10569635b2039d0"} Nov 25 09:39:00 crc kubenswrapper[4776]: I1125 09:39:00.283045 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-p9fdc" Nov 25 09:39:00 crc kubenswrapper[4776]: I1125 09:39:00.303845 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-p9fdc" podStartSLOduration=3.30382636 podStartE2EDuration="3.30382636s" podCreationTimestamp="2025-11-25 09:38:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:39:00.302568898 +0000 UTC m=+885.343628471" watchObservedRunningTime="2025-11-25 09:39:00.30382636 +0000 UTC m=+885.344885913" Nov 25 09:39:08 crc kubenswrapper[4776]: I1125 09:39:08.253567 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-jvm2p" Nov 25 09:39:08 crc kubenswrapper[4776]: I1125 09:39:08.332910 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-dfbc5" event={"ID":"14a3e0bf-32be-4251-9d91-9e41b841d9e2","Type":"ContainerStarted","Data":"5d92ccd479cc0974710a223ae1b2bc2e749323d2e9badf22c879ca4ad0dee779"} Nov 25 09:39:09 crc kubenswrapper[4776]: I1125 09:39:09.339629 4776 generic.go:334] "Generic (PLEG): container finished" podID="6bd80dc6-3a88-459e-bfea-f2fc22aed477" containerID="0b179306633a4176039f129e826994e4451a93c16aac5f6eda5404cf644edcf7" exitCode=0 Nov 25 09:39:09 crc kubenswrapper[4776]: I1125 09:39:09.339697 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hj78r" event={"ID":"6bd80dc6-3a88-459e-bfea-f2fc22aed477","Type":"ContainerDied","Data":"0b179306633a4176039f129e826994e4451a93c16aac5f6eda5404cf644edcf7"} Nov 25 09:39:09 crc kubenswrapper[4776]: I1125 09:39:09.339941 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-dfbc5" Nov 25 09:39:09 crc kubenswrapper[4776]: I1125 09:39:09.365193 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-dfbc5" podStartSLOduration=4.39858653 podStartE2EDuration="12.365173341s" podCreationTimestamp="2025-11-25 09:38:57 +0000 UTC" firstStartedPulling="2025-11-25 09:39:00.04554282 +0000 UTC m=+885.086602373" lastFinishedPulling="2025-11-25 09:39:08.012129631 +0000 UTC m=+893.053189184" observedRunningTime="2025-11-25 09:39:09.35872367 +0000 UTC m=+894.399783243" watchObservedRunningTime="2025-11-25 09:39:09.365173341 +0000 UTC m=+894.406232884" Nov 25 09:39:09 crc kubenswrapper[4776]: I1125 09:39:09.733636 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-p9fdc" Nov 25 09:39:10 crc kubenswrapper[4776]: I1125 09:39:10.348079 4776 generic.go:334] "Generic (PLEG): container finished" podID="6bd80dc6-3a88-459e-bfea-f2fc22aed477" containerID="7a2df454c20b7a7c3518a39b0bede03d15bfe7961deee944c3a589d741c314bf" exitCode=0 Nov 25 09:39:10 crc kubenswrapper[4776]: I1125 09:39:10.348135 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hj78r" event={"ID":"6bd80dc6-3a88-459e-bfea-f2fc22aed477","Type":"ContainerDied","Data":"7a2df454c20b7a7c3518a39b0bede03d15bfe7961deee944c3a589d741c314bf"} Nov 25 09:39:11 crc kubenswrapper[4776]: I1125 09:39:11.359171 4776 generic.go:334] "Generic (PLEG): container finished" podID="6bd80dc6-3a88-459e-bfea-f2fc22aed477" containerID="867499489e2e668b3d83be742c39b2239205595121ef26d36b4f2366daa8e6cb" exitCode=0 Nov 25 09:39:11 crc kubenswrapper[4776]: I1125 09:39:11.359243 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hj78r" event={"ID":"6bd80dc6-3a88-459e-bfea-f2fc22aed477","Type":"ContainerDied","Data":"867499489e2e668b3d83be742c39b2239205595121ef26d36b4f2366daa8e6cb"} Nov 25 09:39:11 crc kubenswrapper[4776]: I1125 09:39:11.563918 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s"] Nov 25 09:39:11 crc kubenswrapper[4776]: I1125 09:39:11.564952 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s" Nov 25 09:39:11 crc kubenswrapper[4776]: I1125 09:39:11.569512 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 25 09:39:11 crc kubenswrapper[4776]: I1125 09:39:11.580806 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s"] Nov 25 09:39:11 crc kubenswrapper[4776]: I1125 09:39:11.712037 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ee805461-014c-4400-9310-f2862d366911-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s\" (UID: \"ee805461-014c-4400-9310-f2862d366911\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s" Nov 25 09:39:11 crc kubenswrapper[4776]: I1125 09:39:11.712350 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9m4fn\" (UniqueName: \"kubernetes.io/projected/ee805461-014c-4400-9310-f2862d366911-kube-api-access-9m4fn\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s\" (UID: \"ee805461-014c-4400-9310-f2862d366911\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s" Nov 25 09:39:11 crc kubenswrapper[4776]: I1125 09:39:11.712383 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ee805461-014c-4400-9310-f2862d366911-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s\" (UID: \"ee805461-014c-4400-9310-f2862d366911\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s" Nov 25 09:39:11 crc kubenswrapper[4776]: I1125 09:39:11.813451 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ee805461-014c-4400-9310-f2862d366911-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s\" (UID: \"ee805461-014c-4400-9310-f2862d366911\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s" Nov 25 09:39:11 crc kubenswrapper[4776]: I1125 09:39:11.813658 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9m4fn\" (UniqueName: \"kubernetes.io/projected/ee805461-014c-4400-9310-f2862d366911-kube-api-access-9m4fn\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s\" (UID: \"ee805461-014c-4400-9310-f2862d366911\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s" Nov 25 09:39:11 crc kubenswrapper[4776]: I1125 09:39:11.813701 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ee805461-014c-4400-9310-f2862d366911-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s\" (UID: \"ee805461-014c-4400-9310-f2862d366911\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s" Nov 25 09:39:11 crc kubenswrapper[4776]: I1125 09:39:11.813882 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ee805461-014c-4400-9310-f2862d366911-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s\" (UID: \"ee805461-014c-4400-9310-f2862d366911\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s" Nov 25 09:39:11 crc kubenswrapper[4776]: I1125 09:39:11.814046 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ee805461-014c-4400-9310-f2862d366911-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s\" (UID: \"ee805461-014c-4400-9310-f2862d366911\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s" Nov 25 09:39:11 crc kubenswrapper[4776]: I1125 09:39:11.843001 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9m4fn\" (UniqueName: \"kubernetes.io/projected/ee805461-014c-4400-9310-f2862d366911-kube-api-access-9m4fn\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s\" (UID: \"ee805461-014c-4400-9310-f2862d366911\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s" Nov 25 09:39:11 crc kubenswrapper[4776]: I1125 09:39:11.878826 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s" Nov 25 09:39:12 crc kubenswrapper[4776]: I1125 09:39:12.292922 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s"] Nov 25 09:39:12 crc kubenswrapper[4776]: W1125 09:39:12.301863 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee805461_014c_4400_9310_f2862d366911.slice/crio-b7405b8ae2103c1448e6b8da325f206f2bfa02569c7770cb53f410de26a52f37 WatchSource:0}: Error finding container b7405b8ae2103c1448e6b8da325f206f2bfa02569c7770cb53f410de26a52f37: Status 404 returned error can't find the container with id b7405b8ae2103c1448e6b8da325f206f2bfa02569c7770cb53f410de26a52f37 Nov 25 09:39:12 crc kubenswrapper[4776]: I1125 09:39:12.370797 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hj78r" event={"ID":"6bd80dc6-3a88-459e-bfea-f2fc22aed477","Type":"ContainerStarted","Data":"667ed3e1aa9a006c5d207f90522d70440cf8bf1863ac8969c30fda9ea08afa3e"} Nov 25 09:39:12 crc kubenswrapper[4776]: I1125 09:39:12.370875 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hj78r" event={"ID":"6bd80dc6-3a88-459e-bfea-f2fc22aed477","Type":"ContainerStarted","Data":"06dc488a6bde3b58ab6e6549097bd309abfe6785dc887f758d5a83fbbb149abc"} Nov 25 09:39:12 crc kubenswrapper[4776]: I1125 09:39:12.370892 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hj78r" event={"ID":"6bd80dc6-3a88-459e-bfea-f2fc22aed477","Type":"ContainerStarted","Data":"29e22efe0a015abdda302789a2a3d4274e727548c328be636753bc74782c6e13"} Nov 25 09:39:12 crc kubenswrapper[4776]: I1125 09:39:12.370901 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hj78r" event={"ID":"6bd80dc6-3a88-459e-bfea-f2fc22aed477","Type":"ContainerStarted","Data":"4181e537768e888274399da29649d2238bbbcbd932c625673ae087a01cbac3dc"} Nov 25 09:39:12 crc kubenswrapper[4776]: I1125 09:39:12.373583 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s" event={"ID":"ee805461-014c-4400-9310-f2862d366911","Type":"ContainerStarted","Data":"b7405b8ae2103c1448e6b8da325f206f2bfa02569c7770cb53f410de26a52f37"} Nov 25 09:39:13 crc kubenswrapper[4776]: I1125 09:39:13.383691 4776 generic.go:334] "Generic (PLEG): container finished" podID="ee805461-014c-4400-9310-f2862d366911" containerID="f06019ab0ec74fe9c31ca41bf764fd09da65b1a1c3e1e87f87fb88dd4d22f53c" exitCode=0 Nov 25 09:39:13 crc kubenswrapper[4776]: I1125 09:39:13.383745 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s" event={"ID":"ee805461-014c-4400-9310-f2862d366911","Type":"ContainerDied","Data":"f06019ab0ec74fe9c31ca41bf764fd09da65b1a1c3e1e87f87fb88dd4d22f53c"} Nov 25 09:39:13 crc kubenswrapper[4776]: I1125 09:39:13.391017 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hj78r" event={"ID":"6bd80dc6-3a88-459e-bfea-f2fc22aed477","Type":"ContainerStarted","Data":"626b148df8715bad51fdad99ccf73c84395878c752e8d960be02ff7d2596de89"} Nov 25 09:39:13 crc kubenswrapper[4776]: I1125 09:39:13.391087 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hj78r" event={"ID":"6bd80dc6-3a88-459e-bfea-f2fc22aed477","Type":"ContainerStarted","Data":"751736b276969d13a7e2b61c25a6ec32c2ba86050f9b28c37d4eebf944af2ab4"} Nov 25 09:39:13 crc kubenswrapper[4776]: I1125 09:39:13.391223 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-hj78r" Nov 25 09:39:13 crc kubenswrapper[4776]: I1125 09:39:13.425811 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-hj78r" podStartSLOduration=8.210012416 podStartE2EDuration="16.425794899s" podCreationTimestamp="2025-11-25 09:38:57 +0000 UTC" firstStartedPulling="2025-11-25 09:38:59.755335811 +0000 UTC m=+884.796395364" lastFinishedPulling="2025-11-25 09:39:07.971118294 +0000 UTC m=+893.012177847" observedRunningTime="2025-11-25 09:39:13.424492716 +0000 UTC m=+898.465552269" watchObservedRunningTime="2025-11-25 09:39:13.425794899 +0000 UTC m=+898.466854452" Nov 25 09:39:14 crc kubenswrapper[4776]: I1125 09:39:14.607300 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-hj78r" Nov 25 09:39:14 crc kubenswrapper[4776]: I1125 09:39:14.657371 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-hj78r" Nov 25 09:39:19 crc kubenswrapper[4776]: I1125 09:39:19.622367 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-dfbc5" Nov 25 09:39:20 crc kubenswrapper[4776]: I1125 09:39:20.444989 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s" event={"ID":"ee805461-014c-4400-9310-f2862d366911","Type":"ContainerStarted","Data":"c508e81c21c8a0e162e7f92a3a15b05fef0d017b99c94442b8f3f0c14fef5e60"} Nov 25 09:39:21 crc kubenswrapper[4776]: I1125 09:39:21.452799 4776 generic.go:334] "Generic (PLEG): container finished" podID="ee805461-014c-4400-9310-f2862d366911" containerID="c508e81c21c8a0e162e7f92a3a15b05fef0d017b99c94442b8f3f0c14fef5e60" exitCode=0 Nov 25 09:39:21 crc kubenswrapper[4776]: I1125 09:39:21.452841 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s" event={"ID":"ee805461-014c-4400-9310-f2862d366911","Type":"ContainerDied","Data":"c508e81c21c8a0e162e7f92a3a15b05fef0d017b99c94442b8f3f0c14fef5e60"} Nov 25 09:39:22 crc kubenswrapper[4776]: I1125 09:39:22.463434 4776 generic.go:334] "Generic (PLEG): container finished" podID="ee805461-014c-4400-9310-f2862d366911" containerID="e26ca6cda9b2167c3786b23363ee17ef6e04d2458c2e849432db1c96697fc707" exitCode=0 Nov 25 09:39:22 crc kubenswrapper[4776]: I1125 09:39:22.463736 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s" event={"ID":"ee805461-014c-4400-9310-f2862d366911","Type":"ContainerDied","Data":"e26ca6cda9b2167c3786b23363ee17ef6e04d2458c2e849432db1c96697fc707"} Nov 25 09:39:23 crc kubenswrapper[4776]: I1125 09:39:23.744659 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s" Nov 25 09:39:23 crc kubenswrapper[4776]: I1125 09:39:23.773346 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ee805461-014c-4400-9310-f2862d366911-bundle\") pod \"ee805461-014c-4400-9310-f2862d366911\" (UID: \"ee805461-014c-4400-9310-f2862d366911\") " Nov 25 09:39:23 crc kubenswrapper[4776]: I1125 09:39:23.773397 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9m4fn\" (UniqueName: \"kubernetes.io/projected/ee805461-014c-4400-9310-f2862d366911-kube-api-access-9m4fn\") pod \"ee805461-014c-4400-9310-f2862d366911\" (UID: \"ee805461-014c-4400-9310-f2862d366911\") " Nov 25 09:39:23 crc kubenswrapper[4776]: I1125 09:39:23.773442 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ee805461-014c-4400-9310-f2862d366911-util\") pod \"ee805461-014c-4400-9310-f2862d366911\" (UID: \"ee805461-014c-4400-9310-f2862d366911\") " Nov 25 09:39:23 crc kubenswrapper[4776]: I1125 09:39:23.775156 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee805461-014c-4400-9310-f2862d366911-bundle" (OuterVolumeSpecName: "bundle") pod "ee805461-014c-4400-9310-f2862d366911" (UID: "ee805461-014c-4400-9310-f2862d366911"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:39:23 crc kubenswrapper[4776]: I1125 09:39:23.778833 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee805461-014c-4400-9310-f2862d366911-kube-api-access-9m4fn" (OuterVolumeSpecName: "kube-api-access-9m4fn") pod "ee805461-014c-4400-9310-f2862d366911" (UID: "ee805461-014c-4400-9310-f2862d366911"). InnerVolumeSpecName "kube-api-access-9m4fn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:39:23 crc kubenswrapper[4776]: I1125 09:39:23.789756 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee805461-014c-4400-9310-f2862d366911-util" (OuterVolumeSpecName: "util") pod "ee805461-014c-4400-9310-f2862d366911" (UID: "ee805461-014c-4400-9310-f2862d366911"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:39:23 crc kubenswrapper[4776]: I1125 09:39:23.874329 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9m4fn\" (UniqueName: \"kubernetes.io/projected/ee805461-014c-4400-9310-f2862d366911-kube-api-access-9m4fn\") on node \"crc\" DevicePath \"\"" Nov 25 09:39:23 crc kubenswrapper[4776]: I1125 09:39:23.874366 4776 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ee805461-014c-4400-9310-f2862d366911-util\") on node \"crc\" DevicePath \"\"" Nov 25 09:39:23 crc kubenswrapper[4776]: I1125 09:39:23.874408 4776 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ee805461-014c-4400-9310-f2862d366911-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:39:24 crc kubenswrapper[4776]: I1125 09:39:24.483568 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s" event={"ID":"ee805461-014c-4400-9310-f2862d366911","Type":"ContainerDied","Data":"b7405b8ae2103c1448e6b8da325f206f2bfa02569c7770cb53f410de26a52f37"} Nov 25 09:39:24 crc kubenswrapper[4776]: I1125 09:39:24.483604 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s" Nov 25 09:39:24 crc kubenswrapper[4776]: I1125 09:39:24.483616 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7405b8ae2103c1448e6b8da325f206f2bfa02569c7770cb53f410de26a52f37" Nov 25 09:39:28 crc kubenswrapper[4776]: I1125 09:39:28.928652 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-v9kp7"] Nov 25 09:39:28 crc kubenswrapper[4776]: E1125 09:39:28.929342 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee805461-014c-4400-9310-f2862d366911" containerName="pull" Nov 25 09:39:28 crc kubenswrapper[4776]: I1125 09:39:28.929353 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee805461-014c-4400-9310-f2862d366911" containerName="pull" Nov 25 09:39:28 crc kubenswrapper[4776]: E1125 09:39:28.929371 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee805461-014c-4400-9310-f2862d366911" containerName="extract" Nov 25 09:39:28 crc kubenswrapper[4776]: I1125 09:39:28.929377 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee805461-014c-4400-9310-f2862d366911" containerName="extract" Nov 25 09:39:28 crc kubenswrapper[4776]: E1125 09:39:28.929392 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee805461-014c-4400-9310-f2862d366911" containerName="util" Nov 25 09:39:28 crc kubenswrapper[4776]: I1125 09:39:28.929399 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee805461-014c-4400-9310-f2862d366911" containerName="util" Nov 25 09:39:28 crc kubenswrapper[4776]: I1125 09:39:28.929508 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee805461-014c-4400-9310-f2862d366911" containerName="extract" Nov 25 09:39:28 crc kubenswrapper[4776]: I1125 09:39:28.929863 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-v9kp7" Nov 25 09:39:28 crc kubenswrapper[4776]: I1125 09:39:28.933912 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Nov 25 09:39:28 crc kubenswrapper[4776]: I1125 09:39:28.935277 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Nov 25 09:39:28 crc kubenswrapper[4776]: I1125 09:39:28.949566 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/9d3d23c4-01c0-4ba8-9ba9-0b79f83ada09-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-v9kp7\" (UID: \"9d3d23c4-01c0-4ba8-9ba9-0b79f83ada09\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-v9kp7" Nov 25 09:39:28 crc kubenswrapper[4776]: I1125 09:39:28.949664 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7hfn\" (UniqueName: \"kubernetes.io/projected/9d3d23c4-01c0-4ba8-9ba9-0b79f83ada09-kube-api-access-n7hfn\") pod \"cert-manager-operator-controller-manager-64cf6dff88-v9kp7\" (UID: \"9d3d23c4-01c0-4ba8-9ba9-0b79f83ada09\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-v9kp7" Nov 25 09:39:28 crc kubenswrapper[4776]: I1125 09:39:28.953288 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-v9kp7"] Nov 25 09:39:28 crc kubenswrapper[4776]: I1125 09:39:28.958431 4776 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-nxtns" Nov 25 09:39:29 crc kubenswrapper[4776]: I1125 09:39:29.051034 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7hfn\" (UniqueName: \"kubernetes.io/projected/9d3d23c4-01c0-4ba8-9ba9-0b79f83ada09-kube-api-access-n7hfn\") pod \"cert-manager-operator-controller-manager-64cf6dff88-v9kp7\" (UID: \"9d3d23c4-01c0-4ba8-9ba9-0b79f83ada09\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-v9kp7" Nov 25 09:39:29 crc kubenswrapper[4776]: I1125 09:39:29.051122 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/9d3d23c4-01c0-4ba8-9ba9-0b79f83ada09-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-v9kp7\" (UID: \"9d3d23c4-01c0-4ba8-9ba9-0b79f83ada09\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-v9kp7" Nov 25 09:39:29 crc kubenswrapper[4776]: I1125 09:39:29.051685 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/9d3d23c4-01c0-4ba8-9ba9-0b79f83ada09-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-v9kp7\" (UID: \"9d3d23c4-01c0-4ba8-9ba9-0b79f83ada09\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-v9kp7" Nov 25 09:39:29 crc kubenswrapper[4776]: I1125 09:39:29.070010 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7hfn\" (UniqueName: \"kubernetes.io/projected/9d3d23c4-01c0-4ba8-9ba9-0b79f83ada09-kube-api-access-n7hfn\") pod \"cert-manager-operator-controller-manager-64cf6dff88-v9kp7\" (UID: \"9d3d23c4-01c0-4ba8-9ba9-0b79f83ada09\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-v9kp7" Nov 25 09:39:29 crc kubenswrapper[4776]: I1125 09:39:29.250665 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-v9kp7" Nov 25 09:39:29 crc kubenswrapper[4776]: I1125 09:39:29.472224 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-v9kp7"] Nov 25 09:39:29 crc kubenswrapper[4776]: I1125 09:39:29.516011 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-v9kp7" event={"ID":"9d3d23c4-01c0-4ba8-9ba9-0b79f83ada09","Type":"ContainerStarted","Data":"428341ab4622f2a000b22310eebf2ab003112c331506ba1369d1f338b9636de6"} Nov 25 09:39:29 crc kubenswrapper[4776]: I1125 09:39:29.613146 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-hj78r" Nov 25 09:39:39 crc kubenswrapper[4776]: I1125 09:39:39.593002 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-v9kp7" event={"ID":"9d3d23c4-01c0-4ba8-9ba9-0b79f83ada09","Type":"ContainerStarted","Data":"92b94cf678369d8d48cbdfa0546a3880dc85379edd614dd62c38219312908e9d"} Nov 25 09:39:42 crc kubenswrapper[4776]: I1125 09:39:42.771095 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-v9kp7" podStartSLOduration=5.170978458 podStartE2EDuration="14.771056564s" podCreationTimestamp="2025-11-25 09:39:28 +0000 UTC" firstStartedPulling="2025-11-25 09:39:29.485219151 +0000 UTC m=+914.526278714" lastFinishedPulling="2025-11-25 09:39:39.085297267 +0000 UTC m=+924.126356820" observedRunningTime="2025-11-25 09:39:39.616961963 +0000 UTC m=+924.658021516" watchObservedRunningTime="2025-11-25 09:39:42.771056564 +0000 UTC m=+927.812116117" Nov 25 09:39:42 crc kubenswrapper[4776]: I1125 09:39:42.774145 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-pt9xh"] Nov 25 09:39:42 crc kubenswrapper[4776]: I1125 09:39:42.774828 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-pt9xh" Nov 25 09:39:42 crc kubenswrapper[4776]: I1125 09:39:42.778151 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 25 09:39:42 crc kubenswrapper[4776]: I1125 09:39:42.778421 4776 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-68fb6" Nov 25 09:39:42 crc kubenswrapper[4776]: I1125 09:39:42.779023 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 25 09:39:42 crc kubenswrapper[4776]: I1125 09:39:42.785221 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-pt9xh"] Nov 25 09:39:42 crc kubenswrapper[4776]: I1125 09:39:42.958513 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/052e47c1-028a-4f6f-bc78-67b4ad351ffe-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-pt9xh\" (UID: \"052e47c1-028a-4f6f-bc78-67b4ad351ffe\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-pt9xh" Nov 25 09:39:42 crc kubenswrapper[4776]: I1125 09:39:42.958635 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6chcp\" (UniqueName: \"kubernetes.io/projected/052e47c1-028a-4f6f-bc78-67b4ad351ffe-kube-api-access-6chcp\") pod \"cert-manager-webhook-f4fb5df64-pt9xh\" (UID: \"052e47c1-028a-4f6f-bc78-67b4ad351ffe\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-pt9xh" Nov 25 09:39:43 crc kubenswrapper[4776]: I1125 09:39:43.059750 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6chcp\" (UniqueName: \"kubernetes.io/projected/052e47c1-028a-4f6f-bc78-67b4ad351ffe-kube-api-access-6chcp\") pod \"cert-manager-webhook-f4fb5df64-pt9xh\" (UID: \"052e47c1-028a-4f6f-bc78-67b4ad351ffe\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-pt9xh" Nov 25 09:39:43 crc kubenswrapper[4776]: I1125 09:39:43.059840 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/052e47c1-028a-4f6f-bc78-67b4ad351ffe-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-pt9xh\" (UID: \"052e47c1-028a-4f6f-bc78-67b4ad351ffe\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-pt9xh" Nov 25 09:39:43 crc kubenswrapper[4776]: I1125 09:39:43.078802 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6chcp\" (UniqueName: \"kubernetes.io/projected/052e47c1-028a-4f6f-bc78-67b4ad351ffe-kube-api-access-6chcp\") pod \"cert-manager-webhook-f4fb5df64-pt9xh\" (UID: \"052e47c1-028a-4f6f-bc78-67b4ad351ffe\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-pt9xh" Nov 25 09:39:43 crc kubenswrapper[4776]: I1125 09:39:43.079818 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/052e47c1-028a-4f6f-bc78-67b4ad351ffe-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-pt9xh\" (UID: \"052e47c1-028a-4f6f-bc78-67b4ad351ffe\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-pt9xh" Nov 25 09:39:43 crc kubenswrapper[4776]: I1125 09:39:43.091216 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-pt9xh" Nov 25 09:39:43 crc kubenswrapper[4776]: I1125 09:39:43.484101 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-pt9xh"] Nov 25 09:39:43 crc kubenswrapper[4776]: I1125 09:39:43.640575 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-pt9xh" event={"ID":"052e47c1-028a-4f6f-bc78-67b4ad351ffe","Type":"ContainerStarted","Data":"b5cb10f7200e7092423118f73ad3b3bf47714fc2d5d077f22aefd4d0e87f4dd6"} Nov 25 09:39:46 crc kubenswrapper[4776]: I1125 09:39:46.305358 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-d9hk5"] Nov 25 09:39:46 crc kubenswrapper[4776]: I1125 09:39:46.306942 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-d9hk5" Nov 25 09:39:46 crc kubenswrapper[4776]: I1125 09:39:46.309021 4776 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-nhg6c" Nov 25 09:39:46 crc kubenswrapper[4776]: I1125 09:39:46.311910 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-d9hk5"] Nov 25 09:39:46 crc kubenswrapper[4776]: I1125 09:39:46.319214 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6w8s\" (UniqueName: \"kubernetes.io/projected/c688fa59-c409-4344-bd47-65cc299fc20e-kube-api-access-r6w8s\") pod \"cert-manager-cainjector-855d9ccff4-d9hk5\" (UID: \"c688fa59-c409-4344-bd47-65cc299fc20e\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-d9hk5" Nov 25 09:39:46 crc kubenswrapper[4776]: I1125 09:39:46.319291 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c688fa59-c409-4344-bd47-65cc299fc20e-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-d9hk5\" (UID: \"c688fa59-c409-4344-bd47-65cc299fc20e\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-d9hk5" Nov 25 09:39:46 crc kubenswrapper[4776]: I1125 09:39:46.420199 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6w8s\" (UniqueName: \"kubernetes.io/projected/c688fa59-c409-4344-bd47-65cc299fc20e-kube-api-access-r6w8s\") pod \"cert-manager-cainjector-855d9ccff4-d9hk5\" (UID: \"c688fa59-c409-4344-bd47-65cc299fc20e\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-d9hk5" Nov 25 09:39:46 crc kubenswrapper[4776]: I1125 09:39:46.420302 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c688fa59-c409-4344-bd47-65cc299fc20e-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-d9hk5\" (UID: \"c688fa59-c409-4344-bd47-65cc299fc20e\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-d9hk5" Nov 25 09:39:46 crc kubenswrapper[4776]: I1125 09:39:46.439464 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c688fa59-c409-4344-bd47-65cc299fc20e-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-d9hk5\" (UID: \"c688fa59-c409-4344-bd47-65cc299fc20e\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-d9hk5" Nov 25 09:39:46 crc kubenswrapper[4776]: I1125 09:39:46.443976 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6w8s\" (UniqueName: \"kubernetes.io/projected/c688fa59-c409-4344-bd47-65cc299fc20e-kube-api-access-r6w8s\") pod \"cert-manager-cainjector-855d9ccff4-d9hk5\" (UID: \"c688fa59-c409-4344-bd47-65cc299fc20e\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-d9hk5" Nov 25 09:39:46 crc kubenswrapper[4776]: I1125 09:39:46.628515 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-d9hk5" Nov 25 09:39:50 crc kubenswrapper[4776]: I1125 09:39:50.969697 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-d9hk5"] Nov 25 09:39:50 crc kubenswrapper[4776]: W1125 09:39:50.971404 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc688fa59_c409_4344_bd47_65cc299fc20e.slice/crio-105cd19d7d1b5aee323f2e22bbf6abcfab46fa820317e589d7b85b00dece4175 WatchSource:0}: Error finding container 105cd19d7d1b5aee323f2e22bbf6abcfab46fa820317e589d7b85b00dece4175: Status 404 returned error can't find the container with id 105cd19d7d1b5aee323f2e22bbf6abcfab46fa820317e589d7b85b00dece4175 Nov 25 09:39:51 crc kubenswrapper[4776]: I1125 09:39:51.692611 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-pt9xh" event={"ID":"052e47c1-028a-4f6f-bc78-67b4ad351ffe","Type":"ContainerStarted","Data":"9272005e3f538070719846e393e43ed3b24aa68a58a66f3c0205ba5784d39cb4"} Nov 25 09:39:51 crc kubenswrapper[4776]: I1125 09:39:51.693545 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-pt9xh" Nov 25 09:39:51 crc kubenswrapper[4776]: I1125 09:39:51.695292 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-d9hk5" event={"ID":"c688fa59-c409-4344-bd47-65cc299fc20e","Type":"ContainerStarted","Data":"804fcbfdae04208652bf2f004fd7fac17358d24b5971cbe64975dfccdd02196f"} Nov 25 09:39:51 crc kubenswrapper[4776]: I1125 09:39:51.695320 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-d9hk5" event={"ID":"c688fa59-c409-4344-bd47-65cc299fc20e","Type":"ContainerStarted","Data":"105cd19d7d1b5aee323f2e22bbf6abcfab46fa820317e589d7b85b00dece4175"} Nov 25 09:39:51 crc kubenswrapper[4776]: I1125 09:39:51.710292 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-pt9xh" podStartSLOduration=2.561937902 podStartE2EDuration="9.710271817s" podCreationTimestamp="2025-11-25 09:39:42 +0000 UTC" firstStartedPulling="2025-11-25 09:39:43.496655279 +0000 UTC m=+928.537714832" lastFinishedPulling="2025-11-25 09:39:50.644989194 +0000 UTC m=+935.686048747" observedRunningTime="2025-11-25 09:39:51.706568834 +0000 UTC m=+936.747628387" watchObservedRunningTime="2025-11-25 09:39:51.710271817 +0000 UTC m=+936.751331370" Nov 25 09:39:51 crc kubenswrapper[4776]: I1125 09:39:51.725933 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-d9hk5" podStartSLOduration=5.725910249 podStartE2EDuration="5.725910249s" podCreationTimestamp="2025-11-25 09:39:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:39:51.722683428 +0000 UTC m=+936.763742981" watchObservedRunningTime="2025-11-25 09:39:51.725910249 +0000 UTC m=+936.766969812" Nov 25 09:39:58 crc kubenswrapper[4776]: I1125 09:39:58.094284 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-pt9xh" Nov 25 09:40:01 crc kubenswrapper[4776]: I1125 09:40:01.669669 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-qzgc4"] Nov 25 09:40:01 crc kubenswrapper[4776]: I1125 09:40:01.670958 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-qzgc4" Nov 25 09:40:01 crc kubenswrapper[4776]: I1125 09:40:01.675667 4776 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-97p4z" Nov 25 09:40:01 crc kubenswrapper[4776]: I1125 09:40:01.686851 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-qzgc4"] Nov 25 09:40:01 crc kubenswrapper[4776]: I1125 09:40:01.844547 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/24ae1849-101b-4e31-8189-eb4db03d8c73-bound-sa-token\") pod \"cert-manager-86cb77c54b-qzgc4\" (UID: \"24ae1849-101b-4e31-8189-eb4db03d8c73\") " pod="cert-manager/cert-manager-86cb77c54b-qzgc4" Nov 25 09:40:01 crc kubenswrapper[4776]: I1125 09:40:01.844633 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8s8xq\" (UniqueName: \"kubernetes.io/projected/24ae1849-101b-4e31-8189-eb4db03d8c73-kube-api-access-8s8xq\") pod \"cert-manager-86cb77c54b-qzgc4\" (UID: \"24ae1849-101b-4e31-8189-eb4db03d8c73\") " pod="cert-manager/cert-manager-86cb77c54b-qzgc4" Nov 25 09:40:01 crc kubenswrapper[4776]: I1125 09:40:01.946465 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/24ae1849-101b-4e31-8189-eb4db03d8c73-bound-sa-token\") pod \"cert-manager-86cb77c54b-qzgc4\" (UID: \"24ae1849-101b-4e31-8189-eb4db03d8c73\") " pod="cert-manager/cert-manager-86cb77c54b-qzgc4" Nov 25 09:40:01 crc kubenswrapper[4776]: I1125 09:40:01.946537 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8s8xq\" (UniqueName: \"kubernetes.io/projected/24ae1849-101b-4e31-8189-eb4db03d8c73-kube-api-access-8s8xq\") pod \"cert-manager-86cb77c54b-qzgc4\" (UID: \"24ae1849-101b-4e31-8189-eb4db03d8c73\") " pod="cert-manager/cert-manager-86cb77c54b-qzgc4" Nov 25 09:40:01 crc kubenswrapper[4776]: I1125 09:40:01.966939 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/24ae1849-101b-4e31-8189-eb4db03d8c73-bound-sa-token\") pod \"cert-manager-86cb77c54b-qzgc4\" (UID: \"24ae1849-101b-4e31-8189-eb4db03d8c73\") " pod="cert-manager/cert-manager-86cb77c54b-qzgc4" Nov 25 09:40:01 crc kubenswrapper[4776]: I1125 09:40:01.973777 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8s8xq\" (UniqueName: \"kubernetes.io/projected/24ae1849-101b-4e31-8189-eb4db03d8c73-kube-api-access-8s8xq\") pod \"cert-manager-86cb77c54b-qzgc4\" (UID: \"24ae1849-101b-4e31-8189-eb4db03d8c73\") " pod="cert-manager/cert-manager-86cb77c54b-qzgc4" Nov 25 09:40:01 crc kubenswrapper[4776]: I1125 09:40:01.988205 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-qzgc4" Nov 25 09:40:02 crc kubenswrapper[4776]: I1125 09:40:02.377663 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-qzgc4"] Nov 25 09:40:02 crc kubenswrapper[4776]: W1125 09:40:02.384367 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24ae1849_101b_4e31_8189_eb4db03d8c73.slice/crio-bdccd946d7df43aa61e1d80d4aa6f5c7f2fd035f9c2254d5eea26dc2aaef313c WatchSource:0}: Error finding container bdccd946d7df43aa61e1d80d4aa6f5c7f2fd035f9c2254d5eea26dc2aaef313c: Status 404 returned error can't find the container with id bdccd946d7df43aa61e1d80d4aa6f5c7f2fd035f9c2254d5eea26dc2aaef313c Nov 25 09:40:02 crc kubenswrapper[4776]: I1125 09:40:02.774015 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-qzgc4" event={"ID":"24ae1849-101b-4e31-8189-eb4db03d8c73","Type":"ContainerStarted","Data":"f77a77a6412f6c4d88b86ddb2fb4c4ae60eb5b20937e9d86010e0e141059021a"} Nov 25 09:40:02 crc kubenswrapper[4776]: I1125 09:40:02.774124 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-qzgc4" event={"ID":"24ae1849-101b-4e31-8189-eb4db03d8c73","Type":"ContainerStarted","Data":"bdccd946d7df43aa61e1d80d4aa6f5c7f2fd035f9c2254d5eea26dc2aaef313c"} Nov 25 09:40:02 crc kubenswrapper[4776]: I1125 09:40:02.807701 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-qzgc4" podStartSLOduration=1.8076717850000001 podStartE2EDuration="1.807671785s" podCreationTimestamp="2025-11-25 09:40:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:40:02.795787998 +0000 UTC m=+947.836847591" watchObservedRunningTime="2025-11-25 09:40:02.807671785 +0000 UTC m=+947.848731378" Nov 25 09:40:11 crc kubenswrapper[4776]: I1125 09:40:11.478821 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-rhcfh"] Nov 25 09:40:11 crc kubenswrapper[4776]: I1125 09:40:11.479993 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rhcfh" Nov 25 09:40:11 crc kubenswrapper[4776]: I1125 09:40:11.482492 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-dpgbg" Nov 25 09:40:11 crc kubenswrapper[4776]: I1125 09:40:11.482790 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 25 09:40:11 crc kubenswrapper[4776]: I1125 09:40:11.482976 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 25 09:40:11 crc kubenswrapper[4776]: I1125 09:40:11.503610 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-rhcfh"] Nov 25 09:40:11 crc kubenswrapper[4776]: I1125 09:40:11.572826 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6tx4\" (UniqueName: \"kubernetes.io/projected/bd355aea-4e3c-489e-9233-46039aa57ec2-kube-api-access-x6tx4\") pod \"openstack-operator-index-rhcfh\" (UID: \"bd355aea-4e3c-489e-9233-46039aa57ec2\") " pod="openstack-operators/openstack-operator-index-rhcfh" Nov 25 09:40:11 crc kubenswrapper[4776]: I1125 09:40:11.673998 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6tx4\" (UniqueName: \"kubernetes.io/projected/bd355aea-4e3c-489e-9233-46039aa57ec2-kube-api-access-x6tx4\") pod \"openstack-operator-index-rhcfh\" (UID: \"bd355aea-4e3c-489e-9233-46039aa57ec2\") " pod="openstack-operators/openstack-operator-index-rhcfh" Nov 25 09:40:11 crc kubenswrapper[4776]: I1125 09:40:11.698653 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6tx4\" (UniqueName: \"kubernetes.io/projected/bd355aea-4e3c-489e-9233-46039aa57ec2-kube-api-access-x6tx4\") pod \"openstack-operator-index-rhcfh\" (UID: \"bd355aea-4e3c-489e-9233-46039aa57ec2\") " pod="openstack-operators/openstack-operator-index-rhcfh" Nov 25 09:40:11 crc kubenswrapper[4776]: I1125 09:40:11.804921 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rhcfh" Nov 25 09:40:12 crc kubenswrapper[4776]: I1125 09:40:12.015326 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-rhcfh"] Nov 25 09:40:12 crc kubenswrapper[4776]: W1125 09:40:12.023044 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd355aea_4e3c_489e_9233_46039aa57ec2.slice/crio-73634acf32d33bec436a3cfb971ed939aba2273f6a91426a75b0cb6e1db91aa6 WatchSource:0}: Error finding container 73634acf32d33bec436a3cfb971ed939aba2273f6a91426a75b0cb6e1db91aa6: Status 404 returned error can't find the container with id 73634acf32d33bec436a3cfb971ed939aba2273f6a91426a75b0cb6e1db91aa6 Nov 25 09:40:12 crc kubenswrapper[4776]: I1125 09:40:12.837223 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rhcfh" event={"ID":"bd355aea-4e3c-489e-9233-46039aa57ec2","Type":"ContainerStarted","Data":"73634acf32d33bec436a3cfb971ed939aba2273f6a91426a75b0cb6e1db91aa6"} Nov 25 09:40:14 crc kubenswrapper[4776]: I1125 09:40:14.660090 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-rhcfh"] Nov 25 09:40:15 crc kubenswrapper[4776]: I1125 09:40:15.282914 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-k6vg9"] Nov 25 09:40:15 crc kubenswrapper[4776]: I1125 09:40:15.284425 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-k6vg9" Nov 25 09:40:15 crc kubenswrapper[4776]: I1125 09:40:15.299837 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-k6vg9"] Nov 25 09:40:15 crc kubenswrapper[4776]: I1125 09:40:15.425676 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twhr8\" (UniqueName: \"kubernetes.io/projected/96957ec4-0d5b-42a9-8ddb-976e00abe710-kube-api-access-twhr8\") pod \"openstack-operator-index-k6vg9\" (UID: \"96957ec4-0d5b-42a9-8ddb-976e00abe710\") " pod="openstack-operators/openstack-operator-index-k6vg9" Nov 25 09:40:15 crc kubenswrapper[4776]: I1125 09:40:15.527433 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twhr8\" (UniqueName: \"kubernetes.io/projected/96957ec4-0d5b-42a9-8ddb-976e00abe710-kube-api-access-twhr8\") pod \"openstack-operator-index-k6vg9\" (UID: \"96957ec4-0d5b-42a9-8ddb-976e00abe710\") " pod="openstack-operators/openstack-operator-index-k6vg9" Nov 25 09:40:15 crc kubenswrapper[4776]: I1125 09:40:15.546999 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twhr8\" (UniqueName: \"kubernetes.io/projected/96957ec4-0d5b-42a9-8ddb-976e00abe710-kube-api-access-twhr8\") pod \"openstack-operator-index-k6vg9\" (UID: \"96957ec4-0d5b-42a9-8ddb-976e00abe710\") " pod="openstack-operators/openstack-operator-index-k6vg9" Nov 25 09:40:15 crc kubenswrapper[4776]: I1125 09:40:15.612618 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-k6vg9" Nov 25 09:40:16 crc kubenswrapper[4776]: I1125 09:40:16.082789 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-k6vg9"] Nov 25 09:40:16 crc kubenswrapper[4776]: W1125 09:40:16.099265 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod96957ec4_0d5b_42a9_8ddb_976e00abe710.slice/crio-8affbe6aeabfce11430924f4ef64400834e725d45f61fa5e110102c97b527748 WatchSource:0}: Error finding container 8affbe6aeabfce11430924f4ef64400834e725d45f61fa5e110102c97b527748: Status 404 returned error can't find the container with id 8affbe6aeabfce11430924f4ef64400834e725d45f61fa5e110102c97b527748 Nov 25 09:40:16 crc kubenswrapper[4776]: I1125 09:40:16.869237 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-k6vg9" event={"ID":"96957ec4-0d5b-42a9-8ddb-976e00abe710","Type":"ContainerStarted","Data":"8affbe6aeabfce11430924f4ef64400834e725d45f61fa5e110102c97b527748"} Nov 25 09:40:18 crc kubenswrapper[4776]: I1125 09:40:18.885160 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-k6vg9" event={"ID":"96957ec4-0d5b-42a9-8ddb-976e00abe710","Type":"ContainerStarted","Data":"4ef7881471c7caaf74fdd12752449d71fbc7dbaec2ecbd59e358f276d76c7744"} Nov 25 09:40:18 crc kubenswrapper[4776]: I1125 09:40:18.886873 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rhcfh" event={"ID":"bd355aea-4e3c-489e-9233-46039aa57ec2","Type":"ContainerStarted","Data":"838b1f8c42fc1cd46db07b22d82261c7956c361c70ab742e401a6b1d165cf5fe"} Nov 25 09:40:18 crc kubenswrapper[4776]: I1125 09:40:18.887008 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-rhcfh" podUID="bd355aea-4e3c-489e-9233-46039aa57ec2" containerName="registry-server" containerID="cri-o://838b1f8c42fc1cd46db07b22d82261c7956c361c70ab742e401a6b1d165cf5fe" gracePeriod=2 Nov 25 09:40:18 crc kubenswrapper[4776]: I1125 09:40:18.898828 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-k6vg9" podStartSLOduration=2.05930255 podStartE2EDuration="3.898813855s" podCreationTimestamp="2025-11-25 09:40:15 +0000 UTC" firstStartedPulling="2025-11-25 09:40:16.101006368 +0000 UTC m=+961.142065921" lastFinishedPulling="2025-11-25 09:40:17.940517673 +0000 UTC m=+962.981577226" observedRunningTime="2025-11-25 09:40:18.898441825 +0000 UTC m=+963.939501378" watchObservedRunningTime="2025-11-25 09:40:18.898813855 +0000 UTC m=+963.939873408" Nov 25 09:40:18 crc kubenswrapper[4776]: I1125 09:40:18.914720 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-rhcfh" podStartSLOduration=1.995904802 podStartE2EDuration="7.914699263s" podCreationTimestamp="2025-11-25 09:40:11 +0000 UTC" firstStartedPulling="2025-11-25 09:40:12.025590969 +0000 UTC m=+957.066650522" lastFinishedPulling="2025-11-25 09:40:17.94438542 +0000 UTC m=+962.985444983" observedRunningTime="2025-11-25 09:40:18.912028706 +0000 UTC m=+963.953088259" watchObservedRunningTime="2025-11-25 09:40:18.914699263 +0000 UTC m=+963.955758816" Nov 25 09:40:19 crc kubenswrapper[4776]: I1125 09:40:19.282642 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rhcfh" Nov 25 09:40:19 crc kubenswrapper[4776]: I1125 09:40:19.390689 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6tx4\" (UniqueName: \"kubernetes.io/projected/bd355aea-4e3c-489e-9233-46039aa57ec2-kube-api-access-x6tx4\") pod \"bd355aea-4e3c-489e-9233-46039aa57ec2\" (UID: \"bd355aea-4e3c-489e-9233-46039aa57ec2\") " Nov 25 09:40:19 crc kubenswrapper[4776]: I1125 09:40:19.398394 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd355aea-4e3c-489e-9233-46039aa57ec2-kube-api-access-x6tx4" (OuterVolumeSpecName: "kube-api-access-x6tx4") pod "bd355aea-4e3c-489e-9233-46039aa57ec2" (UID: "bd355aea-4e3c-489e-9233-46039aa57ec2"). InnerVolumeSpecName "kube-api-access-x6tx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:40:19 crc kubenswrapper[4776]: I1125 09:40:19.492120 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6tx4\" (UniqueName: \"kubernetes.io/projected/bd355aea-4e3c-489e-9233-46039aa57ec2-kube-api-access-x6tx4\") on node \"crc\" DevicePath \"\"" Nov 25 09:40:19 crc kubenswrapper[4776]: I1125 09:40:19.895662 4776 generic.go:334] "Generic (PLEG): container finished" podID="bd355aea-4e3c-489e-9233-46039aa57ec2" containerID="838b1f8c42fc1cd46db07b22d82261c7956c361c70ab742e401a6b1d165cf5fe" exitCode=0 Nov 25 09:40:19 crc kubenswrapper[4776]: I1125 09:40:19.895720 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rhcfh" Nov 25 09:40:19 crc kubenswrapper[4776]: I1125 09:40:19.895760 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rhcfh" event={"ID":"bd355aea-4e3c-489e-9233-46039aa57ec2","Type":"ContainerDied","Data":"838b1f8c42fc1cd46db07b22d82261c7956c361c70ab742e401a6b1d165cf5fe"} Nov 25 09:40:19 crc kubenswrapper[4776]: I1125 09:40:19.895830 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rhcfh" event={"ID":"bd355aea-4e3c-489e-9233-46039aa57ec2","Type":"ContainerDied","Data":"73634acf32d33bec436a3cfb971ed939aba2273f6a91426a75b0cb6e1db91aa6"} Nov 25 09:40:19 crc kubenswrapper[4776]: I1125 09:40:19.895849 4776 scope.go:117] "RemoveContainer" containerID="838b1f8c42fc1cd46db07b22d82261c7956c361c70ab742e401a6b1d165cf5fe" Nov 25 09:40:19 crc kubenswrapper[4776]: I1125 09:40:19.919210 4776 scope.go:117] "RemoveContainer" containerID="838b1f8c42fc1cd46db07b22d82261c7956c361c70ab742e401a6b1d165cf5fe" Nov 25 09:40:19 crc kubenswrapper[4776]: E1125 09:40:19.919708 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"838b1f8c42fc1cd46db07b22d82261c7956c361c70ab742e401a6b1d165cf5fe\": container with ID starting with 838b1f8c42fc1cd46db07b22d82261c7956c361c70ab742e401a6b1d165cf5fe not found: ID does not exist" containerID="838b1f8c42fc1cd46db07b22d82261c7956c361c70ab742e401a6b1d165cf5fe" Nov 25 09:40:19 crc kubenswrapper[4776]: I1125 09:40:19.919743 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"838b1f8c42fc1cd46db07b22d82261c7956c361c70ab742e401a6b1d165cf5fe"} err="failed to get container status \"838b1f8c42fc1cd46db07b22d82261c7956c361c70ab742e401a6b1d165cf5fe\": rpc error: code = NotFound desc = could not find container \"838b1f8c42fc1cd46db07b22d82261c7956c361c70ab742e401a6b1d165cf5fe\": container with ID starting with 838b1f8c42fc1cd46db07b22d82261c7956c361c70ab742e401a6b1d165cf5fe not found: ID does not exist" Nov 25 09:40:19 crc kubenswrapper[4776]: I1125 09:40:19.923101 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-rhcfh"] Nov 25 09:40:19 crc kubenswrapper[4776]: I1125 09:40:19.927198 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-rhcfh"] Nov 25 09:40:21 crc kubenswrapper[4776]: I1125 09:40:21.677815 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd355aea-4e3c-489e-9233-46039aa57ec2" path="/var/lib/kubelet/pods/bd355aea-4e3c-489e-9233-46039aa57ec2/volumes" Nov 25 09:40:25 crc kubenswrapper[4776]: I1125 09:40:25.613287 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-k6vg9" Nov 25 09:40:25 crc kubenswrapper[4776]: I1125 09:40:25.613354 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-k6vg9" Nov 25 09:40:25 crc kubenswrapper[4776]: I1125 09:40:25.643938 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-k6vg9" Nov 25 09:40:25 crc kubenswrapper[4776]: I1125 09:40:25.974477 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-k6vg9" Nov 25 09:40:27 crc kubenswrapper[4776]: I1125 09:40:27.509626 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv"] Nov 25 09:40:27 crc kubenswrapper[4776]: E1125 09:40:27.509931 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd355aea-4e3c-489e-9233-46039aa57ec2" containerName="registry-server" Nov 25 09:40:27 crc kubenswrapper[4776]: I1125 09:40:27.509946 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd355aea-4e3c-489e-9233-46039aa57ec2" containerName="registry-server" Nov 25 09:40:27 crc kubenswrapper[4776]: I1125 09:40:27.510044 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd355aea-4e3c-489e-9233-46039aa57ec2" containerName="registry-server" Nov 25 09:40:27 crc kubenswrapper[4776]: I1125 09:40:27.510896 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv" Nov 25 09:40:27 crc kubenswrapper[4776]: I1125 09:40:27.513276 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-nfqfq" Nov 25 09:40:27 crc kubenswrapper[4776]: I1125 09:40:27.518877 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv"] Nov 25 09:40:27 crc kubenswrapper[4776]: I1125 09:40:27.605351 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8dtt\" (UniqueName: \"kubernetes.io/projected/a13d792f-8c2b-4dce-ba68-c83effbbbd12-kube-api-access-l8dtt\") pod \"bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv\" (UID: \"a13d792f-8c2b-4dce-ba68-c83effbbbd12\") " pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv" Nov 25 09:40:27 crc kubenswrapper[4776]: I1125 09:40:27.605585 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a13d792f-8c2b-4dce-ba68-c83effbbbd12-bundle\") pod \"bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv\" (UID: \"a13d792f-8c2b-4dce-ba68-c83effbbbd12\") " pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv" Nov 25 09:40:27 crc kubenswrapper[4776]: I1125 09:40:27.605632 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a13d792f-8c2b-4dce-ba68-c83effbbbd12-util\") pod \"bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv\" (UID: \"a13d792f-8c2b-4dce-ba68-c83effbbbd12\") " pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv" Nov 25 09:40:27 crc kubenswrapper[4776]: I1125 09:40:27.706856 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8dtt\" (UniqueName: \"kubernetes.io/projected/a13d792f-8c2b-4dce-ba68-c83effbbbd12-kube-api-access-l8dtt\") pod \"bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv\" (UID: \"a13d792f-8c2b-4dce-ba68-c83effbbbd12\") " pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv" Nov 25 09:40:27 crc kubenswrapper[4776]: I1125 09:40:27.707006 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a13d792f-8c2b-4dce-ba68-c83effbbbd12-bundle\") pod \"bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv\" (UID: \"a13d792f-8c2b-4dce-ba68-c83effbbbd12\") " pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv" Nov 25 09:40:27 crc kubenswrapper[4776]: I1125 09:40:27.707181 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a13d792f-8c2b-4dce-ba68-c83effbbbd12-util\") pod \"bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv\" (UID: \"a13d792f-8c2b-4dce-ba68-c83effbbbd12\") " pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv" Nov 25 09:40:27 crc kubenswrapper[4776]: I1125 09:40:27.707555 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a13d792f-8c2b-4dce-ba68-c83effbbbd12-bundle\") pod \"bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv\" (UID: \"a13d792f-8c2b-4dce-ba68-c83effbbbd12\") " pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv" Nov 25 09:40:27 crc kubenswrapper[4776]: I1125 09:40:27.707956 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a13d792f-8c2b-4dce-ba68-c83effbbbd12-util\") pod \"bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv\" (UID: \"a13d792f-8c2b-4dce-ba68-c83effbbbd12\") " pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv" Nov 25 09:40:27 crc kubenswrapper[4776]: I1125 09:40:27.725784 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8dtt\" (UniqueName: \"kubernetes.io/projected/a13d792f-8c2b-4dce-ba68-c83effbbbd12-kube-api-access-l8dtt\") pod \"bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv\" (UID: \"a13d792f-8c2b-4dce-ba68-c83effbbbd12\") " pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv" Nov 25 09:40:27 crc kubenswrapper[4776]: I1125 09:40:27.828955 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv" Nov 25 09:40:28 crc kubenswrapper[4776]: I1125 09:40:28.210499 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv"] Nov 25 09:40:28 crc kubenswrapper[4776]: I1125 09:40:28.958547 4776 generic.go:334] "Generic (PLEG): container finished" podID="a13d792f-8c2b-4dce-ba68-c83effbbbd12" containerID="fa62dc4d3d1304812e2f32d110272fef1eb1f64f0782e9d6e4ef36cf70f2b6d1" exitCode=0 Nov 25 09:40:28 crc kubenswrapper[4776]: I1125 09:40:28.958590 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv" event={"ID":"a13d792f-8c2b-4dce-ba68-c83effbbbd12","Type":"ContainerDied","Data":"fa62dc4d3d1304812e2f32d110272fef1eb1f64f0782e9d6e4ef36cf70f2b6d1"} Nov 25 09:40:28 crc kubenswrapper[4776]: I1125 09:40:28.958618 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv" event={"ID":"a13d792f-8c2b-4dce-ba68-c83effbbbd12","Type":"ContainerStarted","Data":"5da65b114cc5859515b94b220e9b51098f130ce28c9b3e6ec27adfd8781d9c68"} Nov 25 09:40:29 crc kubenswrapper[4776]: I1125 09:40:29.966953 4776 generic.go:334] "Generic (PLEG): container finished" podID="a13d792f-8c2b-4dce-ba68-c83effbbbd12" containerID="b514c4c2233fdabf882e9fbcf4d6929cdcb40475278a5a689f76c40f30ec605c" exitCode=0 Nov 25 09:40:29 crc kubenswrapper[4776]: I1125 09:40:29.967244 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv" event={"ID":"a13d792f-8c2b-4dce-ba68-c83effbbbd12","Type":"ContainerDied","Data":"b514c4c2233fdabf882e9fbcf4d6929cdcb40475278a5a689f76c40f30ec605c"} Nov 25 09:40:30 crc kubenswrapper[4776]: I1125 09:40:30.976162 4776 generic.go:334] "Generic (PLEG): container finished" podID="a13d792f-8c2b-4dce-ba68-c83effbbbd12" containerID="11e1d1da51462a65466c1682a55724a73952fa34adcbaa62e795f0df0db9d03f" exitCode=0 Nov 25 09:40:30 crc kubenswrapper[4776]: I1125 09:40:30.976288 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv" event={"ID":"a13d792f-8c2b-4dce-ba68-c83effbbbd12","Type":"ContainerDied","Data":"11e1d1da51462a65466c1682a55724a73952fa34adcbaa62e795f0df0db9d03f"} Nov 25 09:40:32 crc kubenswrapper[4776]: I1125 09:40:32.251890 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv" Nov 25 09:40:32 crc kubenswrapper[4776]: I1125 09:40:32.378048 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8dtt\" (UniqueName: \"kubernetes.io/projected/a13d792f-8c2b-4dce-ba68-c83effbbbd12-kube-api-access-l8dtt\") pod \"a13d792f-8c2b-4dce-ba68-c83effbbbd12\" (UID: \"a13d792f-8c2b-4dce-ba68-c83effbbbd12\") " Nov 25 09:40:32 crc kubenswrapper[4776]: I1125 09:40:32.378169 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a13d792f-8c2b-4dce-ba68-c83effbbbd12-bundle\") pod \"a13d792f-8c2b-4dce-ba68-c83effbbbd12\" (UID: \"a13d792f-8c2b-4dce-ba68-c83effbbbd12\") " Nov 25 09:40:32 crc kubenswrapper[4776]: I1125 09:40:32.378285 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a13d792f-8c2b-4dce-ba68-c83effbbbd12-util\") pod \"a13d792f-8c2b-4dce-ba68-c83effbbbd12\" (UID: \"a13d792f-8c2b-4dce-ba68-c83effbbbd12\") " Nov 25 09:40:32 crc kubenswrapper[4776]: I1125 09:40:32.378881 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a13d792f-8c2b-4dce-ba68-c83effbbbd12-bundle" (OuterVolumeSpecName: "bundle") pod "a13d792f-8c2b-4dce-ba68-c83effbbbd12" (UID: "a13d792f-8c2b-4dce-ba68-c83effbbbd12"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:40:32 crc kubenswrapper[4776]: I1125 09:40:32.387484 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a13d792f-8c2b-4dce-ba68-c83effbbbd12-kube-api-access-l8dtt" (OuterVolumeSpecName: "kube-api-access-l8dtt") pod "a13d792f-8c2b-4dce-ba68-c83effbbbd12" (UID: "a13d792f-8c2b-4dce-ba68-c83effbbbd12"). InnerVolumeSpecName "kube-api-access-l8dtt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:40:32 crc kubenswrapper[4776]: I1125 09:40:32.397205 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a13d792f-8c2b-4dce-ba68-c83effbbbd12-util" (OuterVolumeSpecName: "util") pod "a13d792f-8c2b-4dce-ba68-c83effbbbd12" (UID: "a13d792f-8c2b-4dce-ba68-c83effbbbd12"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:40:32 crc kubenswrapper[4776]: I1125 09:40:32.479055 4776 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a13d792f-8c2b-4dce-ba68-c83effbbbd12-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:40:32 crc kubenswrapper[4776]: I1125 09:40:32.479097 4776 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a13d792f-8c2b-4dce-ba68-c83effbbbd12-util\") on node \"crc\" DevicePath \"\"" Nov 25 09:40:32 crc kubenswrapper[4776]: I1125 09:40:32.479107 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8dtt\" (UniqueName: \"kubernetes.io/projected/a13d792f-8c2b-4dce-ba68-c83effbbbd12-kube-api-access-l8dtt\") on node \"crc\" DevicePath \"\"" Nov 25 09:40:32 crc kubenswrapper[4776]: I1125 09:40:32.989695 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv" event={"ID":"a13d792f-8c2b-4dce-ba68-c83effbbbd12","Type":"ContainerDied","Data":"5da65b114cc5859515b94b220e9b51098f130ce28c9b3e6ec27adfd8781d9c68"} Nov 25 09:40:32 crc kubenswrapper[4776]: I1125 09:40:32.989731 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv" Nov 25 09:40:32 crc kubenswrapper[4776]: I1125 09:40:32.989736 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5da65b114cc5859515b94b220e9b51098f130ce28c9b3e6ec27adfd8781d9c68" Nov 25 09:40:35 crc kubenswrapper[4776]: I1125 09:40:35.940613 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7b567956b5-q468k"] Nov 25 09:40:35 crc kubenswrapper[4776]: E1125 09:40:35.940931 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a13d792f-8c2b-4dce-ba68-c83effbbbd12" containerName="extract" Nov 25 09:40:35 crc kubenswrapper[4776]: I1125 09:40:35.940947 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a13d792f-8c2b-4dce-ba68-c83effbbbd12" containerName="extract" Nov 25 09:40:35 crc kubenswrapper[4776]: E1125 09:40:35.940963 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a13d792f-8c2b-4dce-ba68-c83effbbbd12" containerName="pull" Nov 25 09:40:35 crc kubenswrapper[4776]: I1125 09:40:35.940970 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a13d792f-8c2b-4dce-ba68-c83effbbbd12" containerName="pull" Nov 25 09:40:35 crc kubenswrapper[4776]: E1125 09:40:35.940978 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a13d792f-8c2b-4dce-ba68-c83effbbbd12" containerName="util" Nov 25 09:40:35 crc kubenswrapper[4776]: I1125 09:40:35.940985 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a13d792f-8c2b-4dce-ba68-c83effbbbd12" containerName="util" Nov 25 09:40:35 crc kubenswrapper[4776]: I1125 09:40:35.941147 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="a13d792f-8c2b-4dce-ba68-c83effbbbd12" containerName="extract" Nov 25 09:40:35 crc kubenswrapper[4776]: I1125 09:40:35.941701 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-q468k" Nov 25 09:40:35 crc kubenswrapper[4776]: I1125 09:40:35.945139 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-bqk8f" Nov 25 09:40:35 crc kubenswrapper[4776]: I1125 09:40:35.963026 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7b567956b5-q468k"] Nov 25 09:40:36 crc kubenswrapper[4776]: I1125 09:40:36.126466 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4zzc\" (UniqueName: \"kubernetes.io/projected/1a1c5930-5dee-4eeb-a09a-842791a33f88-kube-api-access-d4zzc\") pod \"openstack-operator-controller-operator-7b567956b5-q468k\" (UID: \"1a1c5930-5dee-4eeb-a09a-842791a33f88\") " pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-q468k" Nov 25 09:40:36 crc kubenswrapper[4776]: I1125 09:40:36.227838 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4zzc\" (UniqueName: \"kubernetes.io/projected/1a1c5930-5dee-4eeb-a09a-842791a33f88-kube-api-access-d4zzc\") pod \"openstack-operator-controller-operator-7b567956b5-q468k\" (UID: \"1a1c5930-5dee-4eeb-a09a-842791a33f88\") " pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-q468k" Nov 25 09:40:36 crc kubenswrapper[4776]: I1125 09:40:36.263312 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4zzc\" (UniqueName: \"kubernetes.io/projected/1a1c5930-5dee-4eeb-a09a-842791a33f88-kube-api-access-d4zzc\") pod \"openstack-operator-controller-operator-7b567956b5-q468k\" (UID: \"1a1c5930-5dee-4eeb-a09a-842791a33f88\") " pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-q468k" Nov 25 09:40:36 crc kubenswrapper[4776]: I1125 09:40:36.265343 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-q468k" Nov 25 09:40:36 crc kubenswrapper[4776]: I1125 09:40:36.483719 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7b567956b5-q468k"] Nov 25 09:40:37 crc kubenswrapper[4776]: I1125 09:40:37.022523 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-q468k" event={"ID":"1a1c5930-5dee-4eeb-a09a-842791a33f88","Type":"ContainerStarted","Data":"102421fb5e94d5bc23d08084732aea55831e6c72e81448be14e5050edd723032"} Nov 25 09:40:44 crc kubenswrapper[4776]: I1125 09:40:44.079256 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-q468k" event={"ID":"1a1c5930-5dee-4eeb-a09a-842791a33f88","Type":"ContainerStarted","Data":"401d42b653901e957a77b7953373601af97ce493f2cba3f1160692aa85585b4b"} Nov 25 09:40:45 crc kubenswrapper[4776]: I1125 09:40:45.085546 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-q468k" Nov 25 09:40:45 crc kubenswrapper[4776]: I1125 09:40:45.117896 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-q468k" podStartSLOduration=2.962156642 podStartE2EDuration="10.117874756s" podCreationTimestamp="2025-11-25 09:40:35 +0000 UTC" firstStartedPulling="2025-11-25 09:40:36.497602025 +0000 UTC m=+981.538661578" lastFinishedPulling="2025-11-25 09:40:43.653320129 +0000 UTC m=+988.694379692" observedRunningTime="2025-11-25 09:40:45.111228009 +0000 UTC m=+990.152287552" watchObservedRunningTime="2025-11-25 09:40:45.117874756 +0000 UTC m=+990.158934309" Nov 25 09:40:47 crc kubenswrapper[4776]: I1125 09:40:47.818530 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:40:47 crc kubenswrapper[4776]: I1125 09:40:47.818983 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:40:56 crc kubenswrapper[4776]: I1125 09:40:56.267829 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-q468k" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.490884 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-86dc4d89c8-58r67"] Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.492691 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-58r67" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.496424 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-kpsgn" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.504605 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79856dc55c-rnrpd"] Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.505945 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-rnrpd" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.514010 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-86dc4d89c8-58r67"] Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.518846 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-trls8" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.524022 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-7d695c9b56-zwk9j"] Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.525188 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-zwk9j" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.525241 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5fp5\" (UniqueName: \"kubernetes.io/projected/5449efd1-7801-454a-bc48-a5e9e91aba7d-kube-api-access-t5fp5\") pod \"barbican-operator-controller-manager-86dc4d89c8-58r67\" (UID: \"5449efd1-7801-454a-bc48-a5e9e91aba7d\") " pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-58r67" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.542680 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-zzxs5" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.555833 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-7d695c9b56-zwk9j"] Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.559445 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79856dc55c-rnrpd"] Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.589239 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-68b95954c9-h6k9v"] Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.590401 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-h6k9v" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.594878 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-xrb4v" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.614142 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-68b95954c9-h6k9v"] Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.621126 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-774b86978c-gmcp9"] Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.622156 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-774b86978c-gmcp9" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.625165 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-45lmq" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.626009 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bldv6\" (UniqueName: \"kubernetes.io/projected/7a21bb84-b9c5-4e6f-8fbb-3c2bc65dd409-kube-api-access-bldv6\") pod \"glance-operator-controller-manager-68b95954c9-h6k9v\" (UID: \"7a21bb84-b9c5-4e6f-8fbb-3c2bc65dd409\") " pod="openstack-operators/glance-operator-controller-manager-68b95954c9-h6k9v" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.626102 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5fp5\" (UniqueName: \"kubernetes.io/projected/5449efd1-7801-454a-bc48-a5e9e91aba7d-kube-api-access-t5fp5\") pod \"barbican-operator-controller-manager-86dc4d89c8-58r67\" (UID: \"5449efd1-7801-454a-bc48-a5e9e91aba7d\") " pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-58r67" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.626146 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pk8t\" (UniqueName: \"kubernetes.io/projected/2d511feb-5cd8-4ba3-b5af-5d9d01086a82-kube-api-access-7pk8t\") pod \"cinder-operator-controller-manager-79856dc55c-rnrpd\" (UID: \"2d511feb-5cd8-4ba3-b5af-5d9d01086a82\") " pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-rnrpd" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.626225 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkcnc\" (UniqueName: \"kubernetes.io/projected/a590fe89-93f7-4c94-bbdd-82531eedf52f-kube-api-access-dkcnc\") pod \"designate-operator-controller-manager-7d695c9b56-zwk9j\" (UID: \"a590fe89-93f7-4c94-bbdd-82531eedf52f\") " pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-zwk9j" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.668156 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-774b86978c-gmcp9"] Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.684217 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c9694994-khxdw"] Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.686221 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-khxdw" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.687988 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5fp5\" (UniqueName: \"kubernetes.io/projected/5449efd1-7801-454a-bc48-a5e9e91aba7d-kube-api-access-t5fp5\") pod \"barbican-operator-controller-manager-86dc4d89c8-58r67\" (UID: \"5449efd1-7801-454a-bc48-a5e9e91aba7d\") " pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-58r67" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.696642 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-9jslk" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.747965 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pk8t\" (UniqueName: \"kubernetes.io/projected/2d511feb-5cd8-4ba3-b5af-5d9d01086a82-kube-api-access-7pk8t\") pod \"cinder-operator-controller-manager-79856dc55c-rnrpd\" (UID: \"2d511feb-5cd8-4ba3-b5af-5d9d01086a82\") " pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-rnrpd" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.748114 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh54l\" (UniqueName: \"kubernetes.io/projected/187d1384-3de2-4135-abac-45f15155b942-kube-api-access-lh54l\") pod \"horizon-operator-controller-manager-68c9694994-khxdw\" (UID: \"187d1384-3de2-4135-abac-45f15155b942\") " pod="openstack-operators/horizon-operator-controller-manager-68c9694994-khxdw" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.748338 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkcnc\" (UniqueName: \"kubernetes.io/projected/a590fe89-93f7-4c94-bbdd-82531eedf52f-kube-api-access-dkcnc\") pod \"designate-operator-controller-manager-7d695c9b56-zwk9j\" (UID: \"a590fe89-93f7-4c94-bbdd-82531eedf52f\") " pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-zwk9j" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.748466 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bldv6\" (UniqueName: \"kubernetes.io/projected/7a21bb84-b9c5-4e6f-8fbb-3c2bc65dd409-kube-api-access-bldv6\") pod \"glance-operator-controller-manager-68b95954c9-h6k9v\" (UID: \"7a21bb84-b9c5-4e6f-8fbb-3c2bc65dd409\") " pod="openstack-operators/glance-operator-controller-manager-68b95954c9-h6k9v" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.748545 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bkv8\" (UniqueName: \"kubernetes.io/projected/61810388-3bac-4a06-88ea-98cc0337fdee-kube-api-access-4bkv8\") pod \"heat-operator-controller-manager-774b86978c-gmcp9\" (UID: \"61810388-3bac-4a06-88ea-98cc0337fdee\") " pod="openstack-operators/heat-operator-controller-manager-774b86978c-gmcp9" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.801971 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pk8t\" (UniqueName: \"kubernetes.io/projected/2d511feb-5cd8-4ba3-b5af-5d9d01086a82-kube-api-access-7pk8t\") pod \"cinder-operator-controller-manager-79856dc55c-rnrpd\" (UID: \"2d511feb-5cd8-4ba3-b5af-5d9d01086a82\") " pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-rnrpd" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.802348 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bldv6\" (UniqueName: \"kubernetes.io/projected/7a21bb84-b9c5-4e6f-8fbb-3c2bc65dd409-kube-api-access-bldv6\") pod \"glance-operator-controller-manager-68b95954c9-h6k9v\" (UID: \"7a21bb84-b9c5-4e6f-8fbb-3c2bc65dd409\") " pod="openstack-operators/glance-operator-controller-manager-68b95954c9-h6k9v" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.805161 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c9694994-khxdw"] Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.816920 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-58r67" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.819878 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkcnc\" (UniqueName: \"kubernetes.io/projected/a590fe89-93f7-4c94-bbdd-82531eedf52f-kube-api-access-dkcnc\") pod \"designate-operator-controller-manager-7d695c9b56-zwk9j\" (UID: \"a590fe89-93f7-4c94-bbdd-82531eedf52f\") " pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-zwk9j" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.835114 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-rnrpd" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.835845 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-d5cc86f4b-clnxq"] Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.838375 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-clnxq" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.846690 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-zwk9j" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.850245 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.853601 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-jp22w" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.854593 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bkv8\" (UniqueName: \"kubernetes.io/projected/61810388-3bac-4a06-88ea-98cc0337fdee-kube-api-access-4bkv8\") pod \"heat-operator-controller-manager-774b86978c-gmcp9\" (UID: \"61810388-3bac-4a06-88ea-98cc0337fdee\") " pod="openstack-operators/heat-operator-controller-manager-774b86978c-gmcp9" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.854675 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh54l\" (UniqueName: \"kubernetes.io/projected/187d1384-3de2-4135-abac-45f15155b942-kube-api-access-lh54l\") pod \"horizon-operator-controller-manager-68c9694994-khxdw\" (UID: \"187d1384-3de2-4135-abac-45f15155b942\") " pod="openstack-operators/horizon-operator-controller-manager-68c9694994-khxdw" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.862685 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bfcdc958c-cc9wf"] Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.864533 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-cc9wf" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.884476 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-d5cc86f4b-clnxq"] Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.884541 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-748dc6576f-pckfv"] Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.887853 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-pckfv" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.904472 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-jhh9w" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.905270 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-48lxf" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.917519 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bkv8\" (UniqueName: \"kubernetes.io/projected/61810388-3bac-4a06-88ea-98cc0337fdee-kube-api-access-4bkv8\") pod \"heat-operator-controller-manager-774b86978c-gmcp9\" (UID: \"61810388-3bac-4a06-88ea-98cc0337fdee\") " pod="openstack-operators/heat-operator-controller-manager-774b86978c-gmcp9" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.918821 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bfcdc958c-cc9wf"] Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.941713 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh54l\" (UniqueName: \"kubernetes.io/projected/187d1384-3de2-4135-abac-45f15155b942-kube-api-access-lh54l\") pod \"horizon-operator-controller-manager-68c9694994-khxdw\" (UID: \"187d1384-3de2-4135-abac-45f15155b942\") " pod="openstack-operators/horizon-operator-controller-manager-68c9694994-khxdw" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.951101 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-h6k9v" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.961635 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-58bb8d67cc-h57b7"] Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.965612 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnk2q\" (UniqueName: \"kubernetes.io/projected/ebac9e21-7ae6-45e4-a6a8-9728b3670733-kube-api-access-pnk2q\") pod \"ironic-operator-controller-manager-5bfcdc958c-cc9wf\" (UID: \"ebac9e21-7ae6-45e4-a6a8-9728b3670733\") " pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-cc9wf" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.965652 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p48j4\" (UniqueName: \"kubernetes.io/projected/63e1e612-6ab8-4144-b2cc-c8354bd0894a-kube-api-access-p48j4\") pod \"keystone-operator-controller-manager-748dc6576f-pckfv\" (UID: \"63e1e612-6ab8-4144-b2cc-c8354bd0894a\") " pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-pckfv" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.965685 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bf227b94-0dfc-4055-b688-0f73585af089-cert\") pod \"infra-operator-controller-manager-d5cc86f4b-clnxq\" (UID: \"bf227b94-0dfc-4055-b688-0f73585af089\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-clnxq" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.965717 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kp9mc\" (UniqueName: \"kubernetes.io/projected/bf227b94-0dfc-4055-b688-0f73585af089-kube-api-access-kp9mc\") pod \"infra-operator-controller-manager-d5cc86f4b-clnxq\" (UID: \"bf227b94-0dfc-4055-b688-0f73585af089\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-clnxq" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.966222 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-774b86978c-gmcp9" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.971571 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-h57b7" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.977997 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-8rfzb" Nov 25 09:41:12 crc kubenswrapper[4776]: I1125 09:41:12.995098 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-748dc6576f-pckfv"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.016819 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58bb8d67cc-h57b7"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.037161 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-gwj5b"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.038272 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-gwj5b" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.050133 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-85g7d" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.055749 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-jr7jl"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.057001 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-jr7jl" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.061256 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-z858n" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.066985 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnk2q\" (UniqueName: \"kubernetes.io/projected/ebac9e21-7ae6-45e4-a6a8-9728b3670733-kube-api-access-pnk2q\") pod \"ironic-operator-controller-manager-5bfcdc958c-cc9wf\" (UID: \"ebac9e21-7ae6-45e4-a6a8-9728b3670733\") " pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-cc9wf" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.067027 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p48j4\" (UniqueName: \"kubernetes.io/projected/63e1e612-6ab8-4144-b2cc-c8354bd0894a-kube-api-access-p48j4\") pod \"keystone-operator-controller-manager-748dc6576f-pckfv\" (UID: \"63e1e612-6ab8-4144-b2cc-c8354bd0894a\") " pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-pckfv" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.067063 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bf227b94-0dfc-4055-b688-0f73585af089-cert\") pod \"infra-operator-controller-manager-d5cc86f4b-clnxq\" (UID: \"bf227b94-0dfc-4055-b688-0f73585af089\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-clnxq" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.067111 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kp9mc\" (UniqueName: \"kubernetes.io/projected/bf227b94-0dfc-4055-b688-0f73585af089-kube-api-access-kp9mc\") pod \"infra-operator-controller-manager-d5cc86f4b-clnxq\" (UID: \"bf227b94-0dfc-4055-b688-0f73585af089\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-clnxq" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.067150 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzmwn\" (UniqueName: \"kubernetes.io/projected/7af82806-526b-4fe2-8314-65298d1a0539-kube-api-access-mzmwn\") pod \"mariadb-operator-controller-manager-cb6c4fdb7-jr7jl\" (UID: \"7af82806-526b-4fe2-8314-65298d1a0539\") " pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-jr7jl" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.067186 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjwv2\" (UniqueName: \"kubernetes.io/projected/abb2c6a9-56ce-45ff-902f-ac49a249fd81-kube-api-access-bjwv2\") pod \"manila-operator-controller-manager-58bb8d67cc-h57b7\" (UID: \"abb2c6a9-56ce-45ff-902f-ac49a249fd81\") " pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-h57b7" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.067203 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t26gf\" (UniqueName: \"kubernetes.io/projected/88b8f0f1-da6d-43fe-940e-23c48e9248d2-kube-api-access-t26gf\") pod \"neutron-operator-controller-manager-7c57c8bbc4-gwj5b\" (UID: \"88b8f0f1-da6d-43fe-940e-23c48e9248d2\") " pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-gwj5b" Nov 25 09:41:13 crc kubenswrapper[4776]: E1125 09:41:13.067840 4776 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 25 09:41:13 crc kubenswrapper[4776]: E1125 09:41:13.067887 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bf227b94-0dfc-4055-b688-0f73585af089-cert podName:bf227b94-0dfc-4055-b688-0f73585af089 nodeName:}" failed. No retries permitted until 2025-11-25 09:41:13.567868897 +0000 UTC m=+1018.608928450 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/bf227b94-0dfc-4055-b688-0f73585af089-cert") pod "infra-operator-controller-manager-d5cc86f4b-clnxq" (UID: "bf227b94-0dfc-4055-b688-0f73585af089") : secret "infra-operator-webhook-server-cert" not found Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.069060 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-khxdw" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.099274 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-gwj5b"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.105861 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kp9mc\" (UniqueName: \"kubernetes.io/projected/bf227b94-0dfc-4055-b688-0f73585af089-kube-api-access-kp9mc\") pod \"infra-operator-controller-manager-d5cc86f4b-clnxq\" (UID: \"bf227b94-0dfc-4055-b688-0f73585af089\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-clnxq" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.137104 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-jr7jl"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.143128 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-sddh8"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.144087 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-sddh8" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.147156 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-k9jzh" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.148263 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnk2q\" (UniqueName: \"kubernetes.io/projected/ebac9e21-7ae6-45e4-a6a8-9728b3670733-kube-api-access-pnk2q\") pod \"ironic-operator-controller-manager-5bfcdc958c-cc9wf\" (UID: \"ebac9e21-7ae6-45e4-a6a8-9728b3670733\") " pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-cc9wf" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.155837 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p48j4\" (UniqueName: \"kubernetes.io/projected/63e1e612-6ab8-4144-b2cc-c8354bd0894a-kube-api-access-p48j4\") pod \"keystone-operator-controller-manager-748dc6576f-pckfv\" (UID: \"63e1e612-6ab8-4144-b2cc-c8354bd0894a\") " pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-pckfv" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.177047 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-fd75fd47d-h8s6x"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.177667 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jldv\" (UniqueName: \"kubernetes.io/projected/5c262b40-b415-4cca-b1e8-635a9c153d81-kube-api-access-2jldv\") pod \"nova-operator-controller-manager-79556f57fc-sddh8\" (UID: \"5c262b40-b415-4cca-b1e8-635a9c153d81\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-sddh8" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.177803 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzmwn\" (UniqueName: \"kubernetes.io/projected/7af82806-526b-4fe2-8314-65298d1a0539-kube-api-access-mzmwn\") pod \"mariadb-operator-controller-manager-cb6c4fdb7-jr7jl\" (UID: \"7af82806-526b-4fe2-8314-65298d1a0539\") " pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-jr7jl" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.177881 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjwv2\" (UniqueName: \"kubernetes.io/projected/abb2c6a9-56ce-45ff-902f-ac49a249fd81-kube-api-access-bjwv2\") pod \"manila-operator-controller-manager-58bb8d67cc-h57b7\" (UID: \"abb2c6a9-56ce-45ff-902f-ac49a249fd81\") " pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-h57b7" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.177901 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t26gf\" (UniqueName: \"kubernetes.io/projected/88b8f0f1-da6d-43fe-940e-23c48e9248d2-kube-api-access-t26gf\") pod \"neutron-operator-controller-manager-7c57c8bbc4-gwj5b\" (UID: \"88b8f0f1-da6d-43fe-940e-23c48e9248d2\") " pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-gwj5b" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.178472 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-h8s6x" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.189659 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-6dlts" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.201168 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-sddh8"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.214361 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjwv2\" (UniqueName: \"kubernetes.io/projected/abb2c6a9-56ce-45ff-902f-ac49a249fd81-kube-api-access-bjwv2\") pod \"manila-operator-controller-manager-58bb8d67cc-h57b7\" (UID: \"abb2c6a9-56ce-45ff-902f-ac49a249fd81\") " pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-h57b7" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.218978 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzmwn\" (UniqueName: \"kubernetes.io/projected/7af82806-526b-4fe2-8314-65298d1a0539-kube-api-access-mzmwn\") pod \"mariadb-operator-controller-manager-cb6c4fdb7-jr7jl\" (UID: \"7af82806-526b-4fe2-8314-65298d1a0539\") " pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-jr7jl" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.219752 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-fd75fd47d-h8s6x"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.222866 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t26gf\" (UniqueName: \"kubernetes.io/projected/88b8f0f1-da6d-43fe-940e-23c48e9248d2-kube-api-access-t26gf\") pod \"neutron-operator-controller-manager-7c57c8bbc4-gwj5b\" (UID: \"88b8f0f1-da6d-43fe-940e-23c48e9248d2\") " pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-gwj5b" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.227222 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hv6qx"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.231978 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hv6qx" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.240856 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-67bs5" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.241609 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-t9s4l"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.242856 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-t9s4l" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.252138 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hv6qx"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.271655 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-nwbzk" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.271938 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.279209 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msxbh\" (UniqueName: \"kubernetes.io/projected/400f3b2c-a118-4fab-8c8e-2f4d29f5c354-kube-api-access-msxbh\") pod \"ovn-operator-controller-manager-66cf5c67ff-hv6qx\" (UID: \"400f3b2c-a118-4fab-8c8e-2f4d29f5c354\") " pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hv6qx" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.279265 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67n7f\" (UniqueName: \"kubernetes.io/projected/68f49316-eb77-464a-aebb-189800786fa5-kube-api-access-67n7f\") pod \"octavia-operator-controller-manager-fd75fd47d-h8s6x\" (UID: \"68f49316-eb77-464a-aebb-189800786fa5\") " pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-h8s6x" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.279289 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqrtp\" (UniqueName: \"kubernetes.io/projected/ca7c46cd-e5ff-4732-9718-5c2f08b75221-kube-api-access-vqrtp\") pod \"openstack-baremetal-operator-controller-manager-b58f89467-t9s4l\" (UID: \"ca7c46cd-e5ff-4732-9718-5c2f08b75221\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-t9s4l" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.279311 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jldv\" (UniqueName: \"kubernetes.io/projected/5c262b40-b415-4cca-b1e8-635a9c153d81-kube-api-access-2jldv\") pod \"nova-operator-controller-manager-79556f57fc-sddh8\" (UID: \"5c262b40-b415-4cca-b1e8-635a9c153d81\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-sddh8" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.279338 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca7c46cd-e5ff-4732-9718-5c2f08b75221-cert\") pod \"openstack-baremetal-operator-controller-manager-b58f89467-t9s4l\" (UID: \"ca7c46cd-e5ff-4732-9718-5c2f08b75221\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-t9s4l" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.288956 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-cc9wf" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.289447 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5db546f9d9-kbcml"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.300677 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mtqhm"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.301659 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mtqhm" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.301993 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-kbcml" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.307472 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-87r6q" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.307959 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-2jfcd" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.321783 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-pckfv" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.324006 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jldv\" (UniqueName: \"kubernetes.io/projected/5c262b40-b415-4cca-b1e8-635a9c153d81-kube-api-access-2jldv\") pod \"nova-operator-controller-manager-79556f57fc-sddh8\" (UID: \"5c262b40-b415-4cca-b1e8-635a9c153d81\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-sddh8" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.370642 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-t9s4l"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.377837 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-h57b7" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.380597 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msxbh\" (UniqueName: \"kubernetes.io/projected/400f3b2c-a118-4fab-8c8e-2f4d29f5c354-kube-api-access-msxbh\") pod \"ovn-operator-controller-manager-66cf5c67ff-hv6qx\" (UID: \"400f3b2c-a118-4fab-8c8e-2f4d29f5c354\") " pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hv6qx" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.380655 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67n7f\" (UniqueName: \"kubernetes.io/projected/68f49316-eb77-464a-aebb-189800786fa5-kube-api-access-67n7f\") pod \"octavia-operator-controller-manager-fd75fd47d-h8s6x\" (UID: \"68f49316-eb77-464a-aebb-189800786fa5\") " pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-h8s6x" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.380689 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqrtp\" (UniqueName: \"kubernetes.io/projected/ca7c46cd-e5ff-4732-9718-5c2f08b75221-kube-api-access-vqrtp\") pod \"openstack-baremetal-operator-controller-manager-b58f89467-t9s4l\" (UID: \"ca7c46cd-e5ff-4732-9718-5c2f08b75221\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-t9s4l" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.380734 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9j7tp\" (UniqueName: \"kubernetes.io/projected/dac570d6-3425-4117-be42-1c6d30785a4c-kube-api-access-9j7tp\") pod \"placement-operator-controller-manager-5db546f9d9-kbcml\" (UID: \"dac570d6-3425-4117-be42-1c6d30785a4c\") " pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-kbcml" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.380768 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca7c46cd-e5ff-4732-9718-5c2f08b75221-cert\") pod \"openstack-baremetal-operator-controller-manager-b58f89467-t9s4l\" (UID: \"ca7c46cd-e5ff-4732-9718-5c2f08b75221\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-t9s4l" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.380826 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2jh6\" (UniqueName: \"kubernetes.io/projected/9018dbd6-ab49-4231-a172-33f9a0b56d90-kube-api-access-z2jh6\") pod \"swift-operator-controller-manager-6fdc4fcf86-mtqhm\" (UID: \"9018dbd6-ab49-4231-a172-33f9a0b56d90\") " pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mtqhm" Nov 25 09:41:13 crc kubenswrapper[4776]: E1125 09:41:13.381281 4776 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 25 09:41:13 crc kubenswrapper[4776]: E1125 09:41:13.381326 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca7c46cd-e5ff-4732-9718-5c2f08b75221-cert podName:ca7c46cd-e5ff-4732-9718-5c2f08b75221 nodeName:}" failed. No retries permitted until 2025-11-25 09:41:13.881313125 +0000 UTC m=+1018.922372678 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ca7c46cd-e5ff-4732-9718-5c2f08b75221-cert") pod "openstack-baremetal-operator-controller-manager-b58f89467-t9s4l" (UID: "ca7c46cd-e5ff-4732-9718-5c2f08b75221") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.391237 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mtqhm"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.438122 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msxbh\" (UniqueName: \"kubernetes.io/projected/400f3b2c-a118-4fab-8c8e-2f4d29f5c354-kube-api-access-msxbh\") pod \"ovn-operator-controller-manager-66cf5c67ff-hv6qx\" (UID: \"400f3b2c-a118-4fab-8c8e-2f4d29f5c354\") " pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hv6qx" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.453084 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqrtp\" (UniqueName: \"kubernetes.io/projected/ca7c46cd-e5ff-4732-9718-5c2f08b75221-kube-api-access-vqrtp\") pod \"openstack-baremetal-operator-controller-manager-b58f89467-t9s4l\" (UID: \"ca7c46cd-e5ff-4732-9718-5c2f08b75221\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-t9s4l" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.471206 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67n7f\" (UniqueName: \"kubernetes.io/projected/68f49316-eb77-464a-aebb-189800786fa5-kube-api-access-67n7f\") pod \"octavia-operator-controller-manager-fd75fd47d-h8s6x\" (UID: \"68f49316-eb77-464a-aebb-189800786fa5\") " pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-h8s6x" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.477822 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-gwj5b" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.484134 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2jh6\" (UniqueName: \"kubernetes.io/projected/9018dbd6-ab49-4231-a172-33f9a0b56d90-kube-api-access-z2jh6\") pod \"swift-operator-controller-manager-6fdc4fcf86-mtqhm\" (UID: \"9018dbd6-ab49-4231-a172-33f9a0b56d90\") " pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mtqhm" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.484387 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9j7tp\" (UniqueName: \"kubernetes.io/projected/dac570d6-3425-4117-be42-1c6d30785a4c-kube-api-access-9j7tp\") pod \"placement-operator-controller-manager-5db546f9d9-kbcml\" (UID: \"dac570d6-3425-4117-be42-1c6d30785a4c\") " pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-kbcml" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.493837 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-jr7jl" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.510122 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-567f98c9d-ppmfr"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.512665 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-ppmfr" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.515512 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-kjkh4" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.537775 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-h8s6x" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.549039 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2jh6\" (UniqueName: \"kubernetes.io/projected/9018dbd6-ab49-4231-a172-33f9a0b56d90-kube-api-access-z2jh6\") pod \"swift-operator-controller-manager-6fdc4fcf86-mtqhm\" (UID: \"9018dbd6-ab49-4231-a172-33f9a0b56d90\") " pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mtqhm" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.555303 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9j7tp\" (UniqueName: \"kubernetes.io/projected/dac570d6-3425-4117-be42-1c6d30785a4c-kube-api-access-9j7tp\") pod \"placement-operator-controller-manager-5db546f9d9-kbcml\" (UID: \"dac570d6-3425-4117-be42-1c6d30785a4c\") " pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-kbcml" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.599989 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bf227b94-0dfc-4055-b688-0f73585af089-cert\") pod \"infra-operator-controller-manager-d5cc86f4b-clnxq\" (UID: \"bf227b94-0dfc-4055-b688-0f73585af089\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-clnxq" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.600117 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58gbt\" (UniqueName: \"kubernetes.io/projected/3ebf4825-05f2-4fdd-bc60-d35e80e1b294-kube-api-access-58gbt\") pod \"telemetry-operator-controller-manager-567f98c9d-ppmfr\" (UID: \"3ebf4825-05f2-4fdd-bc60-d35e80e1b294\") " pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-ppmfr" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.600596 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-sddh8" Nov 25 09:41:13 crc kubenswrapper[4776]: E1125 09:41:13.601137 4776 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 25 09:41:13 crc kubenswrapper[4776]: E1125 09:41:13.601236 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bf227b94-0dfc-4055-b688-0f73585af089-cert podName:bf227b94-0dfc-4055-b688-0f73585af089 nodeName:}" failed. No retries permitted until 2025-11-25 09:41:14.601207928 +0000 UTC m=+1019.642267641 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/bf227b94-0dfc-4055-b688-0f73585af089-cert") pod "infra-operator-controller-manager-d5cc86f4b-clnxq" (UID: "bf227b94-0dfc-4055-b688-0f73585af089") : secret "infra-operator-webhook-server-cert" not found Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.621204 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5db546f9d9-kbcml"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.632827 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hv6qx" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.662121 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-567f98c9d-ppmfr"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.691862 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mtqhm" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.700799 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cb74df96-v4xgj"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.700915 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58gbt\" (UniqueName: \"kubernetes.io/projected/3ebf4825-05f2-4fdd-bc60-d35e80e1b294-kube-api-access-58gbt\") pod \"telemetry-operator-controller-manager-567f98c9d-ppmfr\" (UID: \"3ebf4825-05f2-4fdd-bc60-d35e80e1b294\") " pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-ppmfr" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.701780 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-864885998-sxmf5"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.701881 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cb74df96-v4xgj" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.703903 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-864885998-sxmf5"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.704017 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-864885998-sxmf5" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.707181 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-5zgtf" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.708799 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cb74df96-v4xgj"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.709672 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-gvnnv" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.738966 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58gbt\" (UniqueName: \"kubernetes.io/projected/3ebf4825-05f2-4fdd-bc60-d35e80e1b294-kube-api-access-58gbt\") pod \"telemetry-operator-controller-manager-567f98c9d-ppmfr\" (UID: \"3ebf4825-05f2-4fdd-bc60-d35e80e1b294\") " pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-ppmfr" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.743352 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-kbcml" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.786556 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7cd5954d9-mnd72"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.787763 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-mnd72" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.789340 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.790618 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.790656 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-zg6cr" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.795815 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7cd5954d9-mnd72"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.801784 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kthmr\" (UniqueName: \"kubernetes.io/projected/f835b8f9-faca-47e0-99ef-8a11a20c5a56-kube-api-access-kthmr\") pod \"openstack-operator-controller-manager-7cd5954d9-mnd72\" (UID: \"f835b8f9-faca-47e0-99ef-8a11a20c5a56\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-mnd72" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.802045 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csskq\" (UniqueName: \"kubernetes.io/projected/27834511-7fe9-454d-b318-efa218524c0d-kube-api-access-csskq\") pod \"watcher-operator-controller-manager-864885998-sxmf5\" (UID: \"27834511-7fe9-454d-b318-efa218524c0d\") " pod="openstack-operators/watcher-operator-controller-manager-864885998-sxmf5" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.802141 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f835b8f9-faca-47e0-99ef-8a11a20c5a56-metrics-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-mnd72\" (UID: \"f835b8f9-faca-47e0-99ef-8a11a20c5a56\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-mnd72" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.802169 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f835b8f9-faca-47e0-99ef-8a11a20c5a56-webhook-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-mnd72\" (UID: \"f835b8f9-faca-47e0-99ef-8a11a20c5a56\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-mnd72" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.802484 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8hwf\" (UniqueName: \"kubernetes.io/projected/c2be0972-c768-499c-95e3-4cb180ad510c-kube-api-access-n8hwf\") pod \"test-operator-controller-manager-5cb74df96-v4xgj\" (UID: \"c2be0972-c768-499c-95e3-4cb180ad510c\") " pod="openstack-operators/test-operator-controller-manager-5cb74df96-v4xgj" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.823875 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-58r67" event={"ID":"5449efd1-7801-454a-bc48-a5e9e91aba7d","Type":"ContainerStarted","Data":"574b185cfc83549ebfd3e848a43e6706c7f68fb5755936b5e903020cfdf1f22a"} Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.825318 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-zwk9j" event={"ID":"a590fe89-93f7-4c94-bbdd-82531eedf52f","Type":"ContainerStarted","Data":"87d06c68507d0fc53b28c1fb6b8fa7c63f6f3d54145be3c5a0cc25ab472b3049"} Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.827936 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-npwhb"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.828914 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-rnrpd" event={"ID":"2d511feb-5cd8-4ba3-b5af-5d9d01086a82","Type":"ContainerStarted","Data":"c17ef0088cafe7a987622f0e53db003e01db75fd6e8a7bc19ba9bf85dd85bce7"} Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.829013 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-npwhb" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.840331 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-xfzrr" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.849245 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-npwhb"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.851840 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79856dc55c-rnrpd"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.870314 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-86dc4d89c8-58r67"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.907809 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kthmr\" (UniqueName: \"kubernetes.io/projected/f835b8f9-faca-47e0-99ef-8a11a20c5a56-kube-api-access-kthmr\") pod \"openstack-operator-controller-manager-7cd5954d9-mnd72\" (UID: \"f835b8f9-faca-47e0-99ef-8a11a20c5a56\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-mnd72" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.907885 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csskq\" (UniqueName: \"kubernetes.io/projected/27834511-7fe9-454d-b318-efa218524c0d-kube-api-access-csskq\") pod \"watcher-operator-controller-manager-864885998-sxmf5\" (UID: \"27834511-7fe9-454d-b318-efa218524c0d\") " pod="openstack-operators/watcher-operator-controller-manager-864885998-sxmf5" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.907919 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f835b8f9-faca-47e0-99ef-8a11a20c5a56-metrics-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-mnd72\" (UID: \"f835b8f9-faca-47e0-99ef-8a11a20c5a56\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-mnd72" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.907941 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f835b8f9-faca-47e0-99ef-8a11a20c5a56-webhook-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-mnd72\" (UID: \"f835b8f9-faca-47e0-99ef-8a11a20c5a56\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-mnd72" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.908004 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca7c46cd-e5ff-4732-9718-5c2f08b75221-cert\") pod \"openstack-baremetal-operator-controller-manager-b58f89467-t9s4l\" (UID: \"ca7c46cd-e5ff-4732-9718-5c2f08b75221\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-t9s4l" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.908050 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8hwf\" (UniqueName: \"kubernetes.io/projected/c2be0972-c768-499c-95e3-4cb180ad510c-kube-api-access-n8hwf\") pod \"test-operator-controller-manager-5cb74df96-v4xgj\" (UID: \"c2be0972-c768-499c-95e3-4cb180ad510c\") " pod="openstack-operators/test-operator-controller-manager-5cb74df96-v4xgj" Nov 25 09:41:13 crc kubenswrapper[4776]: E1125 09:41:13.908876 4776 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 25 09:41:13 crc kubenswrapper[4776]: E1125 09:41:13.908928 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f835b8f9-faca-47e0-99ef-8a11a20c5a56-metrics-certs podName:f835b8f9-faca-47e0-99ef-8a11a20c5a56 nodeName:}" failed. No retries permitted until 2025-11-25 09:41:14.408910572 +0000 UTC m=+1019.449970125 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f835b8f9-faca-47e0-99ef-8a11a20c5a56-metrics-certs") pod "openstack-operator-controller-manager-7cd5954d9-mnd72" (UID: "f835b8f9-faca-47e0-99ef-8a11a20c5a56") : secret "metrics-server-cert" not found Nov 25 09:41:13 crc kubenswrapper[4776]: E1125 09:41:13.909122 4776 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 25 09:41:13 crc kubenswrapper[4776]: E1125 09:41:13.909153 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f835b8f9-faca-47e0-99ef-8a11a20c5a56-webhook-certs podName:f835b8f9-faca-47e0-99ef-8a11a20c5a56 nodeName:}" failed. No retries permitted until 2025-11-25 09:41:14.409143798 +0000 UTC m=+1019.450203351 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/f835b8f9-faca-47e0-99ef-8a11a20c5a56-webhook-certs") pod "openstack-operator-controller-manager-7cd5954d9-mnd72" (UID: "f835b8f9-faca-47e0-99ef-8a11a20c5a56") : secret "webhook-server-cert" not found Nov 25 09:41:13 crc kubenswrapper[4776]: E1125 09:41:13.909204 4776 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 25 09:41:13 crc kubenswrapper[4776]: E1125 09:41:13.909230 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca7c46cd-e5ff-4732-9718-5c2f08b75221-cert podName:ca7c46cd-e5ff-4732-9718-5c2f08b75221 nodeName:}" failed. No retries permitted until 2025-11-25 09:41:14.9092212 +0000 UTC m=+1019.950280763 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ca7c46cd-e5ff-4732-9718-5c2f08b75221-cert") pod "openstack-baremetal-operator-controller-manager-b58f89467-t9s4l" (UID: "ca7c46cd-e5ff-4732-9718-5c2f08b75221") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.924281 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-7d695c9b56-zwk9j"] Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.932309 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csskq\" (UniqueName: \"kubernetes.io/projected/27834511-7fe9-454d-b318-efa218524c0d-kube-api-access-csskq\") pod \"watcher-operator-controller-manager-864885998-sxmf5\" (UID: \"27834511-7fe9-454d-b318-efa218524c0d\") " pod="openstack-operators/watcher-operator-controller-manager-864885998-sxmf5" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.949032 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kthmr\" (UniqueName: \"kubernetes.io/projected/f835b8f9-faca-47e0-99ef-8a11a20c5a56-kube-api-access-kthmr\") pod \"openstack-operator-controller-manager-7cd5954d9-mnd72\" (UID: \"f835b8f9-faca-47e0-99ef-8a11a20c5a56\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-mnd72" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.949373 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-ppmfr" Nov 25 09:41:13 crc kubenswrapper[4776]: I1125 09:41:13.956191 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8hwf\" (UniqueName: \"kubernetes.io/projected/c2be0972-c768-499c-95e3-4cb180ad510c-kube-api-access-n8hwf\") pod \"test-operator-controller-manager-5cb74df96-v4xgj\" (UID: \"c2be0972-c768-499c-95e3-4cb180ad510c\") " pod="openstack-operators/test-operator-controller-manager-5cb74df96-v4xgj" Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.009800 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cb74df96-v4xgj" Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.010896 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22gb8\" (UniqueName: \"kubernetes.io/projected/8f2c9384-1b11-4925-8338-babf955143b1-kube-api-access-22gb8\") pod \"rabbitmq-cluster-operator-manager-668c99d594-npwhb\" (UID: \"8f2c9384-1b11-4925-8338-babf955143b1\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-npwhb" Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.013527 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c9694994-khxdw"] Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.036095 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-864885998-sxmf5" Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.051124 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-774b86978c-gmcp9"] Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.114768 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22gb8\" (UniqueName: \"kubernetes.io/projected/8f2c9384-1b11-4925-8338-babf955143b1-kube-api-access-22gb8\") pod \"rabbitmq-cluster-operator-manager-668c99d594-npwhb\" (UID: \"8f2c9384-1b11-4925-8338-babf955143b1\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-npwhb" Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.138391 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22gb8\" (UniqueName: \"kubernetes.io/projected/8f2c9384-1b11-4925-8338-babf955143b1-kube-api-access-22gb8\") pod \"rabbitmq-cluster-operator-manager-668c99d594-npwhb\" (UID: \"8f2c9384-1b11-4925-8338-babf955143b1\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-npwhb" Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.279686 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58bb8d67cc-h57b7"] Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.295590 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-748dc6576f-pckfv"] Nov 25 09:41:14 crc kubenswrapper[4776]: W1125 09:41:14.297461 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a21bb84_b9c5_4e6f_8fbb_3c2bc65dd409.slice/crio-7b638a4b5e45b70cbfcc1bc4e0461e996dcad765c60bbf8ce19643c8197edbf0 WatchSource:0}: Error finding container 7b638a4b5e45b70cbfcc1bc4e0461e996dcad765c60bbf8ce19643c8197edbf0: Status 404 returned error can't find the container with id 7b638a4b5e45b70cbfcc1bc4e0461e996dcad765c60bbf8ce19643c8197edbf0 Nov 25 09:41:14 crc kubenswrapper[4776]: W1125 09:41:14.299902 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podebac9e21_7ae6_45e4_a6a8_9728b3670733.slice/crio-6caeaec560ddb2d72f94043d5a6dd586975eec3e410a408ed6f13aef69403b3e WatchSource:0}: Error finding container 6caeaec560ddb2d72f94043d5a6dd586975eec3e410a408ed6f13aef69403b3e: Status 404 returned error can't find the container with id 6caeaec560ddb2d72f94043d5a6dd586975eec3e410a408ed6f13aef69403b3e Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.302261 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bfcdc958c-cc9wf"] Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.306767 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-68b95954c9-h6k9v"] Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.379413 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-npwhb" Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.409623 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-gwj5b"] Nov 25 09:41:14 crc kubenswrapper[4776]: W1125 09:41:14.417153 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88b8f0f1_da6d_43fe_940e_23c48e9248d2.slice/crio-8e3f82c95073dbb81eaf831e972c4b581987df31347030ab7ef9479b1ea19aaf WatchSource:0}: Error finding container 8e3f82c95073dbb81eaf831e972c4b581987df31347030ab7ef9479b1ea19aaf: Status 404 returned error can't find the container with id 8e3f82c95073dbb81eaf831e972c4b581987df31347030ab7ef9479b1ea19aaf Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.419723 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f835b8f9-faca-47e0-99ef-8a11a20c5a56-metrics-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-mnd72\" (UID: \"f835b8f9-faca-47e0-99ef-8a11a20c5a56\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-mnd72" Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.419833 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f835b8f9-faca-47e0-99ef-8a11a20c5a56-webhook-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-mnd72\" (UID: \"f835b8f9-faca-47e0-99ef-8a11a20c5a56\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-mnd72" Nov 25 09:41:14 crc kubenswrapper[4776]: E1125 09:41:14.420146 4776 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 25 09:41:14 crc kubenswrapper[4776]: E1125 09:41:14.420267 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f835b8f9-faca-47e0-99ef-8a11a20c5a56-webhook-certs podName:f835b8f9-faca-47e0-99ef-8a11a20c5a56 nodeName:}" failed. No retries permitted until 2025-11-25 09:41:15.420226891 +0000 UTC m=+1020.461286444 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/f835b8f9-faca-47e0-99ef-8a11a20c5a56-webhook-certs") pod "openstack-operator-controller-manager-7cd5954d9-mnd72" (UID: "f835b8f9-faca-47e0-99ef-8a11a20c5a56") : secret "webhook-server-cert" not found Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.425812 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f835b8f9-faca-47e0-99ef-8a11a20c5a56-metrics-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-mnd72\" (UID: \"f835b8f9-faca-47e0-99ef-8a11a20c5a56\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-mnd72" Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.500026 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-jr7jl"] Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.511003 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mtqhm"] Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.519755 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hv6qx"] Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.527018 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-fd75fd47d-h8s6x"] Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.534343 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5db546f9d9-kbcml"] Nov 25 09:41:14 crc kubenswrapper[4776]: W1125 09:41:14.548040 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68f49316_eb77_464a_aebb_189800786fa5.slice/crio-c4dc3df4a620da8d1164a23be69177e6e579d90f5751cad0d3f001e45bd77b56 WatchSource:0}: Error finding container c4dc3df4a620da8d1164a23be69177e6e579d90f5751cad0d3f001e45bd77b56: Status 404 returned error can't find the container with id c4dc3df4a620da8d1164a23be69177e6e579d90f5751cad0d3f001e45bd77b56 Nov 25 09:41:14 crc kubenswrapper[4776]: E1125 09:41:14.552813 4776 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-67n7f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-fd75fd47d-h8s6x_openstack-operators(68f49316-eb77-464a-aebb-189800786fa5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 09:41:14 crc kubenswrapper[4776]: E1125 09:41:14.555887 4776 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-67n7f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-fd75fd47d-h8s6x_openstack-operators(68f49316-eb77-464a-aebb-189800786fa5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 09:41:14 crc kubenswrapper[4776]: E1125 09:41:14.557686 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-h8s6x" podUID="68f49316-eb77-464a-aebb-189800786fa5" Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.626051 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bf227b94-0dfc-4055-b688-0f73585af089-cert\") pod \"infra-operator-controller-manager-d5cc86f4b-clnxq\" (UID: \"bf227b94-0dfc-4055-b688-0f73585af089\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-clnxq" Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.634347 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bf227b94-0dfc-4055-b688-0f73585af089-cert\") pod \"infra-operator-controller-manager-d5cc86f4b-clnxq\" (UID: \"bf227b94-0dfc-4055-b688-0f73585af089\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-clnxq" Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.636698 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-567f98c9d-ppmfr"] Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.647715 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-sddh8"] Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.660724 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-npwhb"] Nov 25 09:41:14 crc kubenswrapper[4776]: E1125 09:41:14.674978 4776 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-58gbt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-567f98c9d-ppmfr_openstack-operators(3ebf4825-05f2-4fdd-bc60-d35e80e1b294): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 09:41:14 crc kubenswrapper[4776]: W1125 09:41:14.677441 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f2c9384_1b11_4925_8338_babf955143b1.slice/crio-b3da3cdfb4e3010198f4247539ca2921669ecc9693acf605ac7050eb99df2e78 WatchSource:0}: Error finding container b3da3cdfb4e3010198f4247539ca2921669ecc9693acf605ac7050eb99df2e78: Status 404 returned error can't find the container with id b3da3cdfb4e3010198f4247539ca2921669ecc9693acf605ac7050eb99df2e78 Nov 25 09:41:14 crc kubenswrapper[4776]: E1125 09:41:14.693623 4776 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-58gbt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-567f98c9d-ppmfr_openstack-operators(3ebf4825-05f2-4fdd-bc60-d35e80e1b294): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 09:41:14 crc kubenswrapper[4776]: E1125 09:41:14.694795 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-ppmfr" podUID="3ebf4825-05f2-4fdd-bc60-d35e80e1b294" Nov 25 09:41:14 crc kubenswrapper[4776]: E1125 09:41:14.704355 4776 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-22gb8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-npwhb_openstack-operators(8f2c9384-1b11-4925-8338-babf955143b1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 09:41:14 crc kubenswrapper[4776]: E1125 09:41:14.705685 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-npwhb" podUID="8f2c9384-1b11-4925-8338-babf955143b1" Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.709604 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-864885998-sxmf5"] Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.714147 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cb74df96-v4xgj"] Nov 25 09:41:14 crc kubenswrapper[4776]: W1125 09:41:14.718682 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27834511_7fe9_454d_b318_efa218524c0d.slice/crio-9d841f0b57d95d6e13bfc1fe8fc59f79680129eaa308cd62d6799dea41e67139 WatchSource:0}: Error finding container 9d841f0b57d95d6e13bfc1fe8fc59f79680129eaa308cd62d6799dea41e67139: Status 404 returned error can't find the container with id 9d841f0b57d95d6e13bfc1fe8fc59f79680129eaa308cd62d6799dea41e67139 Nov 25 09:41:14 crc kubenswrapper[4776]: W1125 09:41:14.720243 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc2be0972_c768_499c_95e3_4cb180ad510c.slice/crio-edc2b611fee51532d63795f0f0a6ddfa4ad84c693072173d42f43efb901d9ac1 WatchSource:0}: Error finding container edc2b611fee51532d63795f0f0a6ddfa4ad84c693072173d42f43efb901d9ac1: Status 404 returned error can't find the container with id edc2b611fee51532d63795f0f0a6ddfa4ad84c693072173d42f43efb901d9ac1 Nov 25 09:41:14 crc kubenswrapper[4776]: E1125 09:41:14.723056 4776 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n8hwf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cb74df96-v4xgj_openstack-operators(c2be0972-c768-499c-95e3-4cb180ad510c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 09:41:14 crc kubenswrapper[4776]: E1125 09:41:14.724725 4776 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n8hwf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cb74df96-v4xgj_openstack-operators(c2be0972-c768-499c-95e3-4cb180ad510c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 09:41:14 crc kubenswrapper[4776]: E1125 09:41:14.725126 4776 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-csskq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-864885998-sxmf5_openstack-operators(27834511-7fe9-454d-b318-efa218524c0d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 09:41:14 crc kubenswrapper[4776]: E1125 09:41:14.726385 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5cb74df96-v4xgj" podUID="c2be0972-c768-499c-95e3-4cb180ad510c" Nov 25 09:41:14 crc kubenswrapper[4776]: E1125 09:41:14.727780 4776 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-csskq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-864885998-sxmf5_openstack-operators(27834511-7fe9-454d-b318-efa218524c0d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 09:41:14 crc kubenswrapper[4776]: E1125 09:41:14.729038 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-864885998-sxmf5" podUID="27834511-7fe9-454d-b318-efa218524c0d" Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.766420 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-clnxq" Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.836033 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-gwj5b" event={"ID":"88b8f0f1-da6d-43fe-940e-23c48e9248d2","Type":"ContainerStarted","Data":"8e3f82c95073dbb81eaf831e972c4b581987df31347030ab7ef9479b1ea19aaf"} Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.837434 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-h6k9v" event={"ID":"7a21bb84-b9c5-4e6f-8fbb-3c2bc65dd409","Type":"ContainerStarted","Data":"7b638a4b5e45b70cbfcc1bc4e0461e996dcad765c60bbf8ce19643c8197edbf0"} Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.838472 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-h57b7" event={"ID":"abb2c6a9-56ce-45ff-902f-ac49a249fd81","Type":"ContainerStarted","Data":"f86905825990758569f72cf40df1a85dd27d6bba74e39b806ae391f1527b6300"} Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.839931 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cb74df96-v4xgj" event={"ID":"c2be0972-c768-499c-95e3-4cb180ad510c","Type":"ContainerStarted","Data":"edc2b611fee51532d63795f0f0a6ddfa4ad84c693072173d42f43efb901d9ac1"} Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.842889 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hv6qx" event={"ID":"400f3b2c-a118-4fab-8c8e-2f4d29f5c354","Type":"ContainerStarted","Data":"53ae5c7565b3d63ebbe7dbd6bdf0912ee47dca7eca5127647e54deb33f37de41"} Nov 25 09:41:14 crc kubenswrapper[4776]: E1125 09:41:14.844095 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5cb74df96-v4xgj" podUID="c2be0972-c768-499c-95e3-4cb180ad510c" Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.844601 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-h8s6x" event={"ID":"68f49316-eb77-464a-aebb-189800786fa5","Type":"ContainerStarted","Data":"c4dc3df4a620da8d1164a23be69177e6e579d90f5751cad0d3f001e45bd77b56"} Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.847426 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-khxdw" event={"ID":"187d1384-3de2-4135-abac-45f15155b942","Type":"ContainerStarted","Data":"51fbdc241f7781f02db8296f9eacd91bc38d3109ffe7ffd2ee7423dbc7ae98e5"} Nov 25 09:41:14 crc kubenswrapper[4776]: E1125 09:41:14.850196 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-h8s6x" podUID="68f49316-eb77-464a-aebb-189800786fa5" Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.864381 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-864885998-sxmf5" event={"ID":"27834511-7fe9-454d-b318-efa218524c0d","Type":"ContainerStarted","Data":"9d841f0b57d95d6e13bfc1fe8fc59f79680129eaa308cd62d6799dea41e67139"} Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.865958 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-jr7jl" event={"ID":"7af82806-526b-4fe2-8314-65298d1a0539","Type":"ContainerStarted","Data":"d37ea961b7789137386e3b56141b7dfea1215a1dc84629ed5defbf1d0a426072"} Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.868082 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mtqhm" event={"ID":"9018dbd6-ab49-4231-a172-33f9a0b56d90","Type":"ContainerStarted","Data":"86018a6444bce022b3f29f1f377d69fa990567c3b90c0d8a789724700c02b678"} Nov 25 09:41:14 crc kubenswrapper[4776]: E1125 09:41:14.869118 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-864885998-sxmf5" podUID="27834511-7fe9-454d-b318-efa218524c0d" Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.869710 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-npwhb" event={"ID":"8f2c9384-1b11-4925-8338-babf955143b1","Type":"ContainerStarted","Data":"b3da3cdfb4e3010198f4247539ca2921669ecc9693acf605ac7050eb99df2e78"} Nov 25 09:41:14 crc kubenswrapper[4776]: E1125 09:41:14.871675 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-npwhb" podUID="8f2c9384-1b11-4925-8338-babf955143b1" Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.873736 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-gmcp9" event={"ID":"61810388-3bac-4a06-88ea-98cc0337fdee","Type":"ContainerStarted","Data":"1097a0ef84cca1870958615aeb51e7dc1d4ad6d4f31242f5dbced42baf2df70c"} Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.879010 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-sddh8" event={"ID":"5c262b40-b415-4cca-b1e8-635a9c153d81","Type":"ContainerStarted","Data":"076ac17d1f409617f33b2300182cd624919b86bc1a86aa182f3e648205f904e8"} Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.881614 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-cc9wf" event={"ID":"ebac9e21-7ae6-45e4-a6a8-9728b3670733","Type":"ContainerStarted","Data":"6caeaec560ddb2d72f94043d5a6dd586975eec3e410a408ed6f13aef69403b3e"} Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.885423 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-kbcml" event={"ID":"dac570d6-3425-4117-be42-1c6d30785a4c","Type":"ContainerStarted","Data":"cf26e95c54a296d5beaccd2465cd71b068848d29283853db86f43e6396934ff3"} Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.888033 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-ppmfr" event={"ID":"3ebf4825-05f2-4fdd-bc60-d35e80e1b294","Type":"ContainerStarted","Data":"af1a89d983524d575334b982b6519ec54d844c21768280c3b1d98d6ace6861d4"} Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.894390 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-pckfv" event={"ID":"63e1e612-6ab8-4144-b2cc-c8354bd0894a","Type":"ContainerStarted","Data":"561994a36d06fae490718ce9d27bb65815b8aa7fc5b9a81b952e0a7081d14894"} Nov 25 09:41:14 crc kubenswrapper[4776]: E1125 09:41:14.894697 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-ppmfr" podUID="3ebf4825-05f2-4fdd-bc60-d35e80e1b294" Nov 25 09:41:14 crc kubenswrapper[4776]: I1125 09:41:14.931344 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca7c46cd-e5ff-4732-9718-5c2f08b75221-cert\") pod \"openstack-baremetal-operator-controller-manager-b58f89467-t9s4l\" (UID: \"ca7c46cd-e5ff-4732-9718-5c2f08b75221\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-t9s4l" Nov 25 09:41:14 crc kubenswrapper[4776]: E1125 09:41:14.932380 4776 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 25 09:41:14 crc kubenswrapper[4776]: E1125 09:41:14.932442 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca7c46cd-e5ff-4732-9718-5c2f08b75221-cert podName:ca7c46cd-e5ff-4732-9718-5c2f08b75221 nodeName:}" failed. No retries permitted until 2025-11-25 09:41:16.932423002 +0000 UTC m=+1021.973482555 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ca7c46cd-e5ff-4732-9718-5c2f08b75221-cert") pod "openstack-baremetal-operator-controller-manager-b58f89467-t9s4l" (UID: "ca7c46cd-e5ff-4732-9718-5c2f08b75221") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 25 09:41:15 crc kubenswrapper[4776]: I1125 09:41:15.209575 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-d5cc86f4b-clnxq"] Nov 25 09:41:15 crc kubenswrapper[4776]: I1125 09:41:15.438383 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f835b8f9-faca-47e0-99ef-8a11a20c5a56-webhook-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-mnd72\" (UID: \"f835b8f9-faca-47e0-99ef-8a11a20c5a56\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-mnd72" Nov 25 09:41:15 crc kubenswrapper[4776]: E1125 09:41:15.438916 4776 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 25 09:41:15 crc kubenswrapper[4776]: E1125 09:41:15.439028 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f835b8f9-faca-47e0-99ef-8a11a20c5a56-webhook-certs podName:f835b8f9-faca-47e0-99ef-8a11a20c5a56 nodeName:}" failed. No retries permitted until 2025-11-25 09:41:17.439006152 +0000 UTC m=+1022.480065735 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/f835b8f9-faca-47e0-99ef-8a11a20c5a56-webhook-certs") pod "openstack-operator-controller-manager-7cd5954d9-mnd72" (UID: "f835b8f9-faca-47e0-99ef-8a11a20c5a56") : secret "webhook-server-cert" not found Nov 25 09:41:15 crc kubenswrapper[4776]: I1125 09:41:15.913530 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-clnxq" event={"ID":"bf227b94-0dfc-4055-b688-0f73585af089","Type":"ContainerStarted","Data":"e4cef4c4ec8f02a47b1679a17b4254aebb73c516cd8742a897c2e5612409acaa"} Nov 25 09:41:15 crc kubenswrapper[4776]: E1125 09:41:15.917981 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-npwhb" podUID="8f2c9384-1b11-4925-8338-babf955143b1" Nov 25 09:41:15 crc kubenswrapper[4776]: E1125 09:41:15.919247 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:442c269d79163f8da75505019c02e9f0815837aaadcaddacb8e6c12df297ca13\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-h8s6x" podUID="68f49316-eb77-464a-aebb-189800786fa5" Nov 25 09:41:15 crc kubenswrapper[4776]: E1125 09:41:15.919266 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-ppmfr" podUID="3ebf4825-05f2-4fdd-bc60-d35e80e1b294" Nov 25 09:41:15 crc kubenswrapper[4776]: E1125 09:41:15.919581 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-864885998-sxmf5" podUID="27834511-7fe9-454d-b318-efa218524c0d" Nov 25 09:41:15 crc kubenswrapper[4776]: E1125 09:41:15.921122 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5cb74df96-v4xgj" podUID="c2be0972-c768-499c-95e3-4cb180ad510c" Nov 25 09:41:16 crc kubenswrapper[4776]: I1125 09:41:16.963986 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca7c46cd-e5ff-4732-9718-5c2f08b75221-cert\") pod \"openstack-baremetal-operator-controller-manager-b58f89467-t9s4l\" (UID: \"ca7c46cd-e5ff-4732-9718-5c2f08b75221\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-t9s4l" Nov 25 09:41:16 crc kubenswrapper[4776]: I1125 09:41:16.972338 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ca7c46cd-e5ff-4732-9718-5c2f08b75221-cert\") pod \"openstack-baremetal-operator-controller-manager-b58f89467-t9s4l\" (UID: \"ca7c46cd-e5ff-4732-9718-5c2f08b75221\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-t9s4l" Nov 25 09:41:17 crc kubenswrapper[4776]: I1125 09:41:17.260413 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-nwbzk" Nov 25 09:41:17 crc kubenswrapper[4776]: I1125 09:41:17.268773 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-t9s4l" Nov 25 09:41:17 crc kubenswrapper[4776]: I1125 09:41:17.474958 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f835b8f9-faca-47e0-99ef-8a11a20c5a56-webhook-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-mnd72\" (UID: \"f835b8f9-faca-47e0-99ef-8a11a20c5a56\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-mnd72" Nov 25 09:41:17 crc kubenswrapper[4776]: I1125 09:41:17.487038 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f835b8f9-faca-47e0-99ef-8a11a20c5a56-webhook-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-mnd72\" (UID: \"f835b8f9-faca-47e0-99ef-8a11a20c5a56\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-mnd72" Nov 25 09:41:17 crc kubenswrapper[4776]: I1125 09:41:17.655274 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-zg6cr" Nov 25 09:41:17 crc kubenswrapper[4776]: I1125 09:41:17.664956 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-mnd72" Nov 25 09:41:17 crc kubenswrapper[4776]: I1125 09:41:17.818944 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:41:17 crc kubenswrapper[4776]: I1125 09:41:17.819121 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:41:23 crc kubenswrapper[4776]: I1125 09:41:23.802118 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7cd5954d9-mnd72"] Nov 25 09:41:23 crc kubenswrapper[4776]: I1125 09:41:23.861691 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-t9s4l"] Nov 25 09:41:23 crc kubenswrapper[4776]: W1125 09:41:23.938288 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca7c46cd_e5ff_4732_9718_5c2f08b75221.slice/crio-f3723456e934aaf3f8b173816e4bd351834465099d289140320c25387a352d5c WatchSource:0}: Error finding container f3723456e934aaf3f8b173816e4bd351834465099d289140320c25387a352d5c: Status 404 returned error can't find the container with id f3723456e934aaf3f8b173816e4bd351834465099d289140320c25387a352d5c Nov 25 09:41:23 crc kubenswrapper[4776]: E1125 09:41:23.951291 4776 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2jldv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-79556f57fc-sddh8_openstack-operators(5c262b40-b415-4cca-b1e8-635a9c153d81): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 09:41:23 crc kubenswrapper[4776]: E1125 09:41:23.952551 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-sddh8" podUID="5c262b40-b415-4cca-b1e8-635a9c153d81" Nov 25 09:41:23 crc kubenswrapper[4776]: I1125 09:41:23.966405 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-kbcml" event={"ID":"dac570d6-3425-4117-be42-1c6d30785a4c","Type":"ContainerStarted","Data":"cf23e70e43e7c8a8b7bdd7a1c028a9020c8022fcd6e938181035104be86b7a30"} Nov 25 09:41:23 crc kubenswrapper[4776]: I1125 09:41:23.974269 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-mnd72" event={"ID":"f835b8f9-faca-47e0-99ef-8a11a20c5a56","Type":"ContainerStarted","Data":"1af044b096132cdb8cb601cf01682c8f75d5db785b57141f66209b362f4bf2ca"} Nov 25 09:41:23 crc kubenswrapper[4776]: I1125 09:41:23.995226 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-h6k9v" event={"ID":"7a21bb84-b9c5-4e6f-8fbb-3c2bc65dd409","Type":"ContainerStarted","Data":"cabdc0ce1c5403a988c4c3ee1e34325a76be14900581f0f64211a7a13edfb542"} Nov 25 09:41:24 crc kubenswrapper[4776]: I1125 09:41:24.010353 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-jr7jl" event={"ID":"7af82806-526b-4fe2-8314-65298d1a0539","Type":"ContainerStarted","Data":"c36dd3753c2efd15b50a312c97869a88af4ae70a7985dfe6455e13559e6d8f5b"} Nov 25 09:41:24 crc kubenswrapper[4776]: I1125 09:41:24.019470 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-h57b7" event={"ID":"abb2c6a9-56ce-45ff-902f-ac49a249fd81","Type":"ContainerStarted","Data":"6093feeffb9769b0f00948aee835b06d246b5c8e740ee4e61a7d467aa36c5265"} Nov 25 09:41:24 crc kubenswrapper[4776]: I1125 09:41:24.028450 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-58r67" event={"ID":"5449efd1-7801-454a-bc48-a5e9e91aba7d","Type":"ContainerStarted","Data":"6604ef8f7b2953c545ebd6560d2fb5cd69306b8c019478f62ef72e1b843889c0"} Nov 25 09:41:24 crc kubenswrapper[4776]: E1125 09:41:24.047080 4776 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-t26gf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-7c57c8bbc4-gwj5b_openstack-operators(88b8f0f1-da6d-43fe-940e-23c48e9248d2): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 09:41:24 crc kubenswrapper[4776]: I1125 09:41:24.047696 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-cc9wf" event={"ID":"ebac9e21-7ae6-45e4-a6a8-9728b3670733","Type":"ContainerStarted","Data":"420c5efc7b875417db8f906c5a8274ec2980e0edbf870b1a2bd38a2f41ae64c8"} Nov 25 09:41:24 crc kubenswrapper[4776]: E1125 09:41:24.053494 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-gwj5b" podUID="88b8f0f1-da6d-43fe-940e-23c48e9248d2" Nov 25 09:41:24 crc kubenswrapper[4776]: I1125 09:41:24.064641 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-pckfv" event={"ID":"63e1e612-6ab8-4144-b2cc-c8354bd0894a","Type":"ContainerStarted","Data":"b2ea8133569d1a1ecfd0788b6ac347ffb97dbebe4981da0ba786875c6eeed85d"} Nov 25 09:41:24 crc kubenswrapper[4776]: I1125 09:41:24.085016 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-clnxq" event={"ID":"bf227b94-0dfc-4055-b688-0f73585af089","Type":"ContainerStarted","Data":"cea0983ddafbc210105240567a5fdcde90ff06fdc556501abbcc87d139349ec6"} Nov 25 09:41:24 crc kubenswrapper[4776]: I1125 09:41:24.109771 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-t9s4l" event={"ID":"ca7c46cd-e5ff-4732-9718-5c2f08b75221","Type":"ContainerStarted","Data":"f3723456e934aaf3f8b173816e4bd351834465099d289140320c25387a352d5c"} Nov 25 09:41:24 crc kubenswrapper[4776]: E1125 09:41:24.116031 4776 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7pk8t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-79856dc55c-rnrpd_openstack-operators(2d511feb-5cd8-4ba3-b5af-5d9d01086a82): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 09:41:24 crc kubenswrapper[4776]: E1125 09:41:24.118620 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-rnrpd" podUID="2d511feb-5cd8-4ba3-b5af-5d9d01086a82" Nov 25 09:41:24 crc kubenswrapper[4776]: E1125 09:41:24.121299 4776 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dkcnc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-7d695c9b56-zwk9j_openstack-operators(a590fe89-93f7-4c94-bbdd-82531eedf52f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 09:41:24 crc kubenswrapper[4776]: E1125 09:41:24.122510 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-zwk9j" podUID="a590fe89-93f7-4c94-bbdd-82531eedf52f" Nov 25 09:41:24 crc kubenswrapper[4776]: I1125 09:41:24.134678 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-sddh8" event={"ID":"5c262b40-b415-4cca-b1e8-635a9c153d81","Type":"ContainerStarted","Data":"1418b243d99a762145ccae73d14259384adaae8ed8d32b475b5f8731342a2a85"} Nov 25 09:41:24 crc kubenswrapper[4776]: I1125 09:41:24.135637 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-sddh8" Nov 25 09:41:24 crc kubenswrapper[4776]: E1125 09:41:24.143360 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-sddh8" podUID="5c262b40-b415-4cca-b1e8-635a9c153d81" Nov 25 09:41:24 crc kubenswrapper[4776]: E1125 09:41:24.160538 4776 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-msxbh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-66cf5c67ff-hv6qx_openstack-operators(400f3b2c-a118-4fab-8c8e-2f4d29f5c354): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 09:41:24 crc kubenswrapper[4776]: E1125 09:41:24.164001 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hv6qx" podUID="400f3b2c-a118-4fab-8c8e-2f4d29f5c354" Nov 25 09:41:24 crc kubenswrapper[4776]: I1125 09:41:24.199240 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mtqhm" event={"ID":"9018dbd6-ab49-4231-a172-33f9a0b56d90","Type":"ContainerStarted","Data":"a300f3d8cb5ecad8d226197f821ead85befd9eb468ae0d9c945f1edb4affecc5"} Nov 25 09:41:24 crc kubenswrapper[4776]: I1125 09:41:24.251287 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-gmcp9" event={"ID":"61810388-3bac-4a06-88ea-98cc0337fdee","Type":"ContainerStarted","Data":"a834ffbaf6ea9efe9b553aa197629909b654913ecf0700320488c9cda46d5fde"} Nov 25 09:41:25 crc kubenswrapper[4776]: I1125 09:41:25.268272 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-khxdw" event={"ID":"187d1384-3de2-4135-abac-45f15155b942","Type":"ContainerStarted","Data":"f307049c7ba04a0c35fc166f771c4a721961a635b502dd2b4c09150d2b6e0ded"} Nov 25 09:41:25 crc kubenswrapper[4776]: I1125 09:41:25.269965 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-mnd72" event={"ID":"f835b8f9-faca-47e0-99ef-8a11a20c5a56","Type":"ContainerStarted","Data":"1ce56954215a0cfa5e8c355ac018c1c680187f358c61682ac88dd27226bc002c"} Nov 25 09:41:25 crc kubenswrapper[4776]: I1125 09:41:25.271098 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-mnd72" Nov 25 09:41:25 crc kubenswrapper[4776]: I1125 09:41:25.273862 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-gwj5b" event={"ID":"88b8f0f1-da6d-43fe-940e-23c48e9248d2","Type":"ContainerStarted","Data":"d52053cc41789ba6e5e412ea0d71ef73081932120c2718faa75513e4b7fcfd20"} Nov 25 09:41:25 crc kubenswrapper[4776]: I1125 09:41:25.274552 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-gwj5b" Nov 25 09:41:25 crc kubenswrapper[4776]: E1125 09:41:25.275626 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-gwj5b" podUID="88b8f0f1-da6d-43fe-940e-23c48e9248d2" Nov 25 09:41:25 crc kubenswrapper[4776]: I1125 09:41:25.277749 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-zwk9j" event={"ID":"a590fe89-93f7-4c94-bbdd-82531eedf52f","Type":"ContainerStarted","Data":"50209a1b138f05ac00b573e57f5ec3fb2acdeb115bf8e47a807817ac3593c1f5"} Nov 25 09:41:25 crc kubenswrapper[4776]: I1125 09:41:25.278670 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-zwk9j" Nov 25 09:41:25 crc kubenswrapper[4776]: E1125 09:41:25.279684 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-zwk9j" podUID="a590fe89-93f7-4c94-bbdd-82531eedf52f" Nov 25 09:41:25 crc kubenswrapper[4776]: I1125 09:41:25.281478 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-rnrpd" event={"ID":"2d511feb-5cd8-4ba3-b5af-5d9d01086a82","Type":"ContainerStarted","Data":"5d6d76d556e8278428f98e5af6a98414c3eee3a5a166207b8a8252bca5da5c19"} Nov 25 09:41:25 crc kubenswrapper[4776]: I1125 09:41:25.282610 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-rnrpd" Nov 25 09:41:25 crc kubenswrapper[4776]: E1125 09:41:25.283291 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-rnrpd" podUID="2d511feb-5cd8-4ba3-b5af-5d9d01086a82" Nov 25 09:41:25 crc kubenswrapper[4776]: I1125 09:41:25.284961 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hv6qx" event={"ID":"400f3b2c-a118-4fab-8c8e-2f4d29f5c354","Type":"ContainerStarted","Data":"c529c7ab1b0d57e5500390b3280904c39647866b01b902b950b3cec7494864bf"} Nov 25 09:41:25 crc kubenswrapper[4776]: I1125 09:41:25.285191 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hv6qx" Nov 25 09:41:25 crc kubenswrapper[4776]: E1125 09:41:25.286006 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hv6qx" podUID="400f3b2c-a118-4fab-8c8e-2f4d29f5c354" Nov 25 09:41:25 crc kubenswrapper[4776]: E1125 09:41:25.286867 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-sddh8" podUID="5c262b40-b415-4cca-b1e8-635a9c153d81" Nov 25 09:41:25 crc kubenswrapper[4776]: I1125 09:41:25.302584 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-mnd72" podStartSLOduration=12.302562742 podStartE2EDuration="12.302562742s" podCreationTimestamp="2025-11-25 09:41:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:41:25.298544501 +0000 UTC m=+1030.339604074" watchObservedRunningTime="2025-11-25 09:41:25.302562742 +0000 UTC m=+1030.343622295" Nov 25 09:41:26 crc kubenswrapper[4776]: E1125 09:41:26.292558 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-gwj5b" podUID="88b8f0f1-da6d-43fe-940e-23c48e9248d2" Nov 25 09:41:26 crc kubenswrapper[4776]: E1125 09:41:26.292885 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hv6qx" podUID="400f3b2c-a118-4fab-8c8e-2f4d29f5c354" Nov 25 09:41:26 crc kubenswrapper[4776]: E1125 09:41:26.294602 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-rnrpd" podUID="2d511feb-5cd8-4ba3-b5af-5d9d01086a82" Nov 25 09:41:26 crc kubenswrapper[4776]: E1125 09:41:26.295158 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-zwk9j" podUID="a590fe89-93f7-4c94-bbdd-82531eedf52f" Nov 25 09:41:30 crc kubenswrapper[4776]: I1125 09:41:30.316231 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-h57b7" event={"ID":"abb2c6a9-56ce-45ff-902f-ac49a249fd81","Type":"ContainerStarted","Data":"9a7f7674231e2f4caded3aeca9c024ce0fe3b8910e8384d709a6c39df748e114"} Nov 25 09:41:30 crc kubenswrapper[4776]: I1125 09:41:30.316791 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-h57b7" Nov 25 09:41:30 crc kubenswrapper[4776]: I1125 09:41:30.318534 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-h57b7" Nov 25 09:41:30 crc kubenswrapper[4776]: I1125 09:41:30.318618 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-ppmfr" event={"ID":"3ebf4825-05f2-4fdd-bc60-d35e80e1b294","Type":"ContainerStarted","Data":"89f5825ad96d2a6c56bb0eabb4fe49912f87311b5d84ab1d9d0d664e8eda696b"} Nov 25 09:41:30 crc kubenswrapper[4776]: I1125 09:41:30.344445 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-h57b7" podStartSLOduration=2.545301667 podStartE2EDuration="18.344423122s" podCreationTimestamp="2025-11-25 09:41:12 +0000 UTC" firstStartedPulling="2025-11-25 09:41:14.287290008 +0000 UTC m=+1019.328349551" lastFinishedPulling="2025-11-25 09:41:30.086411453 +0000 UTC m=+1035.127471006" observedRunningTime="2025-11-25 09:41:30.337915239 +0000 UTC m=+1035.378974812" watchObservedRunningTime="2025-11-25 09:41:30.344423122 +0000 UTC m=+1035.385482695" Nov 25 09:41:31 crc kubenswrapper[4776]: I1125 09:41:31.326545 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-t9s4l" event={"ID":"ca7c46cd-e5ff-4732-9718-5c2f08b75221","Type":"ContainerStarted","Data":"66ea926b64efe0ee4ca027dacc596923a61963508f135222166b1098b3d8f5f2"} Nov 25 09:41:31 crc kubenswrapper[4776]: I1125 09:41:31.326589 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-t9s4l" event={"ID":"ca7c46cd-e5ff-4732-9718-5c2f08b75221","Type":"ContainerStarted","Data":"ddf3569ec24031b5596601b2c67723641624956b6b12371d43c6acabd068944a"} Nov 25 09:41:31 crc kubenswrapper[4776]: I1125 09:41:31.327530 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-t9s4l" Nov 25 09:41:31 crc kubenswrapper[4776]: I1125 09:41:31.328742 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-kbcml" event={"ID":"dac570d6-3425-4117-be42-1c6d30785a4c","Type":"ContainerStarted","Data":"79b344d19194485f16f5b180fbf4971cc2ee4c84588e9470295538dec178fc7c"} Nov 25 09:41:31 crc kubenswrapper[4776]: I1125 09:41:31.329458 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-kbcml" Nov 25 09:41:31 crc kubenswrapper[4776]: I1125 09:41:31.331023 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-kbcml" Nov 25 09:41:31 crc kubenswrapper[4776]: I1125 09:41:31.331144 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-h6k9v" event={"ID":"7a21bb84-b9c5-4e6f-8fbb-3c2bc65dd409","Type":"ContainerStarted","Data":"a24d2c36689803498a9755654c7b55e123ce889e1bf5fde18e9af7d925b2803d"} Nov 25 09:41:31 crc kubenswrapper[4776]: I1125 09:41:31.354670 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-t9s4l" podStartSLOduration=12.973535544 podStartE2EDuration="19.354648879s" podCreationTimestamp="2025-11-25 09:41:12 +0000 UTC" firstStartedPulling="2025-11-25 09:41:23.940979547 +0000 UTC m=+1028.982039100" lastFinishedPulling="2025-11-25 09:41:30.322092882 +0000 UTC m=+1035.363152435" observedRunningTime="2025-11-25 09:41:31.352886654 +0000 UTC m=+1036.393946227" watchObservedRunningTime="2025-11-25 09:41:31.354648879 +0000 UTC m=+1036.395708432" Nov 25 09:41:31 crc kubenswrapper[4776]: I1125 09:41:31.430467 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-h6k9v" podStartSLOduration=3.221429937 podStartE2EDuration="19.430429958s" podCreationTimestamp="2025-11-25 09:41:12 +0000 UTC" firstStartedPulling="2025-11-25 09:41:14.299889654 +0000 UTC m=+1019.340949207" lastFinishedPulling="2025-11-25 09:41:30.508889665 +0000 UTC m=+1035.549949228" observedRunningTime="2025-11-25 09:41:31.412458348 +0000 UTC m=+1036.453517921" watchObservedRunningTime="2025-11-25 09:41:31.430429958 +0000 UTC m=+1036.471489511" Nov 25 09:41:31 crc kubenswrapper[4776]: I1125 09:41:31.435132 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-kbcml" podStartSLOduration=2.849544465 podStartE2EDuration="19.435104556s" podCreationTimestamp="2025-11-25 09:41:12 +0000 UTC" firstStartedPulling="2025-11-25 09:41:14.545300137 +0000 UTC m=+1019.586359680" lastFinishedPulling="2025-11-25 09:41:31.130860208 +0000 UTC m=+1036.171919771" observedRunningTime="2025-11-25 09:41:31.432813768 +0000 UTC m=+1036.473873321" watchObservedRunningTime="2025-11-25 09:41:31.435104556 +0000 UTC m=+1036.476164109" Nov 25 09:41:32 crc kubenswrapper[4776]: I1125 09:41:32.388303 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mtqhm" event={"ID":"9018dbd6-ab49-4231-a172-33f9a0b56d90","Type":"ContainerStarted","Data":"a572fc8b19e740dc89eca712dbee3bfbcf5edba5d73629618dd3ac9e5a03e06f"} Nov 25 09:41:32 crc kubenswrapper[4776]: I1125 09:41:32.388678 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mtqhm" Nov 25 09:41:32 crc kubenswrapper[4776]: I1125 09:41:32.390915 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mtqhm" Nov 25 09:41:32 crc kubenswrapper[4776]: I1125 09:41:32.393280 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-gmcp9" event={"ID":"61810388-3bac-4a06-88ea-98cc0337fdee","Type":"ContainerStarted","Data":"175ab2dca177b49411122d4a9252478c52703da4f41be3449e3fb54578d773cf"} Nov 25 09:41:32 crc kubenswrapper[4776]: I1125 09:41:32.393523 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-774b86978c-gmcp9" Nov 25 09:41:32 crc kubenswrapper[4776]: I1125 09:41:32.396759 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-khxdw" Nov 25 09:41:32 crc kubenswrapper[4776]: I1125 09:41:32.398542 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-khxdw" Nov 25 09:41:32 crc kubenswrapper[4776]: I1125 09:41:32.399251 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-pckfv" event={"ID":"63e1e612-6ab8-4144-b2cc-c8354bd0894a","Type":"ContainerStarted","Data":"3afcff5a2ffc5c5476ecda916751649d00ad4a7cde124131377885826ac8bc67"} Nov 25 09:41:32 crc kubenswrapper[4776]: I1125 09:41:32.400002 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-h6k9v" Nov 25 09:41:32 crc kubenswrapper[4776]: I1125 09:41:32.404305 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-774b86978c-gmcp9" Nov 25 09:41:32 crc kubenswrapper[4776]: I1125 09:41:32.404355 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-h6k9v" Nov 25 09:41:32 crc kubenswrapper[4776]: I1125 09:41:32.414767 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-mtqhm" podStartSLOduration=2.351219681 podStartE2EDuration="19.414751396s" podCreationTimestamp="2025-11-25 09:41:13 +0000 UTC" firstStartedPulling="2025-11-25 09:41:14.543760298 +0000 UTC m=+1019.584819851" lastFinishedPulling="2025-11-25 09:41:31.607292013 +0000 UTC m=+1036.648351566" observedRunningTime="2025-11-25 09:41:32.413549906 +0000 UTC m=+1037.454609459" watchObservedRunningTime="2025-11-25 09:41:32.414751396 +0000 UTC m=+1037.455810949" Nov 25 09:41:32 crc kubenswrapper[4776]: I1125 09:41:32.440090 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-khxdw" podStartSLOduration=2.29773614 podStartE2EDuration="20.44005293s" podCreationTimestamp="2025-11-25 09:41:12 +0000 UTC" firstStartedPulling="2025-11-25 09:41:14.024926831 +0000 UTC m=+1019.065986384" lastFinishedPulling="2025-11-25 09:41:32.167243621 +0000 UTC m=+1037.208303174" observedRunningTime="2025-11-25 09:41:32.428274325 +0000 UTC m=+1037.469333898" watchObservedRunningTime="2025-11-25 09:41:32.44005293 +0000 UTC m=+1037.481112483" Nov 25 09:41:32 crc kubenswrapper[4776]: I1125 09:41:32.452911 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-774b86978c-gmcp9" podStartSLOduration=2.4123261830000002 podStartE2EDuration="20.452894062s" podCreationTimestamp="2025-11-25 09:41:12 +0000 UTC" firstStartedPulling="2025-11-25 09:41:14.025341071 +0000 UTC m=+1019.066400624" lastFinishedPulling="2025-11-25 09:41:32.06590895 +0000 UTC m=+1037.106968503" observedRunningTime="2025-11-25 09:41:32.443540438 +0000 UTC m=+1037.484600011" watchObservedRunningTime="2025-11-25 09:41:32.452894062 +0000 UTC m=+1037.493953615" Nov 25 09:41:32 crc kubenswrapper[4776]: I1125 09:41:32.500101 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-pckfv" podStartSLOduration=2.636023501 podStartE2EDuration="20.500055944s" podCreationTimestamp="2025-11-25 09:41:12 +0000 UTC" firstStartedPulling="2025-11-25 09:41:14.290307584 +0000 UTC m=+1019.331367137" lastFinishedPulling="2025-11-25 09:41:32.154340037 +0000 UTC m=+1037.195399580" observedRunningTime="2025-11-25 09:41:32.475004386 +0000 UTC m=+1037.516063939" watchObservedRunningTime="2025-11-25 09:41:32.500055944 +0000 UTC m=+1037.541115497" Nov 25 09:41:32 crc kubenswrapper[4776]: I1125 09:41:32.839278 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-rnrpd" Nov 25 09:41:32 crc kubenswrapper[4776]: I1125 09:41:32.853120 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-zwk9j" Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.322879 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-pckfv" Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.325084 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-pckfv" Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.409576 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-cc9wf" event={"ID":"ebac9e21-7ae6-45e4-a6a8-9728b3670733","Type":"ContainerStarted","Data":"f705408ae5b603a085effc973f368f1397e0a9abc47a49ff270e0a9dae868d8b"} Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.410144 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-cc9wf" Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.418595 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-cc9wf" Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.419563 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-ppmfr" event={"ID":"3ebf4825-05f2-4fdd-bc60-d35e80e1b294","Type":"ContainerStarted","Data":"dcabb4764aea43ecc51f88cebf9ace294d6dbd9eddc8601338b4aa8c841d7d1f"} Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.419720 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-ppmfr" Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.421964 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-jr7jl" event={"ID":"7af82806-526b-4fe2-8314-65298d1a0539","Type":"ContainerStarted","Data":"a581b2dba3544278fc1557efb1e35596322ab4544cc094d7e0c5bde2ff44af31"} Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.422053 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-jr7jl" Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.425035 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-jr7jl" Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.428341 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-cc9wf" podStartSLOduration=3.352906574 podStartE2EDuration="21.428318686s" podCreationTimestamp="2025-11-25 09:41:12 +0000 UTC" firstStartedPulling="2025-11-25 09:41:14.304611203 +0000 UTC m=+1019.345670756" lastFinishedPulling="2025-11-25 09:41:32.380023315 +0000 UTC m=+1037.421082868" observedRunningTime="2025-11-25 09:41:33.425764112 +0000 UTC m=+1038.466823665" watchObservedRunningTime="2025-11-25 09:41:33.428318686 +0000 UTC m=+1038.469378239" Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.428759 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-zwk9j" event={"ID":"a590fe89-93f7-4c94-bbdd-82531eedf52f","Type":"ContainerStarted","Data":"e5748296c8ce526b637fdfbf94ac3fcfb720e0e498c742b60e244d0fe90e39f9"} Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.433809 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-clnxq" event={"ID":"bf227b94-0dfc-4055-b688-0f73585af089","Type":"ContainerStarted","Data":"7a33da45eca90a7703790b3c2c1483e37a3f18c7d52808ac4419b616d6e522db"} Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.434025 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-clnxq" Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.440785 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-clnxq" Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.446457 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-rnrpd" event={"ID":"2d511feb-5cd8-4ba3-b5af-5d9d01086a82","Type":"ContainerStarted","Data":"d6927da5983fcac1155357ddc620c04892f8c65b3568c7df8a642cade94f24a9"} Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.449546 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-58r67" event={"ID":"5449efd1-7801-454a-bc48-a5e9e91aba7d","Type":"ContainerStarted","Data":"581852003f21529ffc2b6cdac4d0490eca34ba931287e714182a53cab6d781cf"} Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.449716 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-jr7jl" podStartSLOduration=3.499187751 podStartE2EDuration="21.449697212s" podCreationTimestamp="2025-11-25 09:41:12 +0000 UTC" firstStartedPulling="2025-11-25 09:41:14.524305491 +0000 UTC m=+1019.565365044" lastFinishedPulling="2025-11-25 09:41:32.474814952 +0000 UTC m=+1037.515874505" observedRunningTime="2025-11-25 09:41:33.447488787 +0000 UTC m=+1038.488548340" watchObservedRunningTime="2025-11-25 09:41:33.449697212 +0000 UTC m=+1038.490756765" Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.449847 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-58r67" Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.465157 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-58r67" Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.465824 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-khxdw" event={"ID":"187d1384-3de2-4135-abac-45f15155b942","Type":"ContainerStarted","Data":"d647a896f4e96b8de7ba8788b7fb3b98b694c6a8cea2559ce692126f3d292e2d"} Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.473109 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-ppmfr" podStartSLOduration=2.5592361869999998 podStartE2EDuration="20.473084449s" podCreationTimestamp="2025-11-25 09:41:13 +0000 UTC" firstStartedPulling="2025-11-25 09:41:14.674753602 +0000 UTC m=+1019.715813155" lastFinishedPulling="2025-11-25 09:41:32.588601864 +0000 UTC m=+1037.629661417" observedRunningTime="2025-11-25 09:41:33.467878328 +0000 UTC m=+1038.508937881" watchObservedRunningTime="2025-11-25 09:41:33.473084449 +0000 UTC m=+1038.514144002" Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.481675 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-gwj5b" Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.514417 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-zwk9j" podStartSLOduration=12.002883231 podStartE2EDuration="21.514390844s" podCreationTimestamp="2025-11-25 09:41:12 +0000 UTC" firstStartedPulling="2025-11-25 09:41:13.768958234 +0000 UTC m=+1018.810017787" lastFinishedPulling="2025-11-25 09:41:23.280465847 +0000 UTC m=+1028.321525400" observedRunningTime="2025-11-25 09:41:33.506607049 +0000 UTC m=+1038.547666602" watchObservedRunningTime="2025-11-25 09:41:33.514390844 +0000 UTC m=+1038.555450397" Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.541955 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-clnxq" podStartSLOduration=4.522380433 podStartE2EDuration="21.541931485s" podCreationTimestamp="2025-11-25 09:41:12 +0000 UTC" firstStartedPulling="2025-11-25 09:41:15.224410852 +0000 UTC m=+1020.265470405" lastFinishedPulling="2025-11-25 09:41:32.243961904 +0000 UTC m=+1037.285021457" observedRunningTime="2025-11-25 09:41:33.531610376 +0000 UTC m=+1038.572669939" watchObservedRunningTime="2025-11-25 09:41:33.541931485 +0000 UTC m=+1038.582991038" Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.591670 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-58r67" podStartSLOduration=2.978370044 podStartE2EDuration="21.591650681s" podCreationTimestamp="2025-11-25 09:41:12 +0000 UTC" firstStartedPulling="2025-11-25 09:41:13.662600937 +0000 UTC m=+1018.703660490" lastFinishedPulling="2025-11-25 09:41:32.275881574 +0000 UTC m=+1037.316941127" observedRunningTime="2025-11-25 09:41:33.558001748 +0000 UTC m=+1038.599061301" watchObservedRunningTime="2025-11-25 09:41:33.591650681 +0000 UTC m=+1038.632710234" Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.603428 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-sddh8" Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.628561 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-rnrpd" podStartSLOduration=11.979506074 podStartE2EDuration="21.627267374s" podCreationTimestamp="2025-11-25 09:41:12 +0000 UTC" firstStartedPulling="2025-11-25 09:41:13.64715572 +0000 UTC m=+1018.688215273" lastFinishedPulling="2025-11-25 09:41:23.29491702 +0000 UTC m=+1028.335976573" observedRunningTime="2025-11-25 09:41:33.610167955 +0000 UTC m=+1038.651227508" watchObservedRunningTime="2025-11-25 09:41:33.627267374 +0000 UTC m=+1038.668326927" Nov 25 09:41:33 crc kubenswrapper[4776]: I1125 09:41:33.636923 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hv6qx" Nov 25 09:41:34 crc kubenswrapper[4776]: I1125 09:41:34.476323 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-ppmfr" Nov 25 09:41:35 crc kubenswrapper[4776]: I1125 09:41:35.480045 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-gwj5b" event={"ID":"88b8f0f1-da6d-43fe-940e-23c48e9248d2","Type":"ContainerStarted","Data":"7ab25385653a195ad5322bcf2aa433bbfe8f4bc8b1253125e7a6a4e6ae154355"} Nov 25 09:41:35 crc kubenswrapper[4776]: I1125 09:41:35.502515 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-gwj5b" podStartSLOduration=14.632071985 podStartE2EDuration="23.502493036s" podCreationTimestamp="2025-11-25 09:41:12 +0000 UTC" firstStartedPulling="2025-11-25 09:41:14.422272763 +0000 UTC m=+1019.463332316" lastFinishedPulling="2025-11-25 09:41:23.292693814 +0000 UTC m=+1028.333753367" observedRunningTime="2025-11-25 09:41:35.494583987 +0000 UTC m=+1040.535643540" watchObservedRunningTime="2025-11-25 09:41:35.502493036 +0000 UTC m=+1040.543552589" Nov 25 09:41:37 crc kubenswrapper[4776]: I1125 09:41:37.276782 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-t9s4l" Nov 25 09:41:37 crc kubenswrapper[4776]: I1125 09:41:37.673905 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-mnd72" Nov 25 09:41:44 crc kubenswrapper[4776]: I1125 09:41:44.540388 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hv6qx" event={"ID":"400f3b2c-a118-4fab-8c8e-2f4d29f5c354","Type":"ContainerStarted","Data":"715077b75ad6ccdf83ef7d19fc6870bcf7920d558dd7a07a8c697b3563c9d1bd"} Nov 25 09:41:44 crc kubenswrapper[4776]: I1125 09:41:44.542425 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-sddh8" event={"ID":"5c262b40-b415-4cca-b1e8-635a9c153d81","Type":"ContainerStarted","Data":"804a39281171fb4c8c3888ad215f64b1de50955b8654bebc592ce09fb2be1c2d"} Nov 25 09:41:44 crc kubenswrapper[4776]: I1125 09:41:44.545049 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-npwhb" event={"ID":"8f2c9384-1b11-4925-8338-babf955143b1","Type":"ContainerStarted","Data":"400506dbf383a1599199723898a31b831947957eee0f76bcefe9e0e234c82b3d"} Nov 25 09:41:44 crc kubenswrapper[4776]: I1125 09:41:44.551212 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-864885998-sxmf5" event={"ID":"27834511-7fe9-454d-b318-efa218524c0d","Type":"ContainerStarted","Data":"e4336521546477ce0403f55bfbb5cf6409f6757565119991e6a4f6c3103600ee"} Nov 25 09:41:44 crc kubenswrapper[4776]: I1125 09:41:44.551256 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-864885998-sxmf5" event={"ID":"27834511-7fe9-454d-b318-efa218524c0d","Type":"ContainerStarted","Data":"cad4b9e45173174aa8be4882069fa1f386638e928a8047ee5912b3e22e561c93"} Nov 25 09:41:44 crc kubenswrapper[4776]: I1125 09:41:44.551871 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-864885998-sxmf5" Nov 25 09:41:44 crc kubenswrapper[4776]: I1125 09:41:44.559103 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-h8s6x" event={"ID":"68f49316-eb77-464a-aebb-189800786fa5","Type":"ContainerStarted","Data":"f33667b460ba6cbf968437ee0cba40547438712e2bbe3d41575317e2753b39ed"} Nov 25 09:41:44 crc kubenswrapper[4776]: I1125 09:41:44.559155 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-h8s6x" event={"ID":"68f49316-eb77-464a-aebb-189800786fa5","Type":"ContainerStarted","Data":"eee85faad9653b7017a69515f272293f3af973365a2512b5a70edb5e607b87e1"} Nov 25 09:41:44 crc kubenswrapper[4776]: I1125 09:41:44.559884 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-h8s6x" Nov 25 09:41:44 crc kubenswrapper[4776]: I1125 09:41:44.563964 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cb74df96-v4xgj" event={"ID":"c2be0972-c768-499c-95e3-4cb180ad510c","Type":"ContainerStarted","Data":"c28714d9125f58db16f243e3c2d4512764692c61241f71b1f23e375a605c9b05"} Nov 25 09:41:44 crc kubenswrapper[4776]: I1125 09:41:44.564002 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cb74df96-v4xgj" event={"ID":"c2be0972-c768-499c-95e3-4cb180ad510c","Type":"ContainerStarted","Data":"4f3bd5ad527391e65c393bb6566e6cbf978c5be6e5d109863afbfeef8439e039"} Nov 25 09:41:44 crc kubenswrapper[4776]: I1125 09:41:44.564209 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cb74df96-v4xgj" Nov 25 09:41:44 crc kubenswrapper[4776]: I1125 09:41:44.568148 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hv6qx" podStartSLOduration=23.8233219 podStartE2EDuration="32.568129183s" podCreationTimestamp="2025-11-25 09:41:12 +0000 UTC" firstStartedPulling="2025-11-25 09:41:14.551863501 +0000 UTC m=+1019.592923054" lastFinishedPulling="2025-11-25 09:41:23.296670784 +0000 UTC m=+1028.337730337" observedRunningTime="2025-11-25 09:41:44.557453125 +0000 UTC m=+1049.598512688" watchObservedRunningTime="2025-11-25 09:41:44.568129183 +0000 UTC m=+1049.609188736" Nov 25 09:41:44 crc kubenswrapper[4776]: I1125 09:41:44.620850 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-864885998-sxmf5" podStartSLOduration=2.674642869 podStartE2EDuration="31.620829554s" podCreationTimestamp="2025-11-25 09:41:13 +0000 UTC" firstStartedPulling="2025-11-25 09:41:14.725020793 +0000 UTC m=+1019.766080346" lastFinishedPulling="2025-11-25 09:41:43.671207478 +0000 UTC m=+1048.712267031" observedRunningTime="2025-11-25 09:41:44.617103611 +0000 UTC m=+1049.658163164" watchObservedRunningTime="2025-11-25 09:41:44.620829554 +0000 UTC m=+1049.661889097" Nov 25 09:41:44 crc kubenswrapper[4776]: I1125 09:41:44.624243 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-npwhb" podStartSLOduration=2.644747869 podStartE2EDuration="31.624228589s" podCreationTimestamp="2025-11-25 09:41:13 +0000 UTC" firstStartedPulling="2025-11-25 09:41:14.704235881 +0000 UTC m=+1019.745295434" lastFinishedPulling="2025-11-25 09:41:43.683716601 +0000 UTC m=+1048.724776154" observedRunningTime="2025-11-25 09:41:44.597819207 +0000 UTC m=+1049.638878760" watchObservedRunningTime="2025-11-25 09:41:44.624228589 +0000 UTC m=+1049.665288142" Nov 25 09:41:44 crc kubenswrapper[4776]: I1125 09:41:44.644254 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-sddh8" podStartSLOduration=24.001359364 podStartE2EDuration="32.644232801s" podCreationTimestamp="2025-11-25 09:41:12 +0000 UTC" firstStartedPulling="2025-11-25 09:41:14.650534795 +0000 UTC m=+1019.691594338" lastFinishedPulling="2025-11-25 09:41:23.293408222 +0000 UTC m=+1028.334467775" observedRunningTime="2025-11-25 09:41:44.636657671 +0000 UTC m=+1049.677717234" watchObservedRunningTime="2025-11-25 09:41:44.644232801 +0000 UTC m=+1049.685292364" Nov 25 09:41:44 crc kubenswrapper[4776]: I1125 09:41:44.695833 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cb74df96-v4xgj" podStartSLOduration=2.74887383 podStartE2EDuration="31.695813184s" podCreationTimestamp="2025-11-25 09:41:13 +0000 UTC" firstStartedPulling="2025-11-25 09:41:14.722888679 +0000 UTC m=+1019.763948232" lastFinishedPulling="2025-11-25 09:41:43.669828023 +0000 UTC m=+1048.710887586" observedRunningTime="2025-11-25 09:41:44.668116699 +0000 UTC m=+1049.709176252" watchObservedRunningTime="2025-11-25 09:41:44.695813184 +0000 UTC m=+1049.736872737" Nov 25 09:41:44 crc kubenswrapper[4776]: I1125 09:41:44.697861 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-h8s6x" podStartSLOduration=3.577895624 podStartE2EDuration="32.697849365s" podCreationTimestamp="2025-11-25 09:41:12 +0000 UTC" firstStartedPulling="2025-11-25 09:41:14.552521578 +0000 UTC m=+1019.593581131" lastFinishedPulling="2025-11-25 09:41:43.672475319 +0000 UTC m=+1048.713534872" observedRunningTime="2025-11-25 09:41:44.69446005 +0000 UTC m=+1049.735519613" watchObservedRunningTime="2025-11-25 09:41:44.697849365 +0000 UTC m=+1049.738908928" Nov 25 09:41:47 crc kubenswrapper[4776]: I1125 09:41:47.818776 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:41:47 crc kubenswrapper[4776]: I1125 09:41:47.819174 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:41:47 crc kubenswrapper[4776]: I1125 09:41:47.819232 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 09:41:47 crc kubenswrapper[4776]: I1125 09:41:47.820030 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6132d3cd67d80b93c28c890778e005f6af8e158a1aed030b0062e3edb3f984f8"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 09:41:47 crc kubenswrapper[4776]: I1125 09:41:47.820116 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://6132d3cd67d80b93c28c890778e005f6af8e158a1aed030b0062e3edb3f984f8" gracePeriod=600 Nov 25 09:41:48 crc kubenswrapper[4776]: I1125 09:41:48.597871 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="6132d3cd67d80b93c28c890778e005f6af8e158a1aed030b0062e3edb3f984f8" exitCode=0 Nov 25 09:41:48 crc kubenswrapper[4776]: I1125 09:41:48.597938 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"6132d3cd67d80b93c28c890778e005f6af8e158a1aed030b0062e3edb3f984f8"} Nov 25 09:41:48 crc kubenswrapper[4776]: I1125 09:41:48.598199 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"bbf70cb3f31b21dd9b2d356f05e12d08a7b4f65263789554ac47b2758bc077be"} Nov 25 09:41:48 crc kubenswrapper[4776]: I1125 09:41:48.598220 4776 scope.go:117] "RemoveContainer" containerID="dc4f980209957fcb2e8415b8fd0ea1aba1b60952959f82bbdfd9dd9f2991ea9c" Nov 25 09:41:53 crc kubenswrapper[4776]: I1125 09:41:53.542314 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-h8s6x" Nov 25 09:41:54 crc kubenswrapper[4776]: I1125 09:41:54.012268 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cb74df96-v4xgj" Nov 25 09:41:54 crc kubenswrapper[4776]: I1125 09:41:54.038711 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-864885998-sxmf5" Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.107039 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-sclc2"] Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.108715 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdd77c89-sclc2" Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.112160 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.112561 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.112870 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.113271 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-rfp57" Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.118657 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-sclc2"] Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.199059 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24-config\") pod \"dnsmasq-dns-7bdd77c89-sclc2\" (UID: \"bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24\") " pod="openstack/dnsmasq-dns-7bdd77c89-sclc2" Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.199230 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lq5d\" (UniqueName: \"kubernetes.io/projected/bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24-kube-api-access-2lq5d\") pod \"dnsmasq-dns-7bdd77c89-sclc2\" (UID: \"bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24\") " pod="openstack/dnsmasq-dns-7bdd77c89-sclc2" Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.205568 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6584b49599-v697v"] Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.206990 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6584b49599-v697v" Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.208934 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.227252 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-v697v"] Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.300267 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lq5d\" (UniqueName: \"kubernetes.io/projected/bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24-kube-api-access-2lq5d\") pod \"dnsmasq-dns-7bdd77c89-sclc2\" (UID: \"bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24\") " pod="openstack/dnsmasq-dns-7bdd77c89-sclc2" Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.300630 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjbgr\" (UniqueName: \"kubernetes.io/projected/70048934-5af8-43a5-a7e4-b78a6bc70fea-kube-api-access-vjbgr\") pod \"dnsmasq-dns-6584b49599-v697v\" (UID: \"70048934-5af8-43a5-a7e4-b78a6bc70fea\") " pod="openstack/dnsmasq-dns-6584b49599-v697v" Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.300663 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70048934-5af8-43a5-a7e4-b78a6bc70fea-dns-svc\") pod \"dnsmasq-dns-6584b49599-v697v\" (UID: \"70048934-5af8-43a5-a7e4-b78a6bc70fea\") " pod="openstack/dnsmasq-dns-6584b49599-v697v" Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.300683 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24-config\") pod \"dnsmasq-dns-7bdd77c89-sclc2\" (UID: \"bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24\") " pod="openstack/dnsmasq-dns-7bdd77c89-sclc2" Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.300723 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70048934-5af8-43a5-a7e4-b78a6bc70fea-config\") pod \"dnsmasq-dns-6584b49599-v697v\" (UID: \"70048934-5af8-43a5-a7e4-b78a6bc70fea\") " pod="openstack/dnsmasq-dns-6584b49599-v697v" Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.301533 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24-config\") pod \"dnsmasq-dns-7bdd77c89-sclc2\" (UID: \"bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24\") " pod="openstack/dnsmasq-dns-7bdd77c89-sclc2" Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.319331 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lq5d\" (UniqueName: \"kubernetes.io/projected/bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24-kube-api-access-2lq5d\") pod \"dnsmasq-dns-7bdd77c89-sclc2\" (UID: \"bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24\") " pod="openstack/dnsmasq-dns-7bdd77c89-sclc2" Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.401747 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjbgr\" (UniqueName: \"kubernetes.io/projected/70048934-5af8-43a5-a7e4-b78a6bc70fea-kube-api-access-vjbgr\") pod \"dnsmasq-dns-6584b49599-v697v\" (UID: \"70048934-5af8-43a5-a7e4-b78a6bc70fea\") " pod="openstack/dnsmasq-dns-6584b49599-v697v" Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.401803 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70048934-5af8-43a5-a7e4-b78a6bc70fea-dns-svc\") pod \"dnsmasq-dns-6584b49599-v697v\" (UID: \"70048934-5af8-43a5-a7e4-b78a6bc70fea\") " pod="openstack/dnsmasq-dns-6584b49599-v697v" Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.401833 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70048934-5af8-43a5-a7e4-b78a6bc70fea-config\") pod \"dnsmasq-dns-6584b49599-v697v\" (UID: \"70048934-5af8-43a5-a7e4-b78a6bc70fea\") " pod="openstack/dnsmasq-dns-6584b49599-v697v" Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.402866 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70048934-5af8-43a5-a7e4-b78a6bc70fea-config\") pod \"dnsmasq-dns-6584b49599-v697v\" (UID: \"70048934-5af8-43a5-a7e4-b78a6bc70fea\") " pod="openstack/dnsmasq-dns-6584b49599-v697v" Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.402918 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70048934-5af8-43a5-a7e4-b78a6bc70fea-dns-svc\") pod \"dnsmasq-dns-6584b49599-v697v\" (UID: \"70048934-5af8-43a5-a7e4-b78a6bc70fea\") " pod="openstack/dnsmasq-dns-6584b49599-v697v" Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.427774 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdd77c89-sclc2" Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.432788 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjbgr\" (UniqueName: \"kubernetes.io/projected/70048934-5af8-43a5-a7e4-b78a6bc70fea-kube-api-access-vjbgr\") pod \"dnsmasq-dns-6584b49599-v697v\" (UID: \"70048934-5af8-43a5-a7e4-b78a6bc70fea\") " pod="openstack/dnsmasq-dns-6584b49599-v697v" Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.525160 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6584b49599-v697v" Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.842752 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-sclc2"] Nov 25 09:42:08 crc kubenswrapper[4776]: W1125 09:42:08.849938 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb59d0a0_c4a0_41e7_b7a9_aaf9f13b0d24.slice/crio-123ff647f45d05b0940b78b948deaf2d04e5ba4ee28e3fb19394f01658b09194 WatchSource:0}: Error finding container 123ff647f45d05b0940b78b948deaf2d04e5ba4ee28e3fb19394f01658b09194: Status 404 returned error can't find the container with id 123ff647f45d05b0940b78b948deaf2d04e5ba4ee28e3fb19394f01658b09194 Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.852135 4776 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 09:42:08 crc kubenswrapper[4776]: I1125 09:42:08.924704 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-v697v"] Nov 25 09:42:08 crc kubenswrapper[4776]: W1125 09:42:08.929609 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70048934_5af8_43a5_a7e4_b78a6bc70fea.slice/crio-a5e6b816dfdd0cfa8b46cd98ab3055b84aa519299aae28d53c313ff23ad83d50 WatchSource:0}: Error finding container a5e6b816dfdd0cfa8b46cd98ab3055b84aa519299aae28d53c313ff23ad83d50: Status 404 returned error can't find the container with id a5e6b816dfdd0cfa8b46cd98ab3055b84aa519299aae28d53c313ff23ad83d50 Nov 25 09:42:09 crc kubenswrapper[4776]: I1125 09:42:09.176816 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-v697v"] Nov 25 09:42:09 crc kubenswrapper[4776]: I1125 09:42:09.205915 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-swg9r"] Nov 25 09:42:09 crc kubenswrapper[4776]: I1125 09:42:09.207015 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6d9948dc-swg9r" Nov 25 09:42:09 crc kubenswrapper[4776]: I1125 09:42:09.213892 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b732643a-7578-4b49-802e-7f04b5d3793c-config\") pod \"dnsmasq-dns-7c6d9948dc-swg9r\" (UID: \"b732643a-7578-4b49-802e-7f04b5d3793c\") " pod="openstack/dnsmasq-dns-7c6d9948dc-swg9r" Nov 25 09:42:09 crc kubenswrapper[4776]: I1125 09:42:09.215575 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-swg9r"] Nov 25 09:42:09 crc kubenswrapper[4776]: I1125 09:42:09.316029 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b732643a-7578-4b49-802e-7f04b5d3793c-dns-svc\") pod \"dnsmasq-dns-7c6d9948dc-swg9r\" (UID: \"b732643a-7578-4b49-802e-7f04b5d3793c\") " pod="openstack/dnsmasq-dns-7c6d9948dc-swg9r" Nov 25 09:42:09 crc kubenswrapper[4776]: I1125 09:42:09.316088 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5stb\" (UniqueName: \"kubernetes.io/projected/b732643a-7578-4b49-802e-7f04b5d3793c-kube-api-access-x5stb\") pod \"dnsmasq-dns-7c6d9948dc-swg9r\" (UID: \"b732643a-7578-4b49-802e-7f04b5d3793c\") " pod="openstack/dnsmasq-dns-7c6d9948dc-swg9r" Nov 25 09:42:09 crc kubenswrapper[4776]: I1125 09:42:09.316302 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b732643a-7578-4b49-802e-7f04b5d3793c-config\") pod \"dnsmasq-dns-7c6d9948dc-swg9r\" (UID: \"b732643a-7578-4b49-802e-7f04b5d3793c\") " pod="openstack/dnsmasq-dns-7c6d9948dc-swg9r" Nov 25 09:42:09 crc kubenswrapper[4776]: I1125 09:42:09.317148 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b732643a-7578-4b49-802e-7f04b5d3793c-config\") pod \"dnsmasq-dns-7c6d9948dc-swg9r\" (UID: \"b732643a-7578-4b49-802e-7f04b5d3793c\") " pod="openstack/dnsmasq-dns-7c6d9948dc-swg9r" Nov 25 09:42:09 crc kubenswrapper[4776]: I1125 09:42:09.417494 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b732643a-7578-4b49-802e-7f04b5d3793c-dns-svc\") pod \"dnsmasq-dns-7c6d9948dc-swg9r\" (UID: \"b732643a-7578-4b49-802e-7f04b5d3793c\") " pod="openstack/dnsmasq-dns-7c6d9948dc-swg9r" Nov 25 09:42:09 crc kubenswrapper[4776]: I1125 09:42:09.417942 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5stb\" (UniqueName: \"kubernetes.io/projected/b732643a-7578-4b49-802e-7f04b5d3793c-kube-api-access-x5stb\") pod \"dnsmasq-dns-7c6d9948dc-swg9r\" (UID: \"b732643a-7578-4b49-802e-7f04b5d3793c\") " pod="openstack/dnsmasq-dns-7c6d9948dc-swg9r" Nov 25 09:42:09 crc kubenswrapper[4776]: I1125 09:42:09.419894 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b732643a-7578-4b49-802e-7f04b5d3793c-dns-svc\") pod \"dnsmasq-dns-7c6d9948dc-swg9r\" (UID: \"b732643a-7578-4b49-802e-7f04b5d3793c\") " pod="openstack/dnsmasq-dns-7c6d9948dc-swg9r" Nov 25 09:42:09 crc kubenswrapper[4776]: I1125 09:42:09.437021 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5stb\" (UniqueName: \"kubernetes.io/projected/b732643a-7578-4b49-802e-7f04b5d3793c-kube-api-access-x5stb\") pod \"dnsmasq-dns-7c6d9948dc-swg9r\" (UID: \"b732643a-7578-4b49-802e-7f04b5d3793c\") " pod="openstack/dnsmasq-dns-7c6d9948dc-swg9r" Nov 25 09:42:09 crc kubenswrapper[4776]: I1125 09:42:09.521667 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6d9948dc-swg9r" Nov 25 09:42:09 crc kubenswrapper[4776]: I1125 09:42:09.770570 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6584b49599-v697v" event={"ID":"70048934-5af8-43a5-a7e4-b78a6bc70fea","Type":"ContainerStarted","Data":"a5e6b816dfdd0cfa8b46cd98ab3055b84aa519299aae28d53c313ff23ad83d50"} Nov 25 09:42:09 crc kubenswrapper[4776]: I1125 09:42:09.774178 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdd77c89-sclc2" event={"ID":"bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24","Type":"ContainerStarted","Data":"123ff647f45d05b0940b78b948deaf2d04e5ba4ee28e3fb19394f01658b09194"} Nov 25 09:42:09 crc kubenswrapper[4776]: I1125 09:42:09.952401 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-swg9r"] Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.172100 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-sclc2"] Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.202205 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-4l8fb"] Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.203782 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6486446b9f-4l8fb" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.207868 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-4l8fb"] Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.231002 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhtx4\" (UniqueName: \"kubernetes.io/projected/f40cef2d-0d59-4300-9336-fc5a793b9cbf-kube-api-access-nhtx4\") pod \"dnsmasq-dns-6486446b9f-4l8fb\" (UID: \"f40cef2d-0d59-4300-9336-fc5a793b9cbf\") " pod="openstack/dnsmasq-dns-6486446b9f-4l8fb" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.231101 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f40cef2d-0d59-4300-9336-fc5a793b9cbf-dns-svc\") pod \"dnsmasq-dns-6486446b9f-4l8fb\" (UID: \"f40cef2d-0d59-4300-9336-fc5a793b9cbf\") " pod="openstack/dnsmasq-dns-6486446b9f-4l8fb" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.231204 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f40cef2d-0d59-4300-9336-fc5a793b9cbf-config\") pod \"dnsmasq-dns-6486446b9f-4l8fb\" (UID: \"f40cef2d-0d59-4300-9336-fc5a793b9cbf\") " pod="openstack/dnsmasq-dns-6486446b9f-4l8fb" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.325007 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.328775 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.332385 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhtx4\" (UniqueName: \"kubernetes.io/projected/f40cef2d-0d59-4300-9336-fc5a793b9cbf-kube-api-access-nhtx4\") pod \"dnsmasq-dns-6486446b9f-4l8fb\" (UID: \"f40cef2d-0d59-4300-9336-fc5a793b9cbf\") " pod="openstack/dnsmasq-dns-6486446b9f-4l8fb" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.332453 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f40cef2d-0d59-4300-9336-fc5a793b9cbf-dns-svc\") pod \"dnsmasq-dns-6486446b9f-4l8fb\" (UID: \"f40cef2d-0d59-4300-9336-fc5a793b9cbf\") " pod="openstack/dnsmasq-dns-6486446b9f-4l8fb" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.332516 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f40cef2d-0d59-4300-9336-fc5a793b9cbf-config\") pod \"dnsmasq-dns-6486446b9f-4l8fb\" (UID: \"f40cef2d-0d59-4300-9336-fc5a793b9cbf\") " pod="openstack/dnsmasq-dns-6486446b9f-4l8fb" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.333698 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f40cef2d-0d59-4300-9336-fc5a793b9cbf-dns-svc\") pod \"dnsmasq-dns-6486446b9f-4l8fb\" (UID: \"f40cef2d-0d59-4300-9336-fc5a793b9cbf\") " pod="openstack/dnsmasq-dns-6486446b9f-4l8fb" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.333905 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f40cef2d-0d59-4300-9336-fc5a793b9cbf-config\") pod \"dnsmasq-dns-6486446b9f-4l8fb\" (UID: \"f40cef2d-0d59-4300-9336-fc5a793b9cbf\") " pod="openstack/dnsmasq-dns-6486446b9f-4l8fb" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.335193 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-6rj55" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.335253 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.335543 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.335315 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.335389 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.335415 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.336955 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.349042 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.375347 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhtx4\" (UniqueName: \"kubernetes.io/projected/f40cef2d-0d59-4300-9336-fc5a793b9cbf-kube-api-access-nhtx4\") pod \"dnsmasq-dns-6486446b9f-4l8fb\" (UID: \"f40cef2d-0d59-4300-9336-fc5a793b9cbf\") " pod="openstack/dnsmasq-dns-6486446b9f-4l8fb" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.433514 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4512cf4e-6c75-493e-a80a-4b0accd750a8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.433571 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4512cf4e-6c75-493e-a80a-4b0accd750a8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.433603 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4512cf4e-6c75-493e-a80a-4b0accd750a8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.433627 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4512cf4e-6c75-493e-a80a-4b0accd750a8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.433684 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4512cf4e-6c75-493e-a80a-4b0accd750a8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.433717 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g6h4\" (UniqueName: \"kubernetes.io/projected/4512cf4e-6c75-493e-a80a-4b0accd750a8-kube-api-access-5g6h4\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.433743 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4512cf4e-6c75-493e-a80a-4b0accd750a8-config-data\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.433762 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4512cf4e-6c75-493e-a80a-4b0accd750a8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.433800 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4512cf4e-6c75-493e-a80a-4b0accd750a8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.433821 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.433849 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4512cf4e-6c75-493e-a80a-4b0accd750a8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.521721 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6486446b9f-4l8fb" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.535252 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4512cf4e-6c75-493e-a80a-4b0accd750a8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.535312 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.535353 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4512cf4e-6c75-493e-a80a-4b0accd750a8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.535411 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4512cf4e-6c75-493e-a80a-4b0accd750a8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.535437 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4512cf4e-6c75-493e-a80a-4b0accd750a8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.535463 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4512cf4e-6c75-493e-a80a-4b0accd750a8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.535513 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4512cf4e-6c75-493e-a80a-4b0accd750a8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.535571 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4512cf4e-6c75-493e-a80a-4b0accd750a8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.535607 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5g6h4\" (UniqueName: \"kubernetes.io/projected/4512cf4e-6c75-493e-a80a-4b0accd750a8-kube-api-access-5g6h4\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.535634 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4512cf4e-6c75-493e-a80a-4b0accd750a8-config-data\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.535653 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4512cf4e-6c75-493e-a80a-4b0accd750a8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.536803 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.537178 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4512cf4e-6c75-493e-a80a-4b0accd750a8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.537233 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4512cf4e-6c75-493e-a80a-4b0accd750a8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.537327 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4512cf4e-6c75-493e-a80a-4b0accd750a8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.537357 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4512cf4e-6c75-493e-a80a-4b0accd750a8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.539760 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4512cf4e-6c75-493e-a80a-4b0accd750a8-config-data\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.540704 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4512cf4e-6c75-493e-a80a-4b0accd750a8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.540754 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4512cf4e-6c75-493e-a80a-4b0accd750a8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.541365 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4512cf4e-6c75-493e-a80a-4b0accd750a8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.541562 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4512cf4e-6c75-493e-a80a-4b0accd750a8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.554044 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5g6h4\" (UniqueName: \"kubernetes.io/projected/4512cf4e-6c75-493e-a80a-4b0accd750a8-kube-api-access-5g6h4\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.570002 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.673491 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 09:42:10 crc kubenswrapper[4776]: I1125 09:42:10.786309 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6d9948dc-swg9r" event={"ID":"b732643a-7578-4b49-802e-7f04b5d3793c","Type":"ContainerStarted","Data":"e17a9c1f772740c03e90d72e07c3081e4557c7a015e1bc8350618c3c11d4fc9b"} Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.018911 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-4l8fb"] Nov 25 09:42:11 crc kubenswrapper[4776]: W1125 09:42:11.042325 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf40cef2d_0d59_4300_9336_fc5a793b9cbf.slice/crio-06f3a6b4b1b8591c301a21c39834f9588b61217f9a63fb21eef31f106ee411d4 WatchSource:0}: Error finding container 06f3a6b4b1b8591c301a21c39834f9588b61217f9a63fb21eef31f106ee411d4: Status 404 returned error can't find the container with id 06f3a6b4b1b8591c301a21c39834f9588b61217f9a63fb21eef31f106ee411d4 Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.300444 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 09:42:11 crc kubenswrapper[4776]: W1125 09:42:11.305998 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4512cf4e_6c75_493e_a80a_4b0accd750a8.slice/crio-43d7aec6fc9f74385bc4e597e383ea2a628e7dda0fd162984cfba0351a6e82b6 WatchSource:0}: Error finding container 43d7aec6fc9f74385bc4e597e383ea2a628e7dda0fd162984cfba0351a6e82b6: Status 404 returned error can't find the container with id 43d7aec6fc9f74385bc4e597e383ea2a628e7dda0fd162984cfba0351a6e82b6 Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.351124 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.352731 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.357733 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.359113 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.359424 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.359551 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.360079 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.360874 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.361019 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-25ssz" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.363331 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.449290 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f32041a0-a8fc-48a5-afab-476baff89e8c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.449625 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.449653 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f32041a0-a8fc-48a5-afab-476baff89e8c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.449697 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfwhg\" (UniqueName: \"kubernetes.io/projected/f32041a0-a8fc-48a5-afab-476baff89e8c-kube-api-access-qfwhg\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.449730 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f32041a0-a8fc-48a5-afab-476baff89e8c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.449761 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f32041a0-a8fc-48a5-afab-476baff89e8c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.449779 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f32041a0-a8fc-48a5-afab-476baff89e8c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.449910 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f32041a0-a8fc-48a5-afab-476baff89e8c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.449932 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f32041a0-a8fc-48a5-afab-476baff89e8c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.449947 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f32041a0-a8fc-48a5-afab-476baff89e8c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.449968 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f32041a0-a8fc-48a5-afab-476baff89e8c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.558586 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfwhg\" (UniqueName: \"kubernetes.io/projected/f32041a0-a8fc-48a5-afab-476baff89e8c-kube-api-access-qfwhg\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.558706 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f32041a0-a8fc-48a5-afab-476baff89e8c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.559703 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f32041a0-a8fc-48a5-afab-476baff89e8c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.559735 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f32041a0-a8fc-48a5-afab-476baff89e8c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.559779 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f32041a0-a8fc-48a5-afab-476baff89e8c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.559805 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f32041a0-a8fc-48a5-afab-476baff89e8c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.559824 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f32041a0-a8fc-48a5-afab-476baff89e8c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.559847 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f32041a0-a8fc-48a5-afab-476baff89e8c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.559893 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f32041a0-a8fc-48a5-afab-476baff89e8c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.559919 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.559948 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f32041a0-a8fc-48a5-afab-476baff89e8c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.560597 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.560934 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f32041a0-a8fc-48a5-afab-476baff89e8c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.561681 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f32041a0-a8fc-48a5-afab-476baff89e8c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.568629 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f32041a0-a8fc-48a5-afab-476baff89e8c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.568790 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f32041a0-a8fc-48a5-afab-476baff89e8c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.587931 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f32041a0-a8fc-48a5-afab-476baff89e8c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.588853 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f32041a0-a8fc-48a5-afab-476baff89e8c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.589318 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f32041a0-a8fc-48a5-afab-476baff89e8c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.589755 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfwhg\" (UniqueName: \"kubernetes.io/projected/f32041a0-a8fc-48a5-afab-476baff89e8c-kube-api-access-qfwhg\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.590638 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f32041a0-a8fc-48a5-afab-476baff89e8c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.591345 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f32041a0-a8fc-48a5-afab-476baff89e8c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.595246 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.711486 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.794333 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4512cf4e-6c75-493e-a80a-4b0accd750a8","Type":"ContainerStarted","Data":"43d7aec6fc9f74385bc4e597e383ea2a628e7dda0fd162984cfba0351a6e82b6"} Nov 25 09:42:11 crc kubenswrapper[4776]: I1125 09:42:11.796724 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6486446b9f-4l8fb" event={"ID":"f40cef2d-0d59-4300-9336-fc5a793b9cbf","Type":"ContainerStarted","Data":"06f3a6b4b1b8591c301a21c39834f9588b61217f9a63fb21eef31f106ee411d4"} Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.335451 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.789678 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.793043 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.796696 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-2t654" Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.796866 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.796877 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.797497 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.798323 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.802231 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.812816 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f32041a0-a8fc-48a5-afab-476baff89e8c","Type":"ContainerStarted","Data":"bd22466d1d55a4b40c3761c1eed08d859f1f3d4ccfc546edb124b702e0b8ba48"} Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.880635 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/167c5bc5-b21e-4c51-8765-1dcbf290294b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " pod="openstack/openstack-galera-0" Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.881140 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/167c5bc5-b21e-4c51-8765-1dcbf290294b-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " pod="openstack/openstack-galera-0" Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.881198 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/167c5bc5-b21e-4c51-8765-1dcbf290294b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " pod="openstack/openstack-galera-0" Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.881301 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/167c5bc5-b21e-4c51-8765-1dcbf290294b-config-data-default\") pod \"openstack-galera-0\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " pod="openstack/openstack-galera-0" Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.881382 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/167c5bc5-b21e-4c51-8765-1dcbf290294b-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " pod="openstack/openstack-galera-0" Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.882846 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/167c5bc5-b21e-4c51-8765-1dcbf290294b-kolla-config\") pod \"openstack-galera-0\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " pod="openstack/openstack-galera-0" Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.882939 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " pod="openstack/openstack-galera-0" Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.883033 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqc28\" (UniqueName: \"kubernetes.io/projected/167c5bc5-b21e-4c51-8765-1dcbf290294b-kube-api-access-zqc28\") pod \"openstack-galera-0\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " pod="openstack/openstack-galera-0" Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.986091 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/167c5bc5-b21e-4c51-8765-1dcbf290294b-kolla-config\") pod \"openstack-galera-0\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " pod="openstack/openstack-galera-0" Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.986147 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " pod="openstack/openstack-galera-0" Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.986202 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqc28\" (UniqueName: \"kubernetes.io/projected/167c5bc5-b21e-4c51-8765-1dcbf290294b-kube-api-access-zqc28\") pod \"openstack-galera-0\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " pod="openstack/openstack-galera-0" Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.986271 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/167c5bc5-b21e-4c51-8765-1dcbf290294b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " pod="openstack/openstack-galera-0" Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.986315 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/167c5bc5-b21e-4c51-8765-1dcbf290294b-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " pod="openstack/openstack-galera-0" Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.986355 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/167c5bc5-b21e-4c51-8765-1dcbf290294b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " pod="openstack/openstack-galera-0" Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.986386 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/167c5bc5-b21e-4c51-8765-1dcbf290294b-config-data-default\") pod \"openstack-galera-0\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " pod="openstack/openstack-galera-0" Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.986411 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/167c5bc5-b21e-4c51-8765-1dcbf290294b-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " pod="openstack/openstack-galera-0" Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.986605 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/openstack-galera-0" Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.987634 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/167c5bc5-b21e-4c51-8765-1dcbf290294b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " pod="openstack/openstack-galera-0" Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.988135 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/167c5bc5-b21e-4c51-8765-1dcbf290294b-config-data-default\") pod \"openstack-galera-0\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " pod="openstack/openstack-galera-0" Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.988161 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/167c5bc5-b21e-4c51-8765-1dcbf290294b-kolla-config\") pod \"openstack-galera-0\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " pod="openstack/openstack-galera-0" Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.988603 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/167c5bc5-b21e-4c51-8765-1dcbf290294b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " pod="openstack/openstack-galera-0" Nov 25 09:42:12 crc kubenswrapper[4776]: I1125 09:42:12.997018 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/167c5bc5-b21e-4c51-8765-1dcbf290294b-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " pod="openstack/openstack-galera-0" Nov 25 09:42:13 crc kubenswrapper[4776]: I1125 09:42:13.004941 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/167c5bc5-b21e-4c51-8765-1dcbf290294b-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " pod="openstack/openstack-galera-0" Nov 25 09:42:13 crc kubenswrapper[4776]: I1125 09:42:13.012692 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqc28\" (UniqueName: \"kubernetes.io/projected/167c5bc5-b21e-4c51-8765-1dcbf290294b-kube-api-access-zqc28\") pod \"openstack-galera-0\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " pod="openstack/openstack-galera-0" Nov 25 09:42:13 crc kubenswrapper[4776]: I1125 09:42:13.065328 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " pod="openstack/openstack-galera-0" Nov 25 09:42:13 crc kubenswrapper[4776]: I1125 09:42:13.113607 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 25 09:42:13 crc kubenswrapper[4776]: I1125 09:42:13.470991 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 25 09:42:13 crc kubenswrapper[4776]: W1125 09:42:13.535022 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod167c5bc5_b21e_4c51_8765_1dcbf290294b.slice/crio-522e57d9f9f345bed75f21b21c222210c137832a7c267a45a4a71baba0244de8 WatchSource:0}: Error finding container 522e57d9f9f345bed75f21b21c222210c137832a7c267a45a4a71baba0244de8: Status 404 returned error can't find the container with id 522e57d9f9f345bed75f21b21c222210c137832a7c267a45a4a71baba0244de8 Nov 25 09:42:13 crc kubenswrapper[4776]: I1125 09:42:13.833393 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"167c5bc5-b21e-4c51-8765-1dcbf290294b","Type":"ContainerStarted","Data":"522e57d9f9f345bed75f21b21c222210c137832a7c267a45a4a71baba0244de8"} Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.211292 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.213038 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.216668 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-vfsxn" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.217330 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.217442 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.217484 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.224878 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.325030 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5920f152-04cd-4366-b479-665a1ce8937c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.325140 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5b4m\" (UniqueName: \"kubernetes.io/projected/5920f152-04cd-4366-b479-665a1ce8937c-kube-api-access-p5b4m\") pod \"openstack-cell1-galera-0\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.325203 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5920f152-04cd-4366-b479-665a1ce8937c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.325241 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5920f152-04cd-4366-b479-665a1ce8937c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.325286 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.326648 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5920f152-04cd-4366-b479-665a1ce8937c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.326801 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5920f152-04cd-4366-b479-665a1ce8937c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.326872 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5920f152-04cd-4366-b479-665a1ce8937c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.428441 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5920f152-04cd-4366-b479-665a1ce8937c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.428554 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5920f152-04cd-4366-b479-665a1ce8937c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.428589 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5920f152-04cd-4366-b479-665a1ce8937c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.428650 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5b4m\" (UniqueName: \"kubernetes.io/projected/5920f152-04cd-4366-b479-665a1ce8937c-kube-api-access-p5b4m\") pod \"openstack-cell1-galera-0\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.428677 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5920f152-04cd-4366-b479-665a1ce8937c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.428716 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5920f152-04cd-4366-b479-665a1ce8937c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.428753 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.428818 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5920f152-04cd-4366-b479-665a1ce8937c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.429574 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5920f152-04cd-4366-b479-665a1ce8937c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.430328 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5920f152-04cd-4366-b479-665a1ce8937c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.430946 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5920f152-04cd-4366-b479-665a1ce8937c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.431088 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5920f152-04cd-4366-b479-665a1ce8937c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.431396 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.455172 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5920f152-04cd-4366-b479-665a1ce8937c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.456717 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5b4m\" (UniqueName: \"kubernetes.io/projected/5920f152-04cd-4366-b479-665a1ce8937c-kube-api-access-p5b4m\") pod \"openstack-cell1-galera-0\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.466811 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5920f152-04cd-4366-b479-665a1ce8937c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.474836 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.575723 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.583197 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.586190 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.586374 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.586489 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-d68ds" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.586945 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.587527 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.733207 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cd5ce048-5a09-45a4-9a7d-66712b326ab7-kolla-config\") pod \"memcached-0\" (UID: \"cd5ce048-5a09-45a4-9a7d-66712b326ab7\") " pod="openstack/memcached-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.733285 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd5ce048-5a09-45a4-9a7d-66712b326ab7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"cd5ce048-5a09-45a4-9a7d-66712b326ab7\") " pod="openstack/memcached-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.733363 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd5ce048-5a09-45a4-9a7d-66712b326ab7-config-data\") pod \"memcached-0\" (UID: \"cd5ce048-5a09-45a4-9a7d-66712b326ab7\") " pod="openstack/memcached-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.733384 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd5ce048-5a09-45a4-9a7d-66712b326ab7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"cd5ce048-5a09-45a4-9a7d-66712b326ab7\") " pod="openstack/memcached-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.733451 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgvgj\" (UniqueName: \"kubernetes.io/projected/cd5ce048-5a09-45a4-9a7d-66712b326ab7-kube-api-access-fgvgj\") pod \"memcached-0\" (UID: \"cd5ce048-5a09-45a4-9a7d-66712b326ab7\") " pod="openstack/memcached-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.835395 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cd5ce048-5a09-45a4-9a7d-66712b326ab7-kolla-config\") pod \"memcached-0\" (UID: \"cd5ce048-5a09-45a4-9a7d-66712b326ab7\") " pod="openstack/memcached-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.835474 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd5ce048-5a09-45a4-9a7d-66712b326ab7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"cd5ce048-5a09-45a4-9a7d-66712b326ab7\") " pod="openstack/memcached-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.835518 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd5ce048-5a09-45a4-9a7d-66712b326ab7-config-data\") pod \"memcached-0\" (UID: \"cd5ce048-5a09-45a4-9a7d-66712b326ab7\") " pod="openstack/memcached-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.835545 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd5ce048-5a09-45a4-9a7d-66712b326ab7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"cd5ce048-5a09-45a4-9a7d-66712b326ab7\") " pod="openstack/memcached-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.835618 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgvgj\" (UniqueName: \"kubernetes.io/projected/cd5ce048-5a09-45a4-9a7d-66712b326ab7-kube-api-access-fgvgj\") pod \"memcached-0\" (UID: \"cd5ce048-5a09-45a4-9a7d-66712b326ab7\") " pod="openstack/memcached-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.836192 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cd5ce048-5a09-45a4-9a7d-66712b326ab7-kolla-config\") pod \"memcached-0\" (UID: \"cd5ce048-5a09-45a4-9a7d-66712b326ab7\") " pod="openstack/memcached-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.837093 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd5ce048-5a09-45a4-9a7d-66712b326ab7-config-data\") pod \"memcached-0\" (UID: \"cd5ce048-5a09-45a4-9a7d-66712b326ab7\") " pod="openstack/memcached-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.841144 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd5ce048-5a09-45a4-9a7d-66712b326ab7-memcached-tls-certs\") pod \"memcached-0\" (UID: \"cd5ce048-5a09-45a4-9a7d-66712b326ab7\") " pod="openstack/memcached-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.841947 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd5ce048-5a09-45a4-9a7d-66712b326ab7-combined-ca-bundle\") pod \"memcached-0\" (UID: \"cd5ce048-5a09-45a4-9a7d-66712b326ab7\") " pod="openstack/memcached-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.859544 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgvgj\" (UniqueName: \"kubernetes.io/projected/cd5ce048-5a09-45a4-9a7d-66712b326ab7-kube-api-access-fgvgj\") pod \"memcached-0\" (UID: \"cd5ce048-5a09-45a4-9a7d-66712b326ab7\") " pod="openstack/memcached-0" Nov 25 09:42:14 crc kubenswrapper[4776]: I1125 09:42:14.930516 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 25 09:42:16 crc kubenswrapper[4776]: I1125 09:42:16.358379 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 09:42:16 crc kubenswrapper[4776]: I1125 09:42:16.359979 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 09:42:16 crc kubenswrapper[4776]: I1125 09:42:16.363013 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-znbh5" Nov 25 09:42:16 crc kubenswrapper[4776]: I1125 09:42:16.364869 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 09:42:16 crc kubenswrapper[4776]: I1125 09:42:16.466735 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g46km\" (UniqueName: \"kubernetes.io/projected/e06b81f6-6161-46ae-9830-cd735c64d2f0-kube-api-access-g46km\") pod \"kube-state-metrics-0\" (UID: \"e06b81f6-6161-46ae-9830-cd735c64d2f0\") " pod="openstack/kube-state-metrics-0" Nov 25 09:42:16 crc kubenswrapper[4776]: I1125 09:42:16.568109 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g46km\" (UniqueName: \"kubernetes.io/projected/e06b81f6-6161-46ae-9830-cd735c64d2f0-kube-api-access-g46km\") pod \"kube-state-metrics-0\" (UID: \"e06b81f6-6161-46ae-9830-cd735c64d2f0\") " pod="openstack/kube-state-metrics-0" Nov 25 09:42:16 crc kubenswrapper[4776]: I1125 09:42:16.588629 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g46km\" (UniqueName: \"kubernetes.io/projected/e06b81f6-6161-46ae-9830-cd735c64d2f0-kube-api-access-g46km\") pod \"kube-state-metrics-0\" (UID: \"e06b81f6-6161-46ae-9830-cd735c64d2f0\") " pod="openstack/kube-state-metrics-0" Nov 25 09:42:16 crc kubenswrapper[4776]: I1125 09:42:16.678500 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.244674 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-2bs2f"] Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.247333 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2bs2f" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.250461 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.250486 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-dcdtj" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.250876 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.261931 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-2bs2f"] Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.303967 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-7fgjf"] Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.312056 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-7fgjf" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.321443 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-7fgjf"] Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.329187 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjvld\" (UniqueName: \"kubernetes.io/projected/c59c3d1e-c5c3-4036-ae28-436585e303ab-kube-api-access-kjvld\") pod \"ovn-controller-2bs2f\" (UID: \"c59c3d1e-c5c3-4036-ae28-436585e303ab\") " pod="openstack/ovn-controller-2bs2f" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.329370 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c59c3d1e-c5c3-4036-ae28-436585e303ab-var-run\") pod \"ovn-controller-2bs2f\" (UID: \"c59c3d1e-c5c3-4036-ae28-436585e303ab\") " pod="openstack/ovn-controller-2bs2f" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.329420 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c59c3d1e-c5c3-4036-ae28-436585e303ab-var-log-ovn\") pod \"ovn-controller-2bs2f\" (UID: \"c59c3d1e-c5c3-4036-ae28-436585e303ab\") " pod="openstack/ovn-controller-2bs2f" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.329471 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c59c3d1e-c5c3-4036-ae28-436585e303ab-scripts\") pod \"ovn-controller-2bs2f\" (UID: \"c59c3d1e-c5c3-4036-ae28-436585e303ab\") " pod="openstack/ovn-controller-2bs2f" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.329633 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c59c3d1e-c5c3-4036-ae28-436585e303ab-var-run-ovn\") pod \"ovn-controller-2bs2f\" (UID: \"c59c3d1e-c5c3-4036-ae28-436585e303ab\") " pod="openstack/ovn-controller-2bs2f" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.329714 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/c59c3d1e-c5c3-4036-ae28-436585e303ab-ovn-controller-tls-certs\") pod \"ovn-controller-2bs2f\" (UID: \"c59c3d1e-c5c3-4036-ae28-436585e303ab\") " pod="openstack/ovn-controller-2bs2f" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.329752 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c59c3d1e-c5c3-4036-ae28-436585e303ab-combined-ca-bundle\") pod \"ovn-controller-2bs2f\" (UID: \"c59c3d1e-c5c3-4036-ae28-436585e303ab\") " pod="openstack/ovn-controller-2bs2f" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.430601 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c59c3d1e-c5c3-4036-ae28-436585e303ab-var-run-ovn\") pod \"ovn-controller-2bs2f\" (UID: \"c59c3d1e-c5c3-4036-ae28-436585e303ab\") " pod="openstack/ovn-controller-2bs2f" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.430661 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/10fe3dc2-2bde-4380-8ad3-d71832e31095-etc-ovs\") pod \"ovn-controller-ovs-7fgjf\" (UID: \"10fe3dc2-2bde-4380-8ad3-d71832e31095\") " pod="openstack/ovn-controller-ovs-7fgjf" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.430680 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwjwq\" (UniqueName: \"kubernetes.io/projected/10fe3dc2-2bde-4380-8ad3-d71832e31095-kube-api-access-cwjwq\") pod \"ovn-controller-ovs-7fgjf\" (UID: \"10fe3dc2-2bde-4380-8ad3-d71832e31095\") " pod="openstack/ovn-controller-ovs-7fgjf" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.430704 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/c59c3d1e-c5c3-4036-ae28-436585e303ab-ovn-controller-tls-certs\") pod \"ovn-controller-2bs2f\" (UID: \"c59c3d1e-c5c3-4036-ae28-436585e303ab\") " pod="openstack/ovn-controller-2bs2f" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.430739 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c59c3d1e-c5c3-4036-ae28-436585e303ab-combined-ca-bundle\") pod \"ovn-controller-2bs2f\" (UID: \"c59c3d1e-c5c3-4036-ae28-436585e303ab\") " pod="openstack/ovn-controller-2bs2f" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.430781 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjvld\" (UniqueName: \"kubernetes.io/projected/c59c3d1e-c5c3-4036-ae28-436585e303ab-kube-api-access-kjvld\") pod \"ovn-controller-2bs2f\" (UID: \"c59c3d1e-c5c3-4036-ae28-436585e303ab\") " pod="openstack/ovn-controller-2bs2f" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.430813 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c59c3d1e-c5c3-4036-ae28-436585e303ab-var-run\") pod \"ovn-controller-2bs2f\" (UID: \"c59c3d1e-c5c3-4036-ae28-436585e303ab\") " pod="openstack/ovn-controller-2bs2f" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.430838 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c59c3d1e-c5c3-4036-ae28-436585e303ab-var-log-ovn\") pod \"ovn-controller-2bs2f\" (UID: \"c59c3d1e-c5c3-4036-ae28-436585e303ab\") " pod="openstack/ovn-controller-2bs2f" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.430855 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c59c3d1e-c5c3-4036-ae28-436585e303ab-scripts\") pod \"ovn-controller-2bs2f\" (UID: \"c59c3d1e-c5c3-4036-ae28-436585e303ab\") " pod="openstack/ovn-controller-2bs2f" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.430880 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/10fe3dc2-2bde-4380-8ad3-d71832e31095-var-run\") pod \"ovn-controller-ovs-7fgjf\" (UID: \"10fe3dc2-2bde-4380-8ad3-d71832e31095\") " pod="openstack/ovn-controller-ovs-7fgjf" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.430903 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10fe3dc2-2bde-4380-8ad3-d71832e31095-scripts\") pod \"ovn-controller-ovs-7fgjf\" (UID: \"10fe3dc2-2bde-4380-8ad3-d71832e31095\") " pod="openstack/ovn-controller-ovs-7fgjf" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.430925 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/10fe3dc2-2bde-4380-8ad3-d71832e31095-var-lib\") pod \"ovn-controller-ovs-7fgjf\" (UID: \"10fe3dc2-2bde-4380-8ad3-d71832e31095\") " pod="openstack/ovn-controller-ovs-7fgjf" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.430954 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/10fe3dc2-2bde-4380-8ad3-d71832e31095-var-log\") pod \"ovn-controller-ovs-7fgjf\" (UID: \"10fe3dc2-2bde-4380-8ad3-d71832e31095\") " pod="openstack/ovn-controller-ovs-7fgjf" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.431506 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c59c3d1e-c5c3-4036-ae28-436585e303ab-var-run-ovn\") pod \"ovn-controller-2bs2f\" (UID: \"c59c3d1e-c5c3-4036-ae28-436585e303ab\") " pod="openstack/ovn-controller-2bs2f" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.431575 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c59c3d1e-c5c3-4036-ae28-436585e303ab-var-run\") pod \"ovn-controller-2bs2f\" (UID: \"c59c3d1e-c5c3-4036-ae28-436585e303ab\") " pod="openstack/ovn-controller-2bs2f" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.431619 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c59c3d1e-c5c3-4036-ae28-436585e303ab-var-log-ovn\") pod \"ovn-controller-2bs2f\" (UID: \"c59c3d1e-c5c3-4036-ae28-436585e303ab\") " pod="openstack/ovn-controller-2bs2f" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.433725 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c59c3d1e-c5c3-4036-ae28-436585e303ab-scripts\") pod \"ovn-controller-2bs2f\" (UID: \"c59c3d1e-c5c3-4036-ae28-436585e303ab\") " pod="openstack/ovn-controller-2bs2f" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.438398 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/c59c3d1e-c5c3-4036-ae28-436585e303ab-ovn-controller-tls-certs\") pod \"ovn-controller-2bs2f\" (UID: \"c59c3d1e-c5c3-4036-ae28-436585e303ab\") " pod="openstack/ovn-controller-2bs2f" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.438617 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c59c3d1e-c5c3-4036-ae28-436585e303ab-combined-ca-bundle\") pod \"ovn-controller-2bs2f\" (UID: \"c59c3d1e-c5c3-4036-ae28-436585e303ab\") " pod="openstack/ovn-controller-2bs2f" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.453624 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjvld\" (UniqueName: \"kubernetes.io/projected/c59c3d1e-c5c3-4036-ae28-436585e303ab-kube-api-access-kjvld\") pod \"ovn-controller-2bs2f\" (UID: \"c59c3d1e-c5c3-4036-ae28-436585e303ab\") " pod="openstack/ovn-controller-2bs2f" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.532518 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/10fe3dc2-2bde-4380-8ad3-d71832e31095-etc-ovs\") pod \"ovn-controller-ovs-7fgjf\" (UID: \"10fe3dc2-2bde-4380-8ad3-d71832e31095\") " pod="openstack/ovn-controller-ovs-7fgjf" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.532982 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/10fe3dc2-2bde-4380-8ad3-d71832e31095-etc-ovs\") pod \"ovn-controller-ovs-7fgjf\" (UID: \"10fe3dc2-2bde-4380-8ad3-d71832e31095\") " pod="openstack/ovn-controller-ovs-7fgjf" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.533012 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwjwq\" (UniqueName: \"kubernetes.io/projected/10fe3dc2-2bde-4380-8ad3-d71832e31095-kube-api-access-cwjwq\") pod \"ovn-controller-ovs-7fgjf\" (UID: \"10fe3dc2-2bde-4380-8ad3-d71832e31095\") " pod="openstack/ovn-controller-ovs-7fgjf" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.533392 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/10fe3dc2-2bde-4380-8ad3-d71832e31095-var-run\") pod \"ovn-controller-ovs-7fgjf\" (UID: \"10fe3dc2-2bde-4380-8ad3-d71832e31095\") " pod="openstack/ovn-controller-ovs-7fgjf" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.533455 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10fe3dc2-2bde-4380-8ad3-d71832e31095-scripts\") pod \"ovn-controller-ovs-7fgjf\" (UID: \"10fe3dc2-2bde-4380-8ad3-d71832e31095\") " pod="openstack/ovn-controller-ovs-7fgjf" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.533508 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/10fe3dc2-2bde-4380-8ad3-d71832e31095-var-lib\") pod \"ovn-controller-ovs-7fgjf\" (UID: \"10fe3dc2-2bde-4380-8ad3-d71832e31095\") " pod="openstack/ovn-controller-ovs-7fgjf" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.533593 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/10fe3dc2-2bde-4380-8ad3-d71832e31095-var-log\") pod \"ovn-controller-ovs-7fgjf\" (UID: \"10fe3dc2-2bde-4380-8ad3-d71832e31095\") " pod="openstack/ovn-controller-ovs-7fgjf" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.533814 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/10fe3dc2-2bde-4380-8ad3-d71832e31095-var-run\") pod \"ovn-controller-ovs-7fgjf\" (UID: \"10fe3dc2-2bde-4380-8ad3-d71832e31095\") " pod="openstack/ovn-controller-ovs-7fgjf" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.533836 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/10fe3dc2-2bde-4380-8ad3-d71832e31095-var-log\") pod \"ovn-controller-ovs-7fgjf\" (UID: \"10fe3dc2-2bde-4380-8ad3-d71832e31095\") " pod="openstack/ovn-controller-ovs-7fgjf" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.533948 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/10fe3dc2-2bde-4380-8ad3-d71832e31095-var-lib\") pod \"ovn-controller-ovs-7fgjf\" (UID: \"10fe3dc2-2bde-4380-8ad3-d71832e31095\") " pod="openstack/ovn-controller-ovs-7fgjf" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.536185 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10fe3dc2-2bde-4380-8ad3-d71832e31095-scripts\") pod \"ovn-controller-ovs-7fgjf\" (UID: \"10fe3dc2-2bde-4380-8ad3-d71832e31095\") " pod="openstack/ovn-controller-ovs-7fgjf" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.548296 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwjwq\" (UniqueName: \"kubernetes.io/projected/10fe3dc2-2bde-4380-8ad3-d71832e31095-kube-api-access-cwjwq\") pod \"ovn-controller-ovs-7fgjf\" (UID: \"10fe3dc2-2bde-4380-8ad3-d71832e31095\") " pod="openstack/ovn-controller-ovs-7fgjf" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.575901 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2bs2f" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.626632 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-7fgjf" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.792855 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.795847 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.800832 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.801328 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.801820 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.802178 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.802210 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-h5rzz" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.807025 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.839129 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.839182 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.839211 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvrvj\" (UniqueName: \"kubernetes.io/projected/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-kube-api-access-wvrvj\") pod \"ovsdbserver-nb-0\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.839246 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.839288 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-config\") pod \"ovsdbserver-nb-0\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.839327 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.839351 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.839388 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.940651 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.940726 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.940750 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvrvj\" (UniqueName: \"kubernetes.io/projected/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-kube-api-access-wvrvj\") pod \"ovsdbserver-nb-0\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.940791 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.940841 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-config\") pod \"ovsdbserver-nb-0\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.940878 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.940925 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.940975 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.941235 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.941996 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.948127 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.952814 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-config\") pod \"ovsdbserver-nb-0\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.960559 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.962995 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.963156 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvrvj\" (UniqueName: \"kubernetes.io/projected/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-kube-api-access-wvrvj\") pod \"ovsdbserver-nb-0\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.966766 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:20 crc kubenswrapper[4776]: I1125 09:42:20.986529 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:21 crc kubenswrapper[4776]: I1125 09:42:21.126746 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.226410 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.228269 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.230334 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.230380 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.233573 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.234178 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-8shd5" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.236123 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.279435 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.279764 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.279787 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.279849 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxscj\" (UniqueName: \"kubernetes.io/projected/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-kube-api-access-pxscj\") pod \"ovsdbserver-sb-0\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.279869 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.279978 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-config\") pod \"ovsdbserver-sb-0\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.280037 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.280095 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.381467 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxscj\" (UniqueName: \"kubernetes.io/projected/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-kube-api-access-pxscj\") pod \"ovsdbserver-sb-0\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.381521 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.381594 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-config\") pod \"ovsdbserver-sb-0\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.381620 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.381649 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.382000 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.382104 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.382137 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.382516 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.382773 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.383484 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-config\") pod \"ovsdbserver-sb-0\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.384517 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.398388 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.399216 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxscj\" (UniqueName: \"kubernetes.io/projected/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-kube-api-access-pxscj\") pod \"ovsdbserver-sb-0\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.400055 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.410571 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.427255 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:23 crc kubenswrapper[4776]: I1125 09:42:23.559781 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:27 crc kubenswrapper[4776]: E1125 09:42:27.054667 4776 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:95d67f51dfedd5bd3ec785b488425295b2d8c41feae3e6386ef471615381809b" Nov 25 09:42:27 crc kubenswrapper[4776]: E1125 09:42:27.055478 4776 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:95d67f51dfedd5bd3ec785b488425295b2d8c41feae3e6386ef471615381809b,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5g6h4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(4512cf4e-6c75-493e-a80a-4b0accd750a8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 09:42:27 crc kubenswrapper[4776]: E1125 09:42:27.056751 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="4512cf4e-6c75-493e-a80a-4b0accd750a8" Nov 25 09:42:27 crc kubenswrapper[4776]: E1125 09:42:27.836410 4776 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba" Nov 25 09:42:27 crc kubenswrapper[4776]: E1125 09:42:27.836606 4776 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2lq5d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7bdd77c89-sclc2_openstack(bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 09:42:27 crc kubenswrapper[4776]: E1125 09:42:27.837837 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7bdd77c89-sclc2" podUID="bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24" Nov 25 09:42:29 crc kubenswrapper[4776]: E1125 09:42:29.444436 4776 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba" Nov 25 09:42:29 crc kubenswrapper[4776]: E1125 09:42:29.445106 4776 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vjbgr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6584b49599-v697v_openstack(70048934-5af8-43a5-a7e4-b78a6bc70fea): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 09:42:29 crc kubenswrapper[4776]: E1125 09:42:29.446504 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-6584b49599-v697v" podUID="70048934-5af8-43a5-a7e4-b78a6bc70fea" Nov 25 09:42:29 crc kubenswrapper[4776]: E1125 09:42:29.495688 4776 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba" Nov 25 09:42:29 crc kubenswrapper[4776]: E1125 09:42:29.500113 4776 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x5stb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7c6d9948dc-swg9r_openstack(b732643a-7578-4b49-802e-7f04b5d3793c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 09:42:29 crc kubenswrapper[4776]: E1125 09:42:29.502293 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7c6d9948dc-swg9r" podUID="b732643a-7578-4b49-802e-7f04b5d3793c" Nov 25 09:42:29 crc kubenswrapper[4776]: E1125 09:42:29.512718 4776 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba" Nov 25 09:42:29 crc kubenswrapper[4776]: E1125 09:42:29.512954 4776 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nhtx4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6486446b9f-4l8fb_openstack(f40cef2d-0d59-4300-9336-fc5a793b9cbf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 09:42:29 crc kubenswrapper[4776]: E1125 09:42:29.514050 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-6486446b9f-4l8fb" podUID="f40cef2d-0d59-4300-9336-fc5a793b9cbf" Nov 25 09:42:29 crc kubenswrapper[4776]: I1125 09:42:29.579550 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdd77c89-sclc2" Nov 25 09:42:29 crc kubenswrapper[4776]: I1125 09:42:29.678761 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24-config\") pod \"bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24\" (UID: \"bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24\") " Nov 25 09:42:29 crc kubenswrapper[4776]: I1125 09:42:29.679899 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lq5d\" (UniqueName: \"kubernetes.io/projected/bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24-kube-api-access-2lq5d\") pod \"bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24\" (UID: \"bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24\") " Nov 25 09:42:29 crc kubenswrapper[4776]: I1125 09:42:29.679326 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24-config" (OuterVolumeSpecName: "config") pod "bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24" (UID: "bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:42:29 crc kubenswrapper[4776]: I1125 09:42:29.688189 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24-kube-api-access-2lq5d" (OuterVolumeSpecName: "kube-api-access-2lq5d") pod "bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24" (UID: "bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24"). InnerVolumeSpecName "kube-api-access-2lq5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:42:29 crc kubenswrapper[4776]: I1125 09:42:29.781498 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:42:29 crc kubenswrapper[4776]: I1125 09:42:29.781528 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lq5d\" (UniqueName: \"kubernetes.io/projected/bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24-kube-api-access-2lq5d\") on node \"crc\" DevicePath \"\"" Nov 25 09:42:29 crc kubenswrapper[4776]: I1125 09:42:29.996228 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 09:42:29 crc kubenswrapper[4776]: I1125 09:42:29.999558 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"167c5bc5-b21e-4c51-8765-1dcbf290294b","Type":"ContainerStarted","Data":"f655f33165dfd7482166f931fb775ba2b424e5b9c57973932d34b6255eda7162"} Nov 25 09:42:30 crc kubenswrapper[4776]: I1125 09:42:30.004752 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdd77c89-sclc2" Nov 25 09:42:30 crc kubenswrapper[4776]: E1125 09:42:30.007668 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba\\\"\"" pod="openstack/dnsmasq-dns-6486446b9f-4l8fb" podUID="f40cef2d-0d59-4300-9336-fc5a793b9cbf" Nov 25 09:42:30 crc kubenswrapper[4776]: I1125 09:42:30.008096 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdd77c89-sclc2" event={"ID":"bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24","Type":"ContainerDied","Data":"123ff647f45d05b0940b78b948deaf2d04e5ba4ee28e3fb19394f01658b09194"} Nov 25 09:42:30 crc kubenswrapper[4776]: E1125 09:42:30.010420 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba\\\"\"" pod="openstack/dnsmasq-dns-7c6d9948dc-swg9r" podUID="b732643a-7578-4b49-802e-7f04b5d3793c" Nov 25 09:42:30 crc kubenswrapper[4776]: I1125 09:42:30.012489 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 25 09:42:30 crc kubenswrapper[4776]: W1125 09:42:30.015994 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5920f152_04cd_4366_b479_665a1ce8937c.slice/crio-e7a84137f7ca4e701b70532e3ecb35e1f3005afd8011d822ceb5e2e2fba6c8cb WatchSource:0}: Error finding container e7a84137f7ca4e701b70532e3ecb35e1f3005afd8011d822ceb5e2e2fba6c8cb: Status 404 returned error can't find the container with id e7a84137f7ca4e701b70532e3ecb35e1f3005afd8011d822ceb5e2e2fba6c8cb Nov 25 09:42:30 crc kubenswrapper[4776]: I1125 09:42:30.026276 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 25 09:42:30 crc kubenswrapper[4776]: I1125 09:42:30.044889 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-2bs2f"] Nov 25 09:42:30 crc kubenswrapper[4776]: I1125 09:42:30.137583 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-sclc2"] Nov 25 09:42:30 crc kubenswrapper[4776]: I1125 09:42:30.142612 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bdd77c89-sclc2"] Nov 25 09:42:30 crc kubenswrapper[4776]: I1125 09:42:30.363888 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6584b49599-v697v" Nov 25 09:42:30 crc kubenswrapper[4776]: I1125 09:42:30.365992 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 25 09:42:30 crc kubenswrapper[4776]: I1125 09:42:30.389744 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70048934-5af8-43a5-a7e4-b78a6bc70fea-config\") pod \"70048934-5af8-43a5-a7e4-b78a6bc70fea\" (UID: \"70048934-5af8-43a5-a7e4-b78a6bc70fea\") " Nov 25 09:42:30 crc kubenswrapper[4776]: I1125 09:42:30.390296 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjbgr\" (UniqueName: \"kubernetes.io/projected/70048934-5af8-43a5-a7e4-b78a6bc70fea-kube-api-access-vjbgr\") pod \"70048934-5af8-43a5-a7e4-b78a6bc70fea\" (UID: \"70048934-5af8-43a5-a7e4-b78a6bc70fea\") " Nov 25 09:42:30 crc kubenswrapper[4776]: I1125 09:42:30.390231 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70048934-5af8-43a5-a7e4-b78a6bc70fea-config" (OuterVolumeSpecName: "config") pod "70048934-5af8-43a5-a7e4-b78a6bc70fea" (UID: "70048934-5af8-43a5-a7e4-b78a6bc70fea"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:42:30 crc kubenswrapper[4776]: I1125 09:42:30.390494 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70048934-5af8-43a5-a7e4-b78a6bc70fea-dns-svc\") pod \"70048934-5af8-43a5-a7e4-b78a6bc70fea\" (UID: \"70048934-5af8-43a5-a7e4-b78a6bc70fea\") " Nov 25 09:42:30 crc kubenswrapper[4776]: I1125 09:42:30.390764 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70048934-5af8-43a5-a7e4-b78a6bc70fea-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "70048934-5af8-43a5-a7e4-b78a6bc70fea" (UID: "70048934-5af8-43a5-a7e4-b78a6bc70fea"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:42:30 crc kubenswrapper[4776]: I1125 09:42:30.391097 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70048934-5af8-43a5-a7e4-b78a6bc70fea-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:42:30 crc kubenswrapper[4776]: I1125 09:42:30.391116 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70048934-5af8-43a5-a7e4-b78a6bc70fea-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 09:42:30 crc kubenswrapper[4776]: I1125 09:42:30.456330 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70048934-5af8-43a5-a7e4-b78a6bc70fea-kube-api-access-vjbgr" (OuterVolumeSpecName: "kube-api-access-vjbgr") pod "70048934-5af8-43a5-a7e4-b78a6bc70fea" (UID: "70048934-5af8-43a5-a7e4-b78a6bc70fea"). InnerVolumeSpecName "kube-api-access-vjbgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:42:30 crc kubenswrapper[4776]: I1125 09:42:30.492472 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjbgr\" (UniqueName: \"kubernetes.io/projected/70048934-5af8-43a5-a7e4-b78a6bc70fea-kube-api-access-vjbgr\") on node \"crc\" DevicePath \"\"" Nov 25 09:42:30 crc kubenswrapper[4776]: I1125 09:42:30.890092 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 25 09:42:31 crc kubenswrapper[4776]: I1125 09:42:31.016330 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e06b81f6-6161-46ae-9830-cd735c64d2f0","Type":"ContainerStarted","Data":"781f22edcf0938a2b04712393256587ccd5b53a5cb9b50329ea8a3dc55ebe642"} Nov 25 09:42:31 crc kubenswrapper[4776]: I1125 09:42:31.018326 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"5920f152-04cd-4366-b479-665a1ce8937c","Type":"ContainerStarted","Data":"061b5a0bcf2039c372d93606dd4f4d9c15e175c9e5f3706f9ed15162829b3321"} Nov 25 09:42:31 crc kubenswrapper[4776]: I1125 09:42:31.018364 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"5920f152-04cd-4366-b479-665a1ce8937c","Type":"ContainerStarted","Data":"e7a84137f7ca4e701b70532e3ecb35e1f3005afd8011d822ceb5e2e2fba6c8cb"} Nov 25 09:42:31 crc kubenswrapper[4776]: I1125 09:42:31.019889 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2bs2f" event={"ID":"c59c3d1e-c5c3-4036-ae28-436585e303ab","Type":"ContainerStarted","Data":"35bee38875dfc28653ef000ae6a9a992ad8fcb6fb75884a44dd02aeef180c4e2"} Nov 25 09:42:31 crc kubenswrapper[4776]: I1125 09:42:31.022754 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4512cf4e-6c75-493e-a80a-4b0accd750a8","Type":"ContainerStarted","Data":"e5d2c432a51c339c288f48c9a58a0ccdecc76c714979d2e84604d3bc00357881"} Nov 25 09:42:31 crc kubenswrapper[4776]: I1125 09:42:31.024705 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6584b49599-v697v" event={"ID":"70048934-5af8-43a5-a7e4-b78a6bc70fea","Type":"ContainerDied","Data":"a5e6b816dfdd0cfa8b46cd98ab3055b84aa519299aae28d53c313ff23ad83d50"} Nov 25 09:42:31 crc kubenswrapper[4776]: I1125 09:42:31.024744 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6584b49599-v697v" Nov 25 09:42:31 crc kubenswrapper[4776]: I1125 09:42:31.026348 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f32041a0-a8fc-48a5-afab-476baff89e8c","Type":"ContainerStarted","Data":"b75205641b6269bc97dcaf11905b1173d0238b983968101c11decb00e38566b3"} Nov 25 09:42:31 crc kubenswrapper[4776]: I1125 09:42:31.027423 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df","Type":"ContainerStarted","Data":"82692ea71c99491423198e874e369d17ef5e820a3bebef5b7b1bc955ce2bf61e"} Nov 25 09:42:31 crc kubenswrapper[4776]: I1125 09:42:31.028622 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"cd5ce048-5a09-45a4-9a7d-66712b326ab7","Type":"ContainerStarted","Data":"aa48600c469e2c4f1b609fa041f9ecd490af4b97a40bc93c0a48c6d44ea6bc8e"} Nov 25 09:42:31 crc kubenswrapper[4776]: I1125 09:42:31.136306 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-v697v"] Nov 25 09:42:31 crc kubenswrapper[4776]: I1125 09:42:31.146058 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6584b49599-v697v"] Nov 25 09:42:31 crc kubenswrapper[4776]: W1125 09:42:31.329736 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e7d9d60_9538_4870_8f62_434fd9f1ab0d.slice/crio-c5c65c194be6d770911d8e35680f49902817102fc8b9409f79be2693b39814bd WatchSource:0}: Error finding container c5c65c194be6d770911d8e35680f49902817102fc8b9409f79be2693b39814bd: Status 404 returned error can't find the container with id c5c65c194be6d770911d8e35680f49902817102fc8b9409f79be2693b39814bd Nov 25 09:42:31 crc kubenswrapper[4776]: I1125 09:42:31.418305 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-7fgjf"] Nov 25 09:42:31 crc kubenswrapper[4776]: W1125 09:42:31.483169 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10fe3dc2_2bde_4380_8ad3_d71832e31095.slice/crio-39cdf7b2812d7e93af5c98ff938ad95f0c62191b3c2d0b2154912df748de511b WatchSource:0}: Error finding container 39cdf7b2812d7e93af5c98ff938ad95f0c62191b3c2d0b2154912df748de511b: Status 404 returned error can't find the container with id 39cdf7b2812d7e93af5c98ff938ad95f0c62191b3c2d0b2154912df748de511b Nov 25 09:42:31 crc kubenswrapper[4776]: I1125 09:42:31.672025 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70048934-5af8-43a5-a7e4-b78a6bc70fea" path="/var/lib/kubelet/pods/70048934-5af8-43a5-a7e4-b78a6bc70fea/volumes" Nov 25 09:42:31 crc kubenswrapper[4776]: I1125 09:42:31.672547 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24" path="/var/lib/kubelet/pods/bb59d0a0-c4a0-41e7-b7a9-aaf9f13b0d24/volumes" Nov 25 09:42:32 crc kubenswrapper[4776]: I1125 09:42:32.038596 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7fgjf" event={"ID":"10fe3dc2-2bde-4380-8ad3-d71832e31095","Type":"ContainerStarted","Data":"39cdf7b2812d7e93af5c98ff938ad95f0c62191b3c2d0b2154912df748de511b"} Nov 25 09:42:32 crc kubenswrapper[4776]: I1125 09:42:32.040306 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6e7d9d60-9538-4870-8f62-434fd9f1ab0d","Type":"ContainerStarted","Data":"c5c65c194be6d770911d8e35680f49902817102fc8b9409f79be2693b39814bd"} Nov 25 09:42:34 crc kubenswrapper[4776]: I1125 09:42:34.061291 4776 generic.go:334] "Generic (PLEG): container finished" podID="5920f152-04cd-4366-b479-665a1ce8937c" containerID="061b5a0bcf2039c372d93606dd4f4d9c15e175c9e5f3706f9ed15162829b3321" exitCode=0 Nov 25 09:42:34 crc kubenswrapper[4776]: I1125 09:42:34.061377 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"5920f152-04cd-4366-b479-665a1ce8937c","Type":"ContainerDied","Data":"061b5a0bcf2039c372d93606dd4f4d9c15e175c9e5f3706f9ed15162829b3321"} Nov 25 09:42:34 crc kubenswrapper[4776]: I1125 09:42:34.063563 4776 generic.go:334] "Generic (PLEG): container finished" podID="167c5bc5-b21e-4c51-8765-1dcbf290294b" containerID="f655f33165dfd7482166f931fb775ba2b424e5b9c57973932d34b6255eda7162" exitCode=0 Nov 25 09:42:34 crc kubenswrapper[4776]: I1125 09:42:34.063586 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"167c5bc5-b21e-4c51-8765-1dcbf290294b","Type":"ContainerDied","Data":"f655f33165dfd7482166f931fb775ba2b424e5b9c57973932d34b6255eda7162"} Nov 25 09:42:35 crc kubenswrapper[4776]: I1125 09:42:35.072225 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df","Type":"ContainerStarted","Data":"215609fe2a7a4ad5088201866359a4ff41f487eaa766e53084fae555576c9ef0"} Nov 25 09:42:35 crc kubenswrapper[4776]: I1125 09:42:35.074849 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"cd5ce048-5a09-45a4-9a7d-66712b326ab7","Type":"ContainerStarted","Data":"6b83a886269bd5ccf82ce338326e0a5b6749a7805141d811c4f9ff42aa7148ca"} Nov 25 09:42:35 crc kubenswrapper[4776]: I1125 09:42:35.075891 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 25 09:42:35 crc kubenswrapper[4776]: I1125 09:42:35.078189 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"167c5bc5-b21e-4c51-8765-1dcbf290294b","Type":"ContainerStarted","Data":"2f383d0debcf43b01564a0db854854b92a294770e346344bcd8dce403c1e0251"} Nov 25 09:42:35 crc kubenswrapper[4776]: I1125 09:42:35.080657 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"5920f152-04cd-4366-b479-665a1ce8937c","Type":"ContainerStarted","Data":"fb166b129931ff672d247f77f4c3b881346f25b0410d683eb6bb28f72d3af0a4"} Nov 25 09:42:35 crc kubenswrapper[4776]: I1125 09:42:35.107662 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=17.160377683 podStartE2EDuration="21.107637081s" podCreationTimestamp="2025-11-25 09:42:14 +0000 UTC" firstStartedPulling="2025-11-25 09:42:30.01063747 +0000 UTC m=+1095.051697023" lastFinishedPulling="2025-11-25 09:42:33.957896868 +0000 UTC m=+1098.998956421" observedRunningTime="2025-11-25 09:42:35.093362984 +0000 UTC m=+1100.134422537" watchObservedRunningTime="2025-11-25 09:42:35.107637081 +0000 UTC m=+1100.148696634" Nov 25 09:42:35 crc kubenswrapper[4776]: I1125 09:42:35.125091 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=22.125075659 podStartE2EDuration="22.125075659s" podCreationTimestamp="2025-11-25 09:42:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:42:35.123964321 +0000 UTC m=+1100.165023874" watchObservedRunningTime="2025-11-25 09:42:35.125075659 +0000 UTC m=+1100.166135202" Nov 25 09:42:35 crc kubenswrapper[4776]: I1125 09:42:35.150840 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=8.181958113 podStartE2EDuration="24.150817864s" podCreationTimestamp="2025-11-25 09:42:11 +0000 UTC" firstStartedPulling="2025-11-25 09:42:13.543054926 +0000 UTC m=+1078.584114469" lastFinishedPulling="2025-11-25 09:42:29.511914667 +0000 UTC m=+1094.552974220" observedRunningTime="2025-11-25 09:42:35.141695865 +0000 UTC m=+1100.182755418" watchObservedRunningTime="2025-11-25 09:42:35.150817864 +0000 UTC m=+1100.191877417" Nov 25 09:42:36 crc kubenswrapper[4776]: I1125 09:42:36.090812 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e06b81f6-6161-46ae-9830-cd735c64d2f0","Type":"ContainerStarted","Data":"51ee7ef170e152f40514f8d16bd62c383594344be3ed28e799c200c3721328e7"} Nov 25 09:42:36 crc kubenswrapper[4776]: I1125 09:42:36.090928 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 25 09:42:36 crc kubenswrapper[4776]: I1125 09:42:36.092684 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2bs2f" event={"ID":"c59c3d1e-c5c3-4036-ae28-436585e303ab","Type":"ContainerStarted","Data":"c1a79c25e3fcf070aa4efe449ff4827dd403b86266f770d2f2b4441e527d7adf"} Nov 25 09:42:36 crc kubenswrapper[4776]: I1125 09:42:36.093081 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-2bs2f" Nov 25 09:42:36 crc kubenswrapper[4776]: I1125 09:42:36.094689 4776 generic.go:334] "Generic (PLEG): container finished" podID="10fe3dc2-2bde-4380-8ad3-d71832e31095" containerID="b43aaf9ab18cf2548b29395d03bbfa83bdd75486872c6906cf0d6deac19222b4" exitCode=0 Nov 25 09:42:36 crc kubenswrapper[4776]: I1125 09:42:36.094997 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7fgjf" event={"ID":"10fe3dc2-2bde-4380-8ad3-d71832e31095","Type":"ContainerDied","Data":"b43aaf9ab18cf2548b29395d03bbfa83bdd75486872c6906cf0d6deac19222b4"} Nov 25 09:42:36 crc kubenswrapper[4776]: I1125 09:42:36.097432 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6e7d9d60-9538-4870-8f62-434fd9f1ab0d","Type":"ContainerStarted","Data":"ce501a13d7bab63cc6c0c2608dcb59e6fdc92a7de550426b2ee3b8ad29264e30"} Nov 25 09:42:36 crc kubenswrapper[4776]: I1125 09:42:36.106918 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=15.362174083 podStartE2EDuration="20.106895853s" podCreationTimestamp="2025-11-25 09:42:16 +0000 UTC" firstStartedPulling="2025-11-25 09:42:30.01064554 +0000 UTC m=+1095.051705083" lastFinishedPulling="2025-11-25 09:42:34.75536729 +0000 UTC m=+1099.796426853" observedRunningTime="2025-11-25 09:42:36.105791716 +0000 UTC m=+1101.146851269" watchObservedRunningTime="2025-11-25 09:42:36.106895853 +0000 UTC m=+1101.147955406" Nov 25 09:42:36 crc kubenswrapper[4776]: I1125 09:42:36.161141 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-2bs2f" podStartSLOduration=11.495220649 podStartE2EDuration="16.161116953s" podCreationTimestamp="2025-11-25 09:42:20 +0000 UTC" firstStartedPulling="2025-11-25 09:42:30.037123784 +0000 UTC m=+1095.078183337" lastFinishedPulling="2025-11-25 09:42:34.703020088 +0000 UTC m=+1099.744079641" observedRunningTime="2025-11-25 09:42:36.153792279 +0000 UTC m=+1101.194851822" watchObservedRunningTime="2025-11-25 09:42:36.161116953 +0000 UTC m=+1101.202176506" Nov 25 09:42:37 crc kubenswrapper[4776]: I1125 09:42:37.105180 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7fgjf" event={"ID":"10fe3dc2-2bde-4380-8ad3-d71832e31095","Type":"ContainerStarted","Data":"5002006f906f0b17afbf25b23648688a842975cfd3b385d57e3191040607fb57"} Nov 25 09:42:37 crc kubenswrapper[4776]: I1125 09:42:37.105516 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7fgjf" event={"ID":"10fe3dc2-2bde-4380-8ad3-d71832e31095","Type":"ContainerStarted","Data":"eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06"} Nov 25 09:42:37 crc kubenswrapper[4776]: I1125 09:42:37.140652 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-7fgjf" podStartSLOduration=13.92008549 podStartE2EDuration="17.140626149s" podCreationTimestamp="2025-11-25 09:42:20 +0000 UTC" firstStartedPulling="2025-11-25 09:42:31.488889659 +0000 UTC m=+1096.529949212" lastFinishedPulling="2025-11-25 09:42:34.709430318 +0000 UTC m=+1099.750489871" observedRunningTime="2025-11-25 09:42:37.128605378 +0000 UTC m=+1102.169664941" watchObservedRunningTime="2025-11-25 09:42:37.140626149 +0000 UTC m=+1102.181685722" Nov 25 09:42:38 crc kubenswrapper[4776]: I1125 09:42:38.111804 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-7fgjf" Nov 25 09:42:38 crc kubenswrapper[4776]: I1125 09:42:38.111859 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-7fgjf" Nov 25 09:42:39 crc kubenswrapper[4776]: I1125 09:42:39.933207 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 25 09:42:41 crc kubenswrapper[4776]: I1125 09:42:41.141891 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6e7d9d60-9538-4870-8f62-434fd9f1ab0d","Type":"ContainerStarted","Data":"17cb0f155d2722e145dfbd65de7d057bce9f1ad8f11fcd4b4347b19f7b5ce218"} Nov 25 09:42:41 crc kubenswrapper[4776]: I1125 09:42:41.146606 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df","Type":"ContainerStarted","Data":"ee2853983a149e403a868c5541f34466a714bc2a4620f7f4820109d6bf8f4f6c"} Nov 25 09:42:41 crc kubenswrapper[4776]: I1125 09:42:41.177403 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=9.808363813 podStartE2EDuration="19.177377417s" podCreationTimestamp="2025-11-25 09:42:22 +0000 UTC" firstStartedPulling="2025-11-25 09:42:31.334972351 +0000 UTC m=+1096.376031904" lastFinishedPulling="2025-11-25 09:42:40.703985935 +0000 UTC m=+1105.745045508" observedRunningTime="2025-11-25 09:42:41.168464404 +0000 UTC m=+1106.209523957" watchObservedRunningTime="2025-11-25 09:42:41.177377417 +0000 UTC m=+1106.218436980" Nov 25 09:42:41 crc kubenswrapper[4776]: I1125 09:42:41.190549 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=11.963120044 podStartE2EDuration="22.190528897s" podCreationTimestamp="2025-11-25 09:42:19 +0000 UTC" firstStartedPulling="2025-11-25 09:42:30.467014562 +0000 UTC m=+1095.508074115" lastFinishedPulling="2025-11-25 09:42:40.694423425 +0000 UTC m=+1105.735482968" observedRunningTime="2025-11-25 09:42:41.186374293 +0000 UTC m=+1106.227433836" watchObservedRunningTime="2025-11-25 09:42:41.190528897 +0000 UTC m=+1106.231588450" Nov 25 09:42:41 crc kubenswrapper[4776]: I1125 09:42:41.560896 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:41 crc kubenswrapper[4776]: I1125 09:42:41.601501 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.127519 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.164448 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.173654 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.207975 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.449631 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-swg9r"] Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.484696 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-65c9b8d4f7-n2mvq"] Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.486530 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65c9b8d4f7-n2mvq" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.489498 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.497846 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65c9b8d4f7-n2mvq"] Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.577204 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-2mts9"] Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.578401 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-2mts9" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.581430 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.596764 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6xwm\" (UniqueName: \"kubernetes.io/projected/a483d333-16fd-4c48-854f-590fe5f0665c-kube-api-access-q6xwm\") pod \"dnsmasq-dns-65c9b8d4f7-n2mvq\" (UID: \"a483d333-16fd-4c48-854f-590fe5f0665c\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-n2mvq" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.596816 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a483d333-16fd-4c48-854f-590fe5f0665c-config\") pod \"dnsmasq-dns-65c9b8d4f7-n2mvq\" (UID: \"a483d333-16fd-4c48-854f-590fe5f0665c\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-n2mvq" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.596873 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a483d333-16fd-4c48-854f-590fe5f0665c-ovsdbserver-sb\") pod \"dnsmasq-dns-65c9b8d4f7-n2mvq\" (UID: \"a483d333-16fd-4c48-854f-590fe5f0665c\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-n2mvq" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.596929 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a483d333-16fd-4c48-854f-590fe5f0665c-dns-svc\") pod \"dnsmasq-dns-65c9b8d4f7-n2mvq\" (UID: \"a483d333-16fd-4c48-854f-590fe5f0665c\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-n2mvq" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.609154 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-2mts9"] Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.698328 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a483d333-16fd-4c48-854f-590fe5f0665c-ovsdbserver-sb\") pod \"dnsmasq-dns-65c9b8d4f7-n2mvq\" (UID: \"a483d333-16fd-4c48-854f-590fe5f0665c\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-n2mvq" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.698391 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-config\") pod \"ovn-controller-metrics-2mts9\" (UID: \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\") " pod="openstack/ovn-controller-metrics-2mts9" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.698455 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a483d333-16fd-4c48-854f-590fe5f0665c-dns-svc\") pod \"dnsmasq-dns-65c9b8d4f7-n2mvq\" (UID: \"a483d333-16fd-4c48-854f-590fe5f0665c\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-n2mvq" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.698491 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xbmh\" (UniqueName: \"kubernetes.io/projected/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-kube-api-access-8xbmh\") pod \"ovn-controller-metrics-2mts9\" (UID: \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\") " pod="openstack/ovn-controller-metrics-2mts9" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.698540 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-ovs-rundir\") pod \"ovn-controller-metrics-2mts9\" (UID: \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\") " pod="openstack/ovn-controller-metrics-2mts9" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.698574 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-2mts9\" (UID: \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\") " pod="openstack/ovn-controller-metrics-2mts9" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.698609 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-combined-ca-bundle\") pod \"ovn-controller-metrics-2mts9\" (UID: \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\") " pod="openstack/ovn-controller-metrics-2mts9" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.698642 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-ovn-rundir\") pod \"ovn-controller-metrics-2mts9\" (UID: \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\") " pod="openstack/ovn-controller-metrics-2mts9" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.698697 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6xwm\" (UniqueName: \"kubernetes.io/projected/a483d333-16fd-4c48-854f-590fe5f0665c-kube-api-access-q6xwm\") pod \"dnsmasq-dns-65c9b8d4f7-n2mvq\" (UID: \"a483d333-16fd-4c48-854f-590fe5f0665c\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-n2mvq" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.698719 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a483d333-16fd-4c48-854f-590fe5f0665c-config\") pod \"dnsmasq-dns-65c9b8d4f7-n2mvq\" (UID: \"a483d333-16fd-4c48-854f-590fe5f0665c\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-n2mvq" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.700882 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a483d333-16fd-4c48-854f-590fe5f0665c-config\") pod \"dnsmasq-dns-65c9b8d4f7-n2mvq\" (UID: \"a483d333-16fd-4c48-854f-590fe5f0665c\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-n2mvq" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.701046 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a483d333-16fd-4c48-854f-590fe5f0665c-dns-svc\") pod \"dnsmasq-dns-65c9b8d4f7-n2mvq\" (UID: \"a483d333-16fd-4c48-854f-590fe5f0665c\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-n2mvq" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.701443 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a483d333-16fd-4c48-854f-590fe5f0665c-ovsdbserver-sb\") pod \"dnsmasq-dns-65c9b8d4f7-n2mvq\" (UID: \"a483d333-16fd-4c48-854f-590fe5f0665c\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-n2mvq" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.722286 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-4l8fb"] Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.738324 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6xwm\" (UniqueName: \"kubernetes.io/projected/a483d333-16fd-4c48-854f-590fe5f0665c-kube-api-access-q6xwm\") pod \"dnsmasq-dns-65c9b8d4f7-n2mvq\" (UID: \"a483d333-16fd-4c48-854f-590fe5f0665c\") " pod="openstack/dnsmasq-dns-65c9b8d4f7-n2mvq" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.750218 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-mjqnd"] Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.759345 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c476d78c5-mjqnd" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.765544 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.774326 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-mjqnd"] Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.799785 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8402005b-c0d2-468d-894e-8fcd7f42ead6-config\") pod \"dnsmasq-dns-5c476d78c5-mjqnd\" (UID: \"8402005b-c0d2-468d-894e-8fcd7f42ead6\") " pod="openstack/dnsmasq-dns-5c476d78c5-mjqnd" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.799853 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8402005b-c0d2-468d-894e-8fcd7f42ead6-dns-svc\") pod \"dnsmasq-dns-5c476d78c5-mjqnd\" (UID: \"8402005b-c0d2-468d-894e-8fcd7f42ead6\") " pod="openstack/dnsmasq-dns-5c476d78c5-mjqnd" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.799917 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8402005b-c0d2-468d-894e-8fcd7f42ead6-ovsdbserver-sb\") pod \"dnsmasq-dns-5c476d78c5-mjqnd\" (UID: \"8402005b-c0d2-468d-894e-8fcd7f42ead6\") " pod="openstack/dnsmasq-dns-5c476d78c5-mjqnd" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.799959 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9j2zc\" (UniqueName: \"kubernetes.io/projected/8402005b-c0d2-468d-894e-8fcd7f42ead6-kube-api-access-9j2zc\") pod \"dnsmasq-dns-5c476d78c5-mjqnd\" (UID: \"8402005b-c0d2-468d-894e-8fcd7f42ead6\") " pod="openstack/dnsmasq-dns-5c476d78c5-mjqnd" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.799998 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-config\") pod \"ovn-controller-metrics-2mts9\" (UID: \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\") " pod="openstack/ovn-controller-metrics-2mts9" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.800057 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xbmh\" (UniqueName: \"kubernetes.io/projected/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-kube-api-access-8xbmh\") pod \"ovn-controller-metrics-2mts9\" (UID: \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\") " pod="openstack/ovn-controller-metrics-2mts9" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.805088 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-config\") pod \"ovn-controller-metrics-2mts9\" (UID: \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\") " pod="openstack/ovn-controller-metrics-2mts9" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.805204 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-ovs-rundir\") pod \"ovn-controller-metrics-2mts9\" (UID: \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\") " pod="openstack/ovn-controller-metrics-2mts9" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.805237 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8402005b-c0d2-468d-894e-8fcd7f42ead6-ovsdbserver-nb\") pod \"dnsmasq-dns-5c476d78c5-mjqnd\" (UID: \"8402005b-c0d2-468d-894e-8fcd7f42ead6\") " pod="openstack/dnsmasq-dns-5c476d78c5-mjqnd" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.805293 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-2mts9\" (UID: \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\") " pod="openstack/ovn-controller-metrics-2mts9" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.805347 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-combined-ca-bundle\") pod \"ovn-controller-metrics-2mts9\" (UID: \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\") " pod="openstack/ovn-controller-metrics-2mts9" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.805399 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-ovn-rundir\") pod \"ovn-controller-metrics-2mts9\" (UID: \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\") " pod="openstack/ovn-controller-metrics-2mts9" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.805751 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-ovn-rundir\") pod \"ovn-controller-metrics-2mts9\" (UID: \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\") " pod="openstack/ovn-controller-metrics-2mts9" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.805811 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-ovs-rundir\") pod \"ovn-controller-metrics-2mts9\" (UID: \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\") " pod="openstack/ovn-controller-metrics-2mts9" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.806052 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65c9b8d4f7-n2mvq" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.809685 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-2mts9\" (UID: \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\") " pod="openstack/ovn-controller-metrics-2mts9" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.810594 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-combined-ca-bundle\") pod \"ovn-controller-metrics-2mts9\" (UID: \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\") " pod="openstack/ovn-controller-metrics-2mts9" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.822710 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xbmh\" (UniqueName: \"kubernetes.io/projected/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-kube-api-access-8xbmh\") pod \"ovn-controller-metrics-2mts9\" (UID: \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\") " pod="openstack/ovn-controller-metrics-2mts9" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.841592 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6d9948dc-swg9r" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.895257 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-2mts9" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.908652 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b732643a-7578-4b49-802e-7f04b5d3793c-config\") pod \"b732643a-7578-4b49-802e-7f04b5d3793c\" (UID: \"b732643a-7578-4b49-802e-7f04b5d3793c\") " Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.908812 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5stb\" (UniqueName: \"kubernetes.io/projected/b732643a-7578-4b49-802e-7f04b5d3793c-kube-api-access-x5stb\") pod \"b732643a-7578-4b49-802e-7f04b5d3793c\" (UID: \"b732643a-7578-4b49-802e-7f04b5d3793c\") " Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.908854 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b732643a-7578-4b49-802e-7f04b5d3793c-dns-svc\") pod \"b732643a-7578-4b49-802e-7f04b5d3793c\" (UID: \"b732643a-7578-4b49-802e-7f04b5d3793c\") " Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.909111 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8402005b-c0d2-468d-894e-8fcd7f42ead6-config\") pod \"dnsmasq-dns-5c476d78c5-mjqnd\" (UID: \"8402005b-c0d2-468d-894e-8fcd7f42ead6\") " pod="openstack/dnsmasq-dns-5c476d78c5-mjqnd" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.909145 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8402005b-c0d2-468d-894e-8fcd7f42ead6-dns-svc\") pod \"dnsmasq-dns-5c476d78c5-mjqnd\" (UID: \"8402005b-c0d2-468d-894e-8fcd7f42ead6\") " pod="openstack/dnsmasq-dns-5c476d78c5-mjqnd" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.909185 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8402005b-c0d2-468d-894e-8fcd7f42ead6-ovsdbserver-sb\") pod \"dnsmasq-dns-5c476d78c5-mjqnd\" (UID: \"8402005b-c0d2-468d-894e-8fcd7f42ead6\") " pod="openstack/dnsmasq-dns-5c476d78c5-mjqnd" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.909206 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9j2zc\" (UniqueName: \"kubernetes.io/projected/8402005b-c0d2-468d-894e-8fcd7f42ead6-kube-api-access-9j2zc\") pod \"dnsmasq-dns-5c476d78c5-mjqnd\" (UID: \"8402005b-c0d2-468d-894e-8fcd7f42ead6\") " pod="openstack/dnsmasq-dns-5c476d78c5-mjqnd" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.909262 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8402005b-c0d2-468d-894e-8fcd7f42ead6-ovsdbserver-nb\") pod \"dnsmasq-dns-5c476d78c5-mjqnd\" (UID: \"8402005b-c0d2-468d-894e-8fcd7f42ead6\") " pod="openstack/dnsmasq-dns-5c476d78c5-mjqnd" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.909912 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b732643a-7578-4b49-802e-7f04b5d3793c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b732643a-7578-4b49-802e-7f04b5d3793c" (UID: "b732643a-7578-4b49-802e-7f04b5d3793c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.910236 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8402005b-c0d2-468d-894e-8fcd7f42ead6-ovsdbserver-nb\") pod \"dnsmasq-dns-5c476d78c5-mjqnd\" (UID: \"8402005b-c0d2-468d-894e-8fcd7f42ead6\") " pod="openstack/dnsmasq-dns-5c476d78c5-mjqnd" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.910331 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8402005b-c0d2-468d-894e-8fcd7f42ead6-dns-svc\") pod \"dnsmasq-dns-5c476d78c5-mjqnd\" (UID: \"8402005b-c0d2-468d-894e-8fcd7f42ead6\") " pod="openstack/dnsmasq-dns-5c476d78c5-mjqnd" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.910769 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8402005b-c0d2-468d-894e-8fcd7f42ead6-config\") pod \"dnsmasq-dns-5c476d78c5-mjqnd\" (UID: \"8402005b-c0d2-468d-894e-8fcd7f42ead6\") " pod="openstack/dnsmasq-dns-5c476d78c5-mjqnd" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.910854 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8402005b-c0d2-468d-894e-8fcd7f42ead6-ovsdbserver-sb\") pod \"dnsmasq-dns-5c476d78c5-mjqnd\" (UID: \"8402005b-c0d2-468d-894e-8fcd7f42ead6\") " pod="openstack/dnsmasq-dns-5c476d78c5-mjqnd" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.914799 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b732643a-7578-4b49-802e-7f04b5d3793c-kube-api-access-x5stb" (OuterVolumeSpecName: "kube-api-access-x5stb") pod "b732643a-7578-4b49-802e-7f04b5d3793c" (UID: "b732643a-7578-4b49-802e-7f04b5d3793c"). InnerVolumeSpecName "kube-api-access-x5stb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.915511 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b732643a-7578-4b49-802e-7f04b5d3793c-config" (OuterVolumeSpecName: "config") pod "b732643a-7578-4b49-802e-7f04b5d3793c" (UID: "b732643a-7578-4b49-802e-7f04b5d3793c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:42:42 crc kubenswrapper[4776]: I1125 09:42:42.940890 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9j2zc\" (UniqueName: \"kubernetes.io/projected/8402005b-c0d2-468d-894e-8fcd7f42ead6-kube-api-access-9j2zc\") pod \"dnsmasq-dns-5c476d78c5-mjqnd\" (UID: \"8402005b-c0d2-468d-894e-8fcd7f42ead6\") " pod="openstack/dnsmasq-dns-5c476d78c5-mjqnd" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.010411 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b732643a-7578-4b49-802e-7f04b5d3793c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.010444 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b732643a-7578-4b49-802e-7f04b5d3793c-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.010454 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5stb\" (UniqueName: \"kubernetes.io/projected/b732643a-7578-4b49-802e-7f04b5d3793c-kube-api-access-x5stb\") on node \"crc\" DevicePath \"\"" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.120059 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.120109 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.149746 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c476d78c5-mjqnd" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.176889 4776 generic.go:334] "Generic (PLEG): container finished" podID="f40cef2d-0d59-4300-9336-fc5a793b9cbf" containerID="9984fa6e44e2ee971eab863a78b789a3cfa5466e0b6b1162697ce0b228d72658" exitCode=0 Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.176969 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6486446b9f-4l8fb" event={"ID":"f40cef2d-0d59-4300-9336-fc5a793b9cbf","Type":"ContainerDied","Data":"9984fa6e44e2ee971eab863a78b789a3cfa5466e0b6b1162697ce0b228d72658"} Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.178876 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c6d9948dc-swg9r" event={"ID":"b732643a-7578-4b49-802e-7f04b5d3793c","Type":"ContainerDied","Data":"e17a9c1f772740c03e90d72e07c3081e4557c7a015e1bc8350618c3c11d4fc9b"} Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.178936 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c6d9948dc-swg9r" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.179521 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.228143 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.233932 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.235638 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-65c9b8d4f7-n2mvq"] Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.332851 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-swg9r"] Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.348204 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c6d9948dc-swg9r"] Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.359355 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-2mts9"] Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.362125 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.547290 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.551326 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.553125 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6486446b9f-4l8fb" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.553966 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.554229 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-dcjxb" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.554240 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.554347 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.567041 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.627157 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\") " pod="openstack/ovn-northd-0" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.627228 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\") " pod="openstack/ovn-northd-0" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.627256 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sn8sd\" (UniqueName: \"kubernetes.io/projected/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-kube-api-access-sn8sd\") pod \"ovn-northd-0\" (UID: \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\") " pod="openstack/ovn-northd-0" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.627292 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-scripts\") pod \"ovn-northd-0\" (UID: \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\") " pod="openstack/ovn-northd-0" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.627331 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\") " pod="openstack/ovn-northd-0" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.627350 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-config\") pod \"ovn-northd-0\" (UID: \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\") " pod="openstack/ovn-northd-0" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.627379 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\") " pod="openstack/ovn-northd-0" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.672181 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b732643a-7578-4b49-802e-7f04b5d3793c" path="/var/lib/kubelet/pods/b732643a-7578-4b49-802e-7f04b5d3793c/volumes" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.728708 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f40cef2d-0d59-4300-9336-fc5a793b9cbf-dns-svc\") pod \"f40cef2d-0d59-4300-9336-fc5a793b9cbf\" (UID: \"f40cef2d-0d59-4300-9336-fc5a793b9cbf\") " Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.728767 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f40cef2d-0d59-4300-9336-fc5a793b9cbf-config\") pod \"f40cef2d-0d59-4300-9336-fc5a793b9cbf\" (UID: \"f40cef2d-0d59-4300-9336-fc5a793b9cbf\") " Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.728804 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhtx4\" (UniqueName: \"kubernetes.io/projected/f40cef2d-0d59-4300-9336-fc5a793b9cbf-kube-api-access-nhtx4\") pod \"f40cef2d-0d59-4300-9336-fc5a793b9cbf\" (UID: \"f40cef2d-0d59-4300-9336-fc5a793b9cbf\") " Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.729115 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\") " pod="openstack/ovn-northd-0" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.729151 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-config\") pod \"ovn-northd-0\" (UID: \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\") " pod="openstack/ovn-northd-0" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.729196 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\") " pod="openstack/ovn-northd-0" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.729224 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\") " pod="openstack/ovn-northd-0" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.729309 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\") " pod="openstack/ovn-northd-0" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.729350 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sn8sd\" (UniqueName: \"kubernetes.io/projected/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-kube-api-access-sn8sd\") pod \"ovn-northd-0\" (UID: \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\") " pod="openstack/ovn-northd-0" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.729406 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-scripts\") pod \"ovn-northd-0\" (UID: \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\") " pod="openstack/ovn-northd-0" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.732650 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\") " pod="openstack/ovn-northd-0" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.736425 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-config\") pod \"ovn-northd-0\" (UID: \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\") " pod="openstack/ovn-northd-0" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.736699 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f40cef2d-0d59-4300-9336-fc5a793b9cbf-kube-api-access-nhtx4" (OuterVolumeSpecName: "kube-api-access-nhtx4") pod "f40cef2d-0d59-4300-9336-fc5a793b9cbf" (UID: "f40cef2d-0d59-4300-9336-fc5a793b9cbf"). InnerVolumeSpecName "kube-api-access-nhtx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.736963 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\") " pod="openstack/ovn-northd-0" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.737021 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\") " pod="openstack/ovn-northd-0" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.737325 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-scripts\") pod \"ovn-northd-0\" (UID: \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\") " pod="openstack/ovn-northd-0" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.737907 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\") " pod="openstack/ovn-northd-0" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.747285 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sn8sd\" (UniqueName: \"kubernetes.io/projected/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-kube-api-access-sn8sd\") pod \"ovn-northd-0\" (UID: \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\") " pod="openstack/ovn-northd-0" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.752957 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f40cef2d-0d59-4300-9336-fc5a793b9cbf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f40cef2d-0d59-4300-9336-fc5a793b9cbf" (UID: "f40cef2d-0d59-4300-9336-fc5a793b9cbf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.764295 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f40cef2d-0d59-4300-9336-fc5a793b9cbf-config" (OuterVolumeSpecName: "config") pod "f40cef2d-0d59-4300-9336-fc5a793b9cbf" (UID: "f40cef2d-0d59-4300-9336-fc5a793b9cbf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.773120 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-mjqnd"] Nov 25 09:42:43 crc kubenswrapper[4776]: W1125 09:42:43.787832 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8402005b_c0d2_468d_894e_8fcd7f42ead6.slice/crio-380205c17f26d56cdfc0c6fdefa1a898515a2e7c353ff35e75874efbc551ac25 WatchSource:0}: Error finding container 380205c17f26d56cdfc0c6fdefa1a898515a2e7c353ff35e75874efbc551ac25: Status 404 returned error can't find the container with id 380205c17f26d56cdfc0c6fdefa1a898515a2e7c353ff35e75874efbc551ac25 Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.835016 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f40cef2d-0d59-4300-9336-fc5a793b9cbf-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.835058 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f40cef2d-0d59-4300-9336-fc5a793b9cbf-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.835084 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhtx4\" (UniqueName: \"kubernetes.io/projected/f40cef2d-0d59-4300-9336-fc5a793b9cbf-kube-api-access-nhtx4\") on node \"crc\" DevicePath \"\"" Nov 25 09:42:43 crc kubenswrapper[4776]: I1125 09:42:43.907519 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.186294 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c476d78c5-mjqnd" event={"ID":"8402005b-c0d2-468d-894e-8fcd7f42ead6","Type":"ContainerStarted","Data":"380205c17f26d56cdfc0c6fdefa1a898515a2e7c353ff35e75874efbc551ac25"} Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.187413 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65c9b8d4f7-n2mvq" event={"ID":"a483d333-16fd-4c48-854f-590fe5f0665c","Type":"ContainerStarted","Data":"dd1955c5ceacd6add7fbb752f8946dba6a2815567d9c2cd0f43b47b7529469d5"} Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.189002 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6486446b9f-4l8fb" event={"ID":"f40cef2d-0d59-4300-9336-fc5a793b9cbf","Type":"ContainerDied","Data":"06f3a6b4b1b8591c301a21c39834f9588b61217f9a63fb21eef31f106ee411d4"} Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.189098 4776 scope.go:117] "RemoveContainer" containerID="9984fa6e44e2ee971eab863a78b789a3cfa5466e0b6b1162697ce0b228d72658" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.189278 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6486446b9f-4l8fb" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.197742 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-2mts9" event={"ID":"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd","Type":"ContainerStarted","Data":"01da0db65f363a5a31d740c6ee7385aca9b9ea4986f9368086fdf9521ab0b01a"} Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.281459 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-4l8fb"] Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.289507 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6486446b9f-4l8fb"] Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.334263 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.588855 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.588912 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.603504 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-b60a-account-create-nf6k7"] Nov 25 09:42:44 crc kubenswrapper[4776]: E1125 09:42:44.603905 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f40cef2d-0d59-4300-9336-fc5a793b9cbf" containerName="init" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.603923 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="f40cef2d-0d59-4300-9336-fc5a793b9cbf" containerName="init" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.604095 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="f40cef2d-0d59-4300-9336-fc5a793b9cbf" containerName="init" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.604652 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b60a-account-create-nf6k7" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.611099 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.630900 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-tzmv5"] Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.631889 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-tzmv5" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.657575 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-tzmv5"] Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.673366 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-b60a-account-create-nf6k7"] Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.685864 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.751245 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1de7f06d-24e7-4114-966b-d62df3366006-operator-scripts\") pod \"keystone-db-create-tzmv5\" (UID: \"1de7f06d-24e7-4114-966b-d62df3366006\") " pod="openstack/keystone-db-create-tzmv5" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.751375 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpwl9\" (UniqueName: \"kubernetes.io/projected/ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e-kube-api-access-jpwl9\") pod \"keystone-b60a-account-create-nf6k7\" (UID: \"ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e\") " pod="openstack/keystone-b60a-account-create-nf6k7" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.751409 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxsms\" (UniqueName: \"kubernetes.io/projected/1de7f06d-24e7-4114-966b-d62df3366006-kube-api-access-kxsms\") pod \"keystone-db-create-tzmv5\" (UID: \"1de7f06d-24e7-4114-966b-d62df3366006\") " pod="openstack/keystone-db-create-tzmv5" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.752047 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e-operator-scripts\") pod \"keystone-b60a-account-create-nf6k7\" (UID: \"ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e\") " pod="openstack/keystone-b60a-account-create-nf6k7" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.800669 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-7ff97"] Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.801726 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-7ff97" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.815232 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-7ff97"] Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.853139 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1de7f06d-24e7-4114-966b-d62df3366006-operator-scripts\") pod \"keystone-db-create-tzmv5\" (UID: \"1de7f06d-24e7-4114-966b-d62df3366006\") " pod="openstack/keystone-db-create-tzmv5" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.853246 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpwl9\" (UniqueName: \"kubernetes.io/projected/ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e-kube-api-access-jpwl9\") pod \"keystone-b60a-account-create-nf6k7\" (UID: \"ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e\") " pod="openstack/keystone-b60a-account-create-nf6k7" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.853277 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxsms\" (UniqueName: \"kubernetes.io/projected/1de7f06d-24e7-4114-966b-d62df3366006-kube-api-access-kxsms\") pod \"keystone-db-create-tzmv5\" (UID: \"1de7f06d-24e7-4114-966b-d62df3366006\") " pod="openstack/keystone-db-create-tzmv5" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.853319 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e-operator-scripts\") pod \"keystone-b60a-account-create-nf6k7\" (UID: \"ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e\") " pod="openstack/keystone-b60a-account-create-nf6k7" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.854207 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1de7f06d-24e7-4114-966b-d62df3366006-operator-scripts\") pod \"keystone-db-create-tzmv5\" (UID: \"1de7f06d-24e7-4114-966b-d62df3366006\") " pod="openstack/keystone-db-create-tzmv5" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.854230 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e-operator-scripts\") pod \"keystone-b60a-account-create-nf6k7\" (UID: \"ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e\") " pod="openstack/keystone-b60a-account-create-nf6k7" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.871253 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxsms\" (UniqueName: \"kubernetes.io/projected/1de7f06d-24e7-4114-966b-d62df3366006-kube-api-access-kxsms\") pod \"keystone-db-create-tzmv5\" (UID: \"1de7f06d-24e7-4114-966b-d62df3366006\") " pod="openstack/keystone-db-create-tzmv5" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.871809 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpwl9\" (UniqueName: \"kubernetes.io/projected/ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e-kube-api-access-jpwl9\") pod \"keystone-b60a-account-create-nf6k7\" (UID: \"ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e\") " pod="openstack/keystone-b60a-account-create-nf6k7" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.913945 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-e176-account-create-5g442"] Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.914964 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-e176-account-create-5g442" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.916906 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.927053 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-e176-account-create-5g442"] Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.927144 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b60a-account-create-nf6k7" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.947643 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-tzmv5" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.954098 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78j5q\" (UniqueName: \"kubernetes.io/projected/f72b0b94-97f1-4bbc-9be4-c8965143cd26-kube-api-access-78j5q\") pod \"placement-db-create-7ff97\" (UID: \"f72b0b94-97f1-4bbc-9be4-c8965143cd26\") " pod="openstack/placement-db-create-7ff97" Nov 25 09:42:44 crc kubenswrapper[4776]: I1125 09:42:44.954164 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f72b0b94-97f1-4bbc-9be4-c8965143cd26-operator-scripts\") pod \"placement-db-create-7ff97\" (UID: \"f72b0b94-97f1-4bbc-9be4-c8965143cd26\") " pod="openstack/placement-db-create-7ff97" Nov 25 09:42:45 crc kubenswrapper[4776]: I1125 09:42:45.056770 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f72b0b94-97f1-4bbc-9be4-c8965143cd26-operator-scripts\") pod \"placement-db-create-7ff97\" (UID: \"f72b0b94-97f1-4bbc-9be4-c8965143cd26\") " pod="openstack/placement-db-create-7ff97" Nov 25 09:42:45 crc kubenswrapper[4776]: I1125 09:42:45.057163 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6466d754-e356-4950-9ce9-8f56e68d6b97-operator-scripts\") pod \"placement-e176-account-create-5g442\" (UID: \"6466d754-e356-4950-9ce9-8f56e68d6b97\") " pod="openstack/placement-e176-account-create-5g442" Nov 25 09:42:45 crc kubenswrapper[4776]: I1125 09:42:45.057215 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9tzk\" (UniqueName: \"kubernetes.io/projected/6466d754-e356-4950-9ce9-8f56e68d6b97-kube-api-access-s9tzk\") pod \"placement-e176-account-create-5g442\" (UID: \"6466d754-e356-4950-9ce9-8f56e68d6b97\") " pod="openstack/placement-e176-account-create-5g442" Nov 25 09:42:45 crc kubenswrapper[4776]: I1125 09:42:45.057262 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78j5q\" (UniqueName: \"kubernetes.io/projected/f72b0b94-97f1-4bbc-9be4-c8965143cd26-kube-api-access-78j5q\") pod \"placement-db-create-7ff97\" (UID: \"f72b0b94-97f1-4bbc-9be4-c8965143cd26\") " pod="openstack/placement-db-create-7ff97" Nov 25 09:42:45 crc kubenswrapper[4776]: I1125 09:42:45.058354 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f72b0b94-97f1-4bbc-9be4-c8965143cd26-operator-scripts\") pod \"placement-db-create-7ff97\" (UID: \"f72b0b94-97f1-4bbc-9be4-c8965143cd26\") " pod="openstack/placement-db-create-7ff97" Nov 25 09:42:45 crc kubenswrapper[4776]: I1125 09:42:45.078987 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78j5q\" (UniqueName: \"kubernetes.io/projected/f72b0b94-97f1-4bbc-9be4-c8965143cd26-kube-api-access-78j5q\") pod \"placement-db-create-7ff97\" (UID: \"f72b0b94-97f1-4bbc-9be4-c8965143cd26\") " pod="openstack/placement-db-create-7ff97" Nov 25 09:42:45 crc kubenswrapper[4776]: I1125 09:42:45.119370 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-7ff97" Nov 25 09:42:45 crc kubenswrapper[4776]: I1125 09:42:45.158535 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6466d754-e356-4950-9ce9-8f56e68d6b97-operator-scripts\") pod \"placement-e176-account-create-5g442\" (UID: \"6466d754-e356-4950-9ce9-8f56e68d6b97\") " pod="openstack/placement-e176-account-create-5g442" Nov 25 09:42:45 crc kubenswrapper[4776]: I1125 09:42:45.158882 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9tzk\" (UniqueName: \"kubernetes.io/projected/6466d754-e356-4950-9ce9-8f56e68d6b97-kube-api-access-s9tzk\") pod \"placement-e176-account-create-5g442\" (UID: \"6466d754-e356-4950-9ce9-8f56e68d6b97\") " pod="openstack/placement-e176-account-create-5g442" Nov 25 09:42:45 crc kubenswrapper[4776]: I1125 09:42:45.164981 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6466d754-e356-4950-9ce9-8f56e68d6b97-operator-scripts\") pod \"placement-e176-account-create-5g442\" (UID: \"6466d754-e356-4950-9ce9-8f56e68d6b97\") " pod="openstack/placement-e176-account-create-5g442" Nov 25 09:42:45 crc kubenswrapper[4776]: I1125 09:42:45.178218 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9tzk\" (UniqueName: \"kubernetes.io/projected/6466d754-e356-4950-9ce9-8f56e68d6b97-kube-api-access-s9tzk\") pod \"placement-e176-account-create-5g442\" (UID: \"6466d754-e356-4950-9ce9-8f56e68d6b97\") " pod="openstack/placement-e176-account-create-5g442" Nov 25 09:42:45 crc kubenswrapper[4776]: I1125 09:42:45.246370 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7dadc68f-0b89-41ba-84f8-e9bcde16ca64","Type":"ContainerStarted","Data":"7dc4a9f6af83ac0f7f0b95466453bfe79cb1509b6af1217c43f42e812f15390f"} Nov 25 09:42:45 crc kubenswrapper[4776]: I1125 09:42:45.250481 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-e176-account-create-5g442" Nov 25 09:42:45 crc kubenswrapper[4776]: I1125 09:42:45.429543 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-tzmv5"] Nov 25 09:42:45 crc kubenswrapper[4776]: I1125 09:42:45.439676 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-b60a-account-create-nf6k7"] Nov 25 09:42:45 crc kubenswrapper[4776]: I1125 09:42:45.442581 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 25 09:42:45 crc kubenswrapper[4776]: I1125 09:42:45.682054 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f40cef2d-0d59-4300-9336-fc5a793b9cbf" path="/var/lib/kubelet/pods/f40cef2d-0d59-4300-9336-fc5a793b9cbf/volumes" Nov 25 09:42:45 crc kubenswrapper[4776]: I1125 09:42:45.758398 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-7ff97"] Nov 25 09:42:45 crc kubenswrapper[4776]: W1125 09:42:45.758482 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf72b0b94_97f1_4bbc_9be4_c8965143cd26.slice/crio-c49668ccfc465a4f3dd4d5c1bf7321c43f6f109f4bb035456c2971bc7546b116 WatchSource:0}: Error finding container c49668ccfc465a4f3dd4d5c1bf7321c43f6f109f4bb035456c2971bc7546b116: Status 404 returned error can't find the container with id c49668ccfc465a4f3dd4d5c1bf7321c43f6f109f4bb035456c2971bc7546b116 Nov 25 09:42:45 crc kubenswrapper[4776]: I1125 09:42:45.869909 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-e176-account-create-5g442"] Nov 25 09:42:46 crc kubenswrapper[4776]: I1125 09:42:46.255142 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-tzmv5" event={"ID":"1de7f06d-24e7-4114-966b-d62df3366006","Type":"ContainerStarted","Data":"0aa1e539ae4e451d9643ac9678b124f3d381c5072c8c681ff8c42257a7300f79"} Nov 25 09:42:46 crc kubenswrapper[4776]: I1125 09:42:46.257196 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-e176-account-create-5g442" event={"ID":"6466d754-e356-4950-9ce9-8f56e68d6b97","Type":"ContainerStarted","Data":"7e0507e78669d3c7a3bb4af4cbf01b661e9de1faf54792143d44658b49a37b9c"} Nov 25 09:42:46 crc kubenswrapper[4776]: I1125 09:42:46.258734 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b60a-account-create-nf6k7" event={"ID":"ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e","Type":"ContainerStarted","Data":"d6d4544fbe24a31ac24119cf9e366a9b5acc886acc1cd72cc8dc360e39fcd546"} Nov 25 09:42:46 crc kubenswrapper[4776]: I1125 09:42:46.259757 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-7ff97" event={"ID":"f72b0b94-97f1-4bbc-9be4-c8965143cd26","Type":"ContainerStarted","Data":"c49668ccfc465a4f3dd4d5c1bf7321c43f6f109f4bb035456c2971bc7546b116"} Nov 25 09:42:46 crc kubenswrapper[4776]: I1125 09:42:46.632196 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-mjqnd"] Nov 25 09:42:46 crc kubenswrapper[4776]: I1125 09:42:46.675334 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9fdb784c-sh56t"] Nov 25 09:42:46 crc kubenswrapper[4776]: I1125 09:42:46.676980 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" Nov 25 09:42:46 crc kubenswrapper[4776]: I1125 09:42:46.689980 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 25 09:42:46 crc kubenswrapper[4776]: I1125 09:42:46.695371 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9fdb784c-sh56t"] Nov 25 09:42:46 crc kubenswrapper[4776]: I1125 09:42:46.790604 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/adadb0a2-c265-4686-98e0-e3d8ff014b35-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9fdb784c-sh56t\" (UID: \"adadb0a2-c265-4686-98e0-e3d8ff014b35\") " pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" Nov 25 09:42:46 crc kubenswrapper[4776]: I1125 09:42:46.790678 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/adadb0a2-c265-4686-98e0-e3d8ff014b35-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9fdb784c-sh56t\" (UID: \"adadb0a2-c265-4686-98e0-e3d8ff014b35\") " pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" Nov 25 09:42:46 crc kubenswrapper[4776]: I1125 09:42:46.790741 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qb7hn\" (UniqueName: \"kubernetes.io/projected/adadb0a2-c265-4686-98e0-e3d8ff014b35-kube-api-access-qb7hn\") pod \"dnsmasq-dns-5c9fdb784c-sh56t\" (UID: \"adadb0a2-c265-4686-98e0-e3d8ff014b35\") " pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" Nov 25 09:42:46 crc kubenswrapper[4776]: I1125 09:42:46.790863 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/adadb0a2-c265-4686-98e0-e3d8ff014b35-dns-svc\") pod \"dnsmasq-dns-5c9fdb784c-sh56t\" (UID: \"adadb0a2-c265-4686-98e0-e3d8ff014b35\") " pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" Nov 25 09:42:46 crc kubenswrapper[4776]: I1125 09:42:46.791039 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adadb0a2-c265-4686-98e0-e3d8ff014b35-config\") pod \"dnsmasq-dns-5c9fdb784c-sh56t\" (UID: \"adadb0a2-c265-4686-98e0-e3d8ff014b35\") " pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" Nov 25 09:42:46 crc kubenswrapper[4776]: I1125 09:42:46.892640 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adadb0a2-c265-4686-98e0-e3d8ff014b35-config\") pod \"dnsmasq-dns-5c9fdb784c-sh56t\" (UID: \"adadb0a2-c265-4686-98e0-e3d8ff014b35\") " pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" Nov 25 09:42:46 crc kubenswrapper[4776]: I1125 09:42:46.892753 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/adadb0a2-c265-4686-98e0-e3d8ff014b35-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9fdb784c-sh56t\" (UID: \"adadb0a2-c265-4686-98e0-e3d8ff014b35\") " pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" Nov 25 09:42:46 crc kubenswrapper[4776]: I1125 09:42:46.892792 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/adadb0a2-c265-4686-98e0-e3d8ff014b35-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9fdb784c-sh56t\" (UID: \"adadb0a2-c265-4686-98e0-e3d8ff014b35\") " pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" Nov 25 09:42:46 crc kubenswrapper[4776]: I1125 09:42:46.892837 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qb7hn\" (UniqueName: \"kubernetes.io/projected/adadb0a2-c265-4686-98e0-e3d8ff014b35-kube-api-access-qb7hn\") pod \"dnsmasq-dns-5c9fdb784c-sh56t\" (UID: \"adadb0a2-c265-4686-98e0-e3d8ff014b35\") " pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" Nov 25 09:42:46 crc kubenswrapper[4776]: I1125 09:42:46.892876 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/adadb0a2-c265-4686-98e0-e3d8ff014b35-dns-svc\") pod \"dnsmasq-dns-5c9fdb784c-sh56t\" (UID: \"adadb0a2-c265-4686-98e0-e3d8ff014b35\") " pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" Nov 25 09:42:46 crc kubenswrapper[4776]: I1125 09:42:46.894085 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/adadb0a2-c265-4686-98e0-e3d8ff014b35-dns-svc\") pod \"dnsmasq-dns-5c9fdb784c-sh56t\" (UID: \"adadb0a2-c265-4686-98e0-e3d8ff014b35\") " pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" Nov 25 09:42:46 crc kubenswrapper[4776]: I1125 09:42:46.894087 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/adadb0a2-c265-4686-98e0-e3d8ff014b35-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9fdb784c-sh56t\" (UID: \"adadb0a2-c265-4686-98e0-e3d8ff014b35\") " pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" Nov 25 09:42:46 crc kubenswrapper[4776]: I1125 09:42:46.894133 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/adadb0a2-c265-4686-98e0-e3d8ff014b35-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9fdb784c-sh56t\" (UID: \"adadb0a2-c265-4686-98e0-e3d8ff014b35\") " pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" Nov 25 09:42:46 crc kubenswrapper[4776]: I1125 09:42:46.894213 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adadb0a2-c265-4686-98e0-e3d8ff014b35-config\") pod \"dnsmasq-dns-5c9fdb784c-sh56t\" (UID: \"adadb0a2-c265-4686-98e0-e3d8ff014b35\") " pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" Nov 25 09:42:46 crc kubenswrapper[4776]: I1125 09:42:46.911260 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qb7hn\" (UniqueName: \"kubernetes.io/projected/adadb0a2-c265-4686-98e0-e3d8ff014b35-kube-api-access-qb7hn\") pod \"dnsmasq-dns-5c9fdb784c-sh56t\" (UID: \"adadb0a2-c265-4686-98e0-e3d8ff014b35\") " pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" Nov 25 09:42:46 crc kubenswrapper[4776]: I1125 09:42:46.998909 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" Nov 25 09:42:47 crc kubenswrapper[4776]: I1125 09:42:47.287831 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9fdb784c-sh56t"] Nov 25 09:42:47 crc kubenswrapper[4776]: W1125 09:42:47.297241 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podadadb0a2_c265_4686_98e0_e3d8ff014b35.slice/crio-03a07fbf5aa81bf4bc6158da4f2dea735f1f8af821c2ed4c661a4fadbe11d491 WatchSource:0}: Error finding container 03a07fbf5aa81bf4bc6158da4f2dea735f1f8af821c2ed4c661a4fadbe11d491: Status 404 returned error can't find the container with id 03a07fbf5aa81bf4bc6158da4f2dea735f1f8af821c2ed4c661a4fadbe11d491 Nov 25 09:42:47 crc kubenswrapper[4776]: I1125 09:42:47.831907 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Nov 25 09:42:47 crc kubenswrapper[4776]: I1125 09:42:47.838869 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 25 09:42:47 crc kubenswrapper[4776]: I1125 09:42:47.840816 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 25 09:42:47 crc kubenswrapper[4776]: I1125 09:42:47.841222 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 25 09:42:47 crc kubenswrapper[4776]: I1125 09:42:47.841763 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-xqlxn" Nov 25 09:42:47 crc kubenswrapper[4776]: I1125 09:42:47.842115 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 25 09:42:47 crc kubenswrapper[4776]: I1125 09:42:47.863353 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 25 09:42:47 crc kubenswrapper[4776]: I1125 09:42:47.909088 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ebd843d5-4a0b-4660-9636-fc19d8fddec6-etc-swift\") pod \"swift-storage-0\" (UID: \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\") " pod="openstack/swift-storage-0" Nov 25 09:42:47 crc kubenswrapper[4776]: I1125 09:42:47.909143 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/ebd843d5-4a0b-4660-9636-fc19d8fddec6-lock\") pod \"swift-storage-0\" (UID: \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\") " pod="openstack/swift-storage-0" Nov 25 09:42:47 crc kubenswrapper[4776]: I1125 09:42:47.909274 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/ebd843d5-4a0b-4660-9636-fc19d8fddec6-cache\") pod \"swift-storage-0\" (UID: \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\") " pod="openstack/swift-storage-0" Nov 25 09:42:47 crc kubenswrapper[4776]: I1125 09:42:47.909313 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\") " pod="openstack/swift-storage-0" Nov 25 09:42:47 crc kubenswrapper[4776]: I1125 09:42:47.909372 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6c49z\" (UniqueName: \"kubernetes.io/projected/ebd843d5-4a0b-4660-9636-fc19d8fddec6-kube-api-access-6c49z\") pod \"swift-storage-0\" (UID: \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\") " pod="openstack/swift-storage-0" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.010729 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/ebd843d5-4a0b-4660-9636-fc19d8fddec6-cache\") pod \"swift-storage-0\" (UID: \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\") " pod="openstack/swift-storage-0" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.010777 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\") " pod="openstack/swift-storage-0" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.010863 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6c49z\" (UniqueName: \"kubernetes.io/projected/ebd843d5-4a0b-4660-9636-fc19d8fddec6-kube-api-access-6c49z\") pod \"swift-storage-0\" (UID: \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\") " pod="openstack/swift-storage-0" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.010890 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ebd843d5-4a0b-4660-9636-fc19d8fddec6-etc-swift\") pod \"swift-storage-0\" (UID: \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\") " pod="openstack/swift-storage-0" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.010914 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/ebd843d5-4a0b-4660-9636-fc19d8fddec6-lock\") pod \"swift-storage-0\" (UID: \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\") " pod="openstack/swift-storage-0" Nov 25 09:42:48 crc kubenswrapper[4776]: E1125 09:42:48.011182 4776 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 25 09:42:48 crc kubenswrapper[4776]: E1125 09:42:48.011209 4776 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.011264 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/swift-storage-0" Nov 25 09:42:48 crc kubenswrapper[4776]: E1125 09:42:48.011289 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ebd843d5-4a0b-4660-9636-fc19d8fddec6-etc-swift podName:ebd843d5-4a0b-4660-9636-fc19d8fddec6 nodeName:}" failed. No retries permitted until 2025-11-25 09:42:48.511269103 +0000 UTC m=+1113.552328656 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/ebd843d5-4a0b-4660-9636-fc19d8fddec6-etc-swift") pod "swift-storage-0" (UID: "ebd843d5-4a0b-4660-9636-fc19d8fddec6") : configmap "swift-ring-files" not found Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.011415 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/ebd843d5-4a0b-4660-9636-fc19d8fddec6-lock\") pod \"swift-storage-0\" (UID: \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\") " pod="openstack/swift-storage-0" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.011817 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/ebd843d5-4a0b-4660-9636-fc19d8fddec6-cache\") pod \"swift-storage-0\" (UID: \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\") " pod="openstack/swift-storage-0" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.030400 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6c49z\" (UniqueName: \"kubernetes.io/projected/ebd843d5-4a0b-4660-9636-fc19d8fddec6-kube-api-access-6c49z\") pod \"swift-storage-0\" (UID: \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\") " pod="openstack/swift-storage-0" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.034616 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\") " pod="openstack/swift-storage-0" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.283974 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c476d78c5-mjqnd" event={"ID":"8402005b-c0d2-468d-894e-8fcd7f42ead6","Type":"ContainerStarted","Data":"4562f38f4f9bc76cf767127b343ee808ac3da4f3bb78368c48cc58882b9c4dca"} Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.285323 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" event={"ID":"adadb0a2-c265-4686-98e0-e3d8ff014b35","Type":"ContainerStarted","Data":"03a07fbf5aa81bf4bc6158da4f2dea735f1f8af821c2ed4c661a4fadbe11d491"} Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.444573 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-4htz7"] Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.445696 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-4htz7" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.447645 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.447731 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.448802 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.455439 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-4htz7"] Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.518736 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ebd843d5-4a0b-4660-9636-fc19d8fddec6-etc-swift\") pod \"swift-storage-0\" (UID: \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\") " pod="openstack/swift-storage-0" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.518796 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4kll\" (UniqueName: \"kubernetes.io/projected/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-kube-api-access-m4kll\") pod \"swift-ring-rebalance-4htz7\" (UID: \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\") " pod="openstack/swift-ring-rebalance-4htz7" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.518861 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-ring-data-devices\") pod \"swift-ring-rebalance-4htz7\" (UID: \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\") " pod="openstack/swift-ring-rebalance-4htz7" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.518909 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-scripts\") pod \"swift-ring-rebalance-4htz7\" (UID: \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\") " pod="openstack/swift-ring-rebalance-4htz7" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.518974 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-dispersionconf\") pod \"swift-ring-rebalance-4htz7\" (UID: \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\") " pod="openstack/swift-ring-rebalance-4htz7" Nov 25 09:42:48 crc kubenswrapper[4776]: E1125 09:42:48.518926 4776 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 25 09:42:48 crc kubenswrapper[4776]: E1125 09:42:48.519016 4776 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.518994 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-etc-swift\") pod \"swift-ring-rebalance-4htz7\" (UID: \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\") " pod="openstack/swift-ring-rebalance-4htz7" Nov 25 09:42:48 crc kubenswrapper[4776]: E1125 09:42:48.519118 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ebd843d5-4a0b-4660-9636-fc19d8fddec6-etc-swift podName:ebd843d5-4a0b-4660-9636-fc19d8fddec6 nodeName:}" failed. No retries permitted until 2025-11-25 09:42:49.519057948 +0000 UTC m=+1114.560117501 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/ebd843d5-4a0b-4660-9636-fc19d8fddec6-etc-swift") pod "swift-storage-0" (UID: "ebd843d5-4a0b-4660-9636-fc19d8fddec6") : configmap "swift-ring-files" not found Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.519146 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-swiftconf\") pod \"swift-ring-rebalance-4htz7\" (UID: \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\") " pod="openstack/swift-ring-rebalance-4htz7" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.519221 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-combined-ca-bundle\") pod \"swift-ring-rebalance-4htz7\" (UID: \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\") " pod="openstack/swift-ring-rebalance-4htz7" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.620470 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-ring-data-devices\") pod \"swift-ring-rebalance-4htz7\" (UID: \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\") " pod="openstack/swift-ring-rebalance-4htz7" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.620546 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-scripts\") pod \"swift-ring-rebalance-4htz7\" (UID: \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\") " pod="openstack/swift-ring-rebalance-4htz7" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.620562 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-dispersionconf\") pod \"swift-ring-rebalance-4htz7\" (UID: \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\") " pod="openstack/swift-ring-rebalance-4htz7" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.620577 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-etc-swift\") pod \"swift-ring-rebalance-4htz7\" (UID: \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\") " pod="openstack/swift-ring-rebalance-4htz7" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.620613 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-swiftconf\") pod \"swift-ring-rebalance-4htz7\" (UID: \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\") " pod="openstack/swift-ring-rebalance-4htz7" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.620637 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-combined-ca-bundle\") pod \"swift-ring-rebalance-4htz7\" (UID: \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\") " pod="openstack/swift-ring-rebalance-4htz7" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.620690 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4kll\" (UniqueName: \"kubernetes.io/projected/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-kube-api-access-m4kll\") pod \"swift-ring-rebalance-4htz7\" (UID: \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\") " pod="openstack/swift-ring-rebalance-4htz7" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.621443 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-ring-data-devices\") pod \"swift-ring-rebalance-4htz7\" (UID: \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\") " pod="openstack/swift-ring-rebalance-4htz7" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.621457 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-scripts\") pod \"swift-ring-rebalance-4htz7\" (UID: \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\") " pod="openstack/swift-ring-rebalance-4htz7" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.621628 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-etc-swift\") pod \"swift-ring-rebalance-4htz7\" (UID: \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\") " pod="openstack/swift-ring-rebalance-4htz7" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.624737 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-combined-ca-bundle\") pod \"swift-ring-rebalance-4htz7\" (UID: \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\") " pod="openstack/swift-ring-rebalance-4htz7" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.625235 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-dispersionconf\") pod \"swift-ring-rebalance-4htz7\" (UID: \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\") " pod="openstack/swift-ring-rebalance-4htz7" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.625276 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-swiftconf\") pod \"swift-ring-rebalance-4htz7\" (UID: \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\") " pod="openstack/swift-ring-rebalance-4htz7" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.639486 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4kll\" (UniqueName: \"kubernetes.io/projected/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-kube-api-access-m4kll\") pod \"swift-ring-rebalance-4htz7\" (UID: \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\") " pod="openstack/swift-ring-rebalance-4htz7" Nov 25 09:42:48 crc kubenswrapper[4776]: I1125 09:42:48.781726 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-4htz7" Nov 25 09:42:49 crc kubenswrapper[4776]: I1125 09:42:49.284723 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-4htz7"] Nov 25 09:42:49 crc kubenswrapper[4776]: I1125 09:42:49.295345 4776 generic.go:334] "Generic (PLEG): container finished" podID="8402005b-c0d2-468d-894e-8fcd7f42ead6" containerID="4562f38f4f9bc76cf767127b343ee808ac3da4f3bb78368c48cc58882b9c4dca" exitCode=0 Nov 25 09:42:49 crc kubenswrapper[4776]: I1125 09:42:49.295399 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c476d78c5-mjqnd" event={"ID":"8402005b-c0d2-468d-894e-8fcd7f42ead6","Type":"ContainerDied","Data":"4562f38f4f9bc76cf767127b343ee808ac3da4f3bb78368c48cc58882b9c4dca"} Nov 25 09:42:49 crc kubenswrapper[4776]: I1125 09:42:49.542040 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ebd843d5-4a0b-4660-9636-fc19d8fddec6-etc-swift\") pod \"swift-storage-0\" (UID: \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\") " pod="openstack/swift-storage-0" Nov 25 09:42:49 crc kubenswrapper[4776]: E1125 09:42:49.542283 4776 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 25 09:42:49 crc kubenswrapper[4776]: E1125 09:42:49.542527 4776 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 25 09:42:49 crc kubenswrapper[4776]: E1125 09:42:49.542590 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ebd843d5-4a0b-4660-9636-fc19d8fddec6-etc-swift podName:ebd843d5-4a0b-4660-9636-fc19d8fddec6 nodeName:}" failed. No retries permitted until 2025-11-25 09:42:51.542571218 +0000 UTC m=+1116.583630771 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/ebd843d5-4a0b-4660-9636-fc19d8fddec6-etc-swift") pod "swift-storage-0" (UID: "ebd843d5-4a0b-4660-9636-fc19d8fddec6") : configmap "swift-ring-files" not found Nov 25 09:42:49 crc kubenswrapper[4776]: I1125 09:42:49.559373 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c476d78c5-mjqnd" Nov 25 09:42:49 crc kubenswrapper[4776]: I1125 09:42:49.643581 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8402005b-c0d2-468d-894e-8fcd7f42ead6-dns-svc\") pod \"8402005b-c0d2-468d-894e-8fcd7f42ead6\" (UID: \"8402005b-c0d2-468d-894e-8fcd7f42ead6\") " Nov 25 09:42:49 crc kubenswrapper[4776]: I1125 09:42:49.643708 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8402005b-c0d2-468d-894e-8fcd7f42ead6-config\") pod \"8402005b-c0d2-468d-894e-8fcd7f42ead6\" (UID: \"8402005b-c0d2-468d-894e-8fcd7f42ead6\") " Nov 25 09:42:49 crc kubenswrapper[4776]: I1125 09:42:49.643766 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8402005b-c0d2-468d-894e-8fcd7f42ead6-ovsdbserver-sb\") pod \"8402005b-c0d2-468d-894e-8fcd7f42ead6\" (UID: \"8402005b-c0d2-468d-894e-8fcd7f42ead6\") " Nov 25 09:42:49 crc kubenswrapper[4776]: I1125 09:42:49.643801 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8402005b-c0d2-468d-894e-8fcd7f42ead6-ovsdbserver-nb\") pod \"8402005b-c0d2-468d-894e-8fcd7f42ead6\" (UID: \"8402005b-c0d2-468d-894e-8fcd7f42ead6\") " Nov 25 09:42:49 crc kubenswrapper[4776]: I1125 09:42:49.643913 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9j2zc\" (UniqueName: \"kubernetes.io/projected/8402005b-c0d2-468d-894e-8fcd7f42ead6-kube-api-access-9j2zc\") pod \"8402005b-c0d2-468d-894e-8fcd7f42ead6\" (UID: \"8402005b-c0d2-468d-894e-8fcd7f42ead6\") " Nov 25 09:42:49 crc kubenswrapper[4776]: I1125 09:42:49.652714 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8402005b-c0d2-468d-894e-8fcd7f42ead6-kube-api-access-9j2zc" (OuterVolumeSpecName: "kube-api-access-9j2zc") pod "8402005b-c0d2-468d-894e-8fcd7f42ead6" (UID: "8402005b-c0d2-468d-894e-8fcd7f42ead6"). InnerVolumeSpecName "kube-api-access-9j2zc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:42:49 crc kubenswrapper[4776]: I1125 09:42:49.666640 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8402005b-c0d2-468d-894e-8fcd7f42ead6-config" (OuterVolumeSpecName: "config") pod "8402005b-c0d2-468d-894e-8fcd7f42ead6" (UID: "8402005b-c0d2-468d-894e-8fcd7f42ead6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:42:49 crc kubenswrapper[4776]: I1125 09:42:49.669089 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8402005b-c0d2-468d-894e-8fcd7f42ead6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8402005b-c0d2-468d-894e-8fcd7f42ead6" (UID: "8402005b-c0d2-468d-894e-8fcd7f42ead6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:42:49 crc kubenswrapper[4776]: I1125 09:42:49.676764 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8402005b-c0d2-468d-894e-8fcd7f42ead6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8402005b-c0d2-468d-894e-8fcd7f42ead6" (UID: "8402005b-c0d2-468d-894e-8fcd7f42ead6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:42:49 crc kubenswrapper[4776]: I1125 09:42:49.679678 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8402005b-c0d2-468d-894e-8fcd7f42ead6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8402005b-c0d2-468d-894e-8fcd7f42ead6" (UID: "8402005b-c0d2-468d-894e-8fcd7f42ead6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:42:49 crc kubenswrapper[4776]: I1125 09:42:49.746771 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8402005b-c0d2-468d-894e-8fcd7f42ead6-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:42:49 crc kubenswrapper[4776]: I1125 09:42:49.746832 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8402005b-c0d2-468d-894e-8fcd7f42ead6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 09:42:49 crc kubenswrapper[4776]: I1125 09:42:49.746861 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8402005b-c0d2-468d-894e-8fcd7f42ead6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 09:42:49 crc kubenswrapper[4776]: I1125 09:42:49.746889 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9j2zc\" (UniqueName: \"kubernetes.io/projected/8402005b-c0d2-468d-894e-8fcd7f42ead6-kube-api-access-9j2zc\") on node \"crc\" DevicePath \"\"" Nov 25 09:42:49 crc kubenswrapper[4776]: I1125 09:42:49.746912 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8402005b-c0d2-468d-894e-8fcd7f42ead6-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.115698 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-5g4ql"] Nov 25 09:42:50 crc kubenswrapper[4776]: E1125 09:42:50.116684 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8402005b-c0d2-468d-894e-8fcd7f42ead6" containerName="init" Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.116709 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="8402005b-c0d2-468d-894e-8fcd7f42ead6" containerName="init" Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.116934 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="8402005b-c0d2-468d-894e-8fcd7f42ead6" containerName="init" Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.117490 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5g4ql" Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.123830 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-5g4ql"] Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.154046 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8kd6\" (UniqueName: \"kubernetes.io/projected/417adc1f-2f5c-4446-8cf3-994edf3b5e55-kube-api-access-m8kd6\") pod \"glance-db-create-5g4ql\" (UID: \"417adc1f-2f5c-4446-8cf3-994edf3b5e55\") " pod="openstack/glance-db-create-5g4ql" Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.154223 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/417adc1f-2f5c-4446-8cf3-994edf3b5e55-operator-scripts\") pod \"glance-db-create-5g4ql\" (UID: \"417adc1f-2f5c-4446-8cf3-994edf3b5e55\") " pod="openstack/glance-db-create-5g4ql" Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.207847 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-b9b5-account-create-ldfzn"] Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.209389 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b9b5-account-create-ldfzn" Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.211151 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.216360 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-b9b5-account-create-ldfzn"] Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.256086 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/417adc1f-2f5c-4446-8cf3-994edf3b5e55-operator-scripts\") pod \"glance-db-create-5g4ql\" (UID: \"417adc1f-2f5c-4446-8cf3-994edf3b5e55\") " pod="openstack/glance-db-create-5g4ql" Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.256153 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3-operator-scripts\") pod \"glance-b9b5-account-create-ldfzn\" (UID: \"00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3\") " pod="openstack/glance-b9b5-account-create-ldfzn" Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.256195 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8kd6\" (UniqueName: \"kubernetes.io/projected/417adc1f-2f5c-4446-8cf3-994edf3b5e55-kube-api-access-m8kd6\") pod \"glance-db-create-5g4ql\" (UID: \"417adc1f-2f5c-4446-8cf3-994edf3b5e55\") " pod="openstack/glance-db-create-5g4ql" Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.256662 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd2zx\" (UniqueName: \"kubernetes.io/projected/00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3-kube-api-access-qd2zx\") pod \"glance-b9b5-account-create-ldfzn\" (UID: \"00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3\") " pod="openstack/glance-b9b5-account-create-ldfzn" Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.256815 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/417adc1f-2f5c-4446-8cf3-994edf3b5e55-operator-scripts\") pod \"glance-db-create-5g4ql\" (UID: \"417adc1f-2f5c-4446-8cf3-994edf3b5e55\") " pod="openstack/glance-db-create-5g4ql" Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.275055 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8kd6\" (UniqueName: \"kubernetes.io/projected/417adc1f-2f5c-4446-8cf3-994edf3b5e55-kube-api-access-m8kd6\") pod \"glance-db-create-5g4ql\" (UID: \"417adc1f-2f5c-4446-8cf3-994edf3b5e55\") " pod="openstack/glance-db-create-5g4ql" Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.303197 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c476d78c5-mjqnd" Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.303623 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c476d78c5-mjqnd" event={"ID":"8402005b-c0d2-468d-894e-8fcd7f42ead6","Type":"ContainerDied","Data":"380205c17f26d56cdfc0c6fdefa1a898515a2e7c353ff35e75874efbc551ac25"} Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.303784 4776 scope.go:117] "RemoveContainer" containerID="4562f38f4f9bc76cf767127b343ee808ac3da4f3bb78368c48cc58882b9c4dca" Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.308967 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-2mts9" event={"ID":"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd","Type":"ContainerStarted","Data":"29696e80dd0a6c71a81e7417db086e3acded64a814a8e862ba2fc208da55d802"} Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.311021 4776 generic.go:334] "Generic (PLEG): container finished" podID="adadb0a2-c265-4686-98e0-e3d8ff014b35" containerID="5914b48ef0c402081ad751368900f9beae2f9a0429c64a51b0f62898b64cf053" exitCode=0 Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.311108 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" event={"ID":"adadb0a2-c265-4686-98e0-e3d8ff014b35","Type":"ContainerDied","Data":"5914b48ef0c402081ad751368900f9beae2f9a0429c64a51b0f62898b64cf053"} Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.319290 4776 generic.go:334] "Generic (PLEG): container finished" podID="1de7f06d-24e7-4114-966b-d62df3366006" containerID="1ee992a9feeeafbc3dad19eff013d51a325f2d13b85873bd8bf629934e586f03" exitCode=0 Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.319412 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-tzmv5" event={"ID":"1de7f06d-24e7-4114-966b-d62df3366006","Type":"ContainerDied","Data":"1ee992a9feeeafbc3dad19eff013d51a325f2d13b85873bd8bf629934e586f03"} Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.322878 4776 generic.go:334] "Generic (PLEG): container finished" podID="6466d754-e356-4950-9ce9-8f56e68d6b97" containerID="99459662f37648298f22a0f72639de05480bb2132930a4572102745915daeb31" exitCode=0 Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.323102 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-e176-account-create-5g442" event={"ID":"6466d754-e356-4950-9ce9-8f56e68d6b97","Type":"ContainerDied","Data":"99459662f37648298f22a0f72639de05480bb2132930a4572102745915daeb31"} Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.329548 4776 generic.go:334] "Generic (PLEG): container finished" podID="ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e" containerID="f87ecf56de084525eec7915cd66a47ade4adc2c2476c6ecb8bae1e7236795cc6" exitCode=0 Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.329714 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b60a-account-create-nf6k7" event={"ID":"ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e","Type":"ContainerDied","Data":"f87ecf56de084525eec7915cd66a47ade4adc2c2476c6ecb8bae1e7236795cc6"} Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.334052 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-2mts9" podStartSLOduration=8.334019338 podStartE2EDuration="8.334019338s" podCreationTimestamp="2025-11-25 09:42:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:42:50.324491799 +0000 UTC m=+1115.365551352" watchObservedRunningTime="2025-11-25 09:42:50.334019338 +0000 UTC m=+1115.375078901" Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.335846 4776 generic.go:334] "Generic (PLEG): container finished" podID="f72b0b94-97f1-4bbc-9be4-c8965143cd26" containerID="3114febb3411856186b21a212016a8980b4f4ed26617cd0c55aadc1ea4fb8615" exitCode=0 Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.336041 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-7ff97" event={"ID":"f72b0b94-97f1-4bbc-9be4-c8965143cd26","Type":"ContainerDied","Data":"3114febb3411856186b21a212016a8980b4f4ed26617cd0c55aadc1ea4fb8615"} Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.349410 4776 generic.go:334] "Generic (PLEG): container finished" podID="a483d333-16fd-4c48-854f-590fe5f0665c" containerID="a69925572190a55956bc810bf3ddd277c1f7ac89b8e51b395d26a92e98b05dde" exitCode=0 Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.350371 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65c9b8d4f7-n2mvq" event={"ID":"a483d333-16fd-4c48-854f-590fe5f0665c","Type":"ContainerDied","Data":"a69925572190a55956bc810bf3ddd277c1f7ac89b8e51b395d26a92e98b05dde"} Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.356817 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-4htz7" event={"ID":"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515","Type":"ContainerStarted","Data":"50acb690971c2fa83afb22d2985d07b82184bcd8c9b729695c2600e3509a606d"} Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.358466 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd2zx\" (UniqueName: \"kubernetes.io/projected/00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3-kube-api-access-qd2zx\") pod \"glance-b9b5-account-create-ldfzn\" (UID: \"00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3\") " pod="openstack/glance-b9b5-account-create-ldfzn" Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.358609 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3-operator-scripts\") pod \"glance-b9b5-account-create-ldfzn\" (UID: \"00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3\") " pod="openstack/glance-b9b5-account-create-ldfzn" Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.364058 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3-operator-scripts\") pod \"glance-b9b5-account-create-ldfzn\" (UID: \"00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3\") " pod="openstack/glance-b9b5-account-create-ldfzn" Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.384220 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd2zx\" (UniqueName: \"kubernetes.io/projected/00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3-kube-api-access-qd2zx\") pod \"glance-b9b5-account-create-ldfzn\" (UID: \"00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3\") " pod="openstack/glance-b9b5-account-create-ldfzn" Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.440326 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5g4ql" Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.612579 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b9b5-account-create-ldfzn" Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.638507 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-mjqnd"] Nov 25 09:42:50 crc kubenswrapper[4776]: I1125 09:42:50.648447 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c476d78c5-mjqnd"] Nov 25 09:42:51 crc kubenswrapper[4776]: I1125 09:42:51.578394 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ebd843d5-4a0b-4660-9636-fc19d8fddec6-etc-swift\") pod \"swift-storage-0\" (UID: \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\") " pod="openstack/swift-storage-0" Nov 25 09:42:51 crc kubenswrapper[4776]: E1125 09:42:51.578626 4776 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 25 09:42:51 crc kubenswrapper[4776]: E1125 09:42:51.579040 4776 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 25 09:42:51 crc kubenswrapper[4776]: E1125 09:42:51.579110 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ebd843d5-4a0b-4660-9636-fc19d8fddec6-etc-swift podName:ebd843d5-4a0b-4660-9636-fc19d8fddec6 nodeName:}" failed. No retries permitted until 2025-11-25 09:42:55.579093365 +0000 UTC m=+1120.620152918 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/ebd843d5-4a0b-4660-9636-fc19d8fddec6-etc-swift") pod "swift-storage-0" (UID: "ebd843d5-4a0b-4660-9636-fc19d8fddec6") : configmap "swift-ring-files" not found Nov 25 09:42:51 crc kubenswrapper[4776]: I1125 09:42:51.685507 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8402005b-c0d2-468d-894e-8fcd7f42ead6" path="/var/lib/kubelet/pods/8402005b-c0d2-468d-894e-8fcd7f42ead6/volumes" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.154545 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-7ff97" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.160445 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-e176-account-create-5g442" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.176363 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b60a-account-create-nf6k7" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.188354 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-tzmv5" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.199410 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6466d754-e356-4950-9ce9-8f56e68d6b97-operator-scripts\") pod \"6466d754-e356-4950-9ce9-8f56e68d6b97\" (UID: \"6466d754-e356-4950-9ce9-8f56e68d6b97\") " Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.199531 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpwl9\" (UniqueName: \"kubernetes.io/projected/ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e-kube-api-access-jpwl9\") pod \"ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e\" (UID: \"ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e\") " Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.199573 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e-operator-scripts\") pod \"ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e\" (UID: \"ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e\") " Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.199612 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78j5q\" (UniqueName: \"kubernetes.io/projected/f72b0b94-97f1-4bbc-9be4-c8965143cd26-kube-api-access-78j5q\") pod \"f72b0b94-97f1-4bbc-9be4-c8965143cd26\" (UID: \"f72b0b94-97f1-4bbc-9be4-c8965143cd26\") " Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.199633 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9tzk\" (UniqueName: \"kubernetes.io/projected/6466d754-e356-4950-9ce9-8f56e68d6b97-kube-api-access-s9tzk\") pod \"6466d754-e356-4950-9ce9-8f56e68d6b97\" (UID: \"6466d754-e356-4950-9ce9-8f56e68d6b97\") " Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.199658 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f72b0b94-97f1-4bbc-9be4-c8965143cd26-operator-scripts\") pod \"f72b0b94-97f1-4bbc-9be4-c8965143cd26\" (UID: \"f72b0b94-97f1-4bbc-9be4-c8965143cd26\") " Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.201223 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e" (UID: "ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.201354 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f72b0b94-97f1-4bbc-9be4-c8965143cd26-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f72b0b94-97f1-4bbc-9be4-c8965143cd26" (UID: "f72b0b94-97f1-4bbc-9be4-c8965143cd26"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.201934 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6466d754-e356-4950-9ce9-8f56e68d6b97-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6466d754-e356-4950-9ce9-8f56e68d6b97" (UID: "6466d754-e356-4950-9ce9-8f56e68d6b97"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.207106 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6466d754-e356-4950-9ce9-8f56e68d6b97-kube-api-access-s9tzk" (OuterVolumeSpecName: "kube-api-access-s9tzk") pod "6466d754-e356-4950-9ce9-8f56e68d6b97" (UID: "6466d754-e356-4950-9ce9-8f56e68d6b97"). InnerVolumeSpecName "kube-api-access-s9tzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.208022 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e-kube-api-access-jpwl9" (OuterVolumeSpecName: "kube-api-access-jpwl9") pod "ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e" (UID: "ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e"). InnerVolumeSpecName "kube-api-access-jpwl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.211627 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f72b0b94-97f1-4bbc-9be4-c8965143cd26-kube-api-access-78j5q" (OuterVolumeSpecName: "kube-api-access-78j5q") pod "f72b0b94-97f1-4bbc-9be4-c8965143cd26" (UID: "f72b0b94-97f1-4bbc-9be4-c8965143cd26"). InnerVolumeSpecName "kube-api-access-78j5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.301541 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxsms\" (UniqueName: \"kubernetes.io/projected/1de7f06d-24e7-4114-966b-d62df3366006-kube-api-access-kxsms\") pod \"1de7f06d-24e7-4114-966b-d62df3366006\" (UID: \"1de7f06d-24e7-4114-966b-d62df3366006\") " Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.301640 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1de7f06d-24e7-4114-966b-d62df3366006-operator-scripts\") pod \"1de7f06d-24e7-4114-966b-d62df3366006\" (UID: \"1de7f06d-24e7-4114-966b-d62df3366006\") " Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.302301 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6466d754-e356-4950-9ce9-8f56e68d6b97-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.302345 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpwl9\" (UniqueName: \"kubernetes.io/projected/ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e-kube-api-access-jpwl9\") on node \"crc\" DevicePath \"\"" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.302362 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.302377 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78j5q\" (UniqueName: \"kubernetes.io/projected/f72b0b94-97f1-4bbc-9be4-c8965143cd26-kube-api-access-78j5q\") on node \"crc\" DevicePath \"\"" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.302391 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9tzk\" (UniqueName: \"kubernetes.io/projected/6466d754-e356-4950-9ce9-8f56e68d6b97-kube-api-access-s9tzk\") on node \"crc\" DevicePath \"\"" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.302427 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f72b0b94-97f1-4bbc-9be4-c8965143cd26-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.302806 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1de7f06d-24e7-4114-966b-d62df3366006-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1de7f06d-24e7-4114-966b-d62df3366006" (UID: "1de7f06d-24e7-4114-966b-d62df3366006"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.305430 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1de7f06d-24e7-4114-966b-d62df3366006-kube-api-access-kxsms" (OuterVolumeSpecName: "kube-api-access-kxsms") pod "1de7f06d-24e7-4114-966b-d62df3366006" (UID: "1de7f06d-24e7-4114-966b-d62df3366006"). InnerVolumeSpecName "kube-api-access-kxsms". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.379028 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-e176-account-create-5g442" event={"ID":"6466d754-e356-4950-9ce9-8f56e68d6b97","Type":"ContainerDied","Data":"7e0507e78669d3c7a3bb4af4cbf01b661e9de1faf54792143d44658b49a37b9c"} Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.379094 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e0507e78669d3c7a3bb4af4cbf01b661e9de1faf54792143d44658b49a37b9c" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.379166 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-e176-account-create-5g442" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.383643 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b60a-account-create-nf6k7" event={"ID":"ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e","Type":"ContainerDied","Data":"d6d4544fbe24a31ac24119cf9e366a9b5acc886acc1cd72cc8dc360e39fcd546"} Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.383680 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6d4544fbe24a31ac24119cf9e366a9b5acc886acc1cd72cc8dc360e39fcd546" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.383731 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b60a-account-create-nf6k7" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.386710 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-7ff97" event={"ID":"f72b0b94-97f1-4bbc-9be4-c8965143cd26","Type":"ContainerDied","Data":"c49668ccfc465a4f3dd4d5c1bf7321c43f6f109f4bb035456c2971bc7546b116"} Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.386732 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c49668ccfc465a4f3dd4d5c1bf7321c43f6f109f4bb035456c2971bc7546b116" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.386800 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-7ff97" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.394944 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-tzmv5" event={"ID":"1de7f06d-24e7-4114-966b-d62df3366006","Type":"ContainerDied","Data":"0aa1e539ae4e451d9643ac9678b124f3d381c5072c8c681ff8c42257a7300f79"} Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.394975 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0aa1e539ae4e451d9643ac9678b124f3d381c5072c8c681ff8c42257a7300f79" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.394992 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-tzmv5" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.404559 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxsms\" (UniqueName: \"kubernetes.io/projected/1de7f06d-24e7-4114-966b-d62df3366006-kube-api-access-kxsms\") on node \"crc\" DevicePath \"\"" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.404579 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1de7f06d-24e7-4114-966b-d62df3366006-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.476700 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-5g4ql"] Nov 25 09:42:52 crc kubenswrapper[4776]: I1125 09:42:52.580906 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-b9b5-account-create-ldfzn"] Nov 25 09:42:53 crc kubenswrapper[4776]: I1125 09:42:53.404301 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-5g4ql" event={"ID":"417adc1f-2f5c-4446-8cf3-994edf3b5e55","Type":"ContainerStarted","Data":"e6c466809d6ca7435c63046c353094af4ecd2a78a88fbbd6c8f05180a4d2c268"} Nov 25 09:42:53 crc kubenswrapper[4776]: I1125 09:42:53.405674 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b9b5-account-create-ldfzn" event={"ID":"00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3","Type":"ContainerStarted","Data":"e77e7c401a0a6fef9f6d18d6325001e1da4b652050313289924dcffcaf2a3531"} Nov 25 09:42:53 crc kubenswrapper[4776]: E1125 09:42:53.658581 4776 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Nov 25 09:42:53 crc kubenswrapper[4776]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/a483d333-16fd-4c48-854f-590fe5f0665c/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 25 09:42:53 crc kubenswrapper[4776]: > podSandboxID="dd1955c5ceacd6add7fbb752f8946dba6a2815567d9c2cd0f43b47b7529469d5" Nov 25 09:42:53 crc kubenswrapper[4776]: E1125 09:42:53.658726 4776 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 25 09:42:53 crc kubenswrapper[4776]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:18f8463fe46fe6081d5682009e92bbcb3df33282b83b0a2857abaece795cf1ba,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n564h5c8h5fbh5ddh5c5h666hbch5f5h66fh68fh87hfdh699h84hcdh589h64dh76h9ch5cfh5f8h56bh89h67h5fh56bhf6h654h556hch9dh657q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-sb,SubPath:ovsdbserver-sb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-q6xwm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-65c9b8d4f7-n2mvq_openstack(a483d333-16fd-4c48-854f-590fe5f0665c): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/a483d333-16fd-4c48-854f-590fe5f0665c/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 25 09:42:53 crc kubenswrapper[4776]: > logger="UnhandledError" Nov 25 09:42:53 crc kubenswrapper[4776]: E1125 09:42:53.660319 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/a483d333-16fd-4c48-854f-590fe5f0665c/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-65c9b8d4f7-n2mvq" podUID="a483d333-16fd-4c48-854f-590fe5f0665c" Nov 25 09:42:54 crc kubenswrapper[4776]: I1125 09:42:54.420017 4776 generic.go:334] "Generic (PLEG): container finished" podID="00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3" containerID="f098fe0a1fe35cd7b0310e84ef175144d5d65e426c5ec2e06ed70195987f5c40" exitCode=0 Nov 25 09:42:54 crc kubenswrapper[4776]: I1125 09:42:54.420163 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b9b5-account-create-ldfzn" event={"ID":"00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3","Type":"ContainerDied","Data":"f098fe0a1fe35cd7b0310e84ef175144d5d65e426c5ec2e06ed70195987f5c40"} Nov 25 09:42:54 crc kubenswrapper[4776]: I1125 09:42:54.422244 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-4htz7" event={"ID":"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515","Type":"ContainerStarted","Data":"0013160d2c55a26010d58e0144d8a89639c5b92e40762472567d0dae6bd635e9"} Nov 25 09:42:54 crc kubenswrapper[4776]: I1125 09:42:54.425410 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" event={"ID":"adadb0a2-c265-4686-98e0-e3d8ff014b35","Type":"ContainerStarted","Data":"b625f8f354e8416dd5bba6f485f43a6b0b7956f3d528292b946646a11591847a"} Nov 25 09:42:54 crc kubenswrapper[4776]: I1125 09:42:54.426089 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" Nov 25 09:42:54 crc kubenswrapper[4776]: I1125 09:42:54.428322 4776 generic.go:334] "Generic (PLEG): container finished" podID="417adc1f-2f5c-4446-8cf3-994edf3b5e55" containerID="3516bb91cfe2308e98f8eae2924c9343a1bcfca0ad8e712dfbdb8076f26a23b2" exitCode=0 Nov 25 09:42:54 crc kubenswrapper[4776]: I1125 09:42:54.428419 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-5g4ql" event={"ID":"417adc1f-2f5c-4446-8cf3-994edf3b5e55","Type":"ContainerDied","Data":"3516bb91cfe2308e98f8eae2924c9343a1bcfca0ad8e712dfbdb8076f26a23b2"} Nov 25 09:42:54 crc kubenswrapper[4776]: I1125 09:42:54.430820 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7dadc68f-0b89-41ba-84f8-e9bcde16ca64","Type":"ContainerStarted","Data":"048ec6577ec600c7421727ef4cb3662a66e77cfa268385a725211a51ef3249cd"} Nov 25 09:42:54 crc kubenswrapper[4776]: I1125 09:42:54.430895 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7dadc68f-0b89-41ba-84f8-e9bcde16ca64","Type":"ContainerStarted","Data":"1ed6008959fa3368ff11289c84925032c1050a0bd2f104218d9763b1be776a55"} Nov 25 09:42:54 crc kubenswrapper[4776]: I1125 09:42:54.469178 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.822519707 podStartE2EDuration="11.469139938s" podCreationTimestamp="2025-11-25 09:42:43 +0000 UTC" firstStartedPulling="2025-11-25 09:42:44.344494188 +0000 UTC m=+1109.385553741" lastFinishedPulling="2025-11-25 09:42:51.991114409 +0000 UTC m=+1117.032173972" observedRunningTime="2025-11-25 09:42:54.459462645 +0000 UTC m=+1119.500522218" watchObservedRunningTime="2025-11-25 09:42:54.469139938 +0000 UTC m=+1119.510199491" Nov 25 09:42:54 crc kubenswrapper[4776]: I1125 09:42:54.486745 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-4htz7" podStartSLOduration=2.466770308 podStartE2EDuration="6.486727819s" podCreationTimestamp="2025-11-25 09:42:48 +0000 UTC" firstStartedPulling="2025-11-25 09:42:49.306275162 +0000 UTC m=+1114.347334735" lastFinishedPulling="2025-11-25 09:42:53.326232693 +0000 UTC m=+1118.367292246" observedRunningTime="2025-11-25 09:42:54.479523738 +0000 UTC m=+1119.520583291" watchObservedRunningTime="2025-11-25 09:42:54.486727819 +0000 UTC m=+1119.527787372" Nov 25 09:42:54 crc kubenswrapper[4776]: I1125 09:42:54.506729 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" podStartSLOduration=8.50670581 podStartE2EDuration="8.50670581s" podCreationTimestamp="2025-11-25 09:42:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:42:54.500304349 +0000 UTC m=+1119.541363902" watchObservedRunningTime="2025-11-25 09:42:54.50670581 +0000 UTC m=+1119.547765363" Nov 25 09:42:55 crc kubenswrapper[4776]: I1125 09:42:55.441134 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65c9b8d4f7-n2mvq" event={"ID":"a483d333-16fd-4c48-854f-590fe5f0665c","Type":"ContainerStarted","Data":"0c8739c2b70b63c2b0054fa0f48c830c5cfb61f3637054674ebcec6164437ba1"} Nov 25 09:42:55 crc kubenswrapper[4776]: I1125 09:42:55.443551 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 25 09:42:55 crc kubenswrapper[4776]: I1125 09:42:55.443619 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-65c9b8d4f7-n2mvq" Nov 25 09:42:55 crc kubenswrapper[4776]: I1125 09:42:55.460028 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-65c9b8d4f7-n2mvq" podStartSLOduration=13.460008179 podStartE2EDuration="13.460008179s" podCreationTimestamp="2025-11-25 09:42:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:42:55.457204998 +0000 UTC m=+1120.498264551" watchObservedRunningTime="2025-11-25 09:42:55.460008179 +0000 UTC m=+1120.501067732" Nov 25 09:42:55 crc kubenswrapper[4776]: I1125 09:42:55.674592 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ebd843d5-4a0b-4660-9636-fc19d8fddec6-etc-swift\") pod \"swift-storage-0\" (UID: \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\") " pod="openstack/swift-storage-0" Nov 25 09:42:55 crc kubenswrapper[4776]: E1125 09:42:55.674744 4776 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 25 09:42:55 crc kubenswrapper[4776]: E1125 09:42:55.675137 4776 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 25 09:42:55 crc kubenswrapper[4776]: E1125 09:42:55.675228 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ebd843d5-4a0b-4660-9636-fc19d8fddec6-etc-swift podName:ebd843d5-4a0b-4660-9636-fc19d8fddec6 nodeName:}" failed. No retries permitted until 2025-11-25 09:43:03.675207446 +0000 UTC m=+1128.716266999 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/ebd843d5-4a0b-4660-9636-fc19d8fddec6-etc-swift") pod "swift-storage-0" (UID: "ebd843d5-4a0b-4660-9636-fc19d8fddec6") : configmap "swift-ring-files" not found Nov 25 09:42:55 crc kubenswrapper[4776]: I1125 09:42:55.858215 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5g4ql" Nov 25 09:42:55 crc kubenswrapper[4776]: I1125 09:42:55.930802 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b9b5-account-create-ldfzn" Nov 25 09:42:55 crc kubenswrapper[4776]: I1125 09:42:55.987179 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8kd6\" (UniqueName: \"kubernetes.io/projected/417adc1f-2f5c-4446-8cf3-994edf3b5e55-kube-api-access-m8kd6\") pod \"417adc1f-2f5c-4446-8cf3-994edf3b5e55\" (UID: \"417adc1f-2f5c-4446-8cf3-994edf3b5e55\") " Nov 25 09:42:55 crc kubenswrapper[4776]: I1125 09:42:55.987343 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3-operator-scripts\") pod \"00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3\" (UID: \"00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3\") " Nov 25 09:42:55 crc kubenswrapper[4776]: I1125 09:42:55.987363 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qd2zx\" (UniqueName: \"kubernetes.io/projected/00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3-kube-api-access-qd2zx\") pod \"00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3\" (UID: \"00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3\") " Nov 25 09:42:55 crc kubenswrapper[4776]: I1125 09:42:55.987411 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/417adc1f-2f5c-4446-8cf3-994edf3b5e55-operator-scripts\") pod \"417adc1f-2f5c-4446-8cf3-994edf3b5e55\" (UID: \"417adc1f-2f5c-4446-8cf3-994edf3b5e55\") " Nov 25 09:42:55 crc kubenswrapper[4776]: I1125 09:42:55.988696 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3" (UID: "00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:42:55 crc kubenswrapper[4776]: I1125 09:42:55.989551 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/417adc1f-2f5c-4446-8cf3-994edf3b5e55-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "417adc1f-2f5c-4446-8cf3-994edf3b5e55" (UID: "417adc1f-2f5c-4446-8cf3-994edf3b5e55"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:42:55 crc kubenswrapper[4776]: I1125 09:42:55.993776 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3-kube-api-access-qd2zx" (OuterVolumeSpecName: "kube-api-access-qd2zx") pod "00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3" (UID: "00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3"). InnerVolumeSpecName "kube-api-access-qd2zx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:42:56 crc kubenswrapper[4776]: I1125 09:42:56.000914 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/417adc1f-2f5c-4446-8cf3-994edf3b5e55-kube-api-access-m8kd6" (OuterVolumeSpecName: "kube-api-access-m8kd6") pod "417adc1f-2f5c-4446-8cf3-994edf3b5e55" (UID: "417adc1f-2f5c-4446-8cf3-994edf3b5e55"). InnerVolumeSpecName "kube-api-access-m8kd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:42:56 crc kubenswrapper[4776]: I1125 09:42:56.089898 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:42:56 crc kubenswrapper[4776]: I1125 09:42:56.089944 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qd2zx\" (UniqueName: \"kubernetes.io/projected/00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3-kube-api-access-qd2zx\") on node \"crc\" DevicePath \"\"" Nov 25 09:42:56 crc kubenswrapper[4776]: I1125 09:42:56.089964 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/417adc1f-2f5c-4446-8cf3-994edf3b5e55-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:42:56 crc kubenswrapper[4776]: I1125 09:42:56.089979 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8kd6\" (UniqueName: \"kubernetes.io/projected/417adc1f-2f5c-4446-8cf3-994edf3b5e55-kube-api-access-m8kd6\") on node \"crc\" DevicePath \"\"" Nov 25 09:42:56 crc kubenswrapper[4776]: I1125 09:42:56.450042 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-5g4ql" event={"ID":"417adc1f-2f5c-4446-8cf3-994edf3b5e55","Type":"ContainerDied","Data":"e6c466809d6ca7435c63046c353094af4ecd2a78a88fbbd6c8f05180a4d2c268"} Nov 25 09:42:56 crc kubenswrapper[4776]: I1125 09:42:56.450111 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6c466809d6ca7435c63046c353094af4ecd2a78a88fbbd6c8f05180a4d2c268" Nov 25 09:42:56 crc kubenswrapper[4776]: I1125 09:42:56.450363 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-5g4ql" Nov 25 09:42:56 crc kubenswrapper[4776]: I1125 09:42:56.456572 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b9b5-account-create-ldfzn" event={"ID":"00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3","Type":"ContainerDied","Data":"e77e7c401a0a6fef9f6d18d6325001e1da4b652050313289924dcffcaf2a3531"} Nov 25 09:42:56 crc kubenswrapper[4776]: I1125 09:42:56.456710 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b9b5-account-create-ldfzn" Nov 25 09:42:56 crc kubenswrapper[4776]: I1125 09:42:56.458106 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e77e7c401a0a6fef9f6d18d6325001e1da4b652050313289924dcffcaf2a3531" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.439394 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-4b82s"] Nov 25 09:43:00 crc kubenswrapper[4776]: E1125 09:43:00.440190 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1de7f06d-24e7-4114-966b-d62df3366006" containerName="mariadb-database-create" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.440204 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="1de7f06d-24e7-4114-966b-d62df3366006" containerName="mariadb-database-create" Nov 25 09:43:00 crc kubenswrapper[4776]: E1125 09:43:00.440213 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6466d754-e356-4950-9ce9-8f56e68d6b97" containerName="mariadb-account-create" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.440219 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="6466d754-e356-4950-9ce9-8f56e68d6b97" containerName="mariadb-account-create" Nov 25 09:43:00 crc kubenswrapper[4776]: E1125 09:43:00.440232 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f72b0b94-97f1-4bbc-9be4-c8965143cd26" containerName="mariadb-database-create" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.440238 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="f72b0b94-97f1-4bbc-9be4-c8965143cd26" containerName="mariadb-database-create" Nov 25 09:43:00 crc kubenswrapper[4776]: E1125 09:43:00.440258 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="417adc1f-2f5c-4446-8cf3-994edf3b5e55" containerName="mariadb-database-create" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.440263 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="417adc1f-2f5c-4446-8cf3-994edf3b5e55" containerName="mariadb-database-create" Nov 25 09:43:00 crc kubenswrapper[4776]: E1125 09:43:00.440270 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e" containerName="mariadb-account-create" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.440275 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e" containerName="mariadb-account-create" Nov 25 09:43:00 crc kubenswrapper[4776]: E1125 09:43:00.440286 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3" containerName="mariadb-account-create" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.440292 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3" containerName="mariadb-account-create" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.440432 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e" containerName="mariadb-account-create" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.440458 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="f72b0b94-97f1-4bbc-9be4-c8965143cd26" containerName="mariadb-database-create" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.440469 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="1de7f06d-24e7-4114-966b-d62df3366006" containerName="mariadb-database-create" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.440489 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="6466d754-e356-4950-9ce9-8f56e68d6b97" containerName="mariadb-account-create" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.440507 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="417adc1f-2f5c-4446-8cf3-994edf3b5e55" containerName="mariadb-database-create" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.440529 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3" containerName="mariadb-account-create" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.440993 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4b82s" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.443236 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-5xkz6" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.444371 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.460044 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-4b82s"] Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.504249 4776 generic.go:334] "Generic (PLEG): container finished" podID="cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515" containerID="0013160d2c55a26010d58e0144d8a89639c5b92e40762472567d0dae6bd635e9" exitCode=0 Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.504309 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-4htz7" event={"ID":"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515","Type":"ContainerDied","Data":"0013160d2c55a26010d58e0144d8a89639c5b92e40762472567d0dae6bd635e9"} Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.564284 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/758472ef-bebc-469a-87a0-e7f1a83ae279-combined-ca-bundle\") pod \"glance-db-sync-4b82s\" (UID: \"758472ef-bebc-469a-87a0-e7f1a83ae279\") " pod="openstack/glance-db-sync-4b82s" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.564371 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gv5fb\" (UniqueName: \"kubernetes.io/projected/758472ef-bebc-469a-87a0-e7f1a83ae279-kube-api-access-gv5fb\") pod \"glance-db-sync-4b82s\" (UID: \"758472ef-bebc-469a-87a0-e7f1a83ae279\") " pod="openstack/glance-db-sync-4b82s" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.564414 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/758472ef-bebc-469a-87a0-e7f1a83ae279-config-data\") pod \"glance-db-sync-4b82s\" (UID: \"758472ef-bebc-469a-87a0-e7f1a83ae279\") " pod="openstack/glance-db-sync-4b82s" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.564447 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/758472ef-bebc-469a-87a0-e7f1a83ae279-db-sync-config-data\") pod \"glance-db-sync-4b82s\" (UID: \"758472ef-bebc-469a-87a0-e7f1a83ae279\") " pod="openstack/glance-db-sync-4b82s" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.665757 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/758472ef-bebc-469a-87a0-e7f1a83ae279-db-sync-config-data\") pod \"glance-db-sync-4b82s\" (UID: \"758472ef-bebc-469a-87a0-e7f1a83ae279\") " pod="openstack/glance-db-sync-4b82s" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.665851 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/758472ef-bebc-469a-87a0-e7f1a83ae279-combined-ca-bundle\") pod \"glance-db-sync-4b82s\" (UID: \"758472ef-bebc-469a-87a0-e7f1a83ae279\") " pod="openstack/glance-db-sync-4b82s" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.665911 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gv5fb\" (UniqueName: \"kubernetes.io/projected/758472ef-bebc-469a-87a0-e7f1a83ae279-kube-api-access-gv5fb\") pod \"glance-db-sync-4b82s\" (UID: \"758472ef-bebc-469a-87a0-e7f1a83ae279\") " pod="openstack/glance-db-sync-4b82s" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.665960 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/758472ef-bebc-469a-87a0-e7f1a83ae279-config-data\") pod \"glance-db-sync-4b82s\" (UID: \"758472ef-bebc-469a-87a0-e7f1a83ae279\") " pod="openstack/glance-db-sync-4b82s" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.672532 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/758472ef-bebc-469a-87a0-e7f1a83ae279-db-sync-config-data\") pod \"glance-db-sync-4b82s\" (UID: \"758472ef-bebc-469a-87a0-e7f1a83ae279\") " pod="openstack/glance-db-sync-4b82s" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.672751 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/758472ef-bebc-469a-87a0-e7f1a83ae279-combined-ca-bundle\") pod \"glance-db-sync-4b82s\" (UID: \"758472ef-bebc-469a-87a0-e7f1a83ae279\") " pod="openstack/glance-db-sync-4b82s" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.674095 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/758472ef-bebc-469a-87a0-e7f1a83ae279-config-data\") pod \"glance-db-sync-4b82s\" (UID: \"758472ef-bebc-469a-87a0-e7f1a83ae279\") " pod="openstack/glance-db-sync-4b82s" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.683769 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gv5fb\" (UniqueName: \"kubernetes.io/projected/758472ef-bebc-469a-87a0-e7f1a83ae279-kube-api-access-gv5fb\") pod \"glance-db-sync-4b82s\" (UID: \"758472ef-bebc-469a-87a0-e7f1a83ae279\") " pod="openstack/glance-db-sync-4b82s" Nov 25 09:43:00 crc kubenswrapper[4776]: I1125 09:43:00.760662 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4b82s" Nov 25 09:43:01 crc kubenswrapper[4776]: I1125 09:43:01.281329 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-4b82s"] Nov 25 09:43:01 crc kubenswrapper[4776]: W1125 09:43:01.285734 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod758472ef_bebc_469a_87a0_e7f1a83ae279.slice/crio-f32b036d1c5f5c7719a4530e76e35d67b4dea497f08a51a791770990966e84bd WatchSource:0}: Error finding container f32b036d1c5f5c7719a4530e76e35d67b4dea497f08a51a791770990966e84bd: Status 404 returned error can't find the container with id f32b036d1c5f5c7719a4530e76e35d67b4dea497f08a51a791770990966e84bd Nov 25 09:43:01 crc kubenswrapper[4776]: I1125 09:43:01.513125 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4b82s" event={"ID":"758472ef-bebc-469a-87a0-e7f1a83ae279","Type":"ContainerStarted","Data":"f32b036d1c5f5c7719a4530e76e35d67b4dea497f08a51a791770990966e84bd"} Nov 25 09:43:01 crc kubenswrapper[4776]: I1125 09:43:01.785535 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-4htz7" Nov 25 09:43:01 crc kubenswrapper[4776]: I1125 09:43:01.885574 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4kll\" (UniqueName: \"kubernetes.io/projected/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-kube-api-access-m4kll\") pod \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\" (UID: \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\") " Nov 25 09:43:01 crc kubenswrapper[4776]: I1125 09:43:01.885898 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-etc-swift\") pod \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\" (UID: \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\") " Nov 25 09:43:01 crc kubenswrapper[4776]: I1125 09:43:01.886092 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-scripts\") pod \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\" (UID: \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\") " Nov 25 09:43:01 crc kubenswrapper[4776]: I1125 09:43:01.886300 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-ring-data-devices\") pod \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\" (UID: \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\") " Nov 25 09:43:01 crc kubenswrapper[4776]: I1125 09:43:01.886911 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515" (UID: "cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:43:01 crc kubenswrapper[4776]: I1125 09:43:01.886963 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515" (UID: "cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:43:01 crc kubenswrapper[4776]: I1125 09:43:01.887211 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-combined-ca-bundle\") pod \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\" (UID: \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\") " Nov 25 09:43:01 crc kubenswrapper[4776]: I1125 09:43:01.887392 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-dispersionconf\") pod \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\" (UID: \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\") " Nov 25 09:43:01 crc kubenswrapper[4776]: I1125 09:43:01.887778 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-swiftconf\") pod \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\" (UID: \"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515\") " Nov 25 09:43:01 crc kubenswrapper[4776]: I1125 09:43:01.888447 4776 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:01 crc kubenswrapper[4776]: I1125 09:43:01.888545 4776 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:01 crc kubenswrapper[4776]: I1125 09:43:01.898433 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-kube-api-access-m4kll" (OuterVolumeSpecName: "kube-api-access-m4kll") pod "cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515" (UID: "cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515"). InnerVolumeSpecName "kube-api-access-m4kll". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:43:01 crc kubenswrapper[4776]: I1125 09:43:01.899303 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515" (UID: "cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:43:01 crc kubenswrapper[4776]: I1125 09:43:01.909298 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-scripts" (OuterVolumeSpecName: "scripts") pod "cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515" (UID: "cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:43:01 crc kubenswrapper[4776]: I1125 09:43:01.912899 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515" (UID: "cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:43:01 crc kubenswrapper[4776]: I1125 09:43:01.914878 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515" (UID: "cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:43:01 crc kubenswrapper[4776]: I1125 09:43:01.990351 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:01 crc kubenswrapper[4776]: I1125 09:43:01.990448 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:01 crc kubenswrapper[4776]: I1125 09:43:01.990464 4776 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:01 crc kubenswrapper[4776]: I1125 09:43:01.990477 4776 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:01 crc kubenswrapper[4776]: I1125 09:43:01.990489 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4kll\" (UniqueName: \"kubernetes.io/projected/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515-kube-api-access-m4kll\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.000599 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.063756 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65c9b8d4f7-n2mvq"] Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.064179 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-65c9b8d4f7-n2mvq" podUID="a483d333-16fd-4c48-854f-590fe5f0665c" containerName="dnsmasq-dns" containerID="cri-o://0c8739c2b70b63c2b0054fa0f48c830c5cfb61f3637054674ebcec6164437ba1" gracePeriod=10 Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.067440 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-65c9b8d4f7-n2mvq" Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.428046 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65c9b8d4f7-n2mvq" Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.498237 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a483d333-16fd-4c48-854f-590fe5f0665c-config\") pod \"a483d333-16fd-4c48-854f-590fe5f0665c\" (UID: \"a483d333-16fd-4c48-854f-590fe5f0665c\") " Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.498314 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a483d333-16fd-4c48-854f-590fe5f0665c-ovsdbserver-sb\") pod \"a483d333-16fd-4c48-854f-590fe5f0665c\" (UID: \"a483d333-16fd-4c48-854f-590fe5f0665c\") " Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.498412 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6xwm\" (UniqueName: \"kubernetes.io/projected/a483d333-16fd-4c48-854f-590fe5f0665c-kube-api-access-q6xwm\") pod \"a483d333-16fd-4c48-854f-590fe5f0665c\" (UID: \"a483d333-16fd-4c48-854f-590fe5f0665c\") " Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.498518 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a483d333-16fd-4c48-854f-590fe5f0665c-dns-svc\") pod \"a483d333-16fd-4c48-854f-590fe5f0665c\" (UID: \"a483d333-16fd-4c48-854f-590fe5f0665c\") " Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.502043 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a483d333-16fd-4c48-854f-590fe5f0665c-kube-api-access-q6xwm" (OuterVolumeSpecName: "kube-api-access-q6xwm") pod "a483d333-16fd-4c48-854f-590fe5f0665c" (UID: "a483d333-16fd-4c48-854f-590fe5f0665c"). InnerVolumeSpecName "kube-api-access-q6xwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.522112 4776 generic.go:334] "Generic (PLEG): container finished" podID="f32041a0-a8fc-48a5-afab-476baff89e8c" containerID="b75205641b6269bc97dcaf11905b1173d0238b983968101c11decb00e38566b3" exitCode=0 Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.522174 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f32041a0-a8fc-48a5-afab-476baff89e8c","Type":"ContainerDied","Data":"b75205641b6269bc97dcaf11905b1173d0238b983968101c11decb00e38566b3"} Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.525616 4776 generic.go:334] "Generic (PLEG): container finished" podID="a483d333-16fd-4c48-854f-590fe5f0665c" containerID="0c8739c2b70b63c2b0054fa0f48c830c5cfb61f3637054674ebcec6164437ba1" exitCode=0 Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.525740 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-65c9b8d4f7-n2mvq" Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.525818 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65c9b8d4f7-n2mvq" event={"ID":"a483d333-16fd-4c48-854f-590fe5f0665c","Type":"ContainerDied","Data":"0c8739c2b70b63c2b0054fa0f48c830c5cfb61f3637054674ebcec6164437ba1"} Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.525867 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-65c9b8d4f7-n2mvq" event={"ID":"a483d333-16fd-4c48-854f-590fe5f0665c","Type":"ContainerDied","Data":"dd1955c5ceacd6add7fbb752f8946dba6a2815567d9c2cd0f43b47b7529469d5"} Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.525892 4776 scope.go:117] "RemoveContainer" containerID="0c8739c2b70b63c2b0054fa0f48c830c5cfb61f3637054674ebcec6164437ba1" Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.546141 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a483d333-16fd-4c48-854f-590fe5f0665c-config" (OuterVolumeSpecName: "config") pod "a483d333-16fd-4c48-854f-590fe5f0665c" (UID: "a483d333-16fd-4c48-854f-590fe5f0665c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.548632 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a483d333-16fd-4c48-854f-590fe5f0665c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a483d333-16fd-4c48-854f-590fe5f0665c" (UID: "a483d333-16fd-4c48-854f-590fe5f0665c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.550652 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-4htz7" event={"ID":"cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515","Type":"ContainerDied","Data":"50acb690971c2fa83afb22d2985d07b82184bcd8c9b729695c2600e3509a606d"} Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.550684 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50acb690971c2fa83afb22d2985d07b82184bcd8c9b729695c2600e3509a606d" Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.550749 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-4htz7" Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.557094 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a483d333-16fd-4c48-854f-590fe5f0665c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a483d333-16fd-4c48-854f-590fe5f0665c" (UID: "a483d333-16fd-4c48-854f-590fe5f0665c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.566537 4776 scope.go:117] "RemoveContainer" containerID="a69925572190a55956bc810bf3ddd277c1f7ac89b8e51b395d26a92e98b05dde" Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.596945 4776 scope.go:117] "RemoveContainer" containerID="0c8739c2b70b63c2b0054fa0f48c830c5cfb61f3637054674ebcec6164437ba1" Nov 25 09:43:02 crc kubenswrapper[4776]: E1125 09:43:02.597367 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c8739c2b70b63c2b0054fa0f48c830c5cfb61f3637054674ebcec6164437ba1\": container with ID starting with 0c8739c2b70b63c2b0054fa0f48c830c5cfb61f3637054674ebcec6164437ba1 not found: ID does not exist" containerID="0c8739c2b70b63c2b0054fa0f48c830c5cfb61f3637054674ebcec6164437ba1" Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.597425 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c8739c2b70b63c2b0054fa0f48c830c5cfb61f3637054674ebcec6164437ba1"} err="failed to get container status \"0c8739c2b70b63c2b0054fa0f48c830c5cfb61f3637054674ebcec6164437ba1\": rpc error: code = NotFound desc = could not find container \"0c8739c2b70b63c2b0054fa0f48c830c5cfb61f3637054674ebcec6164437ba1\": container with ID starting with 0c8739c2b70b63c2b0054fa0f48c830c5cfb61f3637054674ebcec6164437ba1 not found: ID does not exist" Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.597445 4776 scope.go:117] "RemoveContainer" containerID="a69925572190a55956bc810bf3ddd277c1f7ac89b8e51b395d26a92e98b05dde" Nov 25 09:43:02 crc kubenswrapper[4776]: E1125 09:43:02.597772 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a69925572190a55956bc810bf3ddd277c1f7ac89b8e51b395d26a92e98b05dde\": container with ID starting with a69925572190a55956bc810bf3ddd277c1f7ac89b8e51b395d26a92e98b05dde not found: ID does not exist" containerID="a69925572190a55956bc810bf3ddd277c1f7ac89b8e51b395d26a92e98b05dde" Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.597827 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a69925572190a55956bc810bf3ddd277c1f7ac89b8e51b395d26a92e98b05dde"} err="failed to get container status \"a69925572190a55956bc810bf3ddd277c1f7ac89b8e51b395d26a92e98b05dde\": rpc error: code = NotFound desc = could not find container \"a69925572190a55956bc810bf3ddd277c1f7ac89b8e51b395d26a92e98b05dde\": container with ID starting with a69925572190a55956bc810bf3ddd277c1f7ac89b8e51b395d26a92e98b05dde not found: ID does not exist" Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.600399 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a483d333-16fd-4c48-854f-590fe5f0665c-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.600428 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a483d333-16fd-4c48-854f-590fe5f0665c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.600442 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6xwm\" (UniqueName: \"kubernetes.io/projected/a483d333-16fd-4c48-854f-590fe5f0665c-kube-api-access-q6xwm\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.600456 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a483d333-16fd-4c48-854f-590fe5f0665c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.862475 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-65c9b8d4f7-n2mvq"] Nov 25 09:43:02 crc kubenswrapper[4776]: I1125 09:43:02.869411 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-65c9b8d4f7-n2mvq"] Nov 25 09:43:03 crc kubenswrapper[4776]: I1125 09:43:03.574329 4776 generic.go:334] "Generic (PLEG): container finished" podID="4512cf4e-6c75-493e-a80a-4b0accd750a8" containerID="e5d2c432a51c339c288f48c9a58a0ccdecc76c714979d2e84604d3bc00357881" exitCode=0 Nov 25 09:43:03 crc kubenswrapper[4776]: I1125 09:43:03.574444 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4512cf4e-6c75-493e-a80a-4b0accd750a8","Type":"ContainerDied","Data":"e5d2c432a51c339c288f48c9a58a0ccdecc76c714979d2e84604d3bc00357881"} Nov 25 09:43:03 crc kubenswrapper[4776]: I1125 09:43:03.584271 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f32041a0-a8fc-48a5-afab-476baff89e8c","Type":"ContainerStarted","Data":"7b42621ff074baca099513ddb94523b955d55200c16323a7bb88a078926fca25"} Nov 25 09:43:03 crc kubenswrapper[4776]: I1125 09:43:03.584681 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:43:03 crc kubenswrapper[4776]: I1125 09:43:03.676680 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a483d333-16fd-4c48-854f-590fe5f0665c" path="/var/lib/kubelet/pods/a483d333-16fd-4c48-854f-590fe5f0665c/volumes" Nov 25 09:43:03 crc kubenswrapper[4776]: I1125 09:43:03.737087 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ebd843d5-4a0b-4660-9636-fc19d8fddec6-etc-swift\") pod \"swift-storage-0\" (UID: \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\") " pod="openstack/swift-storage-0" Nov 25 09:43:03 crc kubenswrapper[4776]: I1125 09:43:03.743308 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ebd843d5-4a0b-4660-9636-fc19d8fddec6-etc-swift\") pod \"swift-storage-0\" (UID: \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\") " pod="openstack/swift-storage-0" Nov 25 09:43:03 crc kubenswrapper[4776]: I1125 09:43:03.757036 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 25 09:43:03 crc kubenswrapper[4776]: I1125 09:43:03.967973 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 25 09:43:03 crc kubenswrapper[4776]: I1125 09:43:03.993562 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.952214522 podStartE2EDuration="53.993545319s" podCreationTimestamp="2025-11-25 09:42:10 +0000 UTC" firstStartedPulling="2025-11-25 09:42:12.340841927 +0000 UTC m=+1077.381901480" lastFinishedPulling="2025-11-25 09:42:29.382172684 +0000 UTC m=+1094.423232277" observedRunningTime="2025-11-25 09:43:03.632992956 +0000 UTC m=+1128.674052509" watchObservedRunningTime="2025-11-25 09:43:03.993545319 +0000 UTC m=+1129.034604872" Nov 25 09:43:04 crc kubenswrapper[4776]: I1125 09:43:04.321311 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 25 09:43:04 crc kubenswrapper[4776]: I1125 09:43:04.593339 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerStarted","Data":"8d9f1bd809faa88090af0396f6842af5c4e67d29000a1f49bbc2554cee1a30ea"} Nov 25 09:43:04 crc kubenswrapper[4776]: I1125 09:43:04.595903 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4512cf4e-6c75-493e-a80a-4b0accd750a8","Type":"ContainerStarted","Data":"a5fdd3d6b6a35387501cbc30834b6e26eb181b138a16ef46ae65f71952655401"} Nov 25 09:43:04 crc kubenswrapper[4776]: I1125 09:43:04.596206 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 25 09:43:04 crc kubenswrapper[4776]: I1125 09:43:04.627926 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=-9223371981.226866 podStartE2EDuration="55.627909879s" podCreationTimestamp="2025-11-25 09:42:09 +0000 UTC" firstStartedPulling="2025-11-25 09:42:11.311184373 +0000 UTC m=+1076.352243926" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:43:04.624613296 +0000 UTC m=+1129.665672849" watchObservedRunningTime="2025-11-25 09:43:04.627909879 +0000 UTC m=+1129.668969432" Nov 25 09:43:05 crc kubenswrapper[4776]: I1125 09:43:05.614363 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-2bs2f" podUID="c59c3d1e-c5c3-4036-ae28-436585e303ab" containerName="ovn-controller" probeResult="failure" output=< Nov 25 09:43:05 crc kubenswrapper[4776]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 25 09:43:05 crc kubenswrapper[4776]: > Nov 25 09:43:06 crc kubenswrapper[4776]: I1125 09:43:06.616863 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerStarted","Data":"cd073bb8b2d7b566c99b6e569d6dbfa6b16b08d5db2e84e47a8d2d85eb563f1b"} Nov 25 09:43:06 crc kubenswrapper[4776]: I1125 09:43:06.618218 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerStarted","Data":"6ccd151880ed401c0372e024b6b953cdc0eba4d9d9e8cdd47435d059001cbc30"} Nov 25 09:43:06 crc kubenswrapper[4776]: I1125 09:43:06.618237 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerStarted","Data":"05b65f4bed8c7914cf657df0f540c0d19593a2c84c93c9ae6c4f106530470ec4"} Nov 25 09:43:06 crc kubenswrapper[4776]: I1125 09:43:06.618245 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerStarted","Data":"9e7cf1b3a65acbcaaedff95080720317c88059b1b6b759f4434c871973a12bc3"} Nov 25 09:43:10 crc kubenswrapper[4776]: I1125 09:43:10.620969 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-2bs2f" podUID="c59c3d1e-c5c3-4036-ae28-436585e303ab" containerName="ovn-controller" probeResult="failure" output=< Nov 25 09:43:10 crc kubenswrapper[4776]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 25 09:43:10 crc kubenswrapper[4776]: > Nov 25 09:43:10 crc kubenswrapper[4776]: I1125 09:43:10.666913 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-7fgjf" Nov 25 09:43:10 crc kubenswrapper[4776]: I1125 09:43:10.667964 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-7fgjf" Nov 25 09:43:10 crc kubenswrapper[4776]: I1125 09:43:10.895326 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-2bs2f-config-x7wjq"] Nov 25 09:43:10 crc kubenswrapper[4776]: E1125 09:43:10.895746 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515" containerName="swift-ring-rebalance" Nov 25 09:43:10 crc kubenswrapper[4776]: I1125 09:43:10.895768 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515" containerName="swift-ring-rebalance" Nov 25 09:43:10 crc kubenswrapper[4776]: E1125 09:43:10.895803 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a483d333-16fd-4c48-854f-590fe5f0665c" containerName="dnsmasq-dns" Nov 25 09:43:10 crc kubenswrapper[4776]: I1125 09:43:10.895813 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a483d333-16fd-4c48-854f-590fe5f0665c" containerName="dnsmasq-dns" Nov 25 09:43:10 crc kubenswrapper[4776]: E1125 09:43:10.895824 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a483d333-16fd-4c48-854f-590fe5f0665c" containerName="init" Nov 25 09:43:10 crc kubenswrapper[4776]: I1125 09:43:10.895832 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a483d333-16fd-4c48-854f-590fe5f0665c" containerName="init" Nov 25 09:43:10 crc kubenswrapper[4776]: I1125 09:43:10.896039 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515" containerName="swift-ring-rebalance" Nov 25 09:43:10 crc kubenswrapper[4776]: I1125 09:43:10.896089 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="a483d333-16fd-4c48-854f-590fe5f0665c" containerName="dnsmasq-dns" Nov 25 09:43:10 crc kubenswrapper[4776]: I1125 09:43:10.896744 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2bs2f-config-x7wjq" Nov 25 09:43:10 crc kubenswrapper[4776]: I1125 09:43:10.898773 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 25 09:43:10 crc kubenswrapper[4776]: I1125 09:43:10.919787 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-2bs2f-config-x7wjq"] Nov 25 09:43:10 crc kubenswrapper[4776]: I1125 09:43:10.954272 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-var-run-ovn\") pod \"ovn-controller-2bs2f-config-x7wjq\" (UID: \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\") " pod="openstack/ovn-controller-2bs2f-config-x7wjq" Nov 25 09:43:10 crc kubenswrapper[4776]: I1125 09:43:10.954389 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-scripts\") pod \"ovn-controller-2bs2f-config-x7wjq\" (UID: \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\") " pod="openstack/ovn-controller-2bs2f-config-x7wjq" Nov 25 09:43:10 crc kubenswrapper[4776]: I1125 09:43:10.954421 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-additional-scripts\") pod \"ovn-controller-2bs2f-config-x7wjq\" (UID: \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\") " pod="openstack/ovn-controller-2bs2f-config-x7wjq" Nov 25 09:43:10 crc kubenswrapper[4776]: I1125 09:43:10.954470 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-var-log-ovn\") pod \"ovn-controller-2bs2f-config-x7wjq\" (UID: \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\") " pod="openstack/ovn-controller-2bs2f-config-x7wjq" Nov 25 09:43:10 crc kubenswrapper[4776]: I1125 09:43:10.954499 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-var-run\") pod \"ovn-controller-2bs2f-config-x7wjq\" (UID: \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\") " pod="openstack/ovn-controller-2bs2f-config-x7wjq" Nov 25 09:43:10 crc kubenswrapper[4776]: I1125 09:43:10.954555 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k4h5\" (UniqueName: \"kubernetes.io/projected/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-kube-api-access-2k4h5\") pod \"ovn-controller-2bs2f-config-x7wjq\" (UID: \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\") " pod="openstack/ovn-controller-2bs2f-config-x7wjq" Nov 25 09:43:11 crc kubenswrapper[4776]: I1125 09:43:11.056084 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-scripts\") pod \"ovn-controller-2bs2f-config-x7wjq\" (UID: \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\") " pod="openstack/ovn-controller-2bs2f-config-x7wjq" Nov 25 09:43:11 crc kubenswrapper[4776]: I1125 09:43:11.056141 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-additional-scripts\") pod \"ovn-controller-2bs2f-config-x7wjq\" (UID: \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\") " pod="openstack/ovn-controller-2bs2f-config-x7wjq" Nov 25 09:43:11 crc kubenswrapper[4776]: I1125 09:43:11.056192 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-var-log-ovn\") pod \"ovn-controller-2bs2f-config-x7wjq\" (UID: \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\") " pod="openstack/ovn-controller-2bs2f-config-x7wjq" Nov 25 09:43:11 crc kubenswrapper[4776]: I1125 09:43:11.056212 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-var-run\") pod \"ovn-controller-2bs2f-config-x7wjq\" (UID: \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\") " pod="openstack/ovn-controller-2bs2f-config-x7wjq" Nov 25 09:43:11 crc kubenswrapper[4776]: I1125 09:43:11.056265 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k4h5\" (UniqueName: \"kubernetes.io/projected/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-kube-api-access-2k4h5\") pod \"ovn-controller-2bs2f-config-x7wjq\" (UID: \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\") " pod="openstack/ovn-controller-2bs2f-config-x7wjq" Nov 25 09:43:11 crc kubenswrapper[4776]: I1125 09:43:11.056361 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-var-run-ovn\") pod \"ovn-controller-2bs2f-config-x7wjq\" (UID: \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\") " pod="openstack/ovn-controller-2bs2f-config-x7wjq" Nov 25 09:43:11 crc kubenswrapper[4776]: I1125 09:43:11.056554 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-var-run-ovn\") pod \"ovn-controller-2bs2f-config-x7wjq\" (UID: \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\") " pod="openstack/ovn-controller-2bs2f-config-x7wjq" Nov 25 09:43:11 crc kubenswrapper[4776]: I1125 09:43:11.056567 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-var-log-ovn\") pod \"ovn-controller-2bs2f-config-x7wjq\" (UID: \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\") " pod="openstack/ovn-controller-2bs2f-config-x7wjq" Nov 25 09:43:11 crc kubenswrapper[4776]: I1125 09:43:11.056599 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-var-run\") pod \"ovn-controller-2bs2f-config-x7wjq\" (UID: \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\") " pod="openstack/ovn-controller-2bs2f-config-x7wjq" Nov 25 09:43:11 crc kubenswrapper[4776]: I1125 09:43:11.056970 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-additional-scripts\") pod \"ovn-controller-2bs2f-config-x7wjq\" (UID: \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\") " pod="openstack/ovn-controller-2bs2f-config-x7wjq" Nov 25 09:43:11 crc kubenswrapper[4776]: I1125 09:43:11.063943 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-scripts\") pod \"ovn-controller-2bs2f-config-x7wjq\" (UID: \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\") " pod="openstack/ovn-controller-2bs2f-config-x7wjq" Nov 25 09:43:11 crc kubenswrapper[4776]: I1125 09:43:11.077427 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k4h5\" (UniqueName: \"kubernetes.io/projected/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-kube-api-access-2k4h5\") pod \"ovn-controller-2bs2f-config-x7wjq\" (UID: \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\") " pod="openstack/ovn-controller-2bs2f-config-x7wjq" Nov 25 09:43:11 crc kubenswrapper[4776]: I1125 09:43:11.217992 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2bs2f-config-x7wjq" Nov 25 09:43:13 crc kubenswrapper[4776]: I1125 09:43:13.739339 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-2bs2f-config-x7wjq"] Nov 25 09:43:13 crc kubenswrapper[4776]: W1125 09:43:13.964216 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc24cb0b1_8aba_4239_9829_05f2ad0a30d0.slice/crio-10a190f92dd317ee24723dcf578f43b2084fb953ca37e4e24bb2322a3541c0a7 WatchSource:0}: Error finding container 10a190f92dd317ee24723dcf578f43b2084fb953ca37e4e24bb2322a3541c0a7: Status 404 returned error can't find the container with id 10a190f92dd317ee24723dcf578f43b2084fb953ca37e4e24bb2322a3541c0a7 Nov 25 09:43:14 crc kubenswrapper[4776]: I1125 09:43:14.680247 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4b82s" event={"ID":"758472ef-bebc-469a-87a0-e7f1a83ae279","Type":"ContainerStarted","Data":"356e591a2a72a0074a6ee72d621e09ac39b9f3d69f4d225fdbc0fdceddce7b5a"} Nov 25 09:43:14 crc kubenswrapper[4776]: I1125 09:43:14.682423 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2bs2f-config-x7wjq" event={"ID":"c24cb0b1-8aba-4239-9829-05f2ad0a30d0","Type":"ContainerStarted","Data":"2bd361fc4287baa9805ecbc3dbe86ea5dc2b96fd42699533192f2b89fd8fd73c"} Nov 25 09:43:14 crc kubenswrapper[4776]: I1125 09:43:14.682463 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2bs2f-config-x7wjq" event={"ID":"c24cb0b1-8aba-4239-9829-05f2ad0a30d0","Type":"ContainerStarted","Data":"10a190f92dd317ee24723dcf578f43b2084fb953ca37e4e24bb2322a3541c0a7"} Nov 25 09:43:14 crc kubenswrapper[4776]: I1125 09:43:14.694010 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerStarted","Data":"434eabc5a25e34093e38beb759273c9fac1b8758c3a03273e4ddb5f564fcdc08"} Nov 25 09:43:14 crc kubenswrapper[4776]: I1125 09:43:14.694061 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerStarted","Data":"a4e815cd4ca77d4505a242b624a83fcda104ea6e85674eb4c15ddc064562ffd6"} Nov 25 09:43:14 crc kubenswrapper[4776]: I1125 09:43:14.694095 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerStarted","Data":"e2072d3b3f2b295702755794df1a3f5a2a130db157825a15a28d4ac1c3a1c674"} Nov 25 09:43:14 crc kubenswrapper[4776]: I1125 09:43:14.699296 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-4b82s" podStartSLOduration=2.5225314450000003 podStartE2EDuration="14.699276387s" podCreationTimestamp="2025-11-25 09:43:00 +0000 UTC" firstStartedPulling="2025-11-25 09:43:01.287873191 +0000 UTC m=+1126.328932744" lastFinishedPulling="2025-11-25 09:43:13.464618133 +0000 UTC m=+1138.505677686" observedRunningTime="2025-11-25 09:43:14.696422246 +0000 UTC m=+1139.737481799" watchObservedRunningTime="2025-11-25 09:43:14.699276387 +0000 UTC m=+1139.740335940" Nov 25 09:43:15 crc kubenswrapper[4776]: I1125 09:43:15.624360 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-2bs2f" Nov 25 09:43:15 crc kubenswrapper[4776]: I1125 09:43:15.738678 4776 generic.go:334] "Generic (PLEG): container finished" podID="c24cb0b1-8aba-4239-9829-05f2ad0a30d0" containerID="2bd361fc4287baa9805ecbc3dbe86ea5dc2b96fd42699533192f2b89fd8fd73c" exitCode=0 Nov 25 09:43:15 crc kubenswrapper[4776]: I1125 09:43:15.738740 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2bs2f-config-x7wjq" event={"ID":"c24cb0b1-8aba-4239-9829-05f2ad0a30d0","Type":"ContainerDied","Data":"2bd361fc4287baa9805ecbc3dbe86ea5dc2b96fd42699533192f2b89fd8fd73c"} Nov 25 09:43:15 crc kubenswrapper[4776]: I1125 09:43:15.743702 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerStarted","Data":"f0cdb71ed47fdaaa7eabd7f81c93a293616c6c022e1c4d8cef28b88de0dfb8d5"} Nov 25 09:43:19 crc kubenswrapper[4776]: I1125 09:43:19.061497 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2bs2f-config-x7wjq" Nov 25 09:43:19 crc kubenswrapper[4776]: I1125 09:43:19.185984 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-var-log-ovn\") pod \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\" (UID: \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\") " Nov 25 09:43:19 crc kubenswrapper[4776]: I1125 09:43:19.186109 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-additional-scripts\") pod \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\" (UID: \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\") " Nov 25 09:43:19 crc kubenswrapper[4776]: I1125 09:43:19.186144 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2k4h5\" (UniqueName: \"kubernetes.io/projected/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-kube-api-access-2k4h5\") pod \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\" (UID: \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\") " Nov 25 09:43:19 crc kubenswrapper[4776]: I1125 09:43:19.186144 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "c24cb0b1-8aba-4239-9829-05f2ad0a30d0" (UID: "c24cb0b1-8aba-4239-9829-05f2ad0a30d0"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:43:19 crc kubenswrapper[4776]: I1125 09:43:19.186192 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-scripts\") pod \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\" (UID: \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\") " Nov 25 09:43:19 crc kubenswrapper[4776]: I1125 09:43:19.186289 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-var-run\") pod \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\" (UID: \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\") " Nov 25 09:43:19 crc kubenswrapper[4776]: I1125 09:43:19.186373 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-var-run-ovn\") pod \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\" (UID: \"c24cb0b1-8aba-4239-9829-05f2ad0a30d0\") " Nov 25 09:43:19 crc kubenswrapper[4776]: I1125 09:43:19.186424 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-var-run" (OuterVolumeSpecName: "var-run") pod "c24cb0b1-8aba-4239-9829-05f2ad0a30d0" (UID: "c24cb0b1-8aba-4239-9829-05f2ad0a30d0"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:43:19 crc kubenswrapper[4776]: I1125 09:43:19.186478 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "c24cb0b1-8aba-4239-9829-05f2ad0a30d0" (UID: "c24cb0b1-8aba-4239-9829-05f2ad0a30d0"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:43:19 crc kubenswrapper[4776]: I1125 09:43:19.186742 4776 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:19 crc kubenswrapper[4776]: I1125 09:43:19.186767 4776 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:19 crc kubenswrapper[4776]: I1125 09:43:19.186779 4776 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-var-run\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:19 crc kubenswrapper[4776]: I1125 09:43:19.187469 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "c24cb0b1-8aba-4239-9829-05f2ad0a30d0" (UID: "c24cb0b1-8aba-4239-9829-05f2ad0a30d0"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:43:19 crc kubenswrapper[4776]: I1125 09:43:19.187751 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-scripts" (OuterVolumeSpecName: "scripts") pod "c24cb0b1-8aba-4239-9829-05f2ad0a30d0" (UID: "c24cb0b1-8aba-4239-9829-05f2ad0a30d0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:43:19 crc kubenswrapper[4776]: I1125 09:43:19.191679 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-kube-api-access-2k4h5" (OuterVolumeSpecName: "kube-api-access-2k4h5") pod "c24cb0b1-8aba-4239-9829-05f2ad0a30d0" (UID: "c24cb0b1-8aba-4239-9829-05f2ad0a30d0"). InnerVolumeSpecName "kube-api-access-2k4h5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:43:19 crc kubenswrapper[4776]: I1125 09:43:19.287756 4776 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:19 crc kubenswrapper[4776]: I1125 09:43:19.287796 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2k4h5\" (UniqueName: \"kubernetes.io/projected/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-kube-api-access-2k4h5\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:19 crc kubenswrapper[4776]: I1125 09:43:19.287815 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c24cb0b1-8aba-4239-9829-05f2ad0a30d0-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:19 crc kubenswrapper[4776]: I1125 09:43:19.774129 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2bs2f-config-x7wjq" event={"ID":"c24cb0b1-8aba-4239-9829-05f2ad0a30d0","Type":"ContainerDied","Data":"10a190f92dd317ee24723dcf578f43b2084fb953ca37e4e24bb2322a3541c0a7"} Nov 25 09:43:19 crc kubenswrapper[4776]: I1125 09:43:19.774190 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10a190f92dd317ee24723dcf578f43b2084fb953ca37e4e24bb2322a3541c0a7" Nov 25 09:43:19 crc kubenswrapper[4776]: I1125 09:43:19.774199 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2bs2f-config-x7wjq" Nov 25 09:43:20 crc kubenswrapper[4776]: I1125 09:43:20.171208 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-2bs2f-config-x7wjq"] Nov 25 09:43:20 crc kubenswrapper[4776]: I1125 09:43:20.178832 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-2bs2f-config-x7wjq"] Nov 25 09:43:20 crc kubenswrapper[4776]: I1125 09:43:20.677262 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 25 09:43:20 crc kubenswrapper[4776]: I1125 09:43:20.792371 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerStarted","Data":"2c713a54dd52d79687c4437832031c9afbb02bd373076897a51654ae360ff22a"} Nov 25 09:43:20 crc kubenswrapper[4776]: I1125 09:43:20.991826 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-6l262"] Nov 25 09:43:20 crc kubenswrapper[4776]: E1125 09:43:20.997095 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c24cb0b1-8aba-4239-9829-05f2ad0a30d0" containerName="ovn-config" Nov 25 09:43:20 crc kubenswrapper[4776]: I1125 09:43:20.997131 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="c24cb0b1-8aba-4239-9829-05f2ad0a30d0" containerName="ovn-config" Nov 25 09:43:20 crc kubenswrapper[4776]: I1125 09:43:20.997378 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="c24cb0b1-8aba-4239-9829-05f2ad0a30d0" containerName="ovn-config" Nov 25 09:43:20 crc kubenswrapper[4776]: I1125 09:43:20.997929 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-6l262" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.014983 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-6l262"] Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.078266 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-j6tqf"] Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.079507 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-j6tqf" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.100399 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-j6tqf"] Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.110291 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db0b-account-create-hssb4"] Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.111353 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db0b-account-create-hssb4" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.113259 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.118259 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/411d6157-0520-436e-b350-3a573e7ad984-operator-scripts\") pod \"barbican-db-create-6l262\" (UID: \"411d6157-0520-436e-b350-3a573e7ad984\") " pod="openstack/barbican-db-create-6l262" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.118577 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mj6jl\" (UniqueName: \"kubernetes.io/projected/411d6157-0520-436e-b350-3a573e7ad984-kube-api-access-mj6jl\") pod \"barbican-db-create-6l262\" (UID: \"411d6157-0520-436e-b350-3a573e7ad984\") " pod="openstack/barbican-db-create-6l262" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.141795 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db0b-account-create-hssb4"] Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.221640 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/411d6157-0520-436e-b350-3a573e7ad984-operator-scripts\") pod \"barbican-db-create-6l262\" (UID: \"411d6157-0520-436e-b350-3a573e7ad984\") " pod="openstack/barbican-db-create-6l262" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.221741 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89-operator-scripts\") pod \"barbican-db0b-account-create-hssb4\" (UID: \"83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89\") " pod="openstack/barbican-db0b-account-create-hssb4" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.221830 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mj6jl\" (UniqueName: \"kubernetes.io/projected/411d6157-0520-436e-b350-3a573e7ad984-kube-api-access-mj6jl\") pod \"barbican-db-create-6l262\" (UID: \"411d6157-0520-436e-b350-3a573e7ad984\") " pod="openstack/barbican-db-create-6l262" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.221866 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjf5j\" (UniqueName: \"kubernetes.io/projected/83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89-kube-api-access-kjf5j\") pod \"barbican-db0b-account-create-hssb4\" (UID: \"83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89\") " pod="openstack/barbican-db0b-account-create-hssb4" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.221895 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ca5a2dc7-b729-4038-9028-f08d37442b18-operator-scripts\") pod \"cinder-db-create-j6tqf\" (UID: \"ca5a2dc7-b729-4038-9028-f08d37442b18\") " pod="openstack/cinder-db-create-j6tqf" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.221918 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbngn\" (UniqueName: \"kubernetes.io/projected/ca5a2dc7-b729-4038-9028-f08d37442b18-kube-api-access-xbngn\") pod \"cinder-db-create-j6tqf\" (UID: \"ca5a2dc7-b729-4038-9028-f08d37442b18\") " pod="openstack/cinder-db-create-j6tqf" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.222018 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-774d-account-create-vbblg"] Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.222758 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/411d6157-0520-436e-b350-3a573e7ad984-operator-scripts\") pod \"barbican-db-create-6l262\" (UID: \"411d6157-0520-436e-b350-3a573e7ad984\") " pod="openstack/barbican-db-create-6l262" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.223295 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-774d-account-create-vbblg" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.228468 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.234927 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-774d-account-create-vbblg"] Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.263424 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mj6jl\" (UniqueName: \"kubernetes.io/projected/411d6157-0520-436e-b350-3a573e7ad984-kube-api-access-mj6jl\") pod \"barbican-db-create-6l262\" (UID: \"411d6157-0520-436e-b350-3a573e7ad984\") " pod="openstack/barbican-db-create-6l262" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.270484 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-dvd6n"] Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.271743 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-dvd6n" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.282031 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.282623 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gtj9p" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.282937 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.283103 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.291293 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-dvd6n"] Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.323371 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5705e80c-320b-49da-8c10-743627c8dbdb-operator-scripts\") pod \"cinder-774d-account-create-vbblg\" (UID: \"5705e80c-320b-49da-8c10-743627c8dbdb\") " pod="openstack/cinder-774d-account-create-vbblg" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.323473 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjf5j\" (UniqueName: \"kubernetes.io/projected/83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89-kube-api-access-kjf5j\") pod \"barbican-db0b-account-create-hssb4\" (UID: \"83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89\") " pod="openstack/barbican-db0b-account-create-hssb4" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.323499 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ca5a2dc7-b729-4038-9028-f08d37442b18-operator-scripts\") pod \"cinder-db-create-j6tqf\" (UID: \"ca5a2dc7-b729-4038-9028-f08d37442b18\") " pod="openstack/cinder-db-create-j6tqf" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.323520 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbngn\" (UniqueName: \"kubernetes.io/projected/ca5a2dc7-b729-4038-9028-f08d37442b18-kube-api-access-xbngn\") pod \"cinder-db-create-j6tqf\" (UID: \"ca5a2dc7-b729-4038-9028-f08d37442b18\") " pod="openstack/cinder-db-create-j6tqf" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.323588 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6925\" (UniqueName: \"kubernetes.io/projected/5705e80c-320b-49da-8c10-743627c8dbdb-kube-api-access-p6925\") pod \"cinder-774d-account-create-vbblg\" (UID: \"5705e80c-320b-49da-8c10-743627c8dbdb\") " pod="openstack/cinder-774d-account-create-vbblg" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.323674 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89-operator-scripts\") pod \"barbican-db0b-account-create-hssb4\" (UID: \"83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89\") " pod="openstack/barbican-db0b-account-create-hssb4" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.325156 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ca5a2dc7-b729-4038-9028-f08d37442b18-operator-scripts\") pod \"cinder-db-create-j6tqf\" (UID: \"ca5a2dc7-b729-4038-9028-f08d37442b18\") " pod="openstack/cinder-db-create-j6tqf" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.325422 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89-operator-scripts\") pod \"barbican-db0b-account-create-hssb4\" (UID: \"83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89\") " pod="openstack/barbican-db0b-account-create-hssb4" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.330775 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-6l262" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.351111 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjf5j\" (UniqueName: \"kubernetes.io/projected/83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89-kube-api-access-kjf5j\") pod \"barbican-db0b-account-create-hssb4\" (UID: \"83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89\") " pod="openstack/barbican-db0b-account-create-hssb4" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.358801 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbngn\" (UniqueName: \"kubernetes.io/projected/ca5a2dc7-b729-4038-9028-f08d37442b18-kube-api-access-xbngn\") pod \"cinder-db-create-j6tqf\" (UID: \"ca5a2dc7-b729-4038-9028-f08d37442b18\") " pod="openstack/cinder-db-create-j6tqf" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.398510 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-j6tqf" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.405216 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-mfnw2"] Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.406494 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-mfnw2" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.424612 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e107d317-acb4-46a2-8f19-836572e8904e-combined-ca-bundle\") pod \"keystone-db-sync-dvd6n\" (UID: \"e107d317-acb4-46a2-8f19-836572e8904e\") " pod="openstack/keystone-db-sync-dvd6n" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.424692 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfm9w\" (UniqueName: \"kubernetes.io/projected/e107d317-acb4-46a2-8f19-836572e8904e-kube-api-access-sfm9w\") pod \"keystone-db-sync-dvd6n\" (UID: \"e107d317-acb4-46a2-8f19-836572e8904e\") " pod="openstack/keystone-db-sync-dvd6n" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.424734 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6925\" (UniqueName: \"kubernetes.io/projected/5705e80c-320b-49da-8c10-743627c8dbdb-kube-api-access-p6925\") pod \"cinder-774d-account-create-vbblg\" (UID: \"5705e80c-320b-49da-8c10-743627c8dbdb\") " pod="openstack/cinder-774d-account-create-vbblg" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.424838 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5705e80c-320b-49da-8c10-743627c8dbdb-operator-scripts\") pod \"cinder-774d-account-create-vbblg\" (UID: \"5705e80c-320b-49da-8c10-743627c8dbdb\") " pod="openstack/cinder-774d-account-create-vbblg" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.424898 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e107d317-acb4-46a2-8f19-836572e8904e-config-data\") pod \"keystone-db-sync-dvd6n\" (UID: \"e107d317-acb4-46a2-8f19-836572e8904e\") " pod="openstack/keystone-db-sync-dvd6n" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.431047 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5705e80c-320b-49da-8c10-743627c8dbdb-operator-scripts\") pod \"cinder-774d-account-create-vbblg\" (UID: \"5705e80c-320b-49da-8c10-743627c8dbdb\") " pod="openstack/cinder-774d-account-create-vbblg" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.442423 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-mfnw2"] Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.454003 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6925\" (UniqueName: \"kubernetes.io/projected/5705e80c-320b-49da-8c10-743627c8dbdb-kube-api-access-p6925\") pod \"cinder-774d-account-create-vbblg\" (UID: \"5705e80c-320b-49da-8c10-743627c8dbdb\") " pod="openstack/cinder-774d-account-create-vbblg" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.489805 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5151-account-create-866rm"] Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.491217 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5151-account-create-866rm" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.498585 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db0b-account-create-hssb4" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.499176 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.504789 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5151-account-create-866rm"] Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.526539 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e107d317-acb4-46a2-8f19-836572e8904e-config-data\") pod \"keystone-db-sync-dvd6n\" (UID: \"e107d317-acb4-46a2-8f19-836572e8904e\") " pod="openstack/keystone-db-sync-dvd6n" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.526598 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e107d317-acb4-46a2-8f19-836572e8904e-combined-ca-bundle\") pod \"keystone-db-sync-dvd6n\" (UID: \"e107d317-acb4-46a2-8f19-836572e8904e\") " pod="openstack/keystone-db-sync-dvd6n" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.526651 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfm9w\" (UniqueName: \"kubernetes.io/projected/e107d317-acb4-46a2-8f19-836572e8904e-kube-api-access-sfm9w\") pod \"keystone-db-sync-dvd6n\" (UID: \"e107d317-acb4-46a2-8f19-836572e8904e\") " pod="openstack/keystone-db-sync-dvd6n" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.526708 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83e76214-d644-4e9b-ab3a-2cbcae40cb23-operator-scripts\") pod \"neutron-db-create-mfnw2\" (UID: \"83e76214-d644-4e9b-ab3a-2cbcae40cb23\") " pod="openstack/neutron-db-create-mfnw2" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.526775 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxzp6\" (UniqueName: \"kubernetes.io/projected/83e76214-d644-4e9b-ab3a-2cbcae40cb23-kube-api-access-xxzp6\") pod \"neutron-db-create-mfnw2\" (UID: \"83e76214-d644-4e9b-ab3a-2cbcae40cb23\") " pod="openstack/neutron-db-create-mfnw2" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.530968 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e107d317-acb4-46a2-8f19-836572e8904e-config-data\") pod \"keystone-db-sync-dvd6n\" (UID: \"e107d317-acb4-46a2-8f19-836572e8904e\") " pod="openstack/keystone-db-sync-dvd6n" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.532726 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e107d317-acb4-46a2-8f19-836572e8904e-combined-ca-bundle\") pod \"keystone-db-sync-dvd6n\" (UID: \"e107d317-acb4-46a2-8f19-836572e8904e\") " pod="openstack/keystone-db-sync-dvd6n" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.548578 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-774d-account-create-vbblg" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.555678 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfm9w\" (UniqueName: \"kubernetes.io/projected/e107d317-acb4-46a2-8f19-836572e8904e-kube-api-access-sfm9w\") pod \"keystone-db-sync-dvd6n\" (UID: \"e107d317-acb4-46a2-8f19-836572e8904e\") " pod="openstack/keystone-db-sync-dvd6n" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.630109 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzkz9\" (UniqueName: \"kubernetes.io/projected/9e7d1dce-289c-409f-a68e-b214f8f4cc14-kube-api-access-nzkz9\") pod \"neutron-5151-account-create-866rm\" (UID: \"9e7d1dce-289c-409f-a68e-b214f8f4cc14\") " pod="openstack/neutron-5151-account-create-866rm" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.630184 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83e76214-d644-4e9b-ab3a-2cbcae40cb23-operator-scripts\") pod \"neutron-db-create-mfnw2\" (UID: \"83e76214-d644-4e9b-ab3a-2cbcae40cb23\") " pod="openstack/neutron-db-create-mfnw2" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.630209 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e7d1dce-289c-409f-a68e-b214f8f4cc14-operator-scripts\") pod \"neutron-5151-account-create-866rm\" (UID: \"9e7d1dce-289c-409f-a68e-b214f8f4cc14\") " pod="openstack/neutron-5151-account-create-866rm" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.630273 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxzp6\" (UniqueName: \"kubernetes.io/projected/83e76214-d644-4e9b-ab3a-2cbcae40cb23-kube-api-access-xxzp6\") pod \"neutron-db-create-mfnw2\" (UID: \"83e76214-d644-4e9b-ab3a-2cbcae40cb23\") " pod="openstack/neutron-db-create-mfnw2" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.631087 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83e76214-d644-4e9b-ab3a-2cbcae40cb23-operator-scripts\") pod \"neutron-db-create-mfnw2\" (UID: \"83e76214-d644-4e9b-ab3a-2cbcae40cb23\") " pod="openstack/neutron-db-create-mfnw2" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.669454 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-dvd6n" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.688084 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxzp6\" (UniqueName: \"kubernetes.io/projected/83e76214-d644-4e9b-ab3a-2cbcae40cb23-kube-api-access-xxzp6\") pod \"neutron-db-create-mfnw2\" (UID: \"83e76214-d644-4e9b-ab3a-2cbcae40cb23\") " pod="openstack/neutron-db-create-mfnw2" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.693989 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c24cb0b1-8aba-4239-9829-05f2ad0a30d0" path="/var/lib/kubelet/pods/c24cb0b1-8aba-4239-9829-05f2ad0a30d0/volumes" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.723290 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.731793 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzkz9\" (UniqueName: \"kubernetes.io/projected/9e7d1dce-289c-409f-a68e-b214f8f4cc14-kube-api-access-nzkz9\") pod \"neutron-5151-account-create-866rm\" (UID: \"9e7d1dce-289c-409f-a68e-b214f8f4cc14\") " pod="openstack/neutron-5151-account-create-866rm" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.731868 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e7d1dce-289c-409f-a68e-b214f8f4cc14-operator-scripts\") pod \"neutron-5151-account-create-866rm\" (UID: \"9e7d1dce-289c-409f-a68e-b214f8f4cc14\") " pod="openstack/neutron-5151-account-create-866rm" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.732789 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e7d1dce-289c-409f-a68e-b214f8f4cc14-operator-scripts\") pod \"neutron-5151-account-create-866rm\" (UID: \"9e7d1dce-289c-409f-a68e-b214f8f4cc14\") " pod="openstack/neutron-5151-account-create-866rm" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.742025 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-mfnw2" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.769404 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzkz9\" (UniqueName: \"kubernetes.io/projected/9e7d1dce-289c-409f-a68e-b214f8f4cc14-kube-api-access-nzkz9\") pod \"neutron-5151-account-create-866rm\" (UID: \"9e7d1dce-289c-409f-a68e-b214f8f4cc14\") " pod="openstack/neutron-5151-account-create-866rm" Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.875548 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerStarted","Data":"49bb08566118dc9e19698c8156598960f4778c7e09f832e3eeeaa9b6d4fc9a62"} Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.875596 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerStarted","Data":"670e0e074022b9b3b944ddaeb78954a059e4c4b8e2f7af4e2f91000daddc1263"} Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.875608 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerStarted","Data":"a5e5c6ad114abc906cc7868aa09dc3d1a6a83ffb494d14de45deba5107b01b02"} Nov 25 09:43:21 crc kubenswrapper[4776]: I1125 09:43:21.962403 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5151-account-create-866rm" Nov 25 09:43:22 crc kubenswrapper[4776]: I1125 09:43:22.346303 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-6l262"] Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:22.486416 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-j6tqf"] Nov 25 09:43:23 crc kubenswrapper[4776]: W1125 09:43:22.500174 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca5a2dc7_b729_4038_9028_f08d37442b18.slice/crio-af139e177c74f2ba7dd495988f8272c7a96466d5e95ffe6c7aeb10082da073b8 WatchSource:0}: Error finding container af139e177c74f2ba7dd495988f8272c7a96466d5e95ffe6c7aeb10082da073b8: Status 404 returned error can't find the container with id af139e177c74f2ba7dd495988f8272c7a96466d5e95ffe6c7aeb10082da073b8 Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:22.561272 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db0b-account-create-hssb4"] Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:22.598475 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-774d-account-create-vbblg"] Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:22.901979 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerStarted","Data":"faa52779c1f67996100c644db4f91149b383bf2ef821c114ac1e92b3b697987f"} Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:22.902333 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerStarted","Data":"152026f7bf460a5258df8a4110443a6e694c32ee892ab2dcb9469e4391545baa"} Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:22.902349 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerStarted","Data":"8e84b91c7719175c8908ad8fa6987a175f46744663fe91972c8bbb602ae871d8"} Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:22.909621 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-774d-account-create-vbblg" event={"ID":"5705e80c-320b-49da-8c10-743627c8dbdb","Type":"ContainerStarted","Data":"92b79dbb4ed078714aec0c9d66fe7c33b71b68f0384cad8a81c0c460830f785d"} Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:22.909672 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-774d-account-create-vbblg" event={"ID":"5705e80c-320b-49da-8c10-743627c8dbdb","Type":"ContainerStarted","Data":"bc88670ed04f5a10dba4a006f5149f89fd8f00b11eb8923a2b803e581aa1f9ea"} Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:22.914530 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db0b-account-create-hssb4" event={"ID":"83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89","Type":"ContainerStarted","Data":"fcc2df6898ae37ac42d535f5a9578c9c98363157deb46bb5d8844b5320228fca"} Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:22.914602 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db0b-account-create-hssb4" event={"ID":"83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89","Type":"ContainerStarted","Data":"980bf89aefad66fdf27ce8938bf36f86873601454fb8e109d985da8fe70de95c"} Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:22.925623 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-6l262" event={"ID":"411d6157-0520-436e-b350-3a573e7ad984","Type":"ContainerStarted","Data":"fe29f80314d191348c18f899e06138d8db881e961c44e2efa613f33717c12d39"} Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:22.925653 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-6l262" event={"ID":"411d6157-0520-436e-b350-3a573e7ad984","Type":"ContainerStarted","Data":"8c9a2cf0254c23cd1d016e1b65a192bd1a3d19c75b818320e92efe9c567e6ebd"} Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:22.937976 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-j6tqf" event={"ID":"ca5a2dc7-b729-4038-9028-f08d37442b18","Type":"ContainerStarted","Data":"aeee9511209e1b516d5a4fa252549d9534596d3569e1ff114ec890eef5d6d49d"} Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:22.938015 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-j6tqf" event={"ID":"ca5a2dc7-b729-4038-9028-f08d37442b18","Type":"ContainerStarted","Data":"af139e177c74f2ba7dd495988f8272c7a96466d5e95ffe6c7aeb10082da073b8"} Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:22.951822 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=20.829645067 podStartE2EDuration="36.95180656s" podCreationTimestamp="2025-11-25 09:42:46 +0000 UTC" firstStartedPulling="2025-11-25 09:43:04.321778851 +0000 UTC m=+1129.362838404" lastFinishedPulling="2025-11-25 09:43:20.443940354 +0000 UTC m=+1145.484999897" observedRunningTime="2025-11-25 09:43:22.950089497 +0000 UTC m=+1147.991149050" watchObservedRunningTime="2025-11-25 09:43:22.95180656 +0000 UTC m=+1147.992866113" Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:22.994694 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db0b-account-create-hssb4" podStartSLOduration=1.994675416 podStartE2EDuration="1.994675416s" podCreationTimestamp="2025-11-25 09:43:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:43:22.993281831 +0000 UTC m=+1148.034341404" watchObservedRunningTime="2025-11-25 09:43:22.994675416 +0000 UTC m=+1148.035734969" Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.073769 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-774d-account-create-vbblg" podStartSLOduration=2.073751349 podStartE2EDuration="2.073751349s" podCreationTimestamp="2025-11-25 09:43:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:43:23.022675148 +0000 UTC m=+1148.063734701" watchObservedRunningTime="2025-11-25 09:43:23.073751349 +0000 UTC m=+1148.114810902" Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.083157 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-j6tqf" podStartSLOduration=2.083137704 podStartE2EDuration="2.083137704s" podCreationTimestamp="2025-11-25 09:43:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:43:23.045107351 +0000 UTC m=+1148.086166904" watchObservedRunningTime="2025-11-25 09:43:23.083137704 +0000 UTC m=+1148.124197257" Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.088733 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-6l262" podStartSLOduration=3.088717644 podStartE2EDuration="3.088717644s" podCreationTimestamp="2025-11-25 09:43:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:43:23.07737696 +0000 UTC m=+1148.118436513" watchObservedRunningTime="2025-11-25 09:43:23.088717644 +0000 UTC m=+1148.129777197" Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.127240 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-dvd6n"] Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.165004 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-mfnw2"] Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.288786 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56766df65f-gcd25"] Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.290485 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56766df65f-gcd25" Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.294504 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.323117 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56766df65f-gcd25"] Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.368844 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5151-account-create-866rm"] Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.376570 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-dns-svc\") pod \"dnsmasq-dns-56766df65f-gcd25\" (UID: \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\") " pod="openstack/dnsmasq-dns-56766df65f-gcd25" Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.376942 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md6kl\" (UniqueName: \"kubernetes.io/projected/7a2a0530-6fc7-4297-a8ef-0eb391321b15-kube-api-access-md6kl\") pod \"dnsmasq-dns-56766df65f-gcd25\" (UID: \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\") " pod="openstack/dnsmasq-dns-56766df65f-gcd25" Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.376974 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-dns-swift-storage-0\") pod \"dnsmasq-dns-56766df65f-gcd25\" (UID: \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\") " pod="openstack/dnsmasq-dns-56766df65f-gcd25" Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.376998 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-config\") pod \"dnsmasq-dns-56766df65f-gcd25\" (UID: \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\") " pod="openstack/dnsmasq-dns-56766df65f-gcd25" Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.377074 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-ovsdbserver-nb\") pod \"dnsmasq-dns-56766df65f-gcd25\" (UID: \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\") " pod="openstack/dnsmasq-dns-56766df65f-gcd25" Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.377130 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-ovsdbserver-sb\") pod \"dnsmasq-dns-56766df65f-gcd25\" (UID: \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\") " pod="openstack/dnsmasq-dns-56766df65f-gcd25" Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.479035 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md6kl\" (UniqueName: \"kubernetes.io/projected/7a2a0530-6fc7-4297-a8ef-0eb391321b15-kube-api-access-md6kl\") pod \"dnsmasq-dns-56766df65f-gcd25\" (UID: \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\") " pod="openstack/dnsmasq-dns-56766df65f-gcd25" Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.479369 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-dns-swift-storage-0\") pod \"dnsmasq-dns-56766df65f-gcd25\" (UID: \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\") " pod="openstack/dnsmasq-dns-56766df65f-gcd25" Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.480216 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-dns-swift-storage-0\") pod \"dnsmasq-dns-56766df65f-gcd25\" (UID: \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\") " pod="openstack/dnsmasq-dns-56766df65f-gcd25" Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.480399 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-config\") pod \"dnsmasq-dns-56766df65f-gcd25\" (UID: \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\") " pod="openstack/dnsmasq-dns-56766df65f-gcd25" Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.480997 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-config\") pod \"dnsmasq-dns-56766df65f-gcd25\" (UID: \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\") " pod="openstack/dnsmasq-dns-56766df65f-gcd25" Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.481114 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-ovsdbserver-nb\") pod \"dnsmasq-dns-56766df65f-gcd25\" (UID: \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\") " pod="openstack/dnsmasq-dns-56766df65f-gcd25" Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.481658 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-ovsdbserver-nb\") pod \"dnsmasq-dns-56766df65f-gcd25\" (UID: \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\") " pod="openstack/dnsmasq-dns-56766df65f-gcd25" Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.482504 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-ovsdbserver-sb\") pod \"dnsmasq-dns-56766df65f-gcd25\" (UID: \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\") " pod="openstack/dnsmasq-dns-56766df65f-gcd25" Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.483081 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-ovsdbserver-sb\") pod \"dnsmasq-dns-56766df65f-gcd25\" (UID: \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\") " pod="openstack/dnsmasq-dns-56766df65f-gcd25" Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.483742 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-dns-svc\") pod \"dnsmasq-dns-56766df65f-gcd25\" (UID: \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\") " pod="openstack/dnsmasq-dns-56766df65f-gcd25" Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.483167 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-dns-svc\") pod \"dnsmasq-dns-56766df65f-gcd25\" (UID: \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\") " pod="openstack/dnsmasq-dns-56766df65f-gcd25" Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.499363 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md6kl\" (UniqueName: \"kubernetes.io/projected/7a2a0530-6fc7-4297-a8ef-0eb391321b15-kube-api-access-md6kl\") pod \"dnsmasq-dns-56766df65f-gcd25\" (UID: \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\") " pod="openstack/dnsmasq-dns-56766df65f-gcd25" Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.625259 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56766df65f-gcd25" Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.929594 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56766df65f-gcd25"] Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.985717 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-dvd6n" event={"ID":"e107d317-acb4-46a2-8f19-836572e8904e","Type":"ContainerStarted","Data":"94a1444a112f88890fa525e35d110b8d1637c0b8dcb7f6f8afad671025e23b14"} Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.989786 4776 generic.go:334] "Generic (PLEG): container finished" podID="5705e80c-320b-49da-8c10-743627c8dbdb" containerID="92b79dbb4ed078714aec0c9d66fe7c33b71b68f0384cad8a81c0c460830f785d" exitCode=0 Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.989864 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-774d-account-create-vbblg" event={"ID":"5705e80c-320b-49da-8c10-743627c8dbdb","Type":"ContainerDied","Data":"92b79dbb4ed078714aec0c9d66fe7c33b71b68f0384cad8a81c0c460830f785d"} Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.998405 4776 generic.go:334] "Generic (PLEG): container finished" podID="83e76214-d644-4e9b-ab3a-2cbcae40cb23" containerID="bbe973c26582167ba2a33fb740db5cdcd63e979ff3085ac9bf82ee71344b4aff" exitCode=0 Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.998512 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-mfnw2" event={"ID":"83e76214-d644-4e9b-ab3a-2cbcae40cb23","Type":"ContainerDied","Data":"bbe973c26582167ba2a33fb740db5cdcd63e979ff3085ac9bf82ee71344b4aff"} Nov 25 09:43:23 crc kubenswrapper[4776]: I1125 09:43:23.998541 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-mfnw2" event={"ID":"83e76214-d644-4e9b-ab3a-2cbcae40cb23","Type":"ContainerStarted","Data":"bba8accfa0d6d62918f831e0e0b4df8c9c031c65130fbeb38c949abd9195035c"} Nov 25 09:43:24 crc kubenswrapper[4776]: I1125 09:43:24.002556 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56766df65f-gcd25" event={"ID":"7a2a0530-6fc7-4297-a8ef-0eb391321b15","Type":"ContainerStarted","Data":"a556397e5cf4747e94a3fa9e3bb06249880f3dac10018a509aaac6fed08f62a7"} Nov 25 09:43:24 crc kubenswrapper[4776]: I1125 09:43:24.018940 4776 generic.go:334] "Generic (PLEG): container finished" podID="83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89" containerID="fcc2df6898ae37ac42d535f5a9578c9c98363157deb46bb5d8844b5320228fca" exitCode=0 Nov 25 09:43:24 crc kubenswrapper[4776]: I1125 09:43:24.019273 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db0b-account-create-hssb4" event={"ID":"83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89","Type":"ContainerDied","Data":"fcc2df6898ae37ac42d535f5a9578c9c98363157deb46bb5d8844b5320228fca"} Nov 25 09:43:24 crc kubenswrapper[4776]: I1125 09:43:24.021057 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5151-account-create-866rm" event={"ID":"9e7d1dce-289c-409f-a68e-b214f8f4cc14","Type":"ContainerStarted","Data":"6048c18060333e95c2e0be12fac988647d61e6f042d149f41998c255e6566c91"} Nov 25 09:43:24 crc kubenswrapper[4776]: I1125 09:43:24.021094 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5151-account-create-866rm" event={"ID":"9e7d1dce-289c-409f-a68e-b214f8f4cc14","Type":"ContainerStarted","Data":"96eb141651e8e4ffb7a297120c6bad3309fb20a2f39a662c4f6799fe18a79db8"} Nov 25 09:43:24 crc kubenswrapper[4776]: I1125 09:43:24.025597 4776 generic.go:334] "Generic (PLEG): container finished" podID="411d6157-0520-436e-b350-3a573e7ad984" containerID="fe29f80314d191348c18f899e06138d8db881e961c44e2efa613f33717c12d39" exitCode=0 Nov 25 09:43:24 crc kubenswrapper[4776]: I1125 09:43:24.025922 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-6l262" event={"ID":"411d6157-0520-436e-b350-3a573e7ad984","Type":"ContainerDied","Data":"fe29f80314d191348c18f899e06138d8db881e961c44e2efa613f33717c12d39"} Nov 25 09:43:24 crc kubenswrapper[4776]: I1125 09:43:24.038576 4776 generic.go:334] "Generic (PLEG): container finished" podID="ca5a2dc7-b729-4038-9028-f08d37442b18" containerID="aeee9511209e1b516d5a4fa252549d9534596d3569e1ff114ec890eef5d6d49d" exitCode=0 Nov 25 09:43:24 crc kubenswrapper[4776]: I1125 09:43:24.039407 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-j6tqf" event={"ID":"ca5a2dc7-b729-4038-9028-f08d37442b18","Type":"ContainerDied","Data":"aeee9511209e1b516d5a4fa252549d9534596d3569e1ff114ec890eef5d6d49d"} Nov 25 09:43:24 crc kubenswrapper[4776]: I1125 09:43:24.073724 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5151-account-create-866rm" podStartSLOduration=3.073577195 podStartE2EDuration="3.073577195s" podCreationTimestamp="2025-11-25 09:43:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:43:24.072905898 +0000 UTC m=+1149.113965451" watchObservedRunningTime="2025-11-25 09:43:24.073577195 +0000 UTC m=+1149.114636748" Nov 25 09:43:25 crc kubenswrapper[4776]: I1125 09:43:25.048337 4776 generic.go:334] "Generic (PLEG): container finished" podID="9e7d1dce-289c-409f-a68e-b214f8f4cc14" containerID="6048c18060333e95c2e0be12fac988647d61e6f042d149f41998c255e6566c91" exitCode=0 Nov 25 09:43:25 crc kubenswrapper[4776]: I1125 09:43:25.048411 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5151-account-create-866rm" event={"ID":"9e7d1dce-289c-409f-a68e-b214f8f4cc14","Type":"ContainerDied","Data":"6048c18060333e95c2e0be12fac988647d61e6f042d149f41998c255e6566c91"} Nov 25 09:43:25 crc kubenswrapper[4776]: I1125 09:43:25.052047 4776 generic.go:334] "Generic (PLEG): container finished" podID="7a2a0530-6fc7-4297-a8ef-0eb391321b15" containerID="c94a37851183d8b0aa46947d36df2bdc5de0cc80ade7bf5e8378bf93b49ba161" exitCode=0 Nov 25 09:43:25 crc kubenswrapper[4776]: I1125 09:43:25.052358 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56766df65f-gcd25" event={"ID":"7a2a0530-6fc7-4297-a8ef-0eb391321b15","Type":"ContainerDied","Data":"c94a37851183d8b0aa46947d36df2bdc5de0cc80ade7bf5e8378bf93b49ba161"} Nov 25 09:43:25 crc kubenswrapper[4776]: I1125 09:43:25.430273 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-774d-account-create-vbblg" Nov 25 09:43:25 crc kubenswrapper[4776]: I1125 09:43:25.524813 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5705e80c-320b-49da-8c10-743627c8dbdb-operator-scripts\") pod \"5705e80c-320b-49da-8c10-743627c8dbdb\" (UID: \"5705e80c-320b-49da-8c10-743627c8dbdb\") " Nov 25 09:43:25 crc kubenswrapper[4776]: I1125 09:43:25.524959 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6925\" (UniqueName: \"kubernetes.io/projected/5705e80c-320b-49da-8c10-743627c8dbdb-kube-api-access-p6925\") pod \"5705e80c-320b-49da-8c10-743627c8dbdb\" (UID: \"5705e80c-320b-49da-8c10-743627c8dbdb\") " Nov 25 09:43:25 crc kubenswrapper[4776]: I1125 09:43:25.525546 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5705e80c-320b-49da-8c10-743627c8dbdb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5705e80c-320b-49da-8c10-743627c8dbdb" (UID: "5705e80c-320b-49da-8c10-743627c8dbdb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:43:25 crc kubenswrapper[4776]: I1125 09:43:25.540248 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5705e80c-320b-49da-8c10-743627c8dbdb-kube-api-access-p6925" (OuterVolumeSpecName: "kube-api-access-p6925") pod "5705e80c-320b-49da-8c10-743627c8dbdb" (UID: "5705e80c-320b-49da-8c10-743627c8dbdb"). InnerVolumeSpecName "kube-api-access-p6925". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:43:25 crc kubenswrapper[4776]: I1125 09:43:25.629397 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5705e80c-320b-49da-8c10-743627c8dbdb-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:25 crc kubenswrapper[4776]: I1125 09:43:25.629434 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6925\" (UniqueName: \"kubernetes.io/projected/5705e80c-320b-49da-8c10-743627c8dbdb-kube-api-access-p6925\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:26 crc kubenswrapper[4776]: I1125 09:43:26.075180 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-774d-account-create-vbblg" event={"ID":"5705e80c-320b-49da-8c10-743627c8dbdb","Type":"ContainerDied","Data":"bc88670ed04f5a10dba4a006f5149f89fd8f00b11eb8923a2b803e581aa1f9ea"} Nov 25 09:43:26 crc kubenswrapper[4776]: I1125 09:43:26.075228 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc88670ed04f5a10dba4a006f5149f89fd8f00b11eb8923a2b803e581aa1f9ea" Nov 25 09:43:26 crc kubenswrapper[4776]: I1125 09:43:26.075207 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-774d-account-create-vbblg" Nov 25 09:43:26 crc kubenswrapper[4776]: I1125 09:43:26.094919 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56766df65f-gcd25" event={"ID":"7a2a0530-6fc7-4297-a8ef-0eb391321b15","Type":"ContainerStarted","Data":"139977ade9dc8dfdfa2e5d63de9d6408a2e51495e91a19593b0db46fb322b3f5"} Nov 25 09:43:26 crc kubenswrapper[4776]: I1125 09:43:26.094997 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56766df65f-gcd25" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.057889 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-6l262" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.064327 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db0b-account-create-hssb4" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.074260 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-j6tqf" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.076567 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56766df65f-gcd25" podStartSLOduration=5.07654723 podStartE2EDuration="5.07654723s" podCreationTimestamp="2025-11-25 09:43:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:43:26.119867785 +0000 UTC m=+1151.160927358" watchObservedRunningTime="2025-11-25 09:43:28.07654723 +0000 UTC m=+1153.117606773" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.108508 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5151-account-create-866rm" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.114977 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-6l262" event={"ID":"411d6157-0520-436e-b350-3a573e7ad984","Type":"ContainerDied","Data":"8c9a2cf0254c23cd1d016e1b65a192bd1a3d19c75b818320e92efe9c567e6ebd"} Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.115026 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c9a2cf0254c23cd1d016e1b65a192bd1a3d19c75b818320e92efe9c567e6ebd" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.115112 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-6l262" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.120750 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-j6tqf" event={"ID":"ca5a2dc7-b729-4038-9028-f08d37442b18","Type":"ContainerDied","Data":"af139e177c74f2ba7dd495988f8272c7a96466d5e95ffe6c7aeb10082da073b8"} Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.120788 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af139e177c74f2ba7dd495988f8272c7a96466d5e95ffe6c7aeb10082da073b8" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.120846 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-j6tqf" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.121279 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-mfnw2" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.127041 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-mfnw2" event={"ID":"83e76214-d644-4e9b-ab3a-2cbcae40cb23","Type":"ContainerDied","Data":"bba8accfa0d6d62918f831e0e0b4df8c9c031c65130fbeb38c949abd9195035c"} Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.127120 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bba8accfa0d6d62918f831e0e0b4df8c9c031c65130fbeb38c949abd9195035c" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.128919 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5151-account-create-866rm" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.129100 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5151-account-create-866rm" event={"ID":"9e7d1dce-289c-409f-a68e-b214f8f4cc14","Type":"ContainerDied","Data":"96eb141651e8e4ffb7a297120c6bad3309fb20a2f39a662c4f6799fe18a79db8"} Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.129137 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96eb141651e8e4ffb7a297120c6bad3309fb20a2f39a662c4f6799fe18a79db8" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.130697 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db0b-account-create-hssb4" event={"ID":"83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89","Type":"ContainerDied","Data":"980bf89aefad66fdf27ce8938bf36f86873601454fb8e109d985da8fe70de95c"} Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.130727 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="980bf89aefad66fdf27ce8938bf36f86873601454fb8e109d985da8fe70de95c" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.130739 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db0b-account-create-hssb4" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.178766 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ca5a2dc7-b729-4038-9028-f08d37442b18-operator-scripts\") pod \"ca5a2dc7-b729-4038-9028-f08d37442b18\" (UID: \"ca5a2dc7-b729-4038-9028-f08d37442b18\") " Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.178805 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbngn\" (UniqueName: \"kubernetes.io/projected/ca5a2dc7-b729-4038-9028-f08d37442b18-kube-api-access-xbngn\") pod \"ca5a2dc7-b729-4038-9028-f08d37442b18\" (UID: \"ca5a2dc7-b729-4038-9028-f08d37442b18\") " Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.178842 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mj6jl\" (UniqueName: \"kubernetes.io/projected/411d6157-0520-436e-b350-3a573e7ad984-kube-api-access-mj6jl\") pod \"411d6157-0520-436e-b350-3a573e7ad984\" (UID: \"411d6157-0520-436e-b350-3a573e7ad984\") " Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.178975 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e7d1dce-289c-409f-a68e-b214f8f4cc14-operator-scripts\") pod \"9e7d1dce-289c-409f-a68e-b214f8f4cc14\" (UID: \"9e7d1dce-289c-409f-a68e-b214f8f4cc14\") " Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.179032 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxzp6\" (UniqueName: \"kubernetes.io/projected/83e76214-d644-4e9b-ab3a-2cbcae40cb23-kube-api-access-xxzp6\") pod \"83e76214-d644-4e9b-ab3a-2cbcae40cb23\" (UID: \"83e76214-d644-4e9b-ab3a-2cbcae40cb23\") " Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.179104 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89-operator-scripts\") pod \"83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89\" (UID: \"83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89\") " Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.179145 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83e76214-d644-4e9b-ab3a-2cbcae40cb23-operator-scripts\") pod \"83e76214-d644-4e9b-ab3a-2cbcae40cb23\" (UID: \"83e76214-d644-4e9b-ab3a-2cbcae40cb23\") " Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.179188 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjf5j\" (UniqueName: \"kubernetes.io/projected/83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89-kube-api-access-kjf5j\") pod \"83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89\" (UID: \"83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89\") " Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.179262 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/411d6157-0520-436e-b350-3a573e7ad984-operator-scripts\") pod \"411d6157-0520-436e-b350-3a573e7ad984\" (UID: \"411d6157-0520-436e-b350-3a573e7ad984\") " Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.179716 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89" (UID: "83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.179719 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e7d1dce-289c-409f-a68e-b214f8f4cc14-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9e7d1dce-289c-409f-a68e-b214f8f4cc14" (UID: "9e7d1dce-289c-409f-a68e-b214f8f4cc14"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.180121 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83e76214-d644-4e9b-ab3a-2cbcae40cb23-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "83e76214-d644-4e9b-ab3a-2cbcae40cb23" (UID: "83e76214-d644-4e9b-ab3a-2cbcae40cb23"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.180200 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzkz9\" (UniqueName: \"kubernetes.io/projected/9e7d1dce-289c-409f-a68e-b214f8f4cc14-kube-api-access-nzkz9\") pod \"9e7d1dce-289c-409f-a68e-b214f8f4cc14\" (UID: \"9e7d1dce-289c-409f-a68e-b214f8f4cc14\") " Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.179908 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca5a2dc7-b729-4038-9028-f08d37442b18-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ca5a2dc7-b729-4038-9028-f08d37442b18" (UID: "ca5a2dc7-b729-4038-9028-f08d37442b18"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.180219 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/411d6157-0520-436e-b350-3a573e7ad984-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "411d6157-0520-436e-b350-3a573e7ad984" (UID: "411d6157-0520-436e-b350-3a573e7ad984"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.181216 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83e76214-d644-4e9b-ab3a-2cbcae40cb23-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.181235 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/411d6157-0520-436e-b350-3a573e7ad984-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.181244 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ca5a2dc7-b729-4038-9028-f08d37442b18-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.181252 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e7d1dce-289c-409f-a68e-b214f8f4cc14-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.181260 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.185897 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83e76214-d644-4e9b-ab3a-2cbcae40cb23-kube-api-access-xxzp6" (OuterVolumeSpecName: "kube-api-access-xxzp6") pod "83e76214-d644-4e9b-ab3a-2cbcae40cb23" (UID: "83e76214-d644-4e9b-ab3a-2cbcae40cb23"). InnerVolumeSpecName "kube-api-access-xxzp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.186008 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e7d1dce-289c-409f-a68e-b214f8f4cc14-kube-api-access-nzkz9" (OuterVolumeSpecName: "kube-api-access-nzkz9") pod "9e7d1dce-289c-409f-a68e-b214f8f4cc14" (UID: "9e7d1dce-289c-409f-a68e-b214f8f4cc14"). InnerVolumeSpecName "kube-api-access-nzkz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.189308 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/411d6157-0520-436e-b350-3a573e7ad984-kube-api-access-mj6jl" (OuterVolumeSpecName: "kube-api-access-mj6jl") pod "411d6157-0520-436e-b350-3a573e7ad984" (UID: "411d6157-0520-436e-b350-3a573e7ad984"). InnerVolumeSpecName "kube-api-access-mj6jl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.194838 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89-kube-api-access-kjf5j" (OuterVolumeSpecName: "kube-api-access-kjf5j") pod "83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89" (UID: "83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89"). InnerVolumeSpecName "kube-api-access-kjf5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.208783 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca5a2dc7-b729-4038-9028-f08d37442b18-kube-api-access-xbngn" (OuterVolumeSpecName: "kube-api-access-xbngn") pod "ca5a2dc7-b729-4038-9028-f08d37442b18" (UID: "ca5a2dc7-b729-4038-9028-f08d37442b18"). InnerVolumeSpecName "kube-api-access-xbngn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.282260 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxzp6\" (UniqueName: \"kubernetes.io/projected/83e76214-d644-4e9b-ab3a-2cbcae40cb23-kube-api-access-xxzp6\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.282295 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjf5j\" (UniqueName: \"kubernetes.io/projected/83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89-kube-api-access-kjf5j\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.282307 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzkz9\" (UniqueName: \"kubernetes.io/projected/9e7d1dce-289c-409f-a68e-b214f8f4cc14-kube-api-access-nzkz9\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.282318 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbngn\" (UniqueName: \"kubernetes.io/projected/ca5a2dc7-b729-4038-9028-f08d37442b18-kube-api-access-xbngn\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:28 crc kubenswrapper[4776]: I1125 09:43:28.282331 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mj6jl\" (UniqueName: \"kubernetes.io/projected/411d6157-0520-436e-b350-3a573e7ad984-kube-api-access-mj6jl\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:29 crc kubenswrapper[4776]: I1125 09:43:29.144014 4776 generic.go:334] "Generic (PLEG): container finished" podID="758472ef-bebc-469a-87a0-e7f1a83ae279" containerID="356e591a2a72a0074a6ee72d621e09ac39b9f3d69f4d225fdbc0fdceddce7b5a" exitCode=0 Nov 25 09:43:29 crc kubenswrapper[4776]: I1125 09:43:29.144096 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4b82s" event={"ID":"758472ef-bebc-469a-87a0-e7f1a83ae279","Type":"ContainerDied","Data":"356e591a2a72a0074a6ee72d621e09ac39b9f3d69f4d225fdbc0fdceddce7b5a"} Nov 25 09:43:29 crc kubenswrapper[4776]: I1125 09:43:29.145761 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-mfnw2" Nov 25 09:43:29 crc kubenswrapper[4776]: I1125 09:43:29.145874 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-dvd6n" event={"ID":"e107d317-acb4-46a2-8f19-836572e8904e","Type":"ContainerStarted","Data":"89ea5f6d57e939e9a504efd0696cae5f58d4950592a81b46f72127d3204ecdab"} Nov 25 09:43:29 crc kubenswrapper[4776]: I1125 09:43:29.190428 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-dvd6n" podStartSLOduration=3.406668527 podStartE2EDuration="8.190412694s" podCreationTimestamp="2025-11-25 09:43:21 +0000 UTC" firstStartedPulling="2025-11-25 09:43:23.157209322 +0000 UTC m=+1148.198268885" lastFinishedPulling="2025-11-25 09:43:27.940953499 +0000 UTC m=+1152.982013052" observedRunningTime="2025-11-25 09:43:29.17709308 +0000 UTC m=+1154.218152643" watchObservedRunningTime="2025-11-25 09:43:29.190412694 +0000 UTC m=+1154.231472237" Nov 25 09:43:30 crc kubenswrapper[4776]: I1125 09:43:30.537555 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4b82s" Nov 25 09:43:30 crc kubenswrapper[4776]: I1125 09:43:30.617710 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/758472ef-bebc-469a-87a0-e7f1a83ae279-db-sync-config-data\") pod \"758472ef-bebc-469a-87a0-e7f1a83ae279\" (UID: \"758472ef-bebc-469a-87a0-e7f1a83ae279\") " Nov 25 09:43:30 crc kubenswrapper[4776]: I1125 09:43:30.617763 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gv5fb\" (UniqueName: \"kubernetes.io/projected/758472ef-bebc-469a-87a0-e7f1a83ae279-kube-api-access-gv5fb\") pod \"758472ef-bebc-469a-87a0-e7f1a83ae279\" (UID: \"758472ef-bebc-469a-87a0-e7f1a83ae279\") " Nov 25 09:43:30 crc kubenswrapper[4776]: I1125 09:43:30.617818 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/758472ef-bebc-469a-87a0-e7f1a83ae279-combined-ca-bundle\") pod \"758472ef-bebc-469a-87a0-e7f1a83ae279\" (UID: \"758472ef-bebc-469a-87a0-e7f1a83ae279\") " Nov 25 09:43:30 crc kubenswrapper[4776]: I1125 09:43:30.617949 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/758472ef-bebc-469a-87a0-e7f1a83ae279-config-data\") pod \"758472ef-bebc-469a-87a0-e7f1a83ae279\" (UID: \"758472ef-bebc-469a-87a0-e7f1a83ae279\") " Nov 25 09:43:30 crc kubenswrapper[4776]: I1125 09:43:30.623229 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/758472ef-bebc-469a-87a0-e7f1a83ae279-kube-api-access-gv5fb" (OuterVolumeSpecName: "kube-api-access-gv5fb") pod "758472ef-bebc-469a-87a0-e7f1a83ae279" (UID: "758472ef-bebc-469a-87a0-e7f1a83ae279"). InnerVolumeSpecName "kube-api-access-gv5fb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:43:30 crc kubenswrapper[4776]: I1125 09:43:30.623537 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/758472ef-bebc-469a-87a0-e7f1a83ae279-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "758472ef-bebc-469a-87a0-e7f1a83ae279" (UID: "758472ef-bebc-469a-87a0-e7f1a83ae279"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:43:30 crc kubenswrapper[4776]: I1125 09:43:30.650440 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/758472ef-bebc-469a-87a0-e7f1a83ae279-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "758472ef-bebc-469a-87a0-e7f1a83ae279" (UID: "758472ef-bebc-469a-87a0-e7f1a83ae279"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:43:30 crc kubenswrapper[4776]: I1125 09:43:30.667852 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/758472ef-bebc-469a-87a0-e7f1a83ae279-config-data" (OuterVolumeSpecName: "config-data") pod "758472ef-bebc-469a-87a0-e7f1a83ae279" (UID: "758472ef-bebc-469a-87a0-e7f1a83ae279"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:43:30 crc kubenswrapper[4776]: I1125 09:43:30.720374 4776 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/758472ef-bebc-469a-87a0-e7f1a83ae279-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:30 crc kubenswrapper[4776]: I1125 09:43:30.720420 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gv5fb\" (UniqueName: \"kubernetes.io/projected/758472ef-bebc-469a-87a0-e7f1a83ae279-kube-api-access-gv5fb\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:30 crc kubenswrapper[4776]: I1125 09:43:30.720435 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/758472ef-bebc-469a-87a0-e7f1a83ae279-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:30 crc kubenswrapper[4776]: I1125 09:43:30.720447 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/758472ef-bebc-469a-87a0-e7f1a83ae279-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.167970 4776 generic.go:334] "Generic (PLEG): container finished" podID="e107d317-acb4-46a2-8f19-836572e8904e" containerID="89ea5f6d57e939e9a504efd0696cae5f58d4950592a81b46f72127d3204ecdab" exitCode=0 Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.168036 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-dvd6n" event={"ID":"e107d317-acb4-46a2-8f19-836572e8904e","Type":"ContainerDied","Data":"89ea5f6d57e939e9a504efd0696cae5f58d4950592a81b46f72127d3204ecdab"} Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.169847 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4b82s" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.171137 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4b82s" event={"ID":"758472ef-bebc-469a-87a0-e7f1a83ae279","Type":"ContainerDied","Data":"f32b036d1c5f5c7719a4530e76e35d67b4dea497f08a51a791770990966e84bd"} Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.171220 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f32b036d1c5f5c7719a4530e76e35d67b4dea497f08a51a791770990966e84bd" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.492485 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56766df65f-gcd25"] Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.494387 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56766df65f-gcd25" podUID="7a2a0530-6fc7-4297-a8ef-0eb391321b15" containerName="dnsmasq-dns" containerID="cri-o://139977ade9dc8dfdfa2e5d63de9d6408a2e51495e91a19593b0db46fb322b3f5" gracePeriod=10 Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.495262 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56766df65f-gcd25" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.535378 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6856c564b9-zvxvz"] Nov 25 09:43:31 crc kubenswrapper[4776]: E1125 09:43:31.535863 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83e76214-d644-4e9b-ab3a-2cbcae40cb23" containerName="mariadb-database-create" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.535921 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="83e76214-d644-4e9b-ab3a-2cbcae40cb23" containerName="mariadb-database-create" Nov 25 09:43:31 crc kubenswrapper[4776]: E1125 09:43:31.535972 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca5a2dc7-b729-4038-9028-f08d37442b18" containerName="mariadb-database-create" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.536045 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca5a2dc7-b729-4038-9028-f08d37442b18" containerName="mariadb-database-create" Nov 25 09:43:31 crc kubenswrapper[4776]: E1125 09:43:31.536106 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e7d1dce-289c-409f-a68e-b214f8f4cc14" containerName="mariadb-account-create" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.536151 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e7d1dce-289c-409f-a68e-b214f8f4cc14" containerName="mariadb-account-create" Nov 25 09:43:31 crc kubenswrapper[4776]: E1125 09:43:31.536207 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="758472ef-bebc-469a-87a0-e7f1a83ae279" containerName="glance-db-sync" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.536252 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="758472ef-bebc-469a-87a0-e7f1a83ae279" containerName="glance-db-sync" Nov 25 09:43:31 crc kubenswrapper[4776]: E1125 09:43:31.536307 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5705e80c-320b-49da-8c10-743627c8dbdb" containerName="mariadb-account-create" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.536360 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="5705e80c-320b-49da-8c10-743627c8dbdb" containerName="mariadb-account-create" Nov 25 09:43:31 crc kubenswrapper[4776]: E1125 09:43:31.536412 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89" containerName="mariadb-account-create" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.536455 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89" containerName="mariadb-account-create" Nov 25 09:43:31 crc kubenswrapper[4776]: E1125 09:43:31.536512 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="411d6157-0520-436e-b350-3a573e7ad984" containerName="mariadb-database-create" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.536557 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="411d6157-0520-436e-b350-3a573e7ad984" containerName="mariadb-database-create" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.536753 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="411d6157-0520-436e-b350-3a573e7ad984" containerName="mariadb-database-create" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.536812 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89" containerName="mariadb-account-create" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.536865 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="83e76214-d644-4e9b-ab3a-2cbcae40cb23" containerName="mariadb-database-create" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.536914 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e7d1dce-289c-409f-a68e-b214f8f4cc14" containerName="mariadb-account-create" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.536963 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="5705e80c-320b-49da-8c10-743627c8dbdb" containerName="mariadb-account-create" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.537021 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca5a2dc7-b729-4038-9028-f08d37442b18" containerName="mariadb-database-create" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.537104 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="758472ef-bebc-469a-87a0-e7f1a83ae279" containerName="glance-db-sync" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.538050 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.560945 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6856c564b9-zvxvz"] Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.635353 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-dns-swift-storage-0\") pod \"dnsmasq-dns-6856c564b9-zvxvz\" (UID: \"932e50fa-c886-4d45-8372-e671a7719010\") " pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.635393 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-config\") pod \"dnsmasq-dns-6856c564b9-zvxvz\" (UID: \"932e50fa-c886-4d45-8372-e671a7719010\") " pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.635654 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-ovsdbserver-nb\") pod \"dnsmasq-dns-6856c564b9-zvxvz\" (UID: \"932e50fa-c886-4d45-8372-e671a7719010\") " pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.635731 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-dns-svc\") pod \"dnsmasq-dns-6856c564b9-zvxvz\" (UID: \"932e50fa-c886-4d45-8372-e671a7719010\") " pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.635755 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-ovsdbserver-sb\") pod \"dnsmasq-dns-6856c564b9-zvxvz\" (UID: \"932e50fa-c886-4d45-8372-e671a7719010\") " pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.635802 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rm9ng\" (UniqueName: \"kubernetes.io/projected/932e50fa-c886-4d45-8372-e671a7719010-kube-api-access-rm9ng\") pod \"dnsmasq-dns-6856c564b9-zvxvz\" (UID: \"932e50fa-c886-4d45-8372-e671a7719010\") " pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.737305 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-ovsdbserver-nb\") pod \"dnsmasq-dns-6856c564b9-zvxvz\" (UID: \"932e50fa-c886-4d45-8372-e671a7719010\") " pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.737352 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-dns-svc\") pod \"dnsmasq-dns-6856c564b9-zvxvz\" (UID: \"932e50fa-c886-4d45-8372-e671a7719010\") " pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.738051 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-ovsdbserver-sb\") pod \"dnsmasq-dns-6856c564b9-zvxvz\" (UID: \"932e50fa-c886-4d45-8372-e671a7719010\") " pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.738095 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rm9ng\" (UniqueName: \"kubernetes.io/projected/932e50fa-c886-4d45-8372-e671a7719010-kube-api-access-rm9ng\") pod \"dnsmasq-dns-6856c564b9-zvxvz\" (UID: \"932e50fa-c886-4d45-8372-e671a7719010\") " pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.738167 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-dns-swift-storage-0\") pod \"dnsmasq-dns-6856c564b9-zvxvz\" (UID: \"932e50fa-c886-4d45-8372-e671a7719010\") " pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.738184 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-config\") pod \"dnsmasq-dns-6856c564b9-zvxvz\" (UID: \"932e50fa-c886-4d45-8372-e671a7719010\") " pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.738445 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-ovsdbserver-nb\") pod \"dnsmasq-dns-6856c564b9-zvxvz\" (UID: \"932e50fa-c886-4d45-8372-e671a7719010\") " pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.738713 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-ovsdbserver-sb\") pod \"dnsmasq-dns-6856c564b9-zvxvz\" (UID: \"932e50fa-c886-4d45-8372-e671a7719010\") " pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.738884 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-dns-svc\") pod \"dnsmasq-dns-6856c564b9-zvxvz\" (UID: \"932e50fa-c886-4d45-8372-e671a7719010\") " pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.739225 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-dns-swift-storage-0\") pod \"dnsmasq-dns-6856c564b9-zvxvz\" (UID: \"932e50fa-c886-4d45-8372-e671a7719010\") " pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.739544 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-config\") pod \"dnsmasq-dns-6856c564b9-zvxvz\" (UID: \"932e50fa-c886-4d45-8372-e671a7719010\") " pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.764902 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rm9ng\" (UniqueName: \"kubernetes.io/projected/932e50fa-c886-4d45-8372-e671a7719010-kube-api-access-rm9ng\") pod \"dnsmasq-dns-6856c564b9-zvxvz\" (UID: \"932e50fa-c886-4d45-8372-e671a7719010\") " pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.905598 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" Nov 25 09:43:31 crc kubenswrapper[4776]: I1125 09:43:31.924186 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56766df65f-gcd25" Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.046956 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-dns-swift-storage-0\") pod \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\" (UID: \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\") " Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.047006 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-ovsdbserver-nb\") pod \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\" (UID: \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\") " Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.047086 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-config\") pod \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\" (UID: \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\") " Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.047140 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-md6kl\" (UniqueName: \"kubernetes.io/projected/7a2a0530-6fc7-4297-a8ef-0eb391321b15-kube-api-access-md6kl\") pod \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\" (UID: \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\") " Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.047179 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-dns-svc\") pod \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\" (UID: \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\") " Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.047208 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-ovsdbserver-sb\") pod \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\" (UID: \"7a2a0530-6fc7-4297-a8ef-0eb391321b15\") " Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.056296 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a2a0530-6fc7-4297-a8ef-0eb391321b15-kube-api-access-md6kl" (OuterVolumeSpecName: "kube-api-access-md6kl") pod "7a2a0530-6fc7-4297-a8ef-0eb391321b15" (UID: "7a2a0530-6fc7-4297-a8ef-0eb391321b15"). InnerVolumeSpecName "kube-api-access-md6kl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.092418 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7a2a0530-6fc7-4297-a8ef-0eb391321b15" (UID: "7a2a0530-6fc7-4297-a8ef-0eb391321b15"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.093098 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7a2a0530-6fc7-4297-a8ef-0eb391321b15" (UID: "7a2a0530-6fc7-4297-a8ef-0eb391321b15"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.095226 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7a2a0530-6fc7-4297-a8ef-0eb391321b15" (UID: "7a2a0530-6fc7-4297-a8ef-0eb391321b15"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.105692 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-config" (OuterVolumeSpecName: "config") pod "7a2a0530-6fc7-4297-a8ef-0eb391321b15" (UID: "7a2a0530-6fc7-4297-a8ef-0eb391321b15"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.107484 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7a2a0530-6fc7-4297-a8ef-0eb391321b15" (UID: "7a2a0530-6fc7-4297-a8ef-0eb391321b15"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.149940 4776 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.150582 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.153707 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.153733 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-md6kl\" (UniqueName: \"kubernetes.io/projected/7a2a0530-6fc7-4297-a8ef-0eb391321b15-kube-api-access-md6kl\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.153749 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.153762 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a2a0530-6fc7-4297-a8ef-0eb391321b15-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.183279 4776 generic.go:334] "Generic (PLEG): container finished" podID="7a2a0530-6fc7-4297-a8ef-0eb391321b15" containerID="139977ade9dc8dfdfa2e5d63de9d6408a2e51495e91a19593b0db46fb322b3f5" exitCode=0 Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.183517 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56766df65f-gcd25" Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.183681 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56766df65f-gcd25" event={"ID":"7a2a0530-6fc7-4297-a8ef-0eb391321b15","Type":"ContainerDied","Data":"139977ade9dc8dfdfa2e5d63de9d6408a2e51495e91a19593b0db46fb322b3f5"} Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.183745 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56766df65f-gcd25" event={"ID":"7a2a0530-6fc7-4297-a8ef-0eb391321b15","Type":"ContainerDied","Data":"a556397e5cf4747e94a3fa9e3bb06249880f3dac10018a509aaac6fed08f62a7"} Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.183767 4776 scope.go:117] "RemoveContainer" containerID="139977ade9dc8dfdfa2e5d63de9d6408a2e51495e91a19593b0db46fb322b3f5" Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.206039 4776 scope.go:117] "RemoveContainer" containerID="c94a37851183d8b0aa46947d36df2bdc5de0cc80ade7bf5e8378bf93b49ba161" Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.224349 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56766df65f-gcd25"] Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.233020 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56766df65f-gcd25"] Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.240334 4776 scope.go:117] "RemoveContainer" containerID="139977ade9dc8dfdfa2e5d63de9d6408a2e51495e91a19593b0db46fb322b3f5" Nov 25 09:43:32 crc kubenswrapper[4776]: E1125 09:43:32.240786 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"139977ade9dc8dfdfa2e5d63de9d6408a2e51495e91a19593b0db46fb322b3f5\": container with ID starting with 139977ade9dc8dfdfa2e5d63de9d6408a2e51495e91a19593b0db46fb322b3f5 not found: ID does not exist" containerID="139977ade9dc8dfdfa2e5d63de9d6408a2e51495e91a19593b0db46fb322b3f5" Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.240864 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"139977ade9dc8dfdfa2e5d63de9d6408a2e51495e91a19593b0db46fb322b3f5"} err="failed to get container status \"139977ade9dc8dfdfa2e5d63de9d6408a2e51495e91a19593b0db46fb322b3f5\": rpc error: code = NotFound desc = could not find container \"139977ade9dc8dfdfa2e5d63de9d6408a2e51495e91a19593b0db46fb322b3f5\": container with ID starting with 139977ade9dc8dfdfa2e5d63de9d6408a2e51495e91a19593b0db46fb322b3f5 not found: ID does not exist" Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.240893 4776 scope.go:117] "RemoveContainer" containerID="c94a37851183d8b0aa46947d36df2bdc5de0cc80ade7bf5e8378bf93b49ba161" Nov 25 09:43:32 crc kubenswrapper[4776]: E1125 09:43:32.241285 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c94a37851183d8b0aa46947d36df2bdc5de0cc80ade7bf5e8378bf93b49ba161\": container with ID starting with c94a37851183d8b0aa46947d36df2bdc5de0cc80ade7bf5e8378bf93b49ba161 not found: ID does not exist" containerID="c94a37851183d8b0aa46947d36df2bdc5de0cc80ade7bf5e8378bf93b49ba161" Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.241344 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c94a37851183d8b0aa46947d36df2bdc5de0cc80ade7bf5e8378bf93b49ba161"} err="failed to get container status \"c94a37851183d8b0aa46947d36df2bdc5de0cc80ade7bf5e8378bf93b49ba161\": rpc error: code = NotFound desc = could not find container \"c94a37851183d8b0aa46947d36df2bdc5de0cc80ade7bf5e8378bf93b49ba161\": container with ID starting with c94a37851183d8b0aa46947d36df2bdc5de0cc80ade7bf5e8378bf93b49ba161 not found: ID does not exist" Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.365350 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6856c564b9-zvxvz"] Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.405165 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-dvd6n" Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.459548 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e107d317-acb4-46a2-8f19-836572e8904e-config-data\") pod \"e107d317-acb4-46a2-8f19-836572e8904e\" (UID: \"e107d317-acb4-46a2-8f19-836572e8904e\") " Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.459787 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e107d317-acb4-46a2-8f19-836572e8904e-combined-ca-bundle\") pod \"e107d317-acb4-46a2-8f19-836572e8904e\" (UID: \"e107d317-acb4-46a2-8f19-836572e8904e\") " Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.459823 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfm9w\" (UniqueName: \"kubernetes.io/projected/e107d317-acb4-46a2-8f19-836572e8904e-kube-api-access-sfm9w\") pod \"e107d317-acb4-46a2-8f19-836572e8904e\" (UID: \"e107d317-acb4-46a2-8f19-836572e8904e\") " Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.462459 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e107d317-acb4-46a2-8f19-836572e8904e-kube-api-access-sfm9w" (OuterVolumeSpecName: "kube-api-access-sfm9w") pod "e107d317-acb4-46a2-8f19-836572e8904e" (UID: "e107d317-acb4-46a2-8f19-836572e8904e"). InnerVolumeSpecName "kube-api-access-sfm9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.489921 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e107d317-acb4-46a2-8f19-836572e8904e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e107d317-acb4-46a2-8f19-836572e8904e" (UID: "e107d317-acb4-46a2-8f19-836572e8904e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.511531 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e107d317-acb4-46a2-8f19-836572e8904e-config-data" (OuterVolumeSpecName: "config-data") pod "e107d317-acb4-46a2-8f19-836572e8904e" (UID: "e107d317-acb4-46a2-8f19-836572e8904e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.561719 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e107d317-acb4-46a2-8f19-836572e8904e-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.561748 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e107d317-acb4-46a2-8f19-836572e8904e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:32 crc kubenswrapper[4776]: I1125 09:43:32.561759 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfm9w\" (UniqueName: \"kubernetes.io/projected/e107d317-acb4-46a2-8f19-836572e8904e-kube-api-access-sfm9w\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.191854 4776 generic.go:334] "Generic (PLEG): container finished" podID="932e50fa-c886-4d45-8372-e671a7719010" containerID="4b1650f8058627a45e7f620b18a860f042c250a0cb8ce2b2c139288d15aaa398" exitCode=0 Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.191898 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" event={"ID":"932e50fa-c886-4d45-8372-e671a7719010","Type":"ContainerDied","Data":"4b1650f8058627a45e7f620b18a860f042c250a0cb8ce2b2c139288d15aaa398"} Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.192180 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" event={"ID":"932e50fa-c886-4d45-8372-e671a7719010","Type":"ContainerStarted","Data":"67a41c00e9ac113004b0c5695acc37bb220c2eef1c7009d0c58ca1f171be5792"} Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.196785 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-dvd6n" event={"ID":"e107d317-acb4-46a2-8f19-836572e8904e","Type":"ContainerDied","Data":"94a1444a112f88890fa525e35d110b8d1637c0b8dcb7f6f8afad671025e23b14"} Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.196830 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94a1444a112f88890fa525e35d110b8d1637c0b8dcb7f6f8afad671025e23b14" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.196901 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-dvd6n" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.672258 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a2a0530-6fc7-4297-a8ef-0eb391321b15" path="/var/lib/kubelet/pods/7a2a0530-6fc7-4297-a8ef-0eb391321b15/volumes" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.673216 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6856c564b9-zvxvz"] Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.722926 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-dxg4f"] Nov 25 09:43:33 crc kubenswrapper[4776]: E1125 09:43:33.723624 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e107d317-acb4-46a2-8f19-836572e8904e" containerName="keystone-db-sync" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.723644 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="e107d317-acb4-46a2-8f19-836572e8904e" containerName="keystone-db-sync" Nov 25 09:43:33 crc kubenswrapper[4776]: E1125 09:43:33.723688 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a2a0530-6fc7-4297-a8ef-0eb391321b15" containerName="dnsmasq-dns" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.723695 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a2a0530-6fc7-4297-a8ef-0eb391321b15" containerName="dnsmasq-dns" Nov 25 09:43:33 crc kubenswrapper[4776]: E1125 09:43:33.723719 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a2a0530-6fc7-4297-a8ef-0eb391321b15" containerName="init" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.723727 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a2a0530-6fc7-4297-a8ef-0eb391321b15" containerName="init" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.724055 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a2a0530-6fc7-4297-a8ef-0eb391321b15" containerName="dnsmasq-dns" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.724125 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="e107d317-acb4-46a2-8f19-836572e8904e" containerName="keystone-db-sync" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.724872 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dxg4f" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.732046 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.732179 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gtj9p" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.732430 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.732620 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.732834 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.750624 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7dbf8bff67-jjvsj"] Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.771708 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7dbf8bff67-jjvsj" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.781217 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5wf5\" (UniqueName: \"kubernetes.io/projected/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-kube-api-access-q5wf5\") pod \"keystone-bootstrap-dxg4f\" (UID: \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\") " pod="openstack/keystone-bootstrap-dxg4f" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.781291 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-config-data\") pod \"keystone-bootstrap-dxg4f\" (UID: \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\") " pod="openstack/keystone-bootstrap-dxg4f" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.781334 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-combined-ca-bundle\") pod \"keystone-bootstrap-dxg4f\" (UID: \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\") " pod="openstack/keystone-bootstrap-dxg4f" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.781361 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-scripts\") pod \"keystone-bootstrap-dxg4f\" (UID: \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\") " pod="openstack/keystone-bootstrap-dxg4f" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.781405 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-credential-keys\") pod \"keystone-bootstrap-dxg4f\" (UID: \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\") " pod="openstack/keystone-bootstrap-dxg4f" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.781511 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-fernet-keys\") pod \"keystone-bootstrap-dxg4f\" (UID: \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\") " pod="openstack/keystone-bootstrap-dxg4f" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.783114 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dxg4f"] Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.792791 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7dbf8bff67-jjvsj"] Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.883358 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-fernet-keys\") pod \"keystone-bootstrap-dxg4f\" (UID: \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\") " pod="openstack/keystone-bootstrap-dxg4f" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.884591 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-ovsdbserver-nb\") pod \"dnsmasq-dns-7dbf8bff67-jjvsj\" (UID: \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\") " pod="openstack/dnsmasq-dns-7dbf8bff67-jjvsj" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.884774 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5wf5\" (UniqueName: \"kubernetes.io/projected/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-kube-api-access-q5wf5\") pod \"keystone-bootstrap-dxg4f\" (UID: \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\") " pod="openstack/keystone-bootstrap-dxg4f" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.884908 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-ovsdbserver-sb\") pod \"dnsmasq-dns-7dbf8bff67-jjvsj\" (UID: \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\") " pod="openstack/dnsmasq-dns-7dbf8bff67-jjvsj" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.885037 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-config-data\") pod \"keystone-bootstrap-dxg4f\" (UID: \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\") " pod="openstack/keystone-bootstrap-dxg4f" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.885167 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdbqv\" (UniqueName: \"kubernetes.io/projected/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-kube-api-access-bdbqv\") pod \"dnsmasq-dns-7dbf8bff67-jjvsj\" (UID: \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\") " pod="openstack/dnsmasq-dns-7dbf8bff67-jjvsj" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.885292 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-dns-swift-storage-0\") pod \"dnsmasq-dns-7dbf8bff67-jjvsj\" (UID: \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\") " pod="openstack/dnsmasq-dns-7dbf8bff67-jjvsj" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.885395 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-combined-ca-bundle\") pod \"keystone-bootstrap-dxg4f\" (UID: \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\") " pod="openstack/keystone-bootstrap-dxg4f" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.885537 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-scripts\") pod \"keystone-bootstrap-dxg4f\" (UID: \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\") " pod="openstack/keystone-bootstrap-dxg4f" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.885661 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-dns-svc\") pod \"dnsmasq-dns-7dbf8bff67-jjvsj\" (UID: \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\") " pod="openstack/dnsmasq-dns-7dbf8bff67-jjvsj" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.885805 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-credential-keys\") pod \"keystone-bootstrap-dxg4f\" (UID: \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\") " pod="openstack/keystone-bootstrap-dxg4f" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.885937 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-config\") pod \"dnsmasq-dns-7dbf8bff67-jjvsj\" (UID: \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\") " pod="openstack/dnsmasq-dns-7dbf8bff67-jjvsj" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.888996 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-fernet-keys\") pod \"keystone-bootstrap-dxg4f\" (UID: \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\") " pod="openstack/keystone-bootstrap-dxg4f" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.894291 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-scripts\") pod \"keystone-bootstrap-dxg4f\" (UID: \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\") " pod="openstack/keystone-bootstrap-dxg4f" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.894643 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-combined-ca-bundle\") pod \"keystone-bootstrap-dxg4f\" (UID: \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\") " pod="openstack/keystone-bootstrap-dxg4f" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.897711 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-credential-keys\") pod \"keystone-bootstrap-dxg4f\" (UID: \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\") " pod="openstack/keystone-bootstrap-dxg4f" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.901221 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-config-data\") pod \"keystone-bootstrap-dxg4f\" (UID: \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\") " pod="openstack/keystone-bootstrap-dxg4f" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.915928 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-lgdgb"] Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.917034 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-lgdgb" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.930969 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.931625 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-dp8gc" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.933123 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.943400 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-lgdgb"] Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.944169 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5wf5\" (UniqueName: \"kubernetes.io/projected/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-kube-api-access-q5wf5\") pod \"keystone-bootstrap-dxg4f\" (UID: \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\") " pod="openstack/keystone-bootstrap-dxg4f" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.968127 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-djwk2"] Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.969276 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-djwk2" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.971910 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-5lrgb" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.982934 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.983671 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.987010 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/737e5e32-1fa7-4140-bfad-1bc3e10ee390-config\") pod \"neutron-db-sync-lgdgb\" (UID: \"737e5e32-1fa7-4140-bfad-1bc3e10ee390\") " pod="openstack/neutron-db-sync-lgdgb" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.987102 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-ovsdbserver-nb\") pod \"dnsmasq-dns-7dbf8bff67-jjvsj\" (UID: \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\") " pod="openstack/dnsmasq-dns-7dbf8bff67-jjvsj" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.987139 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ll7sn\" (UniqueName: \"kubernetes.io/projected/737e5e32-1fa7-4140-bfad-1bc3e10ee390-kube-api-access-ll7sn\") pod \"neutron-db-sync-lgdgb\" (UID: \"737e5e32-1fa7-4140-bfad-1bc3e10ee390\") " pod="openstack/neutron-db-sync-lgdgb" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.987159 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-ovsdbserver-sb\") pod \"dnsmasq-dns-7dbf8bff67-jjvsj\" (UID: \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\") " pod="openstack/dnsmasq-dns-7dbf8bff67-jjvsj" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.987189 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdbqv\" (UniqueName: \"kubernetes.io/projected/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-kube-api-access-bdbqv\") pod \"dnsmasq-dns-7dbf8bff67-jjvsj\" (UID: \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\") " pod="openstack/dnsmasq-dns-7dbf8bff67-jjvsj" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.987212 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-dns-swift-storage-0\") pod \"dnsmasq-dns-7dbf8bff67-jjvsj\" (UID: \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\") " pod="openstack/dnsmasq-dns-7dbf8bff67-jjvsj" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.987245 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-dns-svc\") pod \"dnsmasq-dns-7dbf8bff67-jjvsj\" (UID: \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\") " pod="openstack/dnsmasq-dns-7dbf8bff67-jjvsj" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.987298 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-config\") pod \"dnsmasq-dns-7dbf8bff67-jjvsj\" (UID: \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\") " pod="openstack/dnsmasq-dns-7dbf8bff67-jjvsj" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.987346 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/737e5e32-1fa7-4140-bfad-1bc3e10ee390-combined-ca-bundle\") pod \"neutron-db-sync-lgdgb\" (UID: \"737e5e32-1fa7-4140-bfad-1bc3e10ee390\") " pod="openstack/neutron-db-sync-lgdgb" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.990329 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-ovsdbserver-sb\") pod \"dnsmasq-dns-7dbf8bff67-jjvsj\" (UID: \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\") " pod="openstack/dnsmasq-dns-7dbf8bff67-jjvsj" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.991842 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-dns-svc\") pod \"dnsmasq-dns-7dbf8bff67-jjvsj\" (UID: \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\") " pod="openstack/dnsmasq-dns-7dbf8bff67-jjvsj" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.992626 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-config\") pod \"dnsmasq-dns-7dbf8bff67-jjvsj\" (UID: \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\") " pod="openstack/dnsmasq-dns-7dbf8bff67-jjvsj" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.993007 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-dns-swift-storage-0\") pod \"dnsmasq-dns-7dbf8bff67-jjvsj\" (UID: \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\") " pod="openstack/dnsmasq-dns-7dbf8bff67-jjvsj" Nov 25 09:43:33 crc kubenswrapper[4776]: I1125 09:43:33.997536 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-djwk2"] Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.000432 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-ovsdbserver-nb\") pod \"dnsmasq-dns-7dbf8bff67-jjvsj\" (UID: \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\") " pod="openstack/dnsmasq-dns-7dbf8bff67-jjvsj" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.037078 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.039214 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.040756 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdbqv\" (UniqueName: \"kubernetes.io/projected/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-kube-api-access-bdbqv\") pod \"dnsmasq-dns-7dbf8bff67-jjvsj\" (UID: \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\") " pod="openstack/dnsmasq-dns-7dbf8bff67-jjvsj" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.046406 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.046550 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.065184 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.074610 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dxg4f" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.079519 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-bdhw6"] Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.081104 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-bdhw6" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.083652 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.084218 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-c8v8c" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.092587 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-log-httpd\") pod \"ceilometer-0\" (UID: \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\") " pod="openstack/ceilometer-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.092638 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rntzg\" (UniqueName: \"kubernetes.io/projected/e522431f-0956-49f7-9c46-b0202824bd56-kube-api-access-rntzg\") pod \"cinder-db-sync-djwk2\" (UID: \"e522431f-0956-49f7-9c46-b0202824bd56\") " pod="openstack/cinder-db-sync-djwk2" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.092658 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94nlr\" (UniqueName: \"kubernetes.io/projected/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-kube-api-access-94nlr\") pod \"ceilometer-0\" (UID: \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\") " pod="openstack/ceilometer-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.092708 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ll7sn\" (UniqueName: \"kubernetes.io/projected/737e5e32-1fa7-4140-bfad-1bc3e10ee390-kube-api-access-ll7sn\") pod \"neutron-db-sync-lgdgb\" (UID: \"737e5e32-1fa7-4140-bfad-1bc3e10ee390\") " pod="openstack/neutron-db-sync-lgdgb" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.092735 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-scripts\") pod \"ceilometer-0\" (UID: \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\") " pod="openstack/ceilometer-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.092750 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\") " pod="openstack/ceilometer-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.092931 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e522431f-0956-49f7-9c46-b0202824bd56-combined-ca-bundle\") pod \"cinder-db-sync-djwk2\" (UID: \"e522431f-0956-49f7-9c46-b0202824bd56\") " pod="openstack/cinder-db-sync-djwk2" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.092964 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e522431f-0956-49f7-9c46-b0202824bd56-db-sync-config-data\") pod \"cinder-db-sync-djwk2\" (UID: \"e522431f-0956-49f7-9c46-b0202824bd56\") " pod="openstack/cinder-db-sync-djwk2" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.093055 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e522431f-0956-49f7-9c46-b0202824bd56-scripts\") pod \"cinder-db-sync-djwk2\" (UID: \"e522431f-0956-49f7-9c46-b0202824bd56\") " pod="openstack/cinder-db-sync-djwk2" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.093091 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e522431f-0956-49f7-9c46-b0202824bd56-config-data\") pod \"cinder-db-sync-djwk2\" (UID: \"e522431f-0956-49f7-9c46-b0202824bd56\") " pod="openstack/cinder-db-sync-djwk2" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.093123 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-config-data\") pod \"ceilometer-0\" (UID: \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\") " pod="openstack/ceilometer-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.093163 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\") " pod="openstack/ceilometer-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.093184 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/737e5e32-1fa7-4140-bfad-1bc3e10ee390-combined-ca-bundle\") pod \"neutron-db-sync-lgdgb\" (UID: \"737e5e32-1fa7-4140-bfad-1bc3e10ee390\") " pod="openstack/neutron-db-sync-lgdgb" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.093201 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/737e5e32-1fa7-4140-bfad-1bc3e10ee390-config\") pod \"neutron-db-sync-lgdgb\" (UID: \"737e5e32-1fa7-4140-bfad-1bc3e10ee390\") " pod="openstack/neutron-db-sync-lgdgb" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.093222 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e522431f-0956-49f7-9c46-b0202824bd56-etc-machine-id\") pod \"cinder-db-sync-djwk2\" (UID: \"e522431f-0956-49f7-9c46-b0202824bd56\") " pod="openstack/cinder-db-sync-djwk2" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.093273 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-run-httpd\") pod \"ceilometer-0\" (UID: \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\") " pod="openstack/ceilometer-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.104155 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7dbf8bff67-jjvsj"] Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.104540 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7dbf8bff67-jjvsj" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.115303 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/737e5e32-1fa7-4140-bfad-1bc3e10ee390-combined-ca-bundle\") pod \"neutron-db-sync-lgdgb\" (UID: \"737e5e32-1fa7-4140-bfad-1bc3e10ee390\") " pod="openstack/neutron-db-sync-lgdgb" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.127473 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ll7sn\" (UniqueName: \"kubernetes.io/projected/737e5e32-1fa7-4140-bfad-1bc3e10ee390-kube-api-access-ll7sn\") pod \"neutron-db-sync-lgdgb\" (UID: \"737e5e32-1fa7-4140-bfad-1bc3e10ee390\") " pod="openstack/neutron-db-sync-lgdgb" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.142976 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-bdhw6"] Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.152014 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76c58b6d97-jbkjb"] Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.153991 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.165334 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/737e5e32-1fa7-4140-bfad-1bc3e10ee390-config\") pod \"neutron-db-sync-lgdgb\" (UID: \"737e5e32-1fa7-4140-bfad-1bc3e10ee390\") " pod="openstack/neutron-db-sync-lgdgb" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.192922 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-ttw4k"] Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.194047 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ttw4k" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.195384 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rntzg\" (UniqueName: \"kubernetes.io/projected/e522431f-0956-49f7-9c46-b0202824bd56-kube-api-access-rntzg\") pod \"cinder-db-sync-djwk2\" (UID: \"e522431f-0956-49f7-9c46-b0202824bd56\") " pod="openstack/cinder-db-sync-djwk2" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.195430 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-dns-swift-storage-0\") pod \"dnsmasq-dns-76c58b6d97-jbkjb\" (UID: \"624c8057-3576-4249-9a63-5bb0dee2da0a\") " pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.195458 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94nlr\" (UniqueName: \"kubernetes.io/projected/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-kube-api-access-94nlr\") pod \"ceilometer-0\" (UID: \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\") " pod="openstack/ceilometer-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.195484 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8c67\" (UniqueName: \"kubernetes.io/projected/624c8057-3576-4249-9a63-5bb0dee2da0a-kube-api-access-x8c67\") pod \"dnsmasq-dns-76c58b6d97-jbkjb\" (UID: \"624c8057-3576-4249-9a63-5bb0dee2da0a\") " pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.195551 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-scripts\") pod \"ceilometer-0\" (UID: \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\") " pod="openstack/ceilometer-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.195571 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\") " pod="openstack/ceilometer-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.195610 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e522431f-0956-49f7-9c46-b0202824bd56-combined-ca-bundle\") pod \"cinder-db-sync-djwk2\" (UID: \"e522431f-0956-49f7-9c46-b0202824bd56\") " pod="openstack/cinder-db-sync-djwk2" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.195638 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ff2a086-8bce-46e2-b294-3b3f31e39b4e-combined-ca-bundle\") pod \"barbican-db-sync-bdhw6\" (UID: \"3ff2a086-8bce-46e2-b294-3b3f31e39b4e\") " pod="openstack/barbican-db-sync-bdhw6" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.195671 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-ovsdbserver-sb\") pod \"dnsmasq-dns-76c58b6d97-jbkjb\" (UID: \"624c8057-3576-4249-9a63-5bb0dee2da0a\") " pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.195698 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e522431f-0956-49f7-9c46-b0202824bd56-db-sync-config-data\") pod \"cinder-db-sync-djwk2\" (UID: \"e522431f-0956-49f7-9c46-b0202824bd56\") " pod="openstack/cinder-db-sync-djwk2" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.195718 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5cf5\" (UniqueName: \"kubernetes.io/projected/3ff2a086-8bce-46e2-b294-3b3f31e39b4e-kube-api-access-b5cf5\") pod \"barbican-db-sync-bdhw6\" (UID: \"3ff2a086-8bce-46e2-b294-3b3f31e39b4e\") " pod="openstack/barbican-db-sync-bdhw6" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.195756 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-config\") pod \"dnsmasq-dns-76c58b6d97-jbkjb\" (UID: \"624c8057-3576-4249-9a63-5bb0dee2da0a\") " pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.195781 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3ff2a086-8bce-46e2-b294-3b3f31e39b4e-db-sync-config-data\") pod \"barbican-db-sync-bdhw6\" (UID: \"3ff2a086-8bce-46e2-b294-3b3f31e39b4e\") " pod="openstack/barbican-db-sync-bdhw6" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.195808 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-ovsdbserver-nb\") pod \"dnsmasq-dns-76c58b6d97-jbkjb\" (UID: \"624c8057-3576-4249-9a63-5bb0dee2da0a\") " pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.195843 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e522431f-0956-49f7-9c46-b0202824bd56-scripts\") pod \"cinder-db-sync-djwk2\" (UID: \"e522431f-0956-49f7-9c46-b0202824bd56\") " pod="openstack/cinder-db-sync-djwk2" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.195864 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e522431f-0956-49f7-9c46-b0202824bd56-config-data\") pod \"cinder-db-sync-djwk2\" (UID: \"e522431f-0956-49f7-9c46-b0202824bd56\") " pod="openstack/cinder-db-sync-djwk2" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.195933 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-config-data\") pod \"ceilometer-0\" (UID: \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\") " pod="openstack/ceilometer-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.195961 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\") " pod="openstack/ceilometer-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.196006 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-dns-svc\") pod \"dnsmasq-dns-76c58b6d97-jbkjb\" (UID: \"624c8057-3576-4249-9a63-5bb0dee2da0a\") " pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.196047 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e522431f-0956-49f7-9c46-b0202824bd56-etc-machine-id\") pod \"cinder-db-sync-djwk2\" (UID: \"e522431f-0956-49f7-9c46-b0202824bd56\") " pod="openstack/cinder-db-sync-djwk2" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.196095 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-run-httpd\") pod \"ceilometer-0\" (UID: \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\") " pod="openstack/ceilometer-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.196146 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-log-httpd\") pod \"ceilometer-0\" (UID: \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\") " pod="openstack/ceilometer-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.196630 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-log-httpd\") pod \"ceilometer-0\" (UID: \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\") " pod="openstack/ceilometer-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.196661 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-pf9p5" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.196972 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.196986 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.198426 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e522431f-0956-49f7-9c46-b0202824bd56-etc-machine-id\") pod \"cinder-db-sync-djwk2\" (UID: \"e522431f-0956-49f7-9c46-b0202824bd56\") " pod="openstack/cinder-db-sync-djwk2" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.198719 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-run-httpd\") pod \"ceilometer-0\" (UID: \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\") " pod="openstack/ceilometer-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.201390 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e522431f-0956-49f7-9c46-b0202824bd56-db-sync-config-data\") pod \"cinder-db-sync-djwk2\" (UID: \"e522431f-0956-49f7-9c46-b0202824bd56\") " pod="openstack/cinder-db-sync-djwk2" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.202951 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\") " pod="openstack/ceilometer-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.208149 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e522431f-0956-49f7-9c46-b0202824bd56-config-data\") pod \"cinder-db-sync-djwk2\" (UID: \"e522431f-0956-49f7-9c46-b0202824bd56\") " pod="openstack/cinder-db-sync-djwk2" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.209862 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\") " pod="openstack/ceilometer-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.214751 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-config-data\") pod \"ceilometer-0\" (UID: \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\") " pod="openstack/ceilometer-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.215362 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76c58b6d97-jbkjb"] Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.215377 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e522431f-0956-49f7-9c46-b0202824bd56-scripts\") pod \"cinder-db-sync-djwk2\" (UID: \"e522431f-0956-49f7-9c46-b0202824bd56\") " pod="openstack/cinder-db-sync-djwk2" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.216035 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-scripts\") pod \"ceilometer-0\" (UID: \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\") " pod="openstack/ceilometer-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.224125 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e522431f-0956-49f7-9c46-b0202824bd56-combined-ca-bundle\") pod \"cinder-db-sync-djwk2\" (UID: \"e522431f-0956-49f7-9c46-b0202824bd56\") " pod="openstack/cinder-db-sync-djwk2" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.225569 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-ttw4k"] Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.225813 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rntzg\" (UniqueName: \"kubernetes.io/projected/e522431f-0956-49f7-9c46-b0202824bd56-kube-api-access-rntzg\") pod \"cinder-db-sync-djwk2\" (UID: \"e522431f-0956-49f7-9c46-b0202824bd56\") " pod="openstack/cinder-db-sync-djwk2" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.230725 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94nlr\" (UniqueName: \"kubernetes.io/projected/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-kube-api-access-94nlr\") pod \"ceilometer-0\" (UID: \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\") " pod="openstack/ceilometer-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.241576 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" event={"ID":"932e50fa-c886-4d45-8372-e671a7719010","Type":"ContainerStarted","Data":"f6caa82b641d6b347437cba0f2afb987e0ce5625f8944257e0067c00afe42059"} Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.242933 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.279545 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" podStartSLOduration=3.27952551 podStartE2EDuration="3.27952551s" podCreationTimestamp="2025-11-25 09:43:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:43:34.273459488 +0000 UTC m=+1159.314519051" watchObservedRunningTime="2025-11-25 09:43:34.27952551 +0000 UTC m=+1159.320585063" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.301584 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-dns-svc\") pod \"dnsmasq-dns-76c58b6d97-jbkjb\" (UID: \"624c8057-3576-4249-9a63-5bb0dee2da0a\") " pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.301841 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gh6b5\" (UniqueName: \"kubernetes.io/projected/7c18590c-cf10-4c05-8a97-b9f1d091868a-kube-api-access-gh6b5\") pod \"placement-db-sync-ttw4k\" (UID: \"7c18590c-cf10-4c05-8a97-b9f1d091868a\") " pod="openstack/placement-db-sync-ttw4k" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.301871 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-dns-swift-storage-0\") pod \"dnsmasq-dns-76c58b6d97-jbkjb\" (UID: \"624c8057-3576-4249-9a63-5bb0dee2da0a\") " pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.301905 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8c67\" (UniqueName: \"kubernetes.io/projected/624c8057-3576-4249-9a63-5bb0dee2da0a-kube-api-access-x8c67\") pod \"dnsmasq-dns-76c58b6d97-jbkjb\" (UID: \"624c8057-3576-4249-9a63-5bb0dee2da0a\") " pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.301929 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c18590c-cf10-4c05-8a97-b9f1d091868a-config-data\") pod \"placement-db-sync-ttw4k\" (UID: \"7c18590c-cf10-4c05-8a97-b9f1d091868a\") " pod="openstack/placement-db-sync-ttw4k" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.301960 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c18590c-cf10-4c05-8a97-b9f1d091868a-scripts\") pod \"placement-db-sync-ttw4k\" (UID: \"7c18590c-cf10-4c05-8a97-b9f1d091868a\") " pod="openstack/placement-db-sync-ttw4k" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.301988 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ff2a086-8bce-46e2-b294-3b3f31e39b4e-combined-ca-bundle\") pod \"barbican-db-sync-bdhw6\" (UID: \"3ff2a086-8bce-46e2-b294-3b3f31e39b4e\") " pod="openstack/barbican-db-sync-bdhw6" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.302012 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-ovsdbserver-sb\") pod \"dnsmasq-dns-76c58b6d97-jbkjb\" (UID: \"624c8057-3576-4249-9a63-5bb0dee2da0a\") " pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.302027 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5cf5\" (UniqueName: \"kubernetes.io/projected/3ff2a086-8bce-46e2-b294-3b3f31e39b4e-kube-api-access-b5cf5\") pod \"barbican-db-sync-bdhw6\" (UID: \"3ff2a086-8bce-46e2-b294-3b3f31e39b4e\") " pod="openstack/barbican-db-sync-bdhw6" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.302042 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-config\") pod \"dnsmasq-dns-76c58b6d97-jbkjb\" (UID: \"624c8057-3576-4249-9a63-5bb0dee2da0a\") " pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.302074 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3ff2a086-8bce-46e2-b294-3b3f31e39b4e-db-sync-config-data\") pod \"barbican-db-sync-bdhw6\" (UID: \"3ff2a086-8bce-46e2-b294-3b3f31e39b4e\") " pod="openstack/barbican-db-sync-bdhw6" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.302095 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-ovsdbserver-nb\") pod \"dnsmasq-dns-76c58b6d97-jbkjb\" (UID: \"624c8057-3576-4249-9a63-5bb0dee2da0a\") " pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.302113 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c18590c-cf10-4c05-8a97-b9f1d091868a-logs\") pod \"placement-db-sync-ttw4k\" (UID: \"7c18590c-cf10-4c05-8a97-b9f1d091868a\") " pod="openstack/placement-db-sync-ttw4k" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.302146 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c18590c-cf10-4c05-8a97-b9f1d091868a-combined-ca-bundle\") pod \"placement-db-sync-ttw4k\" (UID: \"7c18590c-cf10-4c05-8a97-b9f1d091868a\") " pod="openstack/placement-db-sync-ttw4k" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.307332 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-dns-svc\") pod \"dnsmasq-dns-76c58b6d97-jbkjb\" (UID: \"624c8057-3576-4249-9a63-5bb0dee2da0a\") " pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.308012 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-dns-swift-storage-0\") pod \"dnsmasq-dns-76c58b6d97-jbkjb\" (UID: \"624c8057-3576-4249-9a63-5bb0dee2da0a\") " pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.309655 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-config\") pod \"dnsmasq-dns-76c58b6d97-jbkjb\" (UID: \"624c8057-3576-4249-9a63-5bb0dee2da0a\") " pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.310564 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-ovsdbserver-nb\") pod \"dnsmasq-dns-76c58b6d97-jbkjb\" (UID: \"624c8057-3576-4249-9a63-5bb0dee2da0a\") " pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.311841 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-ovsdbserver-sb\") pod \"dnsmasq-dns-76c58b6d97-jbkjb\" (UID: \"624c8057-3576-4249-9a63-5bb0dee2da0a\") " pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.323343 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3ff2a086-8bce-46e2-b294-3b3f31e39b4e-db-sync-config-data\") pod \"barbican-db-sync-bdhw6\" (UID: \"3ff2a086-8bce-46e2-b294-3b3f31e39b4e\") " pod="openstack/barbican-db-sync-bdhw6" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.326239 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ff2a086-8bce-46e2-b294-3b3f31e39b4e-combined-ca-bundle\") pod \"barbican-db-sync-bdhw6\" (UID: \"3ff2a086-8bce-46e2-b294-3b3f31e39b4e\") " pod="openstack/barbican-db-sync-bdhw6" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.335591 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5cf5\" (UniqueName: \"kubernetes.io/projected/3ff2a086-8bce-46e2-b294-3b3f31e39b4e-kube-api-access-b5cf5\") pod \"barbican-db-sync-bdhw6\" (UID: \"3ff2a086-8bce-46e2-b294-3b3f31e39b4e\") " pod="openstack/barbican-db-sync-bdhw6" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.338844 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8c67\" (UniqueName: \"kubernetes.io/projected/624c8057-3576-4249-9a63-5bb0dee2da0a-kube-api-access-x8c67\") pod \"dnsmasq-dns-76c58b6d97-jbkjb\" (UID: \"624c8057-3576-4249-9a63-5bb0dee2da0a\") " pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.359691 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-lgdgb" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.382383 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-djwk2" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.403387 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c18590c-cf10-4c05-8a97-b9f1d091868a-config-data\") pod \"placement-db-sync-ttw4k\" (UID: \"7c18590c-cf10-4c05-8a97-b9f1d091868a\") " pod="openstack/placement-db-sync-ttw4k" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.403459 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c18590c-cf10-4c05-8a97-b9f1d091868a-scripts\") pod \"placement-db-sync-ttw4k\" (UID: \"7c18590c-cf10-4c05-8a97-b9f1d091868a\") " pod="openstack/placement-db-sync-ttw4k" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.403543 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c18590c-cf10-4c05-8a97-b9f1d091868a-logs\") pod \"placement-db-sync-ttw4k\" (UID: \"7c18590c-cf10-4c05-8a97-b9f1d091868a\") " pod="openstack/placement-db-sync-ttw4k" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.403590 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c18590c-cf10-4c05-8a97-b9f1d091868a-combined-ca-bundle\") pod \"placement-db-sync-ttw4k\" (UID: \"7c18590c-cf10-4c05-8a97-b9f1d091868a\") " pod="openstack/placement-db-sync-ttw4k" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.403673 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gh6b5\" (UniqueName: \"kubernetes.io/projected/7c18590c-cf10-4c05-8a97-b9f1d091868a-kube-api-access-gh6b5\") pod \"placement-db-sync-ttw4k\" (UID: \"7c18590c-cf10-4c05-8a97-b9f1d091868a\") " pod="openstack/placement-db-sync-ttw4k" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.403887 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.404635 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c18590c-cf10-4c05-8a97-b9f1d091868a-logs\") pod \"placement-db-sync-ttw4k\" (UID: \"7c18590c-cf10-4c05-8a97-b9f1d091868a\") " pod="openstack/placement-db-sync-ttw4k" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.406344 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c18590c-cf10-4c05-8a97-b9f1d091868a-scripts\") pod \"placement-db-sync-ttw4k\" (UID: \"7c18590c-cf10-4c05-8a97-b9f1d091868a\") " pod="openstack/placement-db-sync-ttw4k" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.406951 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c18590c-cf10-4c05-8a97-b9f1d091868a-config-data\") pod \"placement-db-sync-ttw4k\" (UID: \"7c18590c-cf10-4c05-8a97-b9f1d091868a\") " pod="openstack/placement-db-sync-ttw4k" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.420769 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c18590c-cf10-4c05-8a97-b9f1d091868a-combined-ca-bundle\") pod \"placement-db-sync-ttw4k\" (UID: \"7c18590c-cf10-4c05-8a97-b9f1d091868a\") " pod="openstack/placement-db-sync-ttw4k" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.433286 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gh6b5\" (UniqueName: \"kubernetes.io/projected/7c18590c-cf10-4c05-8a97-b9f1d091868a-kube-api-access-gh6b5\") pod \"placement-db-sync-ttw4k\" (UID: \"7c18590c-cf10-4c05-8a97-b9f1d091868a\") " pod="openstack/placement-db-sync-ttw4k" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.500105 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-bdhw6" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.523110 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.540106 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ttw4k" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.675991 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7dbf8bff67-jjvsj"] Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.818270 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dxg4f"] Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.846395 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.848137 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.854951 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.855052 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-5xkz6" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.855466 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.857884 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.921921 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cd309d14-e0b5-4a84-93eb-4a114d77620e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.921969 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd309d14-e0b5-4a84-93eb-4a114d77620e-config-data\") pod \"glance-default-external-api-0\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.922005 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcd7m\" (UniqueName: \"kubernetes.io/projected/cd309d14-e0b5-4a84-93eb-4a114d77620e-kube-api-access-lcd7m\") pod \"glance-default-external-api-0\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.922029 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.922337 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd309d14-e0b5-4a84-93eb-4a114d77620e-logs\") pod \"glance-default-external-api-0\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.922447 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd309d14-e0b5-4a84-93eb-4a114d77620e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.922479 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd309d14-e0b5-4a84-93eb-4a114d77620e-scripts\") pod \"glance-default-external-api-0\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.986010 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.993843 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 09:43:34 crc kubenswrapper[4776]: I1125 09:43:34.999352 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.002808 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.024315 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cd309d14-e0b5-4a84-93eb-4a114d77620e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.024364 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd309d14-e0b5-4a84-93eb-4a114d77620e-config-data\") pod \"glance-default-external-api-0\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.024413 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcd7m\" (UniqueName: \"kubernetes.io/projected/cd309d14-e0b5-4a84-93eb-4a114d77620e-kube-api-access-lcd7m\") pod \"glance-default-external-api-0\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.024440 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.024511 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd309d14-e0b5-4a84-93eb-4a114d77620e-logs\") pod \"glance-default-external-api-0\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.024558 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd309d14-e0b5-4a84-93eb-4a114d77620e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.024589 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd309d14-e0b5-4a84-93eb-4a114d77620e-scripts\") pod \"glance-default-external-api-0\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.024857 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.024940 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cd309d14-e0b5-4a84-93eb-4a114d77620e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.025164 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd309d14-e0b5-4a84-93eb-4a114d77620e-logs\") pod \"glance-default-external-api-0\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.038721 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd309d14-e0b5-4a84-93eb-4a114d77620e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.042389 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-djwk2"] Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.044647 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd309d14-e0b5-4a84-93eb-4a114d77620e-config-data\") pod \"glance-default-external-api-0\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: W1125 09:43:35.044751 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode522431f_0956_49f7_9c46_b0202824bd56.slice/crio-622a6e97957a89d6ee65cf73b5cdcd0bbdad37b21925e1450350bdaafaf94d2d WatchSource:0}: Error finding container 622a6e97957a89d6ee65cf73b5cdcd0bbdad37b21925e1450350bdaafaf94d2d: Status 404 returned error can't find the container with id 622a6e97957a89d6ee65cf73b5cdcd0bbdad37b21925e1450350bdaafaf94d2d Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.053400 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-lgdgb"] Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.064321 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd309d14-e0b5-4a84-93eb-4a114d77620e-scripts\") pod \"glance-default-external-api-0\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.071011 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.071761 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcd7m\" (UniqueName: \"kubernetes.io/projected/cd309d14-e0b5-4a84-93eb-4a114d77620e-kube-api-access-lcd7m\") pod \"glance-default-external-api-0\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.132595 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.132649 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.132670 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxnnd\" (UniqueName: \"kubernetes.io/projected/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-kube-api-access-zxnnd\") pod \"glance-default-internal-api-0\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.132686 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-logs\") pod \"glance-default-internal-api-0\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.132716 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.132746 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.132781 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.198688 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.212182 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-bdhw6"] Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.235123 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.236553 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.236616 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.236644 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxnnd\" (UniqueName: \"kubernetes.io/projected/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-kube-api-access-zxnnd\") pod \"glance-default-internal-api-0\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.236666 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-logs\") pod \"glance-default-internal-api-0\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.236709 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.236754 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.236800 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.239276 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.239622 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-logs\") pod \"glance-default-internal-api-0\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.243476 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.244498 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.248981 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76c58b6d97-jbkjb"] Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.253769 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.264946 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.274423 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxnnd\" (UniqueName: \"kubernetes.io/projected/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-kube-api-access-zxnnd\") pod \"glance-default-internal-api-0\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.294631 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-djwk2" event={"ID":"e522431f-0956-49f7-9c46-b0202824bd56","Type":"ContainerStarted","Data":"622a6e97957a89d6ee65cf73b5cdcd0bbdad37b21925e1450350bdaafaf94d2d"} Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.311640 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-lgdgb" event={"ID":"737e5e32-1fa7-4140-bfad-1bc3e10ee390","Type":"ContainerStarted","Data":"cb3326ae7455fbf36c1b1a64cf546d6c0ecc75a6352986c6cfd4f9551eca8c67"} Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.322244 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.327931 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-bdhw6" event={"ID":"3ff2a086-8bce-46e2-b294-3b3f31e39b4e","Type":"ContainerStarted","Data":"fe1ef695256ceeb1bb0ed94ba1c35484ff5f94b4ee7e90cb3f5b249ac69580e5"} Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.346612 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7dbf8bff67-jjvsj" event={"ID":"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152","Type":"ContainerStarted","Data":"3146a777e476c78be75e2e2b6981fe5d74cc7193a03d2083c984313c362ebe39"} Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.347298 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-ttw4k"] Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.353812 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" podUID="932e50fa-c886-4d45-8372-e671a7719010" containerName="dnsmasq-dns" containerID="cri-o://f6caa82b641d6b347437cba0f2afb987e0ce5625f8944257e0067c00afe42059" gracePeriod=10 Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.354080 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dxg4f" event={"ID":"58ed0f88-fcdb-491d-8eaa-5c81827ea82a","Type":"ContainerStarted","Data":"2e58b6a555e966ba5e956cf3709f8488b951a4de87d839026d06787ad5013431"} Nov 25 09:43:35 crc kubenswrapper[4776]: I1125 09:43:35.577446 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 09:43:36 crc kubenswrapper[4776]: I1125 09:43:36.364558 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ttw4k" event={"ID":"7c18590c-cf10-4c05-8a97-b9f1d091868a","Type":"ContainerStarted","Data":"f83314b7c339776a7e5269a458852df9d91d292b0aaeb191eaedaee0240c541a"} Nov 25 09:43:36 crc kubenswrapper[4776]: I1125 09:43:36.365682 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54","Type":"ContainerStarted","Data":"052f4d0d7461848ab646011b61a4b893ab6e2e7552c5cc6dc95a3f69ec0d85c7"} Nov 25 09:43:36 crc kubenswrapper[4776]: I1125 09:43:36.367607 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" event={"ID":"624c8057-3576-4249-9a63-5bb0dee2da0a","Type":"ContainerStarted","Data":"81aa9892ab356d20fd260d0a5c2590947e25c2167e81aeecf818a69d278ffc27"} Nov 25 09:43:36 crc kubenswrapper[4776]: I1125 09:43:36.580765 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 09:43:37 crc kubenswrapper[4776]: I1125 09:43:37.042354 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 09:43:37 crc kubenswrapper[4776]: W1125 09:43:37.052845 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd309d14_e0b5_4a84_93eb_4a114d77620e.slice/crio-7e1397c7ac544d19ab3e7715170643d1757fb6f336b5f82dd7f2019c44743dda WatchSource:0}: Error finding container 7e1397c7ac544d19ab3e7715170643d1757fb6f336b5f82dd7f2019c44743dda: Status 404 returned error can't find the container with id 7e1397c7ac544d19ab3e7715170643d1757fb6f336b5f82dd7f2019c44743dda Nov 25 09:43:37 crc kubenswrapper[4776]: I1125 09:43:37.377837 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dxg4f" event={"ID":"58ed0f88-fcdb-491d-8eaa-5c81827ea82a","Type":"ContainerStarted","Data":"32677d2e2acbdc045ea2ec0ca221f8ca335f3a14b393cf46c3db835819bda093"} Nov 25 09:43:37 crc kubenswrapper[4776]: I1125 09:43:37.392113 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cd309d14-e0b5-4a84-93eb-4a114d77620e","Type":"ContainerStarted","Data":"7e1397c7ac544d19ab3e7715170643d1757fb6f336b5f82dd7f2019c44743dda"} Nov 25 09:43:37 crc kubenswrapper[4776]: I1125 09:43:37.400213 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-dxg4f" podStartSLOduration=4.400194986 podStartE2EDuration="4.400194986s" podCreationTimestamp="2025-11-25 09:43:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:43:37.393895748 +0000 UTC m=+1162.434955301" watchObservedRunningTime="2025-11-25 09:43:37.400194986 +0000 UTC m=+1162.441254539" Nov 25 09:43:37 crc kubenswrapper[4776]: I1125 09:43:37.404577 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7","Type":"ContainerStarted","Data":"0d65de2b3517115753e0f758c12d1119aa81e2ec3450051748cdc86b558bb1a5"} Nov 25 09:43:37 crc kubenswrapper[4776]: I1125 09:43:37.404646 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7","Type":"ContainerStarted","Data":"d4e646a2f0e129b8e8732030127cecb6c6fcf937b9dc3d684f6f71d9e3f50ce8"} Nov 25 09:43:37 crc kubenswrapper[4776]: I1125 09:43:37.409172 4776 generic.go:334] "Generic (PLEG): container finished" podID="932e50fa-c886-4d45-8372-e671a7719010" containerID="f6caa82b641d6b347437cba0f2afb987e0ce5625f8944257e0067c00afe42059" exitCode=0 Nov 25 09:43:37 crc kubenswrapper[4776]: I1125 09:43:37.409250 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" event={"ID":"932e50fa-c886-4d45-8372-e671a7719010","Type":"ContainerDied","Data":"f6caa82b641d6b347437cba0f2afb987e0ce5625f8944257e0067c00afe42059"} Nov 25 09:43:37 crc kubenswrapper[4776]: I1125 09:43:37.416170 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-lgdgb" event={"ID":"737e5e32-1fa7-4140-bfad-1bc3e10ee390","Type":"ContainerStarted","Data":"820a46f0a557e2d6d0e23522585d4acb1b79eaa6244a8621f29bee90ba4a6713"} Nov 25 09:43:37 crc kubenswrapper[4776]: I1125 09:43:37.419162 4776 generic.go:334] "Generic (PLEG): container finished" podID="624c8057-3576-4249-9a63-5bb0dee2da0a" containerID="fd389d7b041963d9f2233f5abfaa28dabd63ef7d03b7319bb290986791d92571" exitCode=0 Nov 25 09:43:37 crc kubenswrapper[4776]: I1125 09:43:37.419216 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" event={"ID":"624c8057-3576-4249-9a63-5bb0dee2da0a","Type":"ContainerDied","Data":"fd389d7b041963d9f2233f5abfaa28dabd63ef7d03b7319bb290986791d92571"} Nov 25 09:43:37 crc kubenswrapper[4776]: I1125 09:43:37.423919 4776 generic.go:334] "Generic (PLEG): container finished" podID="bb5fe37b-e2b5-41e1-a60d-5f22e61fc152" containerID="05ed2bef8ed73ce2aec5b5e352201057d386bb62380986d2b7a149341aa808ea" exitCode=0 Nov 25 09:43:37 crc kubenswrapper[4776]: I1125 09:43:37.423962 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7dbf8bff67-jjvsj" event={"ID":"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152","Type":"ContainerDied","Data":"05ed2bef8ed73ce2aec5b5e352201057d386bb62380986d2b7a149341aa808ea"} Nov 25 09:43:37 crc kubenswrapper[4776]: I1125 09:43:37.434782 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-lgdgb" podStartSLOduration=4.434765033 podStartE2EDuration="4.434765033s" podCreationTimestamp="2025-11-25 09:43:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:43:37.432821744 +0000 UTC m=+1162.473881297" watchObservedRunningTime="2025-11-25 09:43:37.434765033 +0000 UTC m=+1162.475824576" Nov 25 09:43:37 crc kubenswrapper[4776]: I1125 09:43:37.934729 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.018164 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.028875 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.440445 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cd309d14-e0b5-4a84-93eb-4a114d77620e","Type":"ContainerStarted","Data":"8af963d7a3a021f44454dcbf16fad1b64d592359129e9d9a2f798ae8a8df4e8f"} Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.442887 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" event={"ID":"932e50fa-c886-4d45-8372-e671a7719010","Type":"ContainerDied","Data":"67a41c00e9ac113004b0c5695acc37bb220c2eef1c7009d0c58ca1f171be5792"} Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.442943 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67a41c00e9ac113004b0c5695acc37bb220c2eef1c7009d0c58ca1f171be5792" Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.540774 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.545051 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7dbf8bff67-jjvsj" Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.736395 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-ovsdbserver-sb\") pod \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\" (UID: \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\") " Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.736726 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-dns-svc\") pod \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\" (UID: \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\") " Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.736752 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-config\") pod \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\" (UID: \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\") " Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.736780 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-dns-swift-storage-0\") pod \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\" (UID: \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\") " Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.736810 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-dns-swift-storage-0\") pod \"932e50fa-c886-4d45-8372-e671a7719010\" (UID: \"932e50fa-c886-4d45-8372-e671a7719010\") " Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.736881 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-config\") pod \"932e50fa-c886-4d45-8372-e671a7719010\" (UID: \"932e50fa-c886-4d45-8372-e671a7719010\") " Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.736931 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-ovsdbserver-sb\") pod \"932e50fa-c886-4d45-8372-e671a7719010\" (UID: \"932e50fa-c886-4d45-8372-e671a7719010\") " Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.736982 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdbqv\" (UniqueName: \"kubernetes.io/projected/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-kube-api-access-bdbqv\") pod \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\" (UID: \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\") " Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.737000 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-ovsdbserver-nb\") pod \"932e50fa-c886-4d45-8372-e671a7719010\" (UID: \"932e50fa-c886-4d45-8372-e671a7719010\") " Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.737036 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-dns-svc\") pod \"932e50fa-c886-4d45-8372-e671a7719010\" (UID: \"932e50fa-c886-4d45-8372-e671a7719010\") " Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.737080 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rm9ng\" (UniqueName: \"kubernetes.io/projected/932e50fa-c886-4d45-8372-e671a7719010-kube-api-access-rm9ng\") pod \"932e50fa-c886-4d45-8372-e671a7719010\" (UID: \"932e50fa-c886-4d45-8372-e671a7719010\") " Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.737109 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-ovsdbserver-nb\") pod \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\" (UID: \"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152\") " Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.742438 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/932e50fa-c886-4d45-8372-e671a7719010-kube-api-access-rm9ng" (OuterVolumeSpecName: "kube-api-access-rm9ng") pod "932e50fa-c886-4d45-8372-e671a7719010" (UID: "932e50fa-c886-4d45-8372-e671a7719010"). InnerVolumeSpecName "kube-api-access-rm9ng". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.750908 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-kube-api-access-bdbqv" (OuterVolumeSpecName: "kube-api-access-bdbqv") pod "bb5fe37b-e2b5-41e1-a60d-5f22e61fc152" (UID: "bb5fe37b-e2b5-41e1-a60d-5f22e61fc152"). InnerVolumeSpecName "kube-api-access-bdbqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.770577 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bb5fe37b-e2b5-41e1-a60d-5f22e61fc152" (UID: "bb5fe37b-e2b5-41e1-a60d-5f22e61fc152"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.774794 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bb5fe37b-e2b5-41e1-a60d-5f22e61fc152" (UID: "bb5fe37b-e2b5-41e1-a60d-5f22e61fc152"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.790224 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bb5fe37b-e2b5-41e1-a60d-5f22e61fc152" (UID: "bb5fe37b-e2b5-41e1-a60d-5f22e61fc152"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.795601 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bb5fe37b-e2b5-41e1-a60d-5f22e61fc152" (UID: "bb5fe37b-e2b5-41e1-a60d-5f22e61fc152"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.811210 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "932e50fa-c886-4d45-8372-e671a7719010" (UID: "932e50fa-c886-4d45-8372-e671a7719010"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.811600 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-config" (OuterVolumeSpecName: "config") pod "bb5fe37b-e2b5-41e1-a60d-5f22e61fc152" (UID: "bb5fe37b-e2b5-41e1-a60d-5f22e61fc152"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.811750 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "932e50fa-c886-4d45-8372-e671a7719010" (UID: "932e50fa-c886-4d45-8372-e671a7719010"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.826521 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "932e50fa-c886-4d45-8372-e671a7719010" (UID: "932e50fa-c886-4d45-8372-e671a7719010"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.829724 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "932e50fa-c886-4d45-8372-e671a7719010" (UID: "932e50fa-c886-4d45-8372-e671a7719010"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.831201 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-config" (OuterVolumeSpecName: "config") pod "932e50fa-c886-4d45-8372-e671a7719010" (UID: "932e50fa-c886-4d45-8372-e671a7719010"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.840959 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rm9ng\" (UniqueName: \"kubernetes.io/projected/932e50fa-c886-4d45-8372-e671a7719010-kube-api-access-rm9ng\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.841042 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.841056 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.841109 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.841146 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.841186 4776 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.841199 4776 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.841210 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.841221 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.841231 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdbqv\" (UniqueName: \"kubernetes.io/projected/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152-kube-api-access-bdbqv\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.841242 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:38 crc kubenswrapper[4776]: I1125 09:43:38.841277 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/932e50fa-c886-4d45-8372-e671a7719010-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:39 crc kubenswrapper[4776]: I1125 09:43:39.459365 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" event={"ID":"624c8057-3576-4249-9a63-5bb0dee2da0a","Type":"ContainerStarted","Data":"da36d0549e285577f46082c15a99da5ebd9c6b191beddf067a2807915ebcfe1c"} Nov 25 09:43:39 crc kubenswrapper[4776]: I1125 09:43:39.459997 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" Nov 25 09:43:39 crc kubenswrapper[4776]: I1125 09:43:39.463109 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7dbf8bff67-jjvsj" event={"ID":"bb5fe37b-e2b5-41e1-a60d-5f22e61fc152","Type":"ContainerDied","Data":"3146a777e476c78be75e2e2b6981fe5d74cc7193a03d2083c984313c362ebe39"} Nov 25 09:43:39 crc kubenswrapper[4776]: I1125 09:43:39.463169 4776 scope.go:117] "RemoveContainer" containerID="05ed2bef8ed73ce2aec5b5e352201057d386bb62380986d2b7a149341aa808ea" Nov 25 09:43:39 crc kubenswrapper[4776]: I1125 09:43:39.463284 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7dbf8bff67-jjvsj" Nov 25 09:43:39 crc kubenswrapper[4776]: I1125 09:43:39.469561 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cd309d14-e0b5-4a84-93eb-4a114d77620e","Type":"ContainerStarted","Data":"699b3fe4188781d3be4d368031c52d516db4577f759cf786de35bbab1b34d653"} Nov 25 09:43:39 crc kubenswrapper[4776]: I1125 09:43:39.469745 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="cd309d14-e0b5-4a84-93eb-4a114d77620e" containerName="glance-log" containerID="cri-o://8af963d7a3a021f44454dcbf16fad1b64d592359129e9d9a2f798ae8a8df4e8f" gracePeriod=30 Nov 25 09:43:39 crc kubenswrapper[4776]: I1125 09:43:39.469862 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="cd309d14-e0b5-4a84-93eb-4a114d77620e" containerName="glance-httpd" containerID="cri-o://699b3fe4188781d3be4d368031c52d516db4577f759cf786de35bbab1b34d653" gracePeriod=30 Nov 25 09:43:39 crc kubenswrapper[4776]: I1125 09:43:39.488689 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6856c564b9-zvxvz" Nov 25 09:43:39 crc kubenswrapper[4776]: I1125 09:43:39.488972 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" podStartSLOduration=5.488952832 podStartE2EDuration="5.488952832s" podCreationTimestamp="2025-11-25 09:43:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:43:39.488808068 +0000 UTC m=+1164.529867621" watchObservedRunningTime="2025-11-25 09:43:39.488952832 +0000 UTC m=+1164.530012385" Nov 25 09:43:39 crc kubenswrapper[4776]: I1125 09:43:39.489207 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="2a648d65-7ff4-4d5d-afa3-96f628d9a4c7" containerName="glance-log" containerID="cri-o://0d65de2b3517115753e0f758c12d1119aa81e2ec3450051748cdc86b558bb1a5" gracePeriod=30 Nov 25 09:43:39 crc kubenswrapper[4776]: I1125 09:43:39.489442 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="2a648d65-7ff4-4d5d-afa3-96f628d9a4c7" containerName="glance-httpd" containerID="cri-o://11d6db9ac2860e643f2c028a77019007a3c406ac1fbdd49185cafdfd3dbf9b0a" gracePeriod=30 Nov 25 09:43:39 crc kubenswrapper[4776]: I1125 09:43:39.489645 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7","Type":"ContainerStarted","Data":"11d6db9ac2860e643f2c028a77019007a3c406ac1fbdd49185cafdfd3dbf9b0a"} Nov 25 09:43:39 crc kubenswrapper[4776]: I1125 09:43:39.525687 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.525666602 podStartE2EDuration="6.525666602s" podCreationTimestamp="2025-11-25 09:43:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:43:39.514690607 +0000 UTC m=+1164.555750160" watchObservedRunningTime="2025-11-25 09:43:39.525666602 +0000 UTC m=+1164.566726155" Nov 25 09:43:39 crc kubenswrapper[4776]: I1125 09:43:39.557024 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.557004368 podStartE2EDuration="6.557004368s" podCreationTimestamp="2025-11-25 09:43:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:43:39.551666774 +0000 UTC m=+1164.592726327" watchObservedRunningTime="2025-11-25 09:43:39.557004368 +0000 UTC m=+1164.598063921" Nov 25 09:43:39 crc kubenswrapper[4776]: I1125 09:43:39.755867 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7dbf8bff67-jjvsj"] Nov 25 09:43:39 crc kubenswrapper[4776]: I1125 09:43:39.765378 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7dbf8bff67-jjvsj"] Nov 25 09:43:39 crc kubenswrapper[4776]: I1125 09:43:39.774657 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6856c564b9-zvxvz"] Nov 25 09:43:39 crc kubenswrapper[4776]: I1125 09:43:39.781918 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6856c564b9-zvxvz"] Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.253655 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.373901 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"cd309d14-e0b5-4a84-93eb-4a114d77620e\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") " Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.374039 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cd309d14-e0b5-4a84-93eb-4a114d77620e-httpd-run\") pod \"cd309d14-e0b5-4a84-93eb-4a114d77620e\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") " Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.374204 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcd7m\" (UniqueName: \"kubernetes.io/projected/cd309d14-e0b5-4a84-93eb-4a114d77620e-kube-api-access-lcd7m\") pod \"cd309d14-e0b5-4a84-93eb-4a114d77620e\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") " Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.374298 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd309d14-e0b5-4a84-93eb-4a114d77620e-scripts\") pod \"cd309d14-e0b5-4a84-93eb-4a114d77620e\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") " Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.374351 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd309d14-e0b5-4a84-93eb-4a114d77620e-logs\") pod \"cd309d14-e0b5-4a84-93eb-4a114d77620e\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") " Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.374542 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd309d14-e0b5-4a84-93eb-4a114d77620e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "cd309d14-e0b5-4a84-93eb-4a114d77620e" (UID: "cd309d14-e0b5-4a84-93eb-4a114d77620e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.374790 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd309d14-e0b5-4a84-93eb-4a114d77620e-config-data\") pod \"cd309d14-e0b5-4a84-93eb-4a114d77620e\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") " Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.374825 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd309d14-e0b5-4a84-93eb-4a114d77620e-combined-ca-bundle\") pod \"cd309d14-e0b5-4a84-93eb-4a114d77620e\" (UID: \"cd309d14-e0b5-4a84-93eb-4a114d77620e\") " Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.375001 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd309d14-e0b5-4a84-93eb-4a114d77620e-logs" (OuterVolumeSpecName: "logs") pod "cd309d14-e0b5-4a84-93eb-4a114d77620e" (UID: "cd309d14-e0b5-4a84-93eb-4a114d77620e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.375837 4776 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cd309d14-e0b5-4a84-93eb-4a114d77620e-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.375868 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd309d14-e0b5-4a84-93eb-4a114d77620e-logs\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.379575 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd309d14-e0b5-4a84-93eb-4a114d77620e-scripts" (OuterVolumeSpecName: "scripts") pod "cd309d14-e0b5-4a84-93eb-4a114d77620e" (UID: "cd309d14-e0b5-4a84-93eb-4a114d77620e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.380243 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd309d14-e0b5-4a84-93eb-4a114d77620e-kube-api-access-lcd7m" (OuterVolumeSpecName: "kube-api-access-lcd7m") pod "cd309d14-e0b5-4a84-93eb-4a114d77620e" (UID: "cd309d14-e0b5-4a84-93eb-4a114d77620e"). InnerVolumeSpecName "kube-api-access-lcd7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.389510 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "cd309d14-e0b5-4a84-93eb-4a114d77620e" (UID: "cd309d14-e0b5-4a84-93eb-4a114d77620e"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.397018 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.405546 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd309d14-e0b5-4a84-93eb-4a114d77620e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd309d14-e0b5-4a84-93eb-4a114d77620e" (UID: "cd309d14-e0b5-4a84-93eb-4a114d77620e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.434751 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd309d14-e0b5-4a84-93eb-4a114d77620e-config-data" (OuterVolumeSpecName: "config-data") pod "cd309d14-e0b5-4a84-93eb-4a114d77620e" (UID: "cd309d14-e0b5-4a84-93eb-4a114d77620e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.478145 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcd7m\" (UniqueName: \"kubernetes.io/projected/cd309d14-e0b5-4a84-93eb-4a114d77620e-kube-api-access-lcd7m\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.478203 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd309d14-e0b5-4a84-93eb-4a114d77620e-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.478216 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd309d14-e0b5-4a84-93eb-4a114d77620e-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.478248 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd309d14-e0b5-4a84-93eb-4a114d77620e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.478290 4776 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.497863 4776 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.504931 4776 generic.go:334] "Generic (PLEG): container finished" podID="cd309d14-e0b5-4a84-93eb-4a114d77620e" containerID="699b3fe4188781d3be4d368031c52d516db4577f759cf786de35bbab1b34d653" exitCode=143 Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.504984 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.504992 4776 generic.go:334] "Generic (PLEG): container finished" podID="cd309d14-e0b5-4a84-93eb-4a114d77620e" containerID="8af963d7a3a021f44454dcbf16fad1b64d592359129e9d9a2f798ae8a8df4e8f" exitCode=143 Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.505027 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cd309d14-e0b5-4a84-93eb-4a114d77620e","Type":"ContainerDied","Data":"699b3fe4188781d3be4d368031c52d516db4577f759cf786de35bbab1b34d653"} Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.505120 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cd309d14-e0b5-4a84-93eb-4a114d77620e","Type":"ContainerDied","Data":"8af963d7a3a021f44454dcbf16fad1b64d592359129e9d9a2f798ae8a8df4e8f"} Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.505136 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cd309d14-e0b5-4a84-93eb-4a114d77620e","Type":"ContainerDied","Data":"7e1397c7ac544d19ab3e7715170643d1757fb6f336b5f82dd7f2019c44743dda"} Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.505159 4776 scope.go:117] "RemoveContainer" containerID="699b3fe4188781d3be4d368031c52d516db4577f759cf786de35bbab1b34d653" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.507756 4776 generic.go:334] "Generic (PLEG): container finished" podID="2a648d65-7ff4-4d5d-afa3-96f628d9a4c7" containerID="11d6db9ac2860e643f2c028a77019007a3c406ac1fbdd49185cafdfd3dbf9b0a" exitCode=0 Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.507782 4776 generic.go:334] "Generic (PLEG): container finished" podID="2a648d65-7ff4-4d5d-afa3-96f628d9a4c7" containerID="0d65de2b3517115753e0f758c12d1119aa81e2ec3450051748cdc86b558bb1a5" exitCode=143 Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.507802 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7","Type":"ContainerDied","Data":"11d6db9ac2860e643f2c028a77019007a3c406ac1fbdd49185cafdfd3dbf9b0a"} Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.507849 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.507867 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7","Type":"ContainerDied","Data":"0d65de2b3517115753e0f758c12d1119aa81e2ec3450051748cdc86b558bb1a5"} Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.507881 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7","Type":"ContainerDied","Data":"d4e646a2f0e129b8e8732030127cecb6c6fcf937b9dc3d684f6f71d9e3f50ce8"} Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.544806 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.558728 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.561915 4776 scope.go:117] "RemoveContainer" containerID="8af963d7a3a021f44454dcbf16fad1b64d592359129e9d9a2f798ae8a8df4e8f" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.573910 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 09:43:40 crc kubenswrapper[4776]: E1125 09:43:40.574371 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb5fe37b-e2b5-41e1-a60d-5f22e61fc152" containerName="init" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.574393 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb5fe37b-e2b5-41e1-a60d-5f22e61fc152" containerName="init" Nov 25 09:43:40 crc kubenswrapper[4776]: E1125 09:43:40.574410 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a648d65-7ff4-4d5d-afa3-96f628d9a4c7" containerName="glance-log" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.574419 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a648d65-7ff4-4d5d-afa3-96f628d9a4c7" containerName="glance-log" Nov 25 09:43:40 crc kubenswrapper[4776]: E1125 09:43:40.574453 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd309d14-e0b5-4a84-93eb-4a114d77620e" containerName="glance-log" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.574460 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd309d14-e0b5-4a84-93eb-4a114d77620e" containerName="glance-log" Nov 25 09:43:40 crc kubenswrapper[4776]: E1125 09:43:40.574472 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a648d65-7ff4-4d5d-afa3-96f628d9a4c7" containerName="glance-httpd" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.574478 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a648d65-7ff4-4d5d-afa3-96f628d9a4c7" containerName="glance-httpd" Nov 25 09:43:40 crc kubenswrapper[4776]: E1125 09:43:40.574492 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="932e50fa-c886-4d45-8372-e671a7719010" containerName="dnsmasq-dns" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.574499 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="932e50fa-c886-4d45-8372-e671a7719010" containerName="dnsmasq-dns" Nov 25 09:43:40 crc kubenswrapper[4776]: E1125 09:43:40.574515 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="932e50fa-c886-4d45-8372-e671a7719010" containerName="init" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.574522 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="932e50fa-c886-4d45-8372-e671a7719010" containerName="init" Nov 25 09:43:40 crc kubenswrapper[4776]: E1125 09:43:40.574542 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd309d14-e0b5-4a84-93eb-4a114d77620e" containerName="glance-httpd" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.574550 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd309d14-e0b5-4a84-93eb-4a114d77620e" containerName="glance-httpd" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.574753 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb5fe37b-e2b5-41e1-a60d-5f22e61fc152" containerName="init" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.574778 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a648d65-7ff4-4d5d-afa3-96f628d9a4c7" containerName="glance-httpd" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.574790 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a648d65-7ff4-4d5d-afa3-96f628d9a4c7" containerName="glance-log" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.574800 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="932e50fa-c886-4d45-8372-e671a7719010" containerName="dnsmasq-dns" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.574814 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd309d14-e0b5-4a84-93eb-4a114d77620e" containerName="glance-httpd" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.574824 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd309d14-e0b5-4a84-93eb-4a114d77620e" containerName="glance-log" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.575957 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.579583 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-scripts\") pod \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") " Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.579681 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-config-data\") pod \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") " Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.579807 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-httpd-run\") pod \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") " Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.579851 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxnnd\" (UniqueName: \"kubernetes.io/projected/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-kube-api-access-zxnnd\") pod \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") " Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.579871 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-combined-ca-bundle\") pod \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") " Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.579908 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") " Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.579943 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-logs\") pod \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\" (UID: \"2a648d65-7ff4-4d5d-afa3-96f628d9a4c7\") " Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.579976 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.580405 4776 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.580657 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-logs" (OuterVolumeSpecName: "logs") pod "2a648d65-7ff4-4d5d-afa3-96f628d9a4c7" (UID: "2a648d65-7ff4-4d5d-afa3-96f628d9a4c7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.582332 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2a648d65-7ff4-4d5d-afa3-96f628d9a4c7" (UID: "2a648d65-7ff4-4d5d-afa3-96f628d9a4c7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.585909 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "2a648d65-7ff4-4d5d-afa3-96f628d9a4c7" (UID: "2a648d65-7ff4-4d5d-afa3-96f628d9a4c7"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.598762 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-scripts" (OuterVolumeSpecName: "scripts") pod "2a648d65-7ff4-4d5d-afa3-96f628d9a4c7" (UID: "2a648d65-7ff4-4d5d-afa3-96f628d9a4c7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.598818 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-kube-api-access-zxnnd" (OuterVolumeSpecName: "kube-api-access-zxnnd") pod "2a648d65-7ff4-4d5d-afa3-96f628d9a4c7" (UID: "2a648d65-7ff4-4d5d-afa3-96f628d9a4c7"). InnerVolumeSpecName "kube-api-access-zxnnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.612011 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.632110 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2a648d65-7ff4-4d5d-afa3-96f628d9a4c7" (UID: "2a648d65-7ff4-4d5d-afa3-96f628d9a4c7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.642516 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-config-data" (OuterVolumeSpecName: "config-data") pod "2a648d65-7ff4-4d5d-afa3-96f628d9a4c7" (UID: "2a648d65-7ff4-4d5d-afa3-96f628d9a4c7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.649619 4776 scope.go:117] "RemoveContainer" containerID="699b3fe4188781d3be4d368031c52d516db4577f759cf786de35bbab1b34d653" Nov 25 09:43:40 crc kubenswrapper[4776]: E1125 09:43:40.650186 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"699b3fe4188781d3be4d368031c52d516db4577f759cf786de35bbab1b34d653\": container with ID starting with 699b3fe4188781d3be4d368031c52d516db4577f759cf786de35bbab1b34d653 not found: ID does not exist" containerID="699b3fe4188781d3be4d368031c52d516db4577f759cf786de35bbab1b34d653" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.650216 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"699b3fe4188781d3be4d368031c52d516db4577f759cf786de35bbab1b34d653"} err="failed to get container status \"699b3fe4188781d3be4d368031c52d516db4577f759cf786de35bbab1b34d653\": rpc error: code = NotFound desc = could not find container \"699b3fe4188781d3be4d368031c52d516db4577f759cf786de35bbab1b34d653\": container with ID starting with 699b3fe4188781d3be4d368031c52d516db4577f759cf786de35bbab1b34d653 not found: ID does not exist" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.650244 4776 scope.go:117] "RemoveContainer" containerID="8af963d7a3a021f44454dcbf16fad1b64d592359129e9d9a2f798ae8a8df4e8f" Nov 25 09:43:40 crc kubenswrapper[4776]: E1125 09:43:40.650559 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8af963d7a3a021f44454dcbf16fad1b64d592359129e9d9a2f798ae8a8df4e8f\": container with ID starting with 8af963d7a3a021f44454dcbf16fad1b64d592359129e9d9a2f798ae8a8df4e8f not found: ID does not exist" containerID="8af963d7a3a021f44454dcbf16fad1b64d592359129e9d9a2f798ae8a8df4e8f" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.650583 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8af963d7a3a021f44454dcbf16fad1b64d592359129e9d9a2f798ae8a8df4e8f"} err="failed to get container status \"8af963d7a3a021f44454dcbf16fad1b64d592359129e9d9a2f798ae8a8df4e8f\": rpc error: code = NotFound desc = could not find container \"8af963d7a3a021f44454dcbf16fad1b64d592359129e9d9a2f798ae8a8df4e8f\": container with ID starting with 8af963d7a3a021f44454dcbf16fad1b64d592359129e9d9a2f798ae8a8df4e8f not found: ID does not exist" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.650597 4776 scope.go:117] "RemoveContainer" containerID="699b3fe4188781d3be4d368031c52d516db4577f759cf786de35bbab1b34d653" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.650911 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"699b3fe4188781d3be4d368031c52d516db4577f759cf786de35bbab1b34d653"} err="failed to get container status \"699b3fe4188781d3be4d368031c52d516db4577f759cf786de35bbab1b34d653\": rpc error: code = NotFound desc = could not find container \"699b3fe4188781d3be4d368031c52d516db4577f759cf786de35bbab1b34d653\": container with ID starting with 699b3fe4188781d3be4d368031c52d516db4577f759cf786de35bbab1b34d653 not found: ID does not exist" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.650933 4776 scope.go:117] "RemoveContainer" containerID="8af963d7a3a021f44454dcbf16fad1b64d592359129e9d9a2f798ae8a8df4e8f" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.651326 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8af963d7a3a021f44454dcbf16fad1b64d592359129e9d9a2f798ae8a8df4e8f"} err="failed to get container status \"8af963d7a3a021f44454dcbf16fad1b64d592359129e9d9a2f798ae8a8df4e8f\": rpc error: code = NotFound desc = could not find container \"8af963d7a3a021f44454dcbf16fad1b64d592359129e9d9a2f798ae8a8df4e8f\": container with ID starting with 8af963d7a3a021f44454dcbf16fad1b64d592359129e9d9a2f798ae8a8df4e8f not found: ID does not exist" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.651346 4776 scope.go:117] "RemoveContainer" containerID="11d6db9ac2860e643f2c028a77019007a3c406ac1fbdd49185cafdfd3dbf9b0a" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.680190 4776 scope.go:117] "RemoveContainer" containerID="0d65de2b3517115753e0f758c12d1119aa81e2ec3450051748cdc86b558bb1a5" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.681857 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/923266d8-bfe1-4683-9806-0caa0719be85-scripts\") pod \"glance-default-external-api-0\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.681904 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/923266d8-bfe1-4683-9806-0caa0719be85-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.681947 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/923266d8-bfe1-4683-9806-0caa0719be85-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.682267 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d86lj\" (UniqueName: \"kubernetes.io/projected/923266d8-bfe1-4683-9806-0caa0719be85-kube-api-access-d86lj\") pod \"glance-default-external-api-0\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.682317 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/923266d8-bfe1-4683-9806-0caa0719be85-config-data\") pod \"glance-default-external-api-0\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.682347 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.682406 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/923266d8-bfe1-4683-9806-0caa0719be85-logs\") pod \"glance-default-external-api-0\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.682448 4776 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.682460 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxnnd\" (UniqueName: \"kubernetes.io/projected/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-kube-api-access-zxnnd\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.682469 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.682489 4776 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.682498 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-logs\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.682506 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.682516 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.718203 4776 scope.go:117] "RemoveContainer" containerID="11d6db9ac2860e643f2c028a77019007a3c406ac1fbdd49185cafdfd3dbf9b0a" Nov 25 09:43:40 crc kubenswrapper[4776]: E1125 09:43:40.720288 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11d6db9ac2860e643f2c028a77019007a3c406ac1fbdd49185cafdfd3dbf9b0a\": container with ID starting with 11d6db9ac2860e643f2c028a77019007a3c406ac1fbdd49185cafdfd3dbf9b0a not found: ID does not exist" containerID="11d6db9ac2860e643f2c028a77019007a3c406ac1fbdd49185cafdfd3dbf9b0a" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.720350 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11d6db9ac2860e643f2c028a77019007a3c406ac1fbdd49185cafdfd3dbf9b0a"} err="failed to get container status \"11d6db9ac2860e643f2c028a77019007a3c406ac1fbdd49185cafdfd3dbf9b0a\": rpc error: code = NotFound desc = could not find container \"11d6db9ac2860e643f2c028a77019007a3c406ac1fbdd49185cafdfd3dbf9b0a\": container with ID starting with 11d6db9ac2860e643f2c028a77019007a3c406ac1fbdd49185cafdfd3dbf9b0a not found: ID does not exist" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.720380 4776 scope.go:117] "RemoveContainer" containerID="0d65de2b3517115753e0f758c12d1119aa81e2ec3450051748cdc86b558bb1a5" Nov 25 09:43:40 crc kubenswrapper[4776]: E1125 09:43:40.721161 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d65de2b3517115753e0f758c12d1119aa81e2ec3450051748cdc86b558bb1a5\": container with ID starting with 0d65de2b3517115753e0f758c12d1119aa81e2ec3450051748cdc86b558bb1a5 not found: ID does not exist" containerID="0d65de2b3517115753e0f758c12d1119aa81e2ec3450051748cdc86b558bb1a5" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.721183 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d65de2b3517115753e0f758c12d1119aa81e2ec3450051748cdc86b558bb1a5"} err="failed to get container status \"0d65de2b3517115753e0f758c12d1119aa81e2ec3450051748cdc86b558bb1a5\": rpc error: code = NotFound desc = could not find container \"0d65de2b3517115753e0f758c12d1119aa81e2ec3450051748cdc86b558bb1a5\": container with ID starting with 0d65de2b3517115753e0f758c12d1119aa81e2ec3450051748cdc86b558bb1a5 not found: ID does not exist" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.721197 4776 scope.go:117] "RemoveContainer" containerID="11d6db9ac2860e643f2c028a77019007a3c406ac1fbdd49185cafdfd3dbf9b0a" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.721484 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11d6db9ac2860e643f2c028a77019007a3c406ac1fbdd49185cafdfd3dbf9b0a"} err="failed to get container status \"11d6db9ac2860e643f2c028a77019007a3c406ac1fbdd49185cafdfd3dbf9b0a\": rpc error: code = NotFound desc = could not find container \"11d6db9ac2860e643f2c028a77019007a3c406ac1fbdd49185cafdfd3dbf9b0a\": container with ID starting with 11d6db9ac2860e643f2c028a77019007a3c406ac1fbdd49185cafdfd3dbf9b0a not found: ID does not exist" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.721504 4776 scope.go:117] "RemoveContainer" containerID="0d65de2b3517115753e0f758c12d1119aa81e2ec3450051748cdc86b558bb1a5" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.722204 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d65de2b3517115753e0f758c12d1119aa81e2ec3450051748cdc86b558bb1a5"} err="failed to get container status \"0d65de2b3517115753e0f758c12d1119aa81e2ec3450051748cdc86b558bb1a5\": rpc error: code = NotFound desc = could not find container \"0d65de2b3517115753e0f758c12d1119aa81e2ec3450051748cdc86b558bb1a5\": container with ID starting with 0d65de2b3517115753e0f758c12d1119aa81e2ec3450051748cdc86b558bb1a5 not found: ID does not exist" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.725388 4776 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.787419 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/923266d8-bfe1-4683-9806-0caa0719be85-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.787513 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d86lj\" (UniqueName: \"kubernetes.io/projected/923266d8-bfe1-4683-9806-0caa0719be85-kube-api-access-d86lj\") pod \"glance-default-external-api-0\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.787554 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/923266d8-bfe1-4683-9806-0caa0719be85-config-data\") pod \"glance-default-external-api-0\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.787621 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.787708 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/923266d8-bfe1-4683-9806-0caa0719be85-logs\") pod \"glance-default-external-api-0\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.787755 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/923266d8-bfe1-4683-9806-0caa0719be85-scripts\") pod \"glance-default-external-api-0\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.787784 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/923266d8-bfe1-4683-9806-0caa0719be85-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.787840 4776 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.787920 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/923266d8-bfe1-4683-9806-0caa0719be85-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.788125 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.791830 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/923266d8-bfe1-4683-9806-0caa0719be85-logs\") pod \"glance-default-external-api-0\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.792702 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/923266d8-bfe1-4683-9806-0caa0719be85-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.795448 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/923266d8-bfe1-4683-9806-0caa0719be85-scripts\") pod \"glance-default-external-api-0\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.797211 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/923266d8-bfe1-4683-9806-0caa0719be85-config-data\") pod \"glance-default-external-api-0\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.819183 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d86lj\" (UniqueName: \"kubernetes.io/projected/923266d8-bfe1-4683-9806-0caa0719be85-kube-api-access-d86lj\") pod \"glance-default-external-api-0\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.848891 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") " pod="openstack/glance-default-external-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.849534 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.865540 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.876165 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.877966 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.880593 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.892060 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.916848 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.991757 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.991827 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3689daf9-7cc3-4752-ab65-84047549c8f7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.991850 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3689daf9-7cc3-4752-ab65-84047549c8f7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.991917 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmfzt\" (UniqueName: \"kubernetes.io/projected/3689daf9-7cc3-4752-ab65-84047549c8f7-kube-api-access-hmfzt\") pod \"glance-default-internal-api-0\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.991946 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3689daf9-7cc3-4752-ab65-84047549c8f7-logs\") pod \"glance-default-internal-api-0\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.991970 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3689daf9-7cc3-4752-ab65-84047549c8f7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:40 crc kubenswrapper[4776]: I1125 09:43:40.992000 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3689daf9-7cc3-4752-ab65-84047549c8f7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:41 crc kubenswrapper[4776]: E1125 09:43:41.067448 4776 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58ed0f88_fcdb_491d_8eaa_5c81827ea82a.slice/crio-conmon-32677d2e2acbdc045ea2ec0ca221f8ca335f3a14b393cf46c3db835819bda093.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a648d65_7ff4_4d5d_afa3_96f628d9a4c7.slice/crio-d4e646a2f0e129b8e8732030127cecb6c6fcf937b9dc3d684f6f71d9e3f50ce8\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a648d65_7ff4_4d5d_afa3_96f628d9a4c7.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58ed0f88_fcdb_491d_8eaa_5c81827ea82a.slice/crio-32677d2e2acbdc045ea2ec0ca221f8ca335f3a14b393cf46c3db835819bda093.scope\": RecentStats: unable to find data in memory cache]" Nov 25 09:43:41 crc kubenswrapper[4776]: I1125 09:43:41.093437 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3689daf9-7cc3-4752-ab65-84047549c8f7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:41 crc kubenswrapper[4776]: I1125 09:43:41.093747 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3689daf9-7cc3-4752-ab65-84047549c8f7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:41 crc kubenswrapper[4776]: I1125 09:43:41.093825 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmfzt\" (UniqueName: \"kubernetes.io/projected/3689daf9-7cc3-4752-ab65-84047549c8f7-kube-api-access-hmfzt\") pod \"glance-default-internal-api-0\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:41 crc kubenswrapper[4776]: I1125 09:43:41.093850 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3689daf9-7cc3-4752-ab65-84047549c8f7-logs\") pod \"glance-default-internal-api-0\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:41 crc kubenswrapper[4776]: I1125 09:43:41.093872 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3689daf9-7cc3-4752-ab65-84047549c8f7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:41 crc kubenswrapper[4776]: I1125 09:43:41.093899 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3689daf9-7cc3-4752-ab65-84047549c8f7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:41 crc kubenswrapper[4776]: I1125 09:43:41.093963 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:41 crc kubenswrapper[4776]: I1125 09:43:41.094464 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Nov 25 09:43:41 crc kubenswrapper[4776]: I1125 09:43:41.095051 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3689daf9-7cc3-4752-ab65-84047549c8f7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:41 crc kubenswrapper[4776]: I1125 09:43:41.095143 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3689daf9-7cc3-4752-ab65-84047549c8f7-logs\") pod \"glance-default-internal-api-0\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:41 crc kubenswrapper[4776]: I1125 09:43:41.100536 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3689daf9-7cc3-4752-ab65-84047549c8f7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:41 crc kubenswrapper[4776]: I1125 09:43:41.103865 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3689daf9-7cc3-4752-ab65-84047549c8f7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:41 crc kubenswrapper[4776]: I1125 09:43:41.113057 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmfzt\" (UniqueName: \"kubernetes.io/projected/3689daf9-7cc3-4752-ab65-84047549c8f7-kube-api-access-hmfzt\") pod \"glance-default-internal-api-0\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:41 crc kubenswrapper[4776]: I1125 09:43:41.113200 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3689daf9-7cc3-4752-ab65-84047549c8f7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:41 crc kubenswrapper[4776]: I1125 09:43:41.138875 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:43:41 crc kubenswrapper[4776]: I1125 09:43:41.210457 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 09:43:41 crc kubenswrapper[4776]: I1125 09:43:41.521595 4776 generic.go:334] "Generic (PLEG): container finished" podID="58ed0f88-fcdb-491d-8eaa-5c81827ea82a" containerID="32677d2e2acbdc045ea2ec0ca221f8ca335f3a14b393cf46c3db835819bda093" exitCode=0 Nov 25 09:43:41 crc kubenswrapper[4776]: I1125 09:43:41.521671 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dxg4f" event={"ID":"58ed0f88-fcdb-491d-8eaa-5c81827ea82a","Type":"ContainerDied","Data":"32677d2e2acbdc045ea2ec0ca221f8ca335f3a14b393cf46c3db835819bda093"} Nov 25 09:43:41 crc kubenswrapper[4776]: I1125 09:43:41.678722 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a648d65-7ff4-4d5d-afa3-96f628d9a4c7" path="/var/lib/kubelet/pods/2a648d65-7ff4-4d5d-afa3-96f628d9a4c7/volumes" Nov 25 09:43:41 crc kubenswrapper[4776]: I1125 09:43:41.679879 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="932e50fa-c886-4d45-8372-e671a7719010" path="/var/lib/kubelet/pods/932e50fa-c886-4d45-8372-e671a7719010/volumes" Nov 25 09:43:41 crc kubenswrapper[4776]: I1125 09:43:41.681162 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb5fe37b-e2b5-41e1-a60d-5f22e61fc152" path="/var/lib/kubelet/pods/bb5fe37b-e2b5-41e1-a60d-5f22e61fc152/volumes" Nov 25 09:43:41 crc kubenswrapper[4776]: I1125 09:43:41.683378 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd309d14-e0b5-4a84-93eb-4a114d77620e" path="/var/lib/kubelet/pods/cd309d14-e0b5-4a84-93eb-4a114d77620e/volumes" Nov 25 09:43:42 crc kubenswrapper[4776]: I1125 09:43:42.117994 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 09:43:42 crc kubenswrapper[4776]: I1125 09:43:42.231258 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 09:43:44 crc kubenswrapper[4776]: I1125 09:43:44.524214 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" Nov 25 09:43:44 crc kubenswrapper[4776]: I1125 09:43:44.588455 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9fdb784c-sh56t"] Nov 25 09:43:44 crc kubenswrapper[4776]: I1125 09:43:44.588695 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" podUID="adadb0a2-c265-4686-98e0-e3d8ff014b35" containerName="dnsmasq-dns" containerID="cri-o://b625f8f354e8416dd5bba6f485f43a6b0b7956f3d528292b946646a11591847a" gracePeriod=10 Nov 25 09:43:45 crc kubenswrapper[4776]: I1125 09:43:45.588869 4776 generic.go:334] "Generic (PLEG): container finished" podID="adadb0a2-c265-4686-98e0-e3d8ff014b35" containerID="b625f8f354e8416dd5bba6f485f43a6b0b7956f3d528292b946646a11591847a" exitCode=0 Nov 25 09:43:45 crc kubenswrapper[4776]: I1125 09:43:45.588992 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" event={"ID":"adadb0a2-c265-4686-98e0-e3d8ff014b35","Type":"ContainerDied","Data":"b625f8f354e8416dd5bba6f485f43a6b0b7956f3d528292b946646a11591847a"} Nov 25 09:43:47 crc kubenswrapper[4776]: I1125 09:43:47.000188 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" podUID="adadb0a2-c265-4686-98e0-e3d8ff014b35" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.121:5353: connect: connection refused" Nov 25 09:43:49 crc kubenswrapper[4776]: I1125 09:43:49.385168 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dxg4f" Nov 25 09:43:49 crc kubenswrapper[4776]: I1125 09:43:49.575874 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-credential-keys\") pod \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\" (UID: \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\") " Nov 25 09:43:49 crc kubenswrapper[4776]: I1125 09:43:49.575923 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5wf5\" (UniqueName: \"kubernetes.io/projected/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-kube-api-access-q5wf5\") pod \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\" (UID: \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\") " Nov 25 09:43:49 crc kubenswrapper[4776]: I1125 09:43:49.575955 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-scripts\") pod \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\" (UID: \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\") " Nov 25 09:43:49 crc kubenswrapper[4776]: I1125 09:43:49.576022 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-config-data\") pod \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\" (UID: \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\") " Nov 25 09:43:49 crc kubenswrapper[4776]: I1125 09:43:49.576997 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-combined-ca-bundle\") pod \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\" (UID: \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\") " Nov 25 09:43:49 crc kubenswrapper[4776]: I1125 09:43:49.577048 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-fernet-keys\") pod \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\" (UID: \"58ed0f88-fcdb-491d-8eaa-5c81827ea82a\") " Nov 25 09:43:49 crc kubenswrapper[4776]: I1125 09:43:49.582733 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "58ed0f88-fcdb-491d-8eaa-5c81827ea82a" (UID: "58ed0f88-fcdb-491d-8eaa-5c81827ea82a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:43:49 crc kubenswrapper[4776]: I1125 09:43:49.582761 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-kube-api-access-q5wf5" (OuterVolumeSpecName: "kube-api-access-q5wf5") pod "58ed0f88-fcdb-491d-8eaa-5c81827ea82a" (UID: "58ed0f88-fcdb-491d-8eaa-5c81827ea82a"). InnerVolumeSpecName "kube-api-access-q5wf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:43:49 crc kubenswrapper[4776]: I1125 09:43:49.583343 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-scripts" (OuterVolumeSpecName: "scripts") pod "58ed0f88-fcdb-491d-8eaa-5c81827ea82a" (UID: "58ed0f88-fcdb-491d-8eaa-5c81827ea82a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:43:49 crc kubenswrapper[4776]: I1125 09:43:49.595028 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "58ed0f88-fcdb-491d-8eaa-5c81827ea82a" (UID: "58ed0f88-fcdb-491d-8eaa-5c81827ea82a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:43:49 crc kubenswrapper[4776]: I1125 09:43:49.604915 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-config-data" (OuterVolumeSpecName: "config-data") pod "58ed0f88-fcdb-491d-8eaa-5c81827ea82a" (UID: "58ed0f88-fcdb-491d-8eaa-5c81827ea82a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:43:49 crc kubenswrapper[4776]: I1125 09:43:49.606344 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58ed0f88-fcdb-491d-8eaa-5c81827ea82a" (UID: "58ed0f88-fcdb-491d-8eaa-5c81827ea82a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:43:49 crc kubenswrapper[4776]: I1125 09:43:49.624568 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dxg4f" event={"ID":"58ed0f88-fcdb-491d-8eaa-5c81827ea82a","Type":"ContainerDied","Data":"2e58b6a555e966ba5e956cf3709f8488b951a4de87d839026d06787ad5013431"} Nov 25 09:43:49 crc kubenswrapper[4776]: I1125 09:43:49.624609 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e58b6a555e966ba5e956cf3709f8488b951a4de87d839026d06787ad5013431" Nov 25 09:43:49 crc kubenswrapper[4776]: I1125 09:43:49.624658 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dxg4f" Nov 25 09:43:49 crc kubenswrapper[4776]: I1125 09:43:49.679691 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:49 crc kubenswrapper[4776]: I1125 09:43:49.679728 4776 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:49 crc kubenswrapper[4776]: I1125 09:43:49.679741 4776 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:49 crc kubenswrapper[4776]: I1125 09:43:49.679754 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5wf5\" (UniqueName: \"kubernetes.io/projected/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-kube-api-access-q5wf5\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:49 crc kubenswrapper[4776]: I1125 09:43:49.679769 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:49 crc kubenswrapper[4776]: I1125 09:43:49.679780 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58ed0f88-fcdb-491d-8eaa-5c81827ea82a-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.468430 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-dxg4f"] Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.478152 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-dxg4f"] Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.576965 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-jswxm"] Nov 25 09:43:50 crc kubenswrapper[4776]: E1125 09:43:50.577708 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58ed0f88-fcdb-491d-8eaa-5c81827ea82a" containerName="keystone-bootstrap" Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.577795 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="58ed0f88-fcdb-491d-8eaa-5c81827ea82a" containerName="keystone-bootstrap" Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.578117 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="58ed0f88-fcdb-491d-8eaa-5c81827ea82a" containerName="keystone-bootstrap" Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.578852 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jswxm" Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.585028 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.585477 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.585804 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.586472 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gtj9p" Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.590207 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.592362 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jswxm"] Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.697697 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-scripts\") pod \"keystone-bootstrap-jswxm\" (UID: \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\") " pod="openstack/keystone-bootstrap-jswxm" Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.698111 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-combined-ca-bundle\") pod \"keystone-bootstrap-jswxm\" (UID: \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\") " pod="openstack/keystone-bootstrap-jswxm" Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.698153 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-fernet-keys\") pod \"keystone-bootstrap-jswxm\" (UID: \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\") " pod="openstack/keystone-bootstrap-jswxm" Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.698176 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-config-data\") pod \"keystone-bootstrap-jswxm\" (UID: \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\") " pod="openstack/keystone-bootstrap-jswxm" Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.698277 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2ptb\" (UniqueName: \"kubernetes.io/projected/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-kube-api-access-t2ptb\") pod \"keystone-bootstrap-jswxm\" (UID: \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\") " pod="openstack/keystone-bootstrap-jswxm" Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.698301 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-credential-keys\") pod \"keystone-bootstrap-jswxm\" (UID: \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\") " pod="openstack/keystone-bootstrap-jswxm" Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.800189 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-combined-ca-bundle\") pod \"keystone-bootstrap-jswxm\" (UID: \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\") " pod="openstack/keystone-bootstrap-jswxm" Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.800248 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-fernet-keys\") pod \"keystone-bootstrap-jswxm\" (UID: \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\") " pod="openstack/keystone-bootstrap-jswxm" Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.800275 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-config-data\") pod \"keystone-bootstrap-jswxm\" (UID: \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\") " pod="openstack/keystone-bootstrap-jswxm" Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.800338 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2ptb\" (UniqueName: \"kubernetes.io/projected/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-kube-api-access-t2ptb\") pod \"keystone-bootstrap-jswxm\" (UID: \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\") " pod="openstack/keystone-bootstrap-jswxm" Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.800364 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-credential-keys\") pod \"keystone-bootstrap-jswxm\" (UID: \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\") " pod="openstack/keystone-bootstrap-jswxm" Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.800417 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-scripts\") pod \"keystone-bootstrap-jswxm\" (UID: \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\") " pod="openstack/keystone-bootstrap-jswxm" Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.806758 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-combined-ca-bundle\") pod \"keystone-bootstrap-jswxm\" (UID: \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\") " pod="openstack/keystone-bootstrap-jswxm" Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.806809 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-fernet-keys\") pod \"keystone-bootstrap-jswxm\" (UID: \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\") " pod="openstack/keystone-bootstrap-jswxm" Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.807476 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-scripts\") pod \"keystone-bootstrap-jswxm\" (UID: \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\") " pod="openstack/keystone-bootstrap-jswxm" Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.807868 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-config-data\") pod \"keystone-bootstrap-jswxm\" (UID: \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\") " pod="openstack/keystone-bootstrap-jswxm" Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.817331 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2ptb\" (UniqueName: \"kubernetes.io/projected/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-kube-api-access-t2ptb\") pod \"keystone-bootstrap-jswxm\" (UID: \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\") " pod="openstack/keystone-bootstrap-jswxm" Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.820574 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-credential-keys\") pod \"keystone-bootstrap-jswxm\" (UID: \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\") " pod="openstack/keystone-bootstrap-jswxm" Nov 25 09:43:50 crc kubenswrapper[4776]: I1125 09:43:50.912946 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jswxm" Nov 25 09:43:51 crc kubenswrapper[4776]: I1125 09:43:51.673458 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58ed0f88-fcdb-491d-8eaa-5c81827ea82a" path="/var/lib/kubelet/pods/58ed0f88-fcdb-491d-8eaa-5c81827ea82a/volumes" Nov 25 09:43:57 crc kubenswrapper[4776]: I1125 09:43:57.000094 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" podUID="adadb0a2-c265-4686-98e0-e3d8ff014b35" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.121:5353: i/o timeout" Nov 25 09:44:02 crc kubenswrapper[4776]: I1125 09:44:02.001749 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" podUID="adadb0a2-c265-4686-98e0-e3d8ff014b35" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.121:5353: i/o timeout" Nov 25 09:44:02 crc kubenswrapper[4776]: I1125 09:44:02.003555 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" Nov 25 09:44:04 crc kubenswrapper[4776]: E1125 09:44:04.964546 4776 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:37d64e0a00c54e71a4c1fcbbbf7e832f6886ffd03c9a02b6ee3ca48fabc30879" Nov 25 09:44:04 crc kubenswrapper[4776]: E1125 09:44:04.965001 4776 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:37d64e0a00c54e71a4c1fcbbbf7e832f6886ffd03c9a02b6ee3ca48fabc30879,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rntzg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-djwk2_openstack(e522431f-0956-49f7-9c46-b0202824bd56): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 09:44:04 crc kubenswrapper[4776]: E1125 09:44:04.966398 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-djwk2" podUID="e522431f-0956-49f7-9c46-b0202824bd56" Nov 25 09:44:05 crc kubenswrapper[4776]: E1125 09:44:05.399587 4776 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:4c93a5cccb9971e24f05daf93b3aa11ba71752bc3469a1a1a2c4906f92f69645" Nov 25 09:44:05 crc kubenswrapper[4776]: E1125 09:44:05.399768 4776 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:4c93a5cccb9971e24f05daf93b3aa11ba71752bc3469a1a1a2c4906f92f69645,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b5cf5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-bdhw6_openstack(3ff2a086-8bce-46e2-b294-3b3f31e39b4e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 09:44:05 crc kubenswrapper[4776]: E1125 09:44:05.401006 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-bdhw6" podUID="3ff2a086-8bce-46e2-b294-3b3f31e39b4e" Nov 25 09:44:05 crc kubenswrapper[4776]: I1125 09:44:05.758494 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" Nov 25 09:44:05 crc kubenswrapper[4776]: I1125 09:44:05.781386 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" Nov 25 09:44:05 crc kubenswrapper[4776]: I1125 09:44:05.781546 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" event={"ID":"adadb0a2-c265-4686-98e0-e3d8ff014b35","Type":"ContainerDied","Data":"03a07fbf5aa81bf4bc6158da4f2dea735f1f8af821c2ed4c661a4fadbe11d491"} Nov 25 09:44:05 crc kubenswrapper[4776]: I1125 09:44:05.781577 4776 scope.go:117] "RemoveContainer" containerID="b625f8f354e8416dd5bba6f485f43a6b0b7956f3d528292b946646a11591847a" Nov 25 09:44:05 crc kubenswrapper[4776]: E1125 09:44:05.790422 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:37d64e0a00c54e71a4c1fcbbbf7e832f6886ffd03c9a02b6ee3ca48fabc30879\\\"\"" pod="openstack/cinder-db-sync-djwk2" podUID="e522431f-0956-49f7-9c46-b0202824bd56" Nov 25 09:44:05 crc kubenswrapper[4776]: E1125 09:44:05.790458 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:4c93a5cccb9971e24f05daf93b3aa11ba71752bc3469a1a1a2c4906f92f69645\\\"\"" pod="openstack/barbican-db-sync-bdhw6" podUID="3ff2a086-8bce-46e2-b294-3b3f31e39b4e" Nov 25 09:44:05 crc kubenswrapper[4776]: I1125 09:44:05.827722 4776 scope.go:117] "RemoveContainer" containerID="5914b48ef0c402081ad751368900f9beae2f9a0429c64a51b0f62898b64cf053" Nov 25 09:44:05 crc kubenswrapper[4776]: I1125 09:44:05.876217 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/adadb0a2-c265-4686-98e0-e3d8ff014b35-ovsdbserver-sb\") pod \"adadb0a2-c265-4686-98e0-e3d8ff014b35\" (UID: \"adadb0a2-c265-4686-98e0-e3d8ff014b35\") " Nov 25 09:44:05 crc kubenswrapper[4776]: I1125 09:44:05.876314 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qb7hn\" (UniqueName: \"kubernetes.io/projected/adadb0a2-c265-4686-98e0-e3d8ff014b35-kube-api-access-qb7hn\") pod \"adadb0a2-c265-4686-98e0-e3d8ff014b35\" (UID: \"adadb0a2-c265-4686-98e0-e3d8ff014b35\") " Nov 25 09:44:05 crc kubenswrapper[4776]: I1125 09:44:05.876371 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/adadb0a2-c265-4686-98e0-e3d8ff014b35-ovsdbserver-nb\") pod \"adadb0a2-c265-4686-98e0-e3d8ff014b35\" (UID: \"adadb0a2-c265-4686-98e0-e3d8ff014b35\") " Nov 25 09:44:05 crc kubenswrapper[4776]: I1125 09:44:05.876401 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adadb0a2-c265-4686-98e0-e3d8ff014b35-config\") pod \"adadb0a2-c265-4686-98e0-e3d8ff014b35\" (UID: \"adadb0a2-c265-4686-98e0-e3d8ff014b35\") " Nov 25 09:44:05 crc kubenswrapper[4776]: I1125 09:44:05.876446 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/adadb0a2-c265-4686-98e0-e3d8ff014b35-dns-svc\") pod \"adadb0a2-c265-4686-98e0-e3d8ff014b35\" (UID: \"adadb0a2-c265-4686-98e0-e3d8ff014b35\") " Nov 25 09:44:05 crc kubenswrapper[4776]: I1125 09:44:05.890906 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adadb0a2-c265-4686-98e0-e3d8ff014b35-kube-api-access-qb7hn" (OuterVolumeSpecName: "kube-api-access-qb7hn") pod "adadb0a2-c265-4686-98e0-e3d8ff014b35" (UID: "adadb0a2-c265-4686-98e0-e3d8ff014b35"). InnerVolumeSpecName "kube-api-access-qb7hn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:44:05 crc kubenswrapper[4776]: I1125 09:44:05.923792 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adadb0a2-c265-4686-98e0-e3d8ff014b35-config" (OuterVolumeSpecName: "config") pod "adadb0a2-c265-4686-98e0-e3d8ff014b35" (UID: "adadb0a2-c265-4686-98e0-e3d8ff014b35"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:44:05 crc kubenswrapper[4776]: I1125 09:44:05.928750 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adadb0a2-c265-4686-98e0-e3d8ff014b35-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "adadb0a2-c265-4686-98e0-e3d8ff014b35" (UID: "adadb0a2-c265-4686-98e0-e3d8ff014b35"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:44:05 crc kubenswrapper[4776]: I1125 09:44:05.933807 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adadb0a2-c265-4686-98e0-e3d8ff014b35-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "adadb0a2-c265-4686-98e0-e3d8ff014b35" (UID: "adadb0a2-c265-4686-98e0-e3d8ff014b35"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:44:05 crc kubenswrapper[4776]: I1125 09:44:05.938470 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adadb0a2-c265-4686-98e0-e3d8ff014b35-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "adadb0a2-c265-4686-98e0-e3d8ff014b35" (UID: "adadb0a2-c265-4686-98e0-e3d8ff014b35"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:44:05 crc kubenswrapper[4776]: I1125 09:44:05.978913 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/adadb0a2-c265-4686-98e0-e3d8ff014b35-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:05 crc kubenswrapper[4776]: I1125 09:44:05.978967 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/adadb0a2-c265-4686-98e0-e3d8ff014b35-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:05 crc kubenswrapper[4776]: I1125 09:44:05.978979 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qb7hn\" (UniqueName: \"kubernetes.io/projected/adadb0a2-c265-4686-98e0-e3d8ff014b35-kube-api-access-qb7hn\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:05 crc kubenswrapper[4776]: I1125 09:44:05.978988 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/adadb0a2-c265-4686-98e0-e3d8ff014b35-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:05 crc kubenswrapper[4776]: I1125 09:44:05.978997 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/adadb0a2-c265-4686-98e0-e3d8ff014b35-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:05 crc kubenswrapper[4776]: I1125 09:44:05.979669 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 09:44:05 crc kubenswrapper[4776]: W1125 09:44:05.991351 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod923266d8_bfe1_4683_9806_0caa0719be85.slice/crio-7810c4916ed1e901aa529db67040fe73b678a0607a97deb164bf6d85721d35c8 WatchSource:0}: Error finding container 7810c4916ed1e901aa529db67040fe73b678a0607a97deb164bf6d85721d35c8: Status 404 returned error can't find the container with id 7810c4916ed1e901aa529db67040fe73b678a0607a97deb164bf6d85721d35c8 Nov 25 09:44:06 crc kubenswrapper[4776]: I1125 09:44:06.027493 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jswxm"] Nov 25 09:44:06 crc kubenswrapper[4776]: I1125 09:44:06.120905 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9fdb784c-sh56t"] Nov 25 09:44:06 crc kubenswrapper[4776]: I1125 09:44:06.128434 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9fdb784c-sh56t"] Nov 25 09:44:06 crc kubenswrapper[4776]: I1125 09:44:06.135965 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 09:44:06 crc kubenswrapper[4776]: W1125 09:44:06.137744 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3689daf9_7cc3_4752_ab65_84047549c8f7.slice/crio-1bb26b3ea109a5b435fbb683bd34cf7fd1982b89a8022abbbf49896d9fd7de45 WatchSource:0}: Error finding container 1bb26b3ea109a5b435fbb683bd34cf7fd1982b89a8022abbbf49896d9fd7de45: Status 404 returned error can't find the container with id 1bb26b3ea109a5b435fbb683bd34cf7fd1982b89a8022abbbf49896d9fd7de45 Nov 25 09:44:06 crc kubenswrapper[4776]: I1125 09:44:06.796325 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ttw4k" event={"ID":"7c18590c-cf10-4c05-8a97-b9f1d091868a","Type":"ContainerStarted","Data":"1edff0bfa155e1f89f986bf9450cc1b87e57f55c9434668ff95b152144e5b031"} Nov 25 09:44:06 crc kubenswrapper[4776]: I1125 09:44:06.798209 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"923266d8-bfe1-4683-9806-0caa0719be85","Type":"ContainerStarted","Data":"93d90c3b3ff5a236b5dfa082b764c6a433395bd4e2c79cddf052a171885b3e77"} Nov 25 09:44:06 crc kubenswrapper[4776]: I1125 09:44:06.798278 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"923266d8-bfe1-4683-9806-0caa0719be85","Type":"ContainerStarted","Data":"7810c4916ed1e901aa529db67040fe73b678a0607a97deb164bf6d85721d35c8"} Nov 25 09:44:06 crc kubenswrapper[4776]: I1125 09:44:06.800863 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3689daf9-7cc3-4752-ab65-84047549c8f7","Type":"ContainerStarted","Data":"a509db9807d8210abe868d28653b50b9648ff9d84c56f032bed7bd13cb5d2fbb"} Nov 25 09:44:06 crc kubenswrapper[4776]: I1125 09:44:06.800902 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3689daf9-7cc3-4752-ab65-84047549c8f7","Type":"ContainerStarted","Data":"1bb26b3ea109a5b435fbb683bd34cf7fd1982b89a8022abbbf49896d9fd7de45"} Nov 25 09:44:06 crc kubenswrapper[4776]: I1125 09:44:06.803053 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54","Type":"ContainerStarted","Data":"866bc21b77e90a1bb28e58f78fe8177df9aa57d681b670c32ee93364c6a6b878"} Nov 25 09:44:06 crc kubenswrapper[4776]: I1125 09:44:06.805448 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jswxm" event={"ID":"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be","Type":"ContainerStarted","Data":"a0cb3a2dd3f2cee8e943eca96c9c023eed24dd0a9a9aafe2008ae9f59ba5d6ee"} Nov 25 09:44:06 crc kubenswrapper[4776]: I1125 09:44:06.805504 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jswxm" event={"ID":"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be","Type":"ContainerStarted","Data":"f4de120f3e0a69ce464659c6cdbf097499ba32ada4927a0ab9522b9870542861"} Nov 25 09:44:06 crc kubenswrapper[4776]: I1125 09:44:06.810460 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-ttw4k" podStartSLOduration=2.746730237 podStartE2EDuration="32.810434361s" podCreationTimestamp="2025-11-25 09:43:34 +0000 UTC" firstStartedPulling="2025-11-25 09:43:35.385087097 +0000 UTC m=+1160.426146650" lastFinishedPulling="2025-11-25 09:44:05.448791221 +0000 UTC m=+1190.489850774" observedRunningTime="2025-11-25 09:44:06.809702782 +0000 UTC m=+1191.850762355" watchObservedRunningTime="2025-11-25 09:44:06.810434361 +0000 UTC m=+1191.851493914" Nov 25 09:44:06 crc kubenswrapper[4776]: I1125 09:44:06.837652 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-jswxm" podStartSLOduration=16.837631473 podStartE2EDuration="16.837631473s" podCreationTimestamp="2025-11-25 09:43:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:44:06.831351225 +0000 UTC m=+1191.872410778" watchObservedRunningTime="2025-11-25 09:44:06.837631473 +0000 UTC m=+1191.878691026" Nov 25 09:44:07 crc kubenswrapper[4776]: I1125 09:44:07.003755 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c9fdb784c-sh56t" podUID="adadb0a2-c265-4686-98e0-e3d8ff014b35" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.121:5353: i/o timeout" Nov 25 09:44:07 crc kubenswrapper[4776]: I1125 09:44:07.673043 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adadb0a2-c265-4686-98e0-e3d8ff014b35" path="/var/lib/kubelet/pods/adadb0a2-c265-4686-98e0-e3d8ff014b35/volumes" Nov 25 09:44:07 crc kubenswrapper[4776]: I1125 09:44:07.821316 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"923266d8-bfe1-4683-9806-0caa0719be85","Type":"ContainerStarted","Data":"f999d6dfa60b7c06dba41d3c50560601453b9759d9497396d516f05e9e4436f2"} Nov 25 09:44:07 crc kubenswrapper[4776]: I1125 09:44:07.821801 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="923266d8-bfe1-4683-9806-0caa0719be85" containerName="glance-log" containerID="cri-o://93d90c3b3ff5a236b5dfa082b764c6a433395bd4e2c79cddf052a171885b3e77" gracePeriod=30 Nov 25 09:44:07 crc kubenswrapper[4776]: I1125 09:44:07.821930 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="923266d8-bfe1-4683-9806-0caa0719be85" containerName="glance-httpd" containerID="cri-o://f999d6dfa60b7c06dba41d3c50560601453b9759d9497396d516f05e9e4436f2" gracePeriod=30 Nov 25 09:44:07 crc kubenswrapper[4776]: I1125 09:44:07.830341 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3689daf9-7cc3-4752-ab65-84047549c8f7","Type":"ContainerStarted","Data":"fcc8e94c319ca78ea7ca5bdee4ec2585e71ffc12fab7687bbc71c43e400691dc"} Nov 25 09:44:07 crc kubenswrapper[4776]: I1125 09:44:07.830524 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="3689daf9-7cc3-4752-ab65-84047549c8f7" containerName="glance-log" containerID="cri-o://a509db9807d8210abe868d28653b50b9648ff9d84c56f032bed7bd13cb5d2fbb" gracePeriod=30 Nov 25 09:44:07 crc kubenswrapper[4776]: I1125 09:44:07.830543 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="3689daf9-7cc3-4752-ab65-84047549c8f7" containerName="glance-httpd" containerID="cri-o://fcc8e94c319ca78ea7ca5bdee4ec2585e71ffc12fab7687bbc71c43e400691dc" gracePeriod=30 Nov 25 09:44:07 crc kubenswrapper[4776]: I1125 09:44:07.836852 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54","Type":"ContainerStarted","Data":"c5a7ff2ecc6cca04725dbb7788030a76d83ca550c67e52f0ca5ef9c4078a782d"} Nov 25 09:44:07 crc kubenswrapper[4776]: I1125 09:44:07.852823 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=27.852800043 podStartE2EDuration="27.852800043s" podCreationTimestamp="2025-11-25 09:43:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:44:07.844369092 +0000 UTC m=+1192.885428665" watchObservedRunningTime="2025-11-25 09:44:07.852800043 +0000 UTC m=+1192.893859586" Nov 25 09:44:07 crc kubenswrapper[4776]: I1125 09:44:07.870465 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=27.870442456 podStartE2EDuration="27.870442456s" podCreationTimestamp="2025-11-25 09:43:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:44:07.864853566 +0000 UTC m=+1192.905913129" watchObservedRunningTime="2025-11-25 09:44:07.870442456 +0000 UTC m=+1192.911502009" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.397407 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.520695 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/923266d8-bfe1-4683-9806-0caa0719be85-httpd-run\") pod \"923266d8-bfe1-4683-9806-0caa0719be85\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") " Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.520800 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d86lj\" (UniqueName: \"kubernetes.io/projected/923266d8-bfe1-4683-9806-0caa0719be85-kube-api-access-d86lj\") pod \"923266d8-bfe1-4683-9806-0caa0719be85\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") " Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.520907 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"923266d8-bfe1-4683-9806-0caa0719be85\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") " Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.520939 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/923266d8-bfe1-4683-9806-0caa0719be85-scripts\") pod \"923266d8-bfe1-4683-9806-0caa0719be85\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") " Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.521032 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/923266d8-bfe1-4683-9806-0caa0719be85-logs\") pod \"923266d8-bfe1-4683-9806-0caa0719be85\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") " Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.521090 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/923266d8-bfe1-4683-9806-0caa0719be85-combined-ca-bundle\") pod \"923266d8-bfe1-4683-9806-0caa0719be85\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") " Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.521118 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/923266d8-bfe1-4683-9806-0caa0719be85-config-data\") pod \"923266d8-bfe1-4683-9806-0caa0719be85\" (UID: \"923266d8-bfe1-4683-9806-0caa0719be85\") " Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.521686 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/923266d8-bfe1-4683-9806-0caa0719be85-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "923266d8-bfe1-4683-9806-0caa0719be85" (UID: "923266d8-bfe1-4683-9806-0caa0719be85"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.521717 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/923266d8-bfe1-4683-9806-0caa0719be85-logs" (OuterVolumeSpecName: "logs") pod "923266d8-bfe1-4683-9806-0caa0719be85" (UID: "923266d8-bfe1-4683-9806-0caa0719be85"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.526771 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/923266d8-bfe1-4683-9806-0caa0719be85-kube-api-access-d86lj" (OuterVolumeSpecName: "kube-api-access-d86lj") pod "923266d8-bfe1-4683-9806-0caa0719be85" (UID: "923266d8-bfe1-4683-9806-0caa0719be85"). InnerVolumeSpecName "kube-api-access-d86lj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.526849 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/923266d8-bfe1-4683-9806-0caa0719be85-scripts" (OuterVolumeSpecName: "scripts") pod "923266d8-bfe1-4683-9806-0caa0719be85" (UID: "923266d8-bfe1-4683-9806-0caa0719be85"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.544574 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "923266d8-bfe1-4683-9806-0caa0719be85" (UID: "923266d8-bfe1-4683-9806-0caa0719be85"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.555910 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/923266d8-bfe1-4683-9806-0caa0719be85-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "923266d8-bfe1-4683-9806-0caa0719be85" (UID: "923266d8-bfe1-4683-9806-0caa0719be85"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.583023 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/923266d8-bfe1-4683-9806-0caa0719be85-config-data" (OuterVolumeSpecName: "config-data") pod "923266d8-bfe1-4683-9806-0caa0719be85" (UID: "923266d8-bfe1-4683-9806-0caa0719be85"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.622839 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d86lj\" (UniqueName: \"kubernetes.io/projected/923266d8-bfe1-4683-9806-0caa0719be85-kube-api-access-d86lj\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.622926 4776 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.622944 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/923266d8-bfe1-4683-9806-0caa0719be85-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.622956 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/923266d8-bfe1-4683-9806-0caa0719be85-logs\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.622972 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/923266d8-bfe1-4683-9806-0caa0719be85-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.622985 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/923266d8-bfe1-4683-9806-0caa0719be85-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.622998 4776 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/923266d8-bfe1-4683-9806-0caa0719be85-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.646023 4776 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.727149 4776 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.852813 4776 generic.go:334] "Generic (PLEG): container finished" podID="923266d8-bfe1-4683-9806-0caa0719be85" containerID="f999d6dfa60b7c06dba41d3c50560601453b9759d9497396d516f05e9e4436f2" exitCode=0 Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.852848 4776 generic.go:334] "Generic (PLEG): container finished" podID="923266d8-bfe1-4683-9806-0caa0719be85" containerID="93d90c3b3ff5a236b5dfa082b764c6a433395bd4e2c79cddf052a171885b3e77" exitCode=143 Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.852939 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.852929 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"923266d8-bfe1-4683-9806-0caa0719be85","Type":"ContainerDied","Data":"f999d6dfa60b7c06dba41d3c50560601453b9759d9497396d516f05e9e4436f2"} Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.853113 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"923266d8-bfe1-4683-9806-0caa0719be85","Type":"ContainerDied","Data":"93d90c3b3ff5a236b5dfa082b764c6a433395bd4e2c79cddf052a171885b3e77"} Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.853151 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"923266d8-bfe1-4683-9806-0caa0719be85","Type":"ContainerDied","Data":"7810c4916ed1e901aa529db67040fe73b678a0607a97deb164bf6d85721d35c8"} Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.853190 4776 scope.go:117] "RemoveContainer" containerID="f999d6dfa60b7c06dba41d3c50560601453b9759d9497396d516f05e9e4436f2" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.855580 4776 generic.go:334] "Generic (PLEG): container finished" podID="3689daf9-7cc3-4752-ab65-84047549c8f7" containerID="fcc8e94c319ca78ea7ca5bdee4ec2585e71ffc12fab7687bbc71c43e400691dc" exitCode=0 Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.855602 4776 generic.go:334] "Generic (PLEG): container finished" podID="3689daf9-7cc3-4752-ab65-84047549c8f7" containerID="a509db9807d8210abe868d28653b50b9648ff9d84c56f032bed7bd13cb5d2fbb" exitCode=143 Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.855647 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3689daf9-7cc3-4752-ab65-84047549c8f7","Type":"ContainerDied","Data":"fcc8e94c319ca78ea7ca5bdee4ec2585e71ffc12fab7687bbc71c43e400691dc"} Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.855663 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3689daf9-7cc3-4752-ab65-84047549c8f7","Type":"ContainerDied","Data":"a509db9807d8210abe868d28653b50b9648ff9d84c56f032bed7bd13cb5d2fbb"} Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.858486 4776 generic.go:334] "Generic (PLEG): container finished" podID="7c18590c-cf10-4c05-8a97-b9f1d091868a" containerID="1edff0bfa155e1f89f986bf9450cc1b87e57f55c9434668ff95b152144e5b031" exitCode=0 Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.858535 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ttw4k" event={"ID":"7c18590c-cf10-4c05-8a97-b9f1d091868a","Type":"ContainerDied","Data":"1edff0bfa155e1f89f986bf9450cc1b87e57f55c9434668ff95b152144e5b031"} Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.888434 4776 scope.go:117] "RemoveContainer" containerID="93d90c3b3ff5a236b5dfa082b764c6a433395bd4e2c79cddf052a171885b3e77" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.901149 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.907220 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.926103 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 09:44:08 crc kubenswrapper[4776]: E1125 09:44:08.927533 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="923266d8-bfe1-4683-9806-0caa0719be85" containerName="glance-log" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.927565 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="923266d8-bfe1-4683-9806-0caa0719be85" containerName="glance-log" Nov 25 09:44:08 crc kubenswrapper[4776]: E1125 09:44:08.927585 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adadb0a2-c265-4686-98e0-e3d8ff014b35" containerName="dnsmasq-dns" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.927595 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="adadb0a2-c265-4686-98e0-e3d8ff014b35" containerName="dnsmasq-dns" Nov 25 09:44:08 crc kubenswrapper[4776]: E1125 09:44:08.927614 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adadb0a2-c265-4686-98e0-e3d8ff014b35" containerName="init" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.927624 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="adadb0a2-c265-4686-98e0-e3d8ff014b35" containerName="init" Nov 25 09:44:08 crc kubenswrapper[4776]: E1125 09:44:08.927650 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="923266d8-bfe1-4683-9806-0caa0719be85" containerName="glance-httpd" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.927658 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="923266d8-bfe1-4683-9806-0caa0719be85" containerName="glance-httpd" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.927962 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="adadb0a2-c265-4686-98e0-e3d8ff014b35" containerName="dnsmasq-dns" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.927977 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="923266d8-bfe1-4683-9806-0caa0719be85" containerName="glance-log" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.928002 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="923266d8-bfe1-4683-9806-0caa0719be85" containerName="glance-httpd" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.929977 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.932048 4776 scope.go:117] "RemoveContainer" containerID="f999d6dfa60b7c06dba41d3c50560601453b9759d9497396d516f05e9e4436f2" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.935479 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.935790 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 25 09:44:08 crc kubenswrapper[4776]: E1125 09:44:08.946260 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f999d6dfa60b7c06dba41d3c50560601453b9759d9497396d516f05e9e4436f2\": container with ID starting with f999d6dfa60b7c06dba41d3c50560601453b9759d9497396d516f05e9e4436f2 not found: ID does not exist" containerID="f999d6dfa60b7c06dba41d3c50560601453b9759d9497396d516f05e9e4436f2" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.946308 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f999d6dfa60b7c06dba41d3c50560601453b9759d9497396d516f05e9e4436f2"} err="failed to get container status \"f999d6dfa60b7c06dba41d3c50560601453b9759d9497396d516f05e9e4436f2\": rpc error: code = NotFound desc = could not find container \"f999d6dfa60b7c06dba41d3c50560601453b9759d9497396d516f05e9e4436f2\": container with ID starting with f999d6dfa60b7c06dba41d3c50560601453b9759d9497396d516f05e9e4436f2 not found: ID does not exist" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.946343 4776 scope.go:117] "RemoveContainer" containerID="93d90c3b3ff5a236b5dfa082b764c6a433395bd4e2c79cddf052a171885b3e77" Nov 25 09:44:08 crc kubenswrapper[4776]: E1125 09:44:08.950772 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93d90c3b3ff5a236b5dfa082b764c6a433395bd4e2c79cddf052a171885b3e77\": container with ID starting with 93d90c3b3ff5a236b5dfa082b764c6a433395bd4e2c79cddf052a171885b3e77 not found: ID does not exist" containerID="93d90c3b3ff5a236b5dfa082b764c6a433395bd4e2c79cddf052a171885b3e77" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.950806 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93d90c3b3ff5a236b5dfa082b764c6a433395bd4e2c79cddf052a171885b3e77"} err="failed to get container status \"93d90c3b3ff5a236b5dfa082b764c6a433395bd4e2c79cddf052a171885b3e77\": rpc error: code = NotFound desc = could not find container \"93d90c3b3ff5a236b5dfa082b764c6a433395bd4e2c79cddf052a171885b3e77\": container with ID starting with 93d90c3b3ff5a236b5dfa082b764c6a433395bd4e2c79cddf052a171885b3e77 not found: ID does not exist" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.950823 4776 scope.go:117] "RemoveContainer" containerID="f999d6dfa60b7c06dba41d3c50560601453b9759d9497396d516f05e9e4436f2" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.951358 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f999d6dfa60b7c06dba41d3c50560601453b9759d9497396d516f05e9e4436f2"} err="failed to get container status \"f999d6dfa60b7c06dba41d3c50560601453b9759d9497396d516f05e9e4436f2\": rpc error: code = NotFound desc = could not find container \"f999d6dfa60b7c06dba41d3c50560601453b9759d9497396d516f05e9e4436f2\": container with ID starting with f999d6dfa60b7c06dba41d3c50560601453b9759d9497396d516f05e9e4436f2 not found: ID does not exist" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.951384 4776 scope.go:117] "RemoveContainer" containerID="93d90c3b3ff5a236b5dfa082b764c6a433395bd4e2c79cddf052a171885b3e77" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.952249 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93d90c3b3ff5a236b5dfa082b764c6a433395bd4e2c79cddf052a171885b3e77"} err="failed to get container status \"93d90c3b3ff5a236b5dfa082b764c6a433395bd4e2c79cddf052a171885b3e77\": rpc error: code = NotFound desc = could not find container \"93d90c3b3ff5a236b5dfa082b764c6a433395bd4e2c79cddf052a171885b3e77\": container with ID starting with 93d90c3b3ff5a236b5dfa082b764c6a433395bd4e2c79cddf052a171885b3e77 not found: ID does not exist" Nov 25 09:44:08 crc kubenswrapper[4776]: I1125 09:44:08.967412 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.033474 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85e1e1f5-a606-427e-a997-55dfad758e2e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.033545 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85e1e1f5-a606-427e-a997-55dfad758e2e-logs\") pod \"glance-default-external-api-0\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.033596 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85e1e1f5-a606-427e-a997-55dfad758e2e-config-data\") pod \"glance-default-external-api-0\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.033616 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85e1e1f5-a606-427e-a997-55dfad758e2e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.033663 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.033703 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8q55\" (UniqueName: \"kubernetes.io/projected/85e1e1f5-a606-427e-a997-55dfad758e2e-kube-api-access-k8q55\") pod \"glance-default-external-api-0\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.033762 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/85e1e1f5-a606-427e-a997-55dfad758e2e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.033836 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85e1e1f5-a606-427e-a997-55dfad758e2e-scripts\") pod \"glance-default-external-api-0\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.135310 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85e1e1f5-a606-427e-a997-55dfad758e2e-scripts\") pod \"glance-default-external-api-0\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.135378 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85e1e1f5-a606-427e-a997-55dfad758e2e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.135411 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85e1e1f5-a606-427e-a997-55dfad758e2e-logs\") pod \"glance-default-external-api-0\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.135452 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85e1e1f5-a606-427e-a997-55dfad758e2e-config-data\") pod \"glance-default-external-api-0\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.135473 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85e1e1f5-a606-427e-a997-55dfad758e2e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.135521 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.135562 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8q55\" (UniqueName: \"kubernetes.io/projected/85e1e1f5-a606-427e-a997-55dfad758e2e-kube-api-access-k8q55\") pod \"glance-default-external-api-0\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.135618 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/85e1e1f5-a606-427e-a997-55dfad758e2e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.136046 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/85e1e1f5-a606-427e-a997-55dfad758e2e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.138587 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.138765 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85e1e1f5-a606-427e-a997-55dfad758e2e-logs\") pod \"glance-default-external-api-0\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.140144 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85e1e1f5-a606-427e-a997-55dfad758e2e-scripts\") pod \"glance-default-external-api-0\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.142587 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85e1e1f5-a606-427e-a997-55dfad758e2e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.150721 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85e1e1f5-a606-427e-a997-55dfad758e2e-config-data\") pod \"glance-default-external-api-0\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.156196 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85e1e1f5-a606-427e-a997-55dfad758e2e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.156225 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8q55\" (UniqueName: \"kubernetes.io/projected/85e1e1f5-a606-427e-a997-55dfad758e2e-kube-api-access-k8q55\") pod \"glance-default-external-api-0\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.167005 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " pod="openstack/glance-default-external-api-0" Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.261623 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.676387 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="923266d8-bfe1-4683-9806-0caa0719be85" path="/var/lib/kubelet/pods/923266d8-bfe1-4683-9806-0caa0719be85/volumes" Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.875083 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.881926 4776 generic.go:334] "Generic (PLEG): container finished" podID="41e5f9c4-a3fc-4d92-bfee-09ee2fade2be" containerID="a0cb3a2dd3f2cee8e943eca96c9c023eed24dd0a9a9aafe2008ae9f59ba5d6ee" exitCode=0 Nov 25 09:44:09 crc kubenswrapper[4776]: I1125 09:44:09.882052 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jswxm" event={"ID":"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be","Type":"ContainerDied","Data":"a0cb3a2dd3f2cee8e943eca96c9c023eed24dd0a9a9aafe2008ae9f59ba5d6ee"} Nov 25 09:44:11 crc kubenswrapper[4776]: I1125 09:44:11.211676 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 25 09:44:11 crc kubenswrapper[4776]: I1125 09:44:11.212133 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 25 09:44:12 crc kubenswrapper[4776]: W1125 09:44:12.223334 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85e1e1f5_a606_427e_a997_55dfad758e2e.slice/crio-4b6a02d7412a1011e58348b8a6e71bada513cbdf4f20fcaa4f6f8b852e55909a WatchSource:0}: Error finding container 4b6a02d7412a1011e58348b8a6e71bada513cbdf4f20fcaa4f6f8b852e55909a: Status 404 returned error can't find the container with id 4b6a02d7412a1011e58348b8a6e71bada513cbdf4f20fcaa4f6f8b852e55909a Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.316357 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.333983 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ttw4k" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.354726 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jswxm" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.398449 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c18590c-cf10-4c05-8a97-b9f1d091868a-logs\") pod \"7c18590c-cf10-4c05-8a97-b9f1d091868a\" (UID: \"7c18590c-cf10-4c05-8a97-b9f1d091868a\") " Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.398512 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c18590c-cf10-4c05-8a97-b9f1d091868a-scripts\") pod \"7c18590c-cf10-4c05-8a97-b9f1d091868a\" (UID: \"7c18590c-cf10-4c05-8a97-b9f1d091868a\") " Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.398555 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-scripts\") pod \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\" (UID: \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\") " Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.398585 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gh6b5\" (UniqueName: \"kubernetes.io/projected/7c18590c-cf10-4c05-8a97-b9f1d091868a-kube-api-access-gh6b5\") pod \"7c18590c-cf10-4c05-8a97-b9f1d091868a\" (UID: \"7c18590c-cf10-4c05-8a97-b9f1d091868a\") " Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.398609 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3689daf9-7cc3-4752-ab65-84047549c8f7-config-data\") pod \"3689daf9-7cc3-4752-ab65-84047549c8f7\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") " Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.398656 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"3689daf9-7cc3-4752-ab65-84047549c8f7\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") " Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.398680 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3689daf9-7cc3-4752-ab65-84047549c8f7-scripts\") pod \"3689daf9-7cc3-4752-ab65-84047549c8f7\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") " Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.398703 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3689daf9-7cc3-4752-ab65-84047549c8f7-combined-ca-bundle\") pod \"3689daf9-7cc3-4752-ab65-84047549c8f7\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") " Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.398732 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3689daf9-7cc3-4752-ab65-84047549c8f7-logs\") pod \"3689daf9-7cc3-4752-ab65-84047549c8f7\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") " Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.398761 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-config-data\") pod \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\" (UID: \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\") " Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.398777 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2ptb\" (UniqueName: \"kubernetes.io/projected/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-kube-api-access-t2ptb\") pod \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\" (UID: \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\") " Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.398793 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c18590c-cf10-4c05-8a97-b9f1d091868a-config-data\") pod \"7c18590c-cf10-4c05-8a97-b9f1d091868a\" (UID: \"7c18590c-cf10-4c05-8a97-b9f1d091868a\") " Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.398850 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmfzt\" (UniqueName: \"kubernetes.io/projected/3689daf9-7cc3-4752-ab65-84047549c8f7-kube-api-access-hmfzt\") pod \"3689daf9-7cc3-4752-ab65-84047549c8f7\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") " Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.398894 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3689daf9-7cc3-4752-ab65-84047549c8f7-httpd-run\") pod \"3689daf9-7cc3-4752-ab65-84047549c8f7\" (UID: \"3689daf9-7cc3-4752-ab65-84047549c8f7\") " Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.398909 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c18590c-cf10-4c05-8a97-b9f1d091868a-combined-ca-bundle\") pod \"7c18590c-cf10-4c05-8a97-b9f1d091868a\" (UID: \"7c18590c-cf10-4c05-8a97-b9f1d091868a\") " Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.398930 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-credential-keys\") pod \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\" (UID: \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\") " Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.398965 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-combined-ca-bundle\") pod \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\" (UID: \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\") " Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.399022 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-fernet-keys\") pod \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\" (UID: \"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be\") " Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.399873 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3689daf9-7cc3-4752-ab65-84047549c8f7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3689daf9-7cc3-4752-ab65-84047549c8f7" (UID: "3689daf9-7cc3-4752-ab65-84047549c8f7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.402687 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3689daf9-7cc3-4752-ab65-84047549c8f7-logs" (OuterVolumeSpecName: "logs") pod "3689daf9-7cc3-4752-ab65-84047549c8f7" (UID: "3689daf9-7cc3-4752-ab65-84047549c8f7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.403337 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c18590c-cf10-4c05-8a97-b9f1d091868a-logs" (OuterVolumeSpecName: "logs") pod "7c18590c-cf10-4c05-8a97-b9f1d091868a" (UID: "7c18590c-cf10-4c05-8a97-b9f1d091868a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.406366 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-kube-api-access-t2ptb" (OuterVolumeSpecName: "kube-api-access-t2ptb") pod "41e5f9c4-a3fc-4d92-bfee-09ee2fade2be" (UID: "41e5f9c4-a3fc-4d92-bfee-09ee2fade2be"). InnerVolumeSpecName "kube-api-access-t2ptb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.406704 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c18590c-cf10-4c05-8a97-b9f1d091868a-scripts" (OuterVolumeSpecName: "scripts") pod "7c18590c-cf10-4c05-8a97-b9f1d091868a" (UID: "7c18590c-cf10-4c05-8a97-b9f1d091868a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.406961 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "41e5f9c4-a3fc-4d92-bfee-09ee2fade2be" (UID: "41e5f9c4-a3fc-4d92-bfee-09ee2fade2be"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.407251 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "3689daf9-7cc3-4752-ab65-84047549c8f7" (UID: "3689daf9-7cc3-4752-ab65-84047549c8f7"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.407376 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "41e5f9c4-a3fc-4d92-bfee-09ee2fade2be" (UID: "41e5f9c4-a3fc-4d92-bfee-09ee2fade2be"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.417656 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3689daf9-7cc3-4752-ab65-84047549c8f7-kube-api-access-hmfzt" (OuterVolumeSpecName: "kube-api-access-hmfzt") pod "3689daf9-7cc3-4752-ab65-84047549c8f7" (UID: "3689daf9-7cc3-4752-ab65-84047549c8f7"). InnerVolumeSpecName "kube-api-access-hmfzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.417713 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c18590c-cf10-4c05-8a97-b9f1d091868a-kube-api-access-gh6b5" (OuterVolumeSpecName: "kube-api-access-gh6b5") pod "7c18590c-cf10-4c05-8a97-b9f1d091868a" (UID: "7c18590c-cf10-4c05-8a97-b9f1d091868a"). InnerVolumeSpecName "kube-api-access-gh6b5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.423301 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-scripts" (OuterVolumeSpecName: "scripts") pod "41e5f9c4-a3fc-4d92-bfee-09ee2fade2be" (UID: "41e5f9c4-a3fc-4d92-bfee-09ee2fade2be"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.425119 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3689daf9-7cc3-4752-ab65-84047549c8f7-scripts" (OuterVolumeSpecName: "scripts") pod "3689daf9-7cc3-4752-ab65-84047549c8f7" (UID: "3689daf9-7cc3-4752-ab65-84047549c8f7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.429847 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3689daf9-7cc3-4752-ab65-84047549c8f7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3689daf9-7cc3-4752-ab65-84047549c8f7" (UID: "3689daf9-7cc3-4752-ab65-84047549c8f7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.433389 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c18590c-cf10-4c05-8a97-b9f1d091868a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7c18590c-cf10-4c05-8a97-b9f1d091868a" (UID: "7c18590c-cf10-4c05-8a97-b9f1d091868a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.438387 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c18590c-cf10-4c05-8a97-b9f1d091868a-config-data" (OuterVolumeSpecName: "config-data") pod "7c18590c-cf10-4c05-8a97-b9f1d091868a" (UID: "7c18590c-cf10-4c05-8a97-b9f1d091868a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.443834 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "41e5f9c4-a3fc-4d92-bfee-09ee2fade2be" (UID: "41e5f9c4-a3fc-4d92-bfee-09ee2fade2be"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.450866 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-config-data" (OuterVolumeSpecName: "config-data") pod "41e5f9c4-a3fc-4d92-bfee-09ee2fade2be" (UID: "41e5f9c4-a3fc-4d92-bfee-09ee2fade2be"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.469342 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3689daf9-7cc3-4752-ab65-84047549c8f7-config-data" (OuterVolumeSpecName: "config-data") pod "3689daf9-7cc3-4752-ab65-84047549c8f7" (UID: "3689daf9-7cc3-4752-ab65-84047549c8f7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.501056 4776 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.501111 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.501122 4776 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.501131 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7c18590c-cf10-4c05-8a97-b9f1d091868a-logs\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.501139 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7c18590c-cf10-4c05-8a97-b9f1d091868a-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.501165 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.501174 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gh6b5\" (UniqueName: \"kubernetes.io/projected/7c18590c-cf10-4c05-8a97-b9f1d091868a-kube-api-access-gh6b5\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.501184 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3689daf9-7cc3-4752-ab65-84047549c8f7-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.501211 4776 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.501218 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3689daf9-7cc3-4752-ab65-84047549c8f7-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.501265 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3689daf9-7cc3-4752-ab65-84047549c8f7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.501274 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3689daf9-7cc3-4752-ab65-84047549c8f7-logs\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.501282 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.501292 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2ptb\" (UniqueName: \"kubernetes.io/projected/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be-kube-api-access-t2ptb\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.501301 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c18590c-cf10-4c05-8a97-b9f1d091868a-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.505487 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmfzt\" (UniqueName: \"kubernetes.io/projected/3689daf9-7cc3-4752-ab65-84047549c8f7-kube-api-access-hmfzt\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.505507 4776 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3689daf9-7cc3-4752-ab65-84047549c8f7-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.505516 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c18590c-cf10-4c05-8a97-b9f1d091868a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.519759 4776 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.606768 4776 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.916231 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3689daf9-7cc3-4752-ab65-84047549c8f7","Type":"ContainerDied","Data":"1bb26b3ea109a5b435fbb683bd34cf7fd1982b89a8022abbbf49896d9fd7de45"} Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.916293 4776 scope.go:117] "RemoveContainer" containerID="fcc8e94c319ca78ea7ca5bdee4ec2585e71ffc12fab7687bbc71c43e400691dc" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.916429 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.926528 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jswxm" event={"ID":"41e5f9c4-a3fc-4d92-bfee-09ee2fade2be","Type":"ContainerDied","Data":"f4de120f3e0a69ce464659c6cdbf097499ba32ada4927a0ab9522b9870542861"} Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.926565 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4de120f3e0a69ce464659c6cdbf097499ba32ada4927a0ab9522b9870542861" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.926616 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jswxm" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.933531 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ttw4k" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.933559 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ttw4k" event={"ID":"7c18590c-cf10-4c05-8a97-b9f1d091868a","Type":"ContainerDied","Data":"f83314b7c339776a7e5269a458852df9d91d292b0aaeb191eaedaee0240c541a"} Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.933734 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f83314b7c339776a7e5269a458852df9d91d292b0aaeb191eaedaee0240c541a" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.951211 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"85e1e1f5-a606-427e-a997-55dfad758e2e","Type":"ContainerStarted","Data":"4b6a02d7412a1011e58348b8a6e71bada513cbdf4f20fcaa4f6f8b852e55909a"} Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.957710 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.969264 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.982596 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 09:44:12 crc kubenswrapper[4776]: E1125 09:44:12.984644 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c18590c-cf10-4c05-8a97-b9f1d091868a" containerName="placement-db-sync" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.984696 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c18590c-cf10-4c05-8a97-b9f1d091868a" containerName="placement-db-sync" Nov 25 09:44:12 crc kubenswrapper[4776]: E1125 09:44:12.984722 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3689daf9-7cc3-4752-ab65-84047549c8f7" containerName="glance-httpd" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.984731 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="3689daf9-7cc3-4752-ab65-84047549c8f7" containerName="glance-httpd" Nov 25 09:44:12 crc kubenswrapper[4776]: E1125 09:44:12.984783 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3689daf9-7cc3-4752-ab65-84047549c8f7" containerName="glance-log" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.984793 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="3689daf9-7cc3-4752-ab65-84047549c8f7" containerName="glance-log" Nov 25 09:44:12 crc kubenswrapper[4776]: E1125 09:44:12.984844 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41e5f9c4-a3fc-4d92-bfee-09ee2fade2be" containerName="keystone-bootstrap" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.984856 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="41e5f9c4-a3fc-4d92-bfee-09ee2fade2be" containerName="keystone-bootstrap" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.985156 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c18590c-cf10-4c05-8a97-b9f1d091868a" containerName="placement-db-sync" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.985180 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="3689daf9-7cc3-4752-ab65-84047549c8f7" containerName="glance-log" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.985195 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="41e5f9c4-a3fc-4d92-bfee-09ee2fade2be" containerName="keystone-bootstrap" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.985207 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="3689daf9-7cc3-4752-ab65-84047549c8f7" containerName="glance-httpd" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.986427 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.989472 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.989754 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 25 09:44:12 crc kubenswrapper[4776]: I1125 09:44:12.992266 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.115739 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c595957f-32bd-455a-b212-8f2536102d7f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.115775 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c595957f-32bd-455a-b212-8f2536102d7f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.115819 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c595957f-32bd-455a-b212-8f2536102d7f-logs\") pod \"glance-default-internal-api-0\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.115835 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c595957f-32bd-455a-b212-8f2536102d7f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.115877 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.115897 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbx2d\" (UniqueName: \"kubernetes.io/projected/c595957f-32bd-455a-b212-8f2536102d7f-kube-api-access-sbx2d\") pod \"glance-default-internal-api-0\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.115950 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c595957f-32bd-455a-b212-8f2536102d7f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.115976 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c595957f-32bd-455a-b212-8f2536102d7f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.217123 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c595957f-32bd-455a-b212-8f2536102d7f-logs\") pod \"glance-default-internal-api-0\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.217175 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c595957f-32bd-455a-b212-8f2536102d7f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.217244 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.217272 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbx2d\" (UniqueName: \"kubernetes.io/projected/c595957f-32bd-455a-b212-8f2536102d7f-kube-api-access-sbx2d\") pod \"glance-default-internal-api-0\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.217343 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c595957f-32bd-455a-b212-8f2536102d7f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.217378 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c595957f-32bd-455a-b212-8f2536102d7f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.217413 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c595957f-32bd-455a-b212-8f2536102d7f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.217437 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c595957f-32bd-455a-b212-8f2536102d7f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.218008 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.218212 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c595957f-32bd-455a-b212-8f2536102d7f-logs\") pod \"glance-default-internal-api-0\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.218491 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c595957f-32bd-455a-b212-8f2536102d7f-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.221974 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c595957f-32bd-455a-b212-8f2536102d7f-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.222416 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c595957f-32bd-455a-b212-8f2536102d7f-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.223173 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c595957f-32bd-455a-b212-8f2536102d7f-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.240852 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbx2d\" (UniqueName: \"kubernetes.io/projected/c595957f-32bd-455a-b212-8f2536102d7f-kube-api-access-sbx2d\") pod \"glance-default-internal-api-0\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.245534 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c595957f-32bd-455a-b212-8f2536102d7f-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.245784 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.316715 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.454842 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-667769d556-lqtlp"] Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.457566 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-667769d556-lqtlp" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.461143 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.461343 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.461472 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.461623 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.461755 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-pf9p5" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.467754 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-667769d556-lqtlp"] Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.521587 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-public-tls-certs\") pod \"placement-667769d556-lqtlp\" (UID: \"a0b97845-4578-4801-896a-281fdafdb351\") " pod="openstack/placement-667769d556-lqtlp" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.521641 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0b97845-4578-4801-896a-281fdafdb351-logs\") pod \"placement-667769d556-lqtlp\" (UID: \"a0b97845-4578-4801-896a-281fdafdb351\") " pod="openstack/placement-667769d556-lqtlp" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.521896 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-combined-ca-bundle\") pod \"placement-667769d556-lqtlp\" (UID: \"a0b97845-4578-4801-896a-281fdafdb351\") " pod="openstack/placement-667769d556-lqtlp" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.521984 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-config-data\") pod \"placement-667769d556-lqtlp\" (UID: \"a0b97845-4578-4801-896a-281fdafdb351\") " pod="openstack/placement-667769d556-lqtlp" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.522014 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-scripts\") pod \"placement-667769d556-lqtlp\" (UID: \"a0b97845-4578-4801-896a-281fdafdb351\") " pod="openstack/placement-667769d556-lqtlp" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.522133 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-internal-tls-certs\") pod \"placement-667769d556-lqtlp\" (UID: \"a0b97845-4578-4801-896a-281fdafdb351\") " pod="openstack/placement-667769d556-lqtlp" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.522173 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-264xx\" (UniqueName: \"kubernetes.io/projected/a0b97845-4578-4801-896a-281fdafdb351-kube-api-access-264xx\") pod \"placement-667769d556-lqtlp\" (UID: \"a0b97845-4578-4801-896a-281fdafdb351\") " pod="openstack/placement-667769d556-lqtlp" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.544250 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6b75d6ddb7-rmtms"] Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.545685 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.550583 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.550842 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.551005 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.555433 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-gtj9p" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.555562 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.555601 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.561735 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6b75d6ddb7-rmtms"] Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.623515 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-264xx\" (UniqueName: \"kubernetes.io/projected/a0b97845-4578-4801-896a-281fdafdb351-kube-api-access-264xx\") pod \"placement-667769d556-lqtlp\" (UID: \"a0b97845-4578-4801-896a-281fdafdb351\") " pod="openstack/placement-667769d556-lqtlp" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.623559 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-public-tls-certs\") pod \"keystone-6b75d6ddb7-rmtms\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.623583 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-combined-ca-bundle\") pod \"keystone-6b75d6ddb7-rmtms\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.623621 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-internal-tls-certs\") pod \"keystone-6b75d6ddb7-rmtms\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.623660 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-config-data\") pod \"keystone-6b75d6ddb7-rmtms\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.623685 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-public-tls-certs\") pod \"placement-667769d556-lqtlp\" (UID: \"a0b97845-4578-4801-896a-281fdafdb351\") " pod="openstack/placement-667769d556-lqtlp" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.623711 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-scripts\") pod \"keystone-6b75d6ddb7-rmtms\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.623734 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0b97845-4578-4801-896a-281fdafdb351-logs\") pod \"placement-667769d556-lqtlp\" (UID: \"a0b97845-4578-4801-896a-281fdafdb351\") " pod="openstack/placement-667769d556-lqtlp" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.623787 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-combined-ca-bundle\") pod \"placement-667769d556-lqtlp\" (UID: \"a0b97845-4578-4801-896a-281fdafdb351\") " pod="openstack/placement-667769d556-lqtlp" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.623815 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-config-data\") pod \"placement-667769d556-lqtlp\" (UID: \"a0b97845-4578-4801-896a-281fdafdb351\") " pod="openstack/placement-667769d556-lqtlp" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.623831 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-scripts\") pod \"placement-667769d556-lqtlp\" (UID: \"a0b97845-4578-4801-896a-281fdafdb351\") " pod="openstack/placement-667769d556-lqtlp" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.623861 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-credential-keys\") pod \"keystone-6b75d6ddb7-rmtms\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.623889 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4466v\" (UniqueName: \"kubernetes.io/projected/83c080b8-d976-4e81-b103-2442bb2eafe1-kube-api-access-4466v\") pod \"keystone-6b75d6ddb7-rmtms\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.623912 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-internal-tls-certs\") pod \"placement-667769d556-lqtlp\" (UID: \"a0b97845-4578-4801-896a-281fdafdb351\") " pod="openstack/placement-667769d556-lqtlp" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.623932 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-fernet-keys\") pod \"keystone-6b75d6ddb7-rmtms\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.625448 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0b97845-4578-4801-896a-281fdafdb351-logs\") pod \"placement-667769d556-lqtlp\" (UID: \"a0b97845-4578-4801-896a-281fdafdb351\") " pod="openstack/placement-667769d556-lqtlp" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.628097 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-internal-tls-certs\") pod \"placement-667769d556-lqtlp\" (UID: \"a0b97845-4578-4801-896a-281fdafdb351\") " pod="openstack/placement-667769d556-lqtlp" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.628139 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-public-tls-certs\") pod \"placement-667769d556-lqtlp\" (UID: \"a0b97845-4578-4801-896a-281fdafdb351\") " pod="openstack/placement-667769d556-lqtlp" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.628369 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-scripts\") pod \"placement-667769d556-lqtlp\" (UID: \"a0b97845-4578-4801-896a-281fdafdb351\") " pod="openstack/placement-667769d556-lqtlp" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.629934 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-combined-ca-bundle\") pod \"placement-667769d556-lqtlp\" (UID: \"a0b97845-4578-4801-896a-281fdafdb351\") " pod="openstack/placement-667769d556-lqtlp" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.631102 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-config-data\") pod \"placement-667769d556-lqtlp\" (UID: \"a0b97845-4578-4801-896a-281fdafdb351\") " pod="openstack/placement-667769d556-lqtlp" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.643549 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-264xx\" (UniqueName: \"kubernetes.io/projected/a0b97845-4578-4801-896a-281fdafdb351-kube-api-access-264xx\") pod \"placement-667769d556-lqtlp\" (UID: \"a0b97845-4578-4801-896a-281fdafdb351\") " pod="openstack/placement-667769d556-lqtlp" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.683489 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3689daf9-7cc3-4752-ab65-84047549c8f7" path="/var/lib/kubelet/pods/3689daf9-7cc3-4752-ab65-84047549c8f7/volumes" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.726911 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-credential-keys\") pod \"keystone-6b75d6ddb7-rmtms\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.726972 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4466v\" (UniqueName: \"kubernetes.io/projected/83c080b8-d976-4e81-b103-2442bb2eafe1-kube-api-access-4466v\") pod \"keystone-6b75d6ddb7-rmtms\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.726997 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-fernet-keys\") pod \"keystone-6b75d6ddb7-rmtms\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.727016 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-public-tls-certs\") pod \"keystone-6b75d6ddb7-rmtms\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.727037 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-combined-ca-bundle\") pod \"keystone-6b75d6ddb7-rmtms\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.727095 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-internal-tls-certs\") pod \"keystone-6b75d6ddb7-rmtms\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.727141 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-config-data\") pod \"keystone-6b75d6ddb7-rmtms\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.727177 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-scripts\") pod \"keystone-6b75d6ddb7-rmtms\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.731503 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-credential-keys\") pod \"keystone-6b75d6ddb7-rmtms\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.732700 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-config-data\") pod \"keystone-6b75d6ddb7-rmtms\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.733120 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-scripts\") pod \"keystone-6b75d6ddb7-rmtms\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.733901 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-fernet-keys\") pod \"keystone-6b75d6ddb7-rmtms\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.736961 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-internal-tls-certs\") pod \"keystone-6b75d6ddb7-rmtms\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.743862 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-public-tls-certs\") pod \"keystone-6b75d6ddb7-rmtms\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.744722 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-combined-ca-bundle\") pod \"keystone-6b75d6ddb7-rmtms\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.745154 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4466v\" (UniqueName: \"kubernetes.io/projected/83c080b8-d976-4e81-b103-2442bb2eafe1-kube-api-access-4466v\") pod \"keystone-6b75d6ddb7-rmtms\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.780934 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-667769d556-lqtlp" Nov 25 09:44:13 crc kubenswrapper[4776]: I1125 09:44:13.873650 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:44:16 crc kubenswrapper[4776]: I1125 09:44:16.343686 4776 scope.go:117] "RemoveContainer" containerID="a509db9807d8210abe868d28653b50b9648ff9d84c56f032bed7bd13cb5d2fbb" Nov 25 09:44:16 crc kubenswrapper[4776]: W1125 09:44:16.898190 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0b97845_4578_4801_896a_281fdafdb351.slice/crio-422d5dbcd2c0da803fdd9482a7272e0f11941f78aaba66e0811b7ff8c628f5f6 WatchSource:0}: Error finding container 422d5dbcd2c0da803fdd9482a7272e0f11941f78aaba66e0811b7ff8c628f5f6: Status 404 returned error can't find the container with id 422d5dbcd2c0da803fdd9482a7272e0f11941f78aaba66e0811b7ff8c628f5f6 Nov 25 09:44:16 crc kubenswrapper[4776]: I1125 09:44:16.898862 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-667769d556-lqtlp"] Nov 25 09:44:17 crc kubenswrapper[4776]: I1125 09:44:17.008661 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54","Type":"ContainerStarted","Data":"4ea77e92eb9f4a8c4a0aea41ded4fa353ada185d27f639d8c33be0c971e1cbf5"} Nov 25 09:44:17 crc kubenswrapper[4776]: I1125 09:44:17.010906 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-667769d556-lqtlp" event={"ID":"a0b97845-4578-4801-896a-281fdafdb351","Type":"ContainerStarted","Data":"422d5dbcd2c0da803fdd9482a7272e0f11941f78aaba66e0811b7ff8c628f5f6"} Nov 25 09:44:17 crc kubenswrapper[4776]: I1125 09:44:17.067666 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6b75d6ddb7-rmtms"] Nov 25 09:44:17 crc kubenswrapper[4776]: I1125 09:44:17.075329 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 09:44:17 crc kubenswrapper[4776]: W1125 09:44:17.078739 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83c080b8_d976_4e81_b103_2442bb2eafe1.slice/crio-2894d4dea69bb1d9b29dcb1eef2460fb6473db26f24cc542599755b582fcb965 WatchSource:0}: Error finding container 2894d4dea69bb1d9b29dcb1eef2460fb6473db26f24cc542599755b582fcb965: Status 404 returned error can't find the container with id 2894d4dea69bb1d9b29dcb1eef2460fb6473db26f24cc542599755b582fcb965 Nov 25 09:44:17 crc kubenswrapper[4776]: W1125 09:44:17.081701 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc595957f_32bd_455a_b212_8f2536102d7f.slice/crio-65ba093605052686fb8026a5ab79fac757efd95fc530300ccdbf7a8e1ef95bd9 WatchSource:0}: Error finding container 65ba093605052686fb8026a5ab79fac757efd95fc530300ccdbf7a8e1ef95bd9: Status 404 returned error can't find the container with id 65ba093605052686fb8026a5ab79fac757efd95fc530300ccdbf7a8e1ef95bd9 Nov 25 09:44:17 crc kubenswrapper[4776]: I1125 09:44:17.819102 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:44:17 crc kubenswrapper[4776]: I1125 09:44:17.819644 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:44:18 crc kubenswrapper[4776]: I1125 09:44:18.025174 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6b75d6ddb7-rmtms" event={"ID":"83c080b8-d976-4e81-b103-2442bb2eafe1","Type":"ContainerStarted","Data":"d35530a8d2f45153bec6fcc975ede0bc99a55fb6d60c6edac19a05f0939799d7"} Nov 25 09:44:18 crc kubenswrapper[4776]: I1125 09:44:18.025225 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6b75d6ddb7-rmtms" event={"ID":"83c080b8-d976-4e81-b103-2442bb2eafe1","Type":"ContainerStarted","Data":"2894d4dea69bb1d9b29dcb1eef2460fb6473db26f24cc542599755b582fcb965"} Nov 25 09:44:18 crc kubenswrapper[4776]: I1125 09:44:18.025247 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:44:18 crc kubenswrapper[4776]: I1125 09:44:18.027758 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"85e1e1f5-a606-427e-a997-55dfad758e2e","Type":"ContainerStarted","Data":"f4038d3a7a822862f0092282cd6f4b6eaa945eeff59e6d0ac36071314e20b11d"} Nov 25 09:44:18 crc kubenswrapper[4776]: I1125 09:44:18.029742 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c595957f-32bd-455a-b212-8f2536102d7f","Type":"ContainerStarted","Data":"7e9a37b45f237cc3a95404fab964097152209924df283c3b016813d90db66b84"} Nov 25 09:44:18 crc kubenswrapper[4776]: I1125 09:44:18.029900 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c595957f-32bd-455a-b212-8f2536102d7f","Type":"ContainerStarted","Data":"65ba093605052686fb8026a5ab79fac757efd95fc530300ccdbf7a8e1ef95bd9"} Nov 25 09:44:18 crc kubenswrapper[4776]: I1125 09:44:18.031421 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-667769d556-lqtlp" event={"ID":"a0b97845-4578-4801-896a-281fdafdb351","Type":"ContainerStarted","Data":"856d610eaa67921f1c8d09ca91236aafc5070c66f02e6f2c2f6ee1982f26930e"} Nov 25 09:44:18 crc kubenswrapper[4776]: I1125 09:44:18.031455 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-667769d556-lqtlp" event={"ID":"a0b97845-4578-4801-896a-281fdafdb351","Type":"ContainerStarted","Data":"5fac5a59b29d2ae80cf45865e0db91319f65183c12c9b736093c4452a98a4355"} Nov 25 09:44:18 crc kubenswrapper[4776]: I1125 09:44:18.031649 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-667769d556-lqtlp" Nov 25 09:44:18 crc kubenswrapper[4776]: I1125 09:44:18.031691 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-667769d556-lqtlp" Nov 25 09:44:18 crc kubenswrapper[4776]: I1125 09:44:18.048022 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6b75d6ddb7-rmtms" podStartSLOduration=5.048007059 podStartE2EDuration="5.048007059s" podCreationTimestamp="2025-11-25 09:44:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:44:18.046725777 +0000 UTC m=+1203.087785330" watchObservedRunningTime="2025-11-25 09:44:18.048007059 +0000 UTC m=+1203.089066602" Nov 25 09:44:18 crc kubenswrapper[4776]: I1125 09:44:18.073754 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-667769d556-lqtlp" podStartSLOduration=5.073727865 podStartE2EDuration="5.073727865s" podCreationTimestamp="2025-11-25 09:44:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:44:18.064967155 +0000 UTC m=+1203.106026718" watchObservedRunningTime="2025-11-25 09:44:18.073727865 +0000 UTC m=+1203.114787418" Nov 25 09:44:19 crc kubenswrapper[4776]: I1125 09:44:19.044000 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c595957f-32bd-455a-b212-8f2536102d7f","Type":"ContainerStarted","Data":"3ca0b83197510418dd82fff89f9a0b35f7b3833f3df2f065cda0aa3374493db9"} Nov 25 09:44:19 crc kubenswrapper[4776]: I1125 09:44:19.048515 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"85e1e1f5-a606-427e-a997-55dfad758e2e","Type":"ContainerStarted","Data":"776925a5c003ac3e724c8cfd6aff1758a1de35d749a1353e22ea01e705238a2e"} Nov 25 09:44:19 crc kubenswrapper[4776]: I1125 09:44:19.101016 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.100994808 podStartE2EDuration="7.100994808s" podCreationTimestamp="2025-11-25 09:44:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:44:19.073492568 +0000 UTC m=+1204.114552121" watchObservedRunningTime="2025-11-25 09:44:19.100994808 +0000 UTC m=+1204.142054361" Nov 25 09:44:19 crc kubenswrapper[4776]: I1125 09:44:19.102918 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=11.102910896 podStartE2EDuration="11.102910896s" podCreationTimestamp="2025-11-25 09:44:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:44:19.097167152 +0000 UTC m=+1204.138226705" watchObservedRunningTime="2025-11-25 09:44:19.102910896 +0000 UTC m=+1204.143970449" Nov 25 09:44:19 crc kubenswrapper[4776]: I1125 09:44:19.261940 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 25 09:44:19 crc kubenswrapper[4776]: I1125 09:44:19.261986 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 25 09:44:19 crc kubenswrapper[4776]: I1125 09:44:19.294910 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 25 09:44:19 crc kubenswrapper[4776]: I1125 09:44:19.309131 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 25 09:44:20 crc kubenswrapper[4776]: I1125 09:44:20.057504 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 25 09:44:20 crc kubenswrapper[4776]: I1125 09:44:20.057876 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 25 09:44:21 crc kubenswrapper[4776]: I1125 09:44:21.080089 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-djwk2" event={"ID":"e522431f-0956-49f7-9c46-b0202824bd56","Type":"ContainerStarted","Data":"5cb968840c90f1277c3bd3ba5789c34ed207afcd5657d28a1ee8b953a59bcbe3"} Nov 25 09:44:21 crc kubenswrapper[4776]: I1125 09:44:21.082627 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-bdhw6" event={"ID":"3ff2a086-8bce-46e2-b294-3b3f31e39b4e","Type":"ContainerStarted","Data":"f6f2a58bc0a49e61eaaf2dd6aaa0193a12cd24b220d82c3d3115c644b6cac73b"} Nov 25 09:44:21 crc kubenswrapper[4776]: I1125 09:44:21.112387 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-djwk2" podStartSLOduration=3.097418504 podStartE2EDuration="48.112335083s" podCreationTimestamp="2025-11-25 09:43:33 +0000 UTC" firstStartedPulling="2025-11-25 09:43:35.050524877 +0000 UTC m=+1160.091584430" lastFinishedPulling="2025-11-25 09:44:20.065441456 +0000 UTC m=+1205.106501009" observedRunningTime="2025-11-25 09:44:21.100685181 +0000 UTC m=+1206.141744734" watchObservedRunningTime="2025-11-25 09:44:21.112335083 +0000 UTC m=+1206.153394636" Nov 25 09:44:21 crc kubenswrapper[4776]: I1125 09:44:21.116284 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-bdhw6" podStartSLOduration=2.192263631 podStartE2EDuration="47.116270931s" podCreationTimestamp="2025-11-25 09:43:34 +0000 UTC" firstStartedPulling="2025-11-25 09:43:35.242739617 +0000 UTC m=+1160.283799170" lastFinishedPulling="2025-11-25 09:44:20.166746917 +0000 UTC m=+1205.207806470" observedRunningTime="2025-11-25 09:44:21.115206685 +0000 UTC m=+1206.156266248" watchObservedRunningTime="2025-11-25 09:44:21.116270931 +0000 UTC m=+1206.157330484" Nov 25 09:44:22 crc kubenswrapper[4776]: I1125 09:44:22.094878 4776 generic.go:334] "Generic (PLEG): container finished" podID="737e5e32-1fa7-4140-bfad-1bc3e10ee390" containerID="820a46f0a557e2d6d0e23522585d4acb1b79eaa6244a8621f29bee90ba4a6713" exitCode=0 Nov 25 09:44:22 crc kubenswrapper[4776]: I1125 09:44:22.094984 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-lgdgb" event={"ID":"737e5e32-1fa7-4140-bfad-1bc3e10ee390","Type":"ContainerDied","Data":"820a46f0a557e2d6d0e23522585d4acb1b79eaa6244a8621f29bee90ba4a6713"} Nov 25 09:44:23 crc kubenswrapper[4776]: I1125 09:44:23.317535 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 25 09:44:23 crc kubenswrapper[4776]: I1125 09:44:23.317802 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 25 09:44:23 crc kubenswrapper[4776]: I1125 09:44:23.364386 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 25 09:44:23 crc kubenswrapper[4776]: I1125 09:44:23.380288 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 25 09:44:24 crc kubenswrapper[4776]: I1125 09:44:24.112819 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 25 09:44:24 crc kubenswrapper[4776]: I1125 09:44:24.112863 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 25 09:44:25 crc kubenswrapper[4776]: I1125 09:44:25.838643 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-lgdgb" Nov 25 09:44:25 crc kubenswrapper[4776]: I1125 09:44:25.974266 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ll7sn\" (UniqueName: \"kubernetes.io/projected/737e5e32-1fa7-4140-bfad-1bc3e10ee390-kube-api-access-ll7sn\") pod \"737e5e32-1fa7-4140-bfad-1bc3e10ee390\" (UID: \"737e5e32-1fa7-4140-bfad-1bc3e10ee390\") " Nov 25 09:44:25 crc kubenswrapper[4776]: I1125 09:44:25.974727 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/737e5e32-1fa7-4140-bfad-1bc3e10ee390-combined-ca-bundle\") pod \"737e5e32-1fa7-4140-bfad-1bc3e10ee390\" (UID: \"737e5e32-1fa7-4140-bfad-1bc3e10ee390\") " Nov 25 09:44:25 crc kubenswrapper[4776]: I1125 09:44:25.974861 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/737e5e32-1fa7-4140-bfad-1bc3e10ee390-config\") pod \"737e5e32-1fa7-4140-bfad-1bc3e10ee390\" (UID: \"737e5e32-1fa7-4140-bfad-1bc3e10ee390\") " Nov 25 09:44:25 crc kubenswrapper[4776]: I1125 09:44:25.986095 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/737e5e32-1fa7-4140-bfad-1bc3e10ee390-kube-api-access-ll7sn" (OuterVolumeSpecName: "kube-api-access-ll7sn") pod "737e5e32-1fa7-4140-bfad-1bc3e10ee390" (UID: "737e5e32-1fa7-4140-bfad-1bc3e10ee390"). InnerVolumeSpecName "kube-api-access-ll7sn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:44:26 crc kubenswrapper[4776]: I1125 09:44:26.006202 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/737e5e32-1fa7-4140-bfad-1bc3e10ee390-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "737e5e32-1fa7-4140-bfad-1bc3e10ee390" (UID: "737e5e32-1fa7-4140-bfad-1bc3e10ee390"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:26 crc kubenswrapper[4776]: I1125 09:44:26.008503 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/737e5e32-1fa7-4140-bfad-1bc3e10ee390-config" (OuterVolumeSpecName: "config") pod "737e5e32-1fa7-4140-bfad-1bc3e10ee390" (UID: "737e5e32-1fa7-4140-bfad-1bc3e10ee390"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:26 crc kubenswrapper[4776]: I1125 09:44:26.077843 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/737e5e32-1fa7-4140-bfad-1bc3e10ee390-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:26 crc kubenswrapper[4776]: I1125 09:44:26.077888 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/737e5e32-1fa7-4140-bfad-1bc3e10ee390-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:26 crc kubenswrapper[4776]: I1125 09:44:26.077901 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ll7sn\" (UniqueName: \"kubernetes.io/projected/737e5e32-1fa7-4140-bfad-1bc3e10ee390-kube-api-access-ll7sn\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:26 crc kubenswrapper[4776]: I1125 09:44:26.132080 4776 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 09:44:26 crc kubenswrapper[4776]: I1125 09:44:26.132103 4776 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 09:44:26 crc kubenswrapper[4776]: I1125 09:44:26.132818 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-lgdgb" Nov 25 09:44:26 crc kubenswrapper[4776]: I1125 09:44:26.134267 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-lgdgb" event={"ID":"737e5e32-1fa7-4140-bfad-1bc3e10ee390","Type":"ContainerDied","Data":"cb3326ae7455fbf36c1b1a64cf546d6c0ecc75a6352986c6cfd4f9551eca8c67"} Nov 25 09:44:26 crc kubenswrapper[4776]: I1125 09:44:26.134318 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb3326ae7455fbf36c1b1a64cf546d6c0ecc75a6352986c6cfd4f9551eca8c67" Nov 25 09:44:26 crc kubenswrapper[4776]: I1125 09:44:26.305549 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 25 09:44:26 crc kubenswrapper[4776]: I1125 09:44:26.306414 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.003258 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c654c9745-whq24"] Nov 25 09:44:27 crc kubenswrapper[4776]: E1125 09:44:27.004474 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="737e5e32-1fa7-4140-bfad-1bc3e10ee390" containerName="neutron-db-sync" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.004489 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="737e5e32-1fa7-4140-bfad-1bc3e10ee390" containerName="neutron-db-sync" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.004658 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="737e5e32-1fa7-4140-bfad-1bc3e10ee390" containerName="neutron-db-sync" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.005651 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c654c9745-whq24" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.032411 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c654c9745-whq24"] Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.078104 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5f4c844d98-nt7dn"] Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.079532 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5f4c844d98-nt7dn" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.084368 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-dp8gc" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.085375 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.088521 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.088699 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.101424 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-ovsdbserver-sb\") pod \"dnsmasq-dns-6c654c9745-whq24\" (UID: \"3d2378da-fcb1-4ca6-8906-ae48da346e95\") " pod="openstack/dnsmasq-dns-6c654c9745-whq24" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.101475 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-config\") pod \"dnsmasq-dns-6c654c9745-whq24\" (UID: \"3d2378da-fcb1-4ca6-8906-ae48da346e95\") " pod="openstack/dnsmasq-dns-6c654c9745-whq24" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.101502 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-dns-swift-storage-0\") pod \"dnsmasq-dns-6c654c9745-whq24\" (UID: \"3d2378da-fcb1-4ca6-8906-ae48da346e95\") " pod="openstack/dnsmasq-dns-6c654c9745-whq24" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.101542 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-ovsdbserver-nb\") pod \"dnsmasq-dns-6c654c9745-whq24\" (UID: \"3d2378da-fcb1-4ca6-8906-ae48da346e95\") " pod="openstack/dnsmasq-dns-6c654c9745-whq24" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.101604 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ls52k\" (UniqueName: \"kubernetes.io/projected/3d2378da-fcb1-4ca6-8906-ae48da346e95-kube-api-access-ls52k\") pod \"dnsmasq-dns-6c654c9745-whq24\" (UID: \"3d2378da-fcb1-4ca6-8906-ae48da346e95\") " pod="openstack/dnsmasq-dns-6c654c9745-whq24" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.101649 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-dns-svc\") pod \"dnsmasq-dns-6c654c9745-whq24\" (UID: \"3d2378da-fcb1-4ca6-8906-ae48da346e95\") " pod="openstack/dnsmasq-dns-6c654c9745-whq24" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.139626 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5f4c844d98-nt7dn"] Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.203374 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-ovsdbserver-sb\") pod \"dnsmasq-dns-6c654c9745-whq24\" (UID: \"3d2378da-fcb1-4ca6-8906-ae48da346e95\") " pod="openstack/dnsmasq-dns-6c654c9745-whq24" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.203423 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-config\") pod \"dnsmasq-dns-6c654c9745-whq24\" (UID: \"3d2378da-fcb1-4ca6-8906-ae48da346e95\") " pod="openstack/dnsmasq-dns-6c654c9745-whq24" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.203454 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-dns-swift-storage-0\") pod \"dnsmasq-dns-6c654c9745-whq24\" (UID: \"3d2378da-fcb1-4ca6-8906-ae48da346e95\") " pod="openstack/dnsmasq-dns-6c654c9745-whq24" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.203484 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-ovsdbserver-nb\") pod \"dnsmasq-dns-6c654c9745-whq24\" (UID: \"3d2378da-fcb1-4ca6-8906-ae48da346e95\") " pod="openstack/dnsmasq-dns-6c654c9745-whq24" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.203518 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-config\") pod \"neutron-5f4c844d98-nt7dn\" (UID: \"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0\") " pod="openstack/neutron-5f4c844d98-nt7dn" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.203549 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-httpd-config\") pod \"neutron-5f4c844d98-nt7dn\" (UID: \"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0\") " pod="openstack/neutron-5f4c844d98-nt7dn" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.203571 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-combined-ca-bundle\") pod \"neutron-5f4c844d98-nt7dn\" (UID: \"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0\") " pod="openstack/neutron-5f4c844d98-nt7dn" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.203587 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xts9\" (UniqueName: \"kubernetes.io/projected/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-kube-api-access-7xts9\") pod \"neutron-5f4c844d98-nt7dn\" (UID: \"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0\") " pod="openstack/neutron-5f4c844d98-nt7dn" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.203610 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-ovndb-tls-certs\") pod \"neutron-5f4c844d98-nt7dn\" (UID: \"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0\") " pod="openstack/neutron-5f4c844d98-nt7dn" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.203627 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ls52k\" (UniqueName: \"kubernetes.io/projected/3d2378da-fcb1-4ca6-8906-ae48da346e95-kube-api-access-ls52k\") pod \"dnsmasq-dns-6c654c9745-whq24\" (UID: \"3d2378da-fcb1-4ca6-8906-ae48da346e95\") " pod="openstack/dnsmasq-dns-6c654c9745-whq24" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.203708 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-dns-svc\") pod \"dnsmasq-dns-6c654c9745-whq24\" (UID: \"3d2378da-fcb1-4ca6-8906-ae48da346e95\") " pod="openstack/dnsmasq-dns-6c654c9745-whq24" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.204401 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-config\") pod \"dnsmasq-dns-6c654c9745-whq24\" (UID: \"3d2378da-fcb1-4ca6-8906-ae48da346e95\") " pod="openstack/dnsmasq-dns-6c654c9745-whq24" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.204477 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-dns-swift-storage-0\") pod \"dnsmasq-dns-6c654c9745-whq24\" (UID: \"3d2378da-fcb1-4ca6-8906-ae48da346e95\") " pod="openstack/dnsmasq-dns-6c654c9745-whq24" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.204656 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-ovsdbserver-nb\") pod \"dnsmasq-dns-6c654c9745-whq24\" (UID: \"3d2378da-fcb1-4ca6-8906-ae48da346e95\") " pod="openstack/dnsmasq-dns-6c654c9745-whq24" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.204656 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-dns-svc\") pod \"dnsmasq-dns-6c654c9745-whq24\" (UID: \"3d2378da-fcb1-4ca6-8906-ae48da346e95\") " pod="openstack/dnsmasq-dns-6c654c9745-whq24" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.205354 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-ovsdbserver-sb\") pod \"dnsmasq-dns-6c654c9745-whq24\" (UID: \"3d2378da-fcb1-4ca6-8906-ae48da346e95\") " pod="openstack/dnsmasq-dns-6c654c9745-whq24" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.239940 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ls52k\" (UniqueName: \"kubernetes.io/projected/3d2378da-fcb1-4ca6-8906-ae48da346e95-kube-api-access-ls52k\") pod \"dnsmasq-dns-6c654c9745-whq24\" (UID: \"3d2378da-fcb1-4ca6-8906-ae48da346e95\") " pod="openstack/dnsmasq-dns-6c654c9745-whq24" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.305753 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-ovndb-tls-certs\") pod \"neutron-5f4c844d98-nt7dn\" (UID: \"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0\") " pod="openstack/neutron-5f4c844d98-nt7dn" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.306432 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-config\") pod \"neutron-5f4c844d98-nt7dn\" (UID: \"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0\") " pod="openstack/neutron-5f4c844d98-nt7dn" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.306471 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-httpd-config\") pod \"neutron-5f4c844d98-nt7dn\" (UID: \"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0\") " pod="openstack/neutron-5f4c844d98-nt7dn" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.306502 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-combined-ca-bundle\") pod \"neutron-5f4c844d98-nt7dn\" (UID: \"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0\") " pod="openstack/neutron-5f4c844d98-nt7dn" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.306517 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xts9\" (UniqueName: \"kubernetes.io/projected/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-kube-api-access-7xts9\") pod \"neutron-5f4c844d98-nt7dn\" (UID: \"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0\") " pod="openstack/neutron-5f4c844d98-nt7dn" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.309983 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-httpd-config\") pod \"neutron-5f4c844d98-nt7dn\" (UID: \"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0\") " pod="openstack/neutron-5f4c844d98-nt7dn" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.311089 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-ovndb-tls-certs\") pod \"neutron-5f4c844d98-nt7dn\" (UID: \"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0\") " pod="openstack/neutron-5f4c844d98-nt7dn" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.311503 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-combined-ca-bundle\") pod \"neutron-5f4c844d98-nt7dn\" (UID: \"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0\") " pod="openstack/neutron-5f4c844d98-nt7dn" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.322582 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-config\") pod \"neutron-5f4c844d98-nt7dn\" (UID: \"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0\") " pod="openstack/neutron-5f4c844d98-nt7dn" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.326713 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xts9\" (UniqueName: \"kubernetes.io/projected/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-kube-api-access-7xts9\") pod \"neutron-5f4c844d98-nt7dn\" (UID: \"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0\") " pod="openstack/neutron-5f4c844d98-nt7dn" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.333218 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c654c9745-whq24" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.408652 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5f4c844d98-nt7dn" Nov 25 09:44:27 crc kubenswrapper[4776]: I1125 09:44:27.926257 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c654c9745-whq24"] Nov 25 09:44:27 crc kubenswrapper[4776]: W1125 09:44:27.939999 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d2378da_fcb1_4ca6_8906_ae48da346e95.slice/crio-c4756ea9775ac58ef3226a2d6d0d3612568a801c8743e71c78aa176436223403 WatchSource:0}: Error finding container c4756ea9775ac58ef3226a2d6d0d3612568a801c8743e71c78aa176436223403: Status 404 returned error can't find the container with id c4756ea9775ac58ef3226a2d6d0d3612568a801c8743e71c78aa176436223403 Nov 25 09:44:28 crc kubenswrapper[4776]: I1125 09:44:28.157701 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c654c9745-whq24" event={"ID":"3d2378da-fcb1-4ca6-8906-ae48da346e95","Type":"ContainerStarted","Data":"c4756ea9775ac58ef3226a2d6d0d3612568a801c8743e71c78aa176436223403"} Nov 25 09:44:28 crc kubenswrapper[4776]: I1125 09:44:28.166480 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54","Type":"ContainerStarted","Data":"96da9294397b2a279bcaa5e6613b6233bff7993839bcf9c297a377f464d05dcb"} Nov 25 09:44:28 crc kubenswrapper[4776]: I1125 09:44:28.166619 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="61cc96f1-6a4d-4bc8-93a1-e92fdb254e54" containerName="ceilometer-central-agent" containerID="cri-o://866bc21b77e90a1bb28e58f78fe8177df9aa57d681b670c32ee93364c6a6b878" gracePeriod=30 Nov 25 09:44:28 crc kubenswrapper[4776]: I1125 09:44:28.167040 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="61cc96f1-6a4d-4bc8-93a1-e92fdb254e54" containerName="proxy-httpd" containerID="cri-o://96da9294397b2a279bcaa5e6613b6233bff7993839bcf9c297a377f464d05dcb" gracePeriod=30 Nov 25 09:44:28 crc kubenswrapper[4776]: I1125 09:44:28.167113 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="61cc96f1-6a4d-4bc8-93a1-e92fdb254e54" containerName="sg-core" containerID="cri-o://4ea77e92eb9f4a8c4a0aea41ded4fa353ada185d27f639d8c33be0c971e1cbf5" gracePeriod=30 Nov 25 09:44:28 crc kubenswrapper[4776]: I1125 09:44:28.167158 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="61cc96f1-6a4d-4bc8-93a1-e92fdb254e54" containerName="ceilometer-notification-agent" containerID="cri-o://c5a7ff2ecc6cca04725dbb7788030a76d83ca550c67e52f0ca5ef9c4078a782d" gracePeriod=30 Nov 25 09:44:28 crc kubenswrapper[4776]: I1125 09:44:28.185685 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5f4c844d98-nt7dn"] Nov 25 09:44:28 crc kubenswrapper[4776]: W1125 09:44:28.190940 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27cb0fd2_6e6d_4e2a_b5d8_a1940c1ef8b0.slice/crio-dd48824ff71693cb62c0c30eae3bf938ed6d90a2c5d5cf3c217fc228e4ebb86f WatchSource:0}: Error finding container dd48824ff71693cb62c0c30eae3bf938ed6d90a2c5d5cf3c217fc228e4ebb86f: Status 404 returned error can't find the container with id dd48824ff71693cb62c0c30eae3bf938ed6d90a2c5d5cf3c217fc228e4ebb86f Nov 25 09:44:28 crc kubenswrapper[4776]: I1125 09:44:28.199672 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.0889392 podStartE2EDuration="55.199656122s" podCreationTimestamp="2025-11-25 09:43:33 +0000 UTC" firstStartedPulling="2025-11-25 09:43:35.278304479 +0000 UTC m=+1160.319364032" lastFinishedPulling="2025-11-25 09:44:27.389021401 +0000 UTC m=+1212.430080954" observedRunningTime="2025-11-25 09:44:28.196743409 +0000 UTC m=+1213.237802962" watchObservedRunningTime="2025-11-25 09:44:28.199656122 +0000 UTC m=+1213.240715675" Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.175407 4776 generic.go:334] "Generic (PLEG): container finished" podID="3d2378da-fcb1-4ca6-8906-ae48da346e95" containerID="98b75f64d71dad3aa5ba54b90468297f0304b42e7e1c65945da0c8d495925174" exitCode=0 Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.175510 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c654c9745-whq24" event={"ID":"3d2378da-fcb1-4ca6-8906-ae48da346e95","Type":"ContainerDied","Data":"98b75f64d71dad3aa5ba54b90468297f0304b42e7e1c65945da0c8d495925174"} Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.178304 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5f4c844d98-nt7dn" event={"ID":"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0","Type":"ContainerStarted","Data":"d8c7c2e271eae150bda7090f4dde04fe427c80ff86bd65b1e36e0f137aac0a52"} Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.178555 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5f4c844d98-nt7dn" event={"ID":"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0","Type":"ContainerStarted","Data":"780469763d9074be88958dd18def678a1f9122fe313a2bc79714b4d658d275e9"} Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.178570 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5f4c844d98-nt7dn" event={"ID":"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0","Type":"ContainerStarted","Data":"dd48824ff71693cb62c0c30eae3bf938ed6d90a2c5d5cf3c217fc228e4ebb86f"} Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.178586 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5f4c844d98-nt7dn" Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.196698 4776 generic.go:334] "Generic (PLEG): container finished" podID="61cc96f1-6a4d-4bc8-93a1-e92fdb254e54" containerID="96da9294397b2a279bcaa5e6613b6233bff7993839bcf9c297a377f464d05dcb" exitCode=0 Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.196730 4776 generic.go:334] "Generic (PLEG): container finished" podID="61cc96f1-6a4d-4bc8-93a1-e92fdb254e54" containerID="4ea77e92eb9f4a8c4a0aea41ded4fa353ada185d27f639d8c33be0c971e1cbf5" exitCode=2 Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.196738 4776 generic.go:334] "Generic (PLEG): container finished" podID="61cc96f1-6a4d-4bc8-93a1-e92fdb254e54" containerID="866bc21b77e90a1bb28e58f78fe8177df9aa57d681b670c32ee93364c6a6b878" exitCode=0 Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.196761 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54","Type":"ContainerDied","Data":"96da9294397b2a279bcaa5e6613b6233bff7993839bcf9c297a377f464d05dcb"} Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.196796 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54","Type":"ContainerDied","Data":"4ea77e92eb9f4a8c4a0aea41ded4fa353ada185d27f639d8c33be0c971e1cbf5"} Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.196806 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54","Type":"ContainerDied","Data":"866bc21b77e90a1bb28e58f78fe8177df9aa57d681b670c32ee93364c6a6b878"} Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.229338 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5f4c844d98-nt7dn" podStartSLOduration=2.229312395 podStartE2EDuration="2.229312395s" podCreationTimestamp="2025-11-25 09:44:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:44:29.222934695 +0000 UTC m=+1214.263994258" watchObservedRunningTime="2025-11-25 09:44:29.229312395 +0000 UTC m=+1214.270371958" Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.498298 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-54dbffbb57-kmqfb"] Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.500145 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-54dbffbb57-kmqfb" Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.502501 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.504003 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.511928 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-54dbffbb57-kmqfb"] Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.547117 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-internal-tls-certs\") pod \"neutron-54dbffbb57-kmqfb\" (UID: \"fe3ff67b-01ff-480f-be91-7c8235593c97\") " pod="openstack/neutron-54dbffbb57-kmqfb" Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.547164 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-httpd-config\") pod \"neutron-54dbffbb57-kmqfb\" (UID: \"fe3ff67b-01ff-480f-be91-7c8235593c97\") " pod="openstack/neutron-54dbffbb57-kmqfb" Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.547196 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f78rw\" (UniqueName: \"kubernetes.io/projected/fe3ff67b-01ff-480f-be91-7c8235593c97-kube-api-access-f78rw\") pod \"neutron-54dbffbb57-kmqfb\" (UID: \"fe3ff67b-01ff-480f-be91-7c8235593c97\") " pod="openstack/neutron-54dbffbb57-kmqfb" Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.547218 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-combined-ca-bundle\") pod \"neutron-54dbffbb57-kmqfb\" (UID: \"fe3ff67b-01ff-480f-be91-7c8235593c97\") " pod="openstack/neutron-54dbffbb57-kmqfb" Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.547279 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-config\") pod \"neutron-54dbffbb57-kmqfb\" (UID: \"fe3ff67b-01ff-480f-be91-7c8235593c97\") " pod="openstack/neutron-54dbffbb57-kmqfb" Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.547548 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-ovndb-tls-certs\") pod \"neutron-54dbffbb57-kmqfb\" (UID: \"fe3ff67b-01ff-480f-be91-7c8235593c97\") " pod="openstack/neutron-54dbffbb57-kmqfb" Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.547644 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-public-tls-certs\") pod \"neutron-54dbffbb57-kmqfb\" (UID: \"fe3ff67b-01ff-480f-be91-7c8235593c97\") " pod="openstack/neutron-54dbffbb57-kmqfb" Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.649472 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-config\") pod \"neutron-54dbffbb57-kmqfb\" (UID: \"fe3ff67b-01ff-480f-be91-7c8235593c97\") " pod="openstack/neutron-54dbffbb57-kmqfb" Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.649553 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-ovndb-tls-certs\") pod \"neutron-54dbffbb57-kmqfb\" (UID: \"fe3ff67b-01ff-480f-be91-7c8235593c97\") " pod="openstack/neutron-54dbffbb57-kmqfb" Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.649594 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-public-tls-certs\") pod \"neutron-54dbffbb57-kmqfb\" (UID: \"fe3ff67b-01ff-480f-be91-7c8235593c97\") " pod="openstack/neutron-54dbffbb57-kmqfb" Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.649671 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-internal-tls-certs\") pod \"neutron-54dbffbb57-kmqfb\" (UID: \"fe3ff67b-01ff-480f-be91-7c8235593c97\") " pod="openstack/neutron-54dbffbb57-kmqfb" Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.649699 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-httpd-config\") pod \"neutron-54dbffbb57-kmqfb\" (UID: \"fe3ff67b-01ff-480f-be91-7c8235593c97\") " pod="openstack/neutron-54dbffbb57-kmqfb" Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.649737 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f78rw\" (UniqueName: \"kubernetes.io/projected/fe3ff67b-01ff-480f-be91-7c8235593c97-kube-api-access-f78rw\") pod \"neutron-54dbffbb57-kmqfb\" (UID: \"fe3ff67b-01ff-480f-be91-7c8235593c97\") " pod="openstack/neutron-54dbffbb57-kmqfb" Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.649763 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-combined-ca-bundle\") pod \"neutron-54dbffbb57-kmqfb\" (UID: \"fe3ff67b-01ff-480f-be91-7c8235593c97\") " pod="openstack/neutron-54dbffbb57-kmqfb" Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.654639 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-internal-tls-certs\") pod \"neutron-54dbffbb57-kmqfb\" (UID: \"fe3ff67b-01ff-480f-be91-7c8235593c97\") " pod="openstack/neutron-54dbffbb57-kmqfb" Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.655172 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-public-tls-certs\") pod \"neutron-54dbffbb57-kmqfb\" (UID: \"fe3ff67b-01ff-480f-be91-7c8235593c97\") " pod="openstack/neutron-54dbffbb57-kmqfb" Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.655977 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-combined-ca-bundle\") pod \"neutron-54dbffbb57-kmqfb\" (UID: \"fe3ff67b-01ff-480f-be91-7c8235593c97\") " pod="openstack/neutron-54dbffbb57-kmqfb" Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.656811 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-httpd-config\") pod \"neutron-54dbffbb57-kmqfb\" (UID: \"fe3ff67b-01ff-480f-be91-7c8235593c97\") " pod="openstack/neutron-54dbffbb57-kmqfb" Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.657856 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-config\") pod \"neutron-54dbffbb57-kmqfb\" (UID: \"fe3ff67b-01ff-480f-be91-7c8235593c97\") " pod="openstack/neutron-54dbffbb57-kmqfb" Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.659549 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-ovndb-tls-certs\") pod \"neutron-54dbffbb57-kmqfb\" (UID: \"fe3ff67b-01ff-480f-be91-7c8235593c97\") " pod="openstack/neutron-54dbffbb57-kmqfb" Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.677122 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f78rw\" (UniqueName: \"kubernetes.io/projected/fe3ff67b-01ff-480f-be91-7c8235593c97-kube-api-access-f78rw\") pod \"neutron-54dbffbb57-kmqfb\" (UID: \"fe3ff67b-01ff-480f-be91-7c8235593c97\") " pod="openstack/neutron-54dbffbb57-kmqfb" Nov 25 09:44:29 crc kubenswrapper[4776]: I1125 09:44:29.898128 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-54dbffbb57-kmqfb" Nov 25 09:44:30 crc kubenswrapper[4776]: I1125 09:44:30.208278 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c654c9745-whq24" event={"ID":"3d2378da-fcb1-4ca6-8906-ae48da346e95","Type":"ContainerStarted","Data":"fdf727eac872d98e72351108b80191b34fb3b7d5ebcee63350880bfe42d78f37"} Nov 25 09:44:30 crc kubenswrapper[4776]: I1125 09:44:30.208725 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6c654c9745-whq24" Nov 25 09:44:30 crc kubenswrapper[4776]: I1125 09:44:30.215482 4776 generic.go:334] "Generic (PLEG): container finished" podID="61cc96f1-6a4d-4bc8-93a1-e92fdb254e54" containerID="c5a7ff2ecc6cca04725dbb7788030a76d83ca550c67e52f0ca5ef9c4078a782d" exitCode=0 Nov 25 09:44:30 crc kubenswrapper[4776]: I1125 09:44:30.215586 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54","Type":"ContainerDied","Data":"c5a7ff2ecc6cca04725dbb7788030a76d83ca550c67e52f0ca5ef9c4078a782d"} Nov 25 09:44:30 crc kubenswrapper[4776]: I1125 09:44:30.236237 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6c654c9745-whq24" podStartSLOduration=4.236220608 podStartE2EDuration="4.236220608s" podCreationTimestamp="2025-11-25 09:44:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:44:30.230415873 +0000 UTC m=+1215.271475446" watchObservedRunningTime="2025-11-25 09:44:30.236220608 +0000 UTC m=+1215.277280151" Nov 25 09:44:30 crc kubenswrapper[4776]: I1125 09:44:30.486858 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 09:44:30 crc kubenswrapper[4776]: I1125 09:44:30.566664 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-log-httpd\") pod \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\" (UID: \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\") " Nov 25 09:44:30 crc kubenswrapper[4776]: I1125 09:44:30.566770 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-scripts\") pod \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\" (UID: \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\") " Nov 25 09:44:30 crc kubenswrapper[4776]: I1125 09:44:30.566859 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94nlr\" (UniqueName: \"kubernetes.io/projected/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-kube-api-access-94nlr\") pod \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\" (UID: \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\") " Nov 25 09:44:30 crc kubenswrapper[4776]: I1125 09:44:30.566899 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-combined-ca-bundle\") pod \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\" (UID: \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\") " Nov 25 09:44:30 crc kubenswrapper[4776]: I1125 09:44:30.566931 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-run-httpd\") pod \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\" (UID: \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\") " Nov 25 09:44:30 crc kubenswrapper[4776]: I1125 09:44:30.566977 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-sg-core-conf-yaml\") pod \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\" (UID: \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\") " Nov 25 09:44:30 crc kubenswrapper[4776]: I1125 09:44:30.567022 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-config-data\") pod \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\" (UID: \"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54\") " Nov 25 09:44:30 crc kubenswrapper[4776]: I1125 09:44:30.568569 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "61cc96f1-6a4d-4bc8-93a1-e92fdb254e54" (UID: "61cc96f1-6a4d-4bc8-93a1-e92fdb254e54"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:44:30 crc kubenswrapper[4776]: I1125 09:44:30.568797 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "61cc96f1-6a4d-4bc8-93a1-e92fdb254e54" (UID: "61cc96f1-6a4d-4bc8-93a1-e92fdb254e54"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:44:30 crc kubenswrapper[4776]: I1125 09:44:30.589440 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-scripts" (OuterVolumeSpecName: "scripts") pod "61cc96f1-6a4d-4bc8-93a1-e92fdb254e54" (UID: "61cc96f1-6a4d-4bc8-93a1-e92fdb254e54"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:30 crc kubenswrapper[4776]: I1125 09:44:30.615278 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-kube-api-access-94nlr" (OuterVolumeSpecName: "kube-api-access-94nlr") pod "61cc96f1-6a4d-4bc8-93a1-e92fdb254e54" (UID: "61cc96f1-6a4d-4bc8-93a1-e92fdb254e54"). InnerVolumeSpecName "kube-api-access-94nlr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:44:30 crc kubenswrapper[4776]: I1125 09:44:30.669366 4776 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:30 crc kubenswrapper[4776]: I1125 09:44:30.669621 4776 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:30 crc kubenswrapper[4776]: I1125 09:44:30.669690 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:30 crc kubenswrapper[4776]: I1125 09:44:30.669821 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94nlr\" (UniqueName: \"kubernetes.io/projected/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-kube-api-access-94nlr\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:30 crc kubenswrapper[4776]: I1125 09:44:30.670139 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "61cc96f1-6a4d-4bc8-93a1-e92fdb254e54" (UID: "61cc96f1-6a4d-4bc8-93a1-e92fdb254e54"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:30 crc kubenswrapper[4776]: I1125 09:44:30.712400 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "61cc96f1-6a4d-4bc8-93a1-e92fdb254e54" (UID: "61cc96f1-6a4d-4bc8-93a1-e92fdb254e54"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:30 crc kubenswrapper[4776]: W1125 09:44:30.728732 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe3ff67b_01ff_480f_be91_7c8235593c97.slice/crio-7d83e210b3c97dae8119bdb60c3219ec6037f6359c74ccc6a38954fcea697f4c WatchSource:0}: Error finding container 7d83e210b3c97dae8119bdb60c3219ec6037f6359c74ccc6a38954fcea697f4c: Status 404 returned error can't find the container with id 7d83e210b3c97dae8119bdb60c3219ec6037f6359c74ccc6a38954fcea697f4c Nov 25 09:44:30 crc kubenswrapper[4776]: I1125 09:44:30.730613 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-config-data" (OuterVolumeSpecName: "config-data") pod "61cc96f1-6a4d-4bc8-93a1-e92fdb254e54" (UID: "61cc96f1-6a4d-4bc8-93a1-e92fdb254e54"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:30 crc kubenswrapper[4776]: I1125 09:44:30.731701 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-54dbffbb57-kmqfb"] Nov 25 09:44:30 crc kubenswrapper[4776]: I1125 09:44:30.771641 4776 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:30 crc kubenswrapper[4776]: I1125 09:44:30.771876 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:30 crc kubenswrapper[4776]: I1125 09:44:30.771984 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.233462 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"61cc96f1-6a4d-4bc8-93a1-e92fdb254e54","Type":"ContainerDied","Data":"052f4d0d7461848ab646011b61a4b893ab6e2e7552c5cc6dc95a3f69ec0d85c7"} Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.234757 4776 scope.go:117] "RemoveContainer" containerID="96da9294397b2a279bcaa5e6613b6233bff7993839bcf9c297a377f464d05dcb" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.233512 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.236947 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54dbffbb57-kmqfb" event={"ID":"fe3ff67b-01ff-480f-be91-7c8235593c97","Type":"ContainerStarted","Data":"5c1129db350942a00317a1332a9a8db16c63fbd966ba3d63890450c144964ddb"} Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.237183 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54dbffbb57-kmqfb" event={"ID":"fe3ff67b-01ff-480f-be91-7c8235593c97","Type":"ContainerStarted","Data":"94e2c5021a1689b990cd8fd474841522443a046e3012b95cb3f5f1f9b90efdd1"} Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.237316 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54dbffbb57-kmqfb" event={"ID":"fe3ff67b-01ff-480f-be91-7c8235593c97","Type":"ContainerStarted","Data":"7d83e210b3c97dae8119bdb60c3219ec6037f6359c74ccc6a38954fcea697f4c"} Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.260126 4776 scope.go:117] "RemoveContainer" containerID="4ea77e92eb9f4a8c4a0aea41ded4fa353ada185d27f639d8c33be0c971e1cbf5" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.264987 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-54dbffbb57-kmqfb" podStartSLOduration=2.264937739 podStartE2EDuration="2.264937739s" podCreationTimestamp="2025-11-25 09:44:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:44:31.261999995 +0000 UTC m=+1216.303059568" watchObservedRunningTime="2025-11-25 09:44:31.264937739 +0000 UTC m=+1216.305997302" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.302142 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.312578 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.332137 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:44:31 crc kubenswrapper[4776]: E1125 09:44:31.332570 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61cc96f1-6a4d-4bc8-93a1-e92fdb254e54" containerName="sg-core" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.332587 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="61cc96f1-6a4d-4bc8-93a1-e92fdb254e54" containerName="sg-core" Nov 25 09:44:31 crc kubenswrapper[4776]: E1125 09:44:31.332608 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61cc96f1-6a4d-4bc8-93a1-e92fdb254e54" containerName="ceilometer-central-agent" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.332615 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="61cc96f1-6a4d-4bc8-93a1-e92fdb254e54" containerName="ceilometer-central-agent" Nov 25 09:44:31 crc kubenswrapper[4776]: E1125 09:44:31.332628 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61cc96f1-6a4d-4bc8-93a1-e92fdb254e54" containerName="proxy-httpd" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.332635 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="61cc96f1-6a4d-4bc8-93a1-e92fdb254e54" containerName="proxy-httpd" Nov 25 09:44:31 crc kubenswrapper[4776]: E1125 09:44:31.332644 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61cc96f1-6a4d-4bc8-93a1-e92fdb254e54" containerName="ceilometer-notification-agent" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.332650 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="61cc96f1-6a4d-4bc8-93a1-e92fdb254e54" containerName="ceilometer-notification-agent" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.332814 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="61cc96f1-6a4d-4bc8-93a1-e92fdb254e54" containerName="ceilometer-notification-agent" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.332836 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="61cc96f1-6a4d-4bc8-93a1-e92fdb254e54" containerName="sg-core" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.332846 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="61cc96f1-6a4d-4bc8-93a1-e92fdb254e54" containerName="proxy-httpd" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.332858 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="61cc96f1-6a4d-4bc8-93a1-e92fdb254e54" containerName="ceilometer-central-agent" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.334355 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.337258 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.337416 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.344590 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.361917 4776 scope.go:117] "RemoveContainer" containerID="c5a7ff2ecc6cca04725dbb7788030a76d83ca550c67e52f0ca5ef9c4078a782d" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.385873 4776 scope.go:117] "RemoveContainer" containerID="866bc21b77e90a1bb28e58f78fe8177df9aa57d681b670c32ee93364c6a6b878" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.391924 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b62eda75-27dc-4a93-9ef5-a729483eb7a0-run-httpd\") pod \"ceilometer-0\" (UID: \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\") " pod="openstack/ceilometer-0" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.392022 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b62eda75-27dc-4a93-9ef5-a729483eb7a0-log-httpd\") pod \"ceilometer-0\" (UID: \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\") " pod="openstack/ceilometer-0" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.392110 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b62eda75-27dc-4a93-9ef5-a729483eb7a0-scripts\") pod \"ceilometer-0\" (UID: \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\") " pod="openstack/ceilometer-0" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.392139 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b62eda75-27dc-4a93-9ef5-a729483eb7a0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\") " pod="openstack/ceilometer-0" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.392193 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b62eda75-27dc-4a93-9ef5-a729483eb7a0-config-data\") pod \"ceilometer-0\" (UID: \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\") " pod="openstack/ceilometer-0" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.392225 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5492f\" (UniqueName: \"kubernetes.io/projected/b62eda75-27dc-4a93-9ef5-a729483eb7a0-kube-api-access-5492f\") pod \"ceilometer-0\" (UID: \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\") " pod="openstack/ceilometer-0" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.392260 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b62eda75-27dc-4a93-9ef5-a729483eb7a0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\") " pod="openstack/ceilometer-0" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.493487 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b62eda75-27dc-4a93-9ef5-a729483eb7a0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\") " pod="openstack/ceilometer-0" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.493799 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b62eda75-27dc-4a93-9ef5-a729483eb7a0-run-httpd\") pod \"ceilometer-0\" (UID: \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\") " pod="openstack/ceilometer-0" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.493875 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b62eda75-27dc-4a93-9ef5-a729483eb7a0-log-httpd\") pod \"ceilometer-0\" (UID: \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\") " pod="openstack/ceilometer-0" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.493912 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b62eda75-27dc-4a93-9ef5-a729483eb7a0-scripts\") pod \"ceilometer-0\" (UID: \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\") " pod="openstack/ceilometer-0" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.493933 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b62eda75-27dc-4a93-9ef5-a729483eb7a0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\") " pod="openstack/ceilometer-0" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.493982 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b62eda75-27dc-4a93-9ef5-a729483eb7a0-config-data\") pod \"ceilometer-0\" (UID: \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\") " pod="openstack/ceilometer-0" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.494014 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5492f\" (UniqueName: \"kubernetes.io/projected/b62eda75-27dc-4a93-9ef5-a729483eb7a0-kube-api-access-5492f\") pod \"ceilometer-0\" (UID: \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\") " pod="openstack/ceilometer-0" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.494637 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b62eda75-27dc-4a93-9ef5-a729483eb7a0-run-httpd\") pod \"ceilometer-0\" (UID: \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\") " pod="openstack/ceilometer-0" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.494720 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b62eda75-27dc-4a93-9ef5-a729483eb7a0-log-httpd\") pod \"ceilometer-0\" (UID: \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\") " pod="openstack/ceilometer-0" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.498468 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b62eda75-27dc-4a93-9ef5-a729483eb7a0-scripts\") pod \"ceilometer-0\" (UID: \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\") " pod="openstack/ceilometer-0" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.503873 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b62eda75-27dc-4a93-9ef5-a729483eb7a0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\") " pod="openstack/ceilometer-0" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.509043 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b62eda75-27dc-4a93-9ef5-a729483eb7a0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\") " pod="openstack/ceilometer-0" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.516221 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5492f\" (UniqueName: \"kubernetes.io/projected/b62eda75-27dc-4a93-9ef5-a729483eb7a0-kube-api-access-5492f\") pod \"ceilometer-0\" (UID: \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\") " pod="openstack/ceilometer-0" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.519294 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b62eda75-27dc-4a93-9ef5-a729483eb7a0-config-data\") pod \"ceilometer-0\" (UID: \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\") " pod="openstack/ceilometer-0" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.662001 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 09:44:31 crc kubenswrapper[4776]: I1125 09:44:31.708098 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61cc96f1-6a4d-4bc8-93a1-e92fdb254e54" path="/var/lib/kubelet/pods/61cc96f1-6a4d-4bc8-93a1-e92fdb254e54/volumes" Nov 25 09:44:32 crc kubenswrapper[4776]: I1125 09:44:32.247472 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-54dbffbb57-kmqfb" Nov 25 09:44:32 crc kubenswrapper[4776]: I1125 09:44:32.273546 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:44:33 crc kubenswrapper[4776]: I1125 09:44:33.256381 4776 generic.go:334] "Generic (PLEG): container finished" podID="3ff2a086-8bce-46e2-b294-3b3f31e39b4e" containerID="f6f2a58bc0a49e61eaaf2dd6aaa0193a12cd24b220d82c3d3115c644b6cac73b" exitCode=0 Nov 25 09:44:33 crc kubenswrapper[4776]: I1125 09:44:33.256444 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-bdhw6" event={"ID":"3ff2a086-8bce-46e2-b294-3b3f31e39b4e","Type":"ContainerDied","Data":"f6f2a58bc0a49e61eaaf2dd6aaa0193a12cd24b220d82c3d3115c644b6cac73b"} Nov 25 09:44:33 crc kubenswrapper[4776]: I1125 09:44:33.259697 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b62eda75-27dc-4a93-9ef5-a729483eb7a0","Type":"ContainerStarted","Data":"26c5b8e121542f7c1cf527f237c648db13a0b1275674c0685082bcf0225acb5e"} Nov 25 09:44:33 crc kubenswrapper[4776]: I1125 09:44:33.259727 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b62eda75-27dc-4a93-9ef5-a729483eb7a0","Type":"ContainerStarted","Data":"291ad566091d112b62fe0169f01df0accd1db826a1ec99183d4c8a28556aa0c4"} Nov 25 09:44:34 crc kubenswrapper[4776]: I1125 09:44:34.635718 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-bdhw6" Nov 25 09:44:34 crc kubenswrapper[4776]: I1125 09:44:34.765935 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5cf5\" (UniqueName: \"kubernetes.io/projected/3ff2a086-8bce-46e2-b294-3b3f31e39b4e-kube-api-access-b5cf5\") pod \"3ff2a086-8bce-46e2-b294-3b3f31e39b4e\" (UID: \"3ff2a086-8bce-46e2-b294-3b3f31e39b4e\") " Nov 25 09:44:34 crc kubenswrapper[4776]: I1125 09:44:34.766015 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3ff2a086-8bce-46e2-b294-3b3f31e39b4e-db-sync-config-data\") pod \"3ff2a086-8bce-46e2-b294-3b3f31e39b4e\" (UID: \"3ff2a086-8bce-46e2-b294-3b3f31e39b4e\") " Nov 25 09:44:34 crc kubenswrapper[4776]: I1125 09:44:34.766186 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ff2a086-8bce-46e2-b294-3b3f31e39b4e-combined-ca-bundle\") pod \"3ff2a086-8bce-46e2-b294-3b3f31e39b4e\" (UID: \"3ff2a086-8bce-46e2-b294-3b3f31e39b4e\") " Nov 25 09:44:34 crc kubenswrapper[4776]: I1125 09:44:34.772925 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ff2a086-8bce-46e2-b294-3b3f31e39b4e-kube-api-access-b5cf5" (OuterVolumeSpecName: "kube-api-access-b5cf5") pod "3ff2a086-8bce-46e2-b294-3b3f31e39b4e" (UID: "3ff2a086-8bce-46e2-b294-3b3f31e39b4e"). InnerVolumeSpecName "kube-api-access-b5cf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:44:34 crc kubenswrapper[4776]: I1125 09:44:34.778029 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ff2a086-8bce-46e2-b294-3b3f31e39b4e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "3ff2a086-8bce-46e2-b294-3b3f31e39b4e" (UID: "3ff2a086-8bce-46e2-b294-3b3f31e39b4e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:34 crc kubenswrapper[4776]: I1125 09:44:34.806155 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ff2a086-8bce-46e2-b294-3b3f31e39b4e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ff2a086-8bce-46e2-b294-3b3f31e39b4e" (UID: "3ff2a086-8bce-46e2-b294-3b3f31e39b4e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:34 crc kubenswrapper[4776]: I1125 09:44:34.869680 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5cf5\" (UniqueName: \"kubernetes.io/projected/3ff2a086-8bce-46e2-b294-3b3f31e39b4e-kube-api-access-b5cf5\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:34 crc kubenswrapper[4776]: I1125 09:44:34.869726 4776 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3ff2a086-8bce-46e2-b294-3b3f31e39b4e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:34 crc kubenswrapper[4776]: I1125 09:44:34.869736 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ff2a086-8bce-46e2-b294-3b3f31e39b4e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.275408 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-bdhw6" event={"ID":"3ff2a086-8bce-46e2-b294-3b3f31e39b4e","Type":"ContainerDied","Data":"fe1ef695256ceeb1bb0ed94ba1c35484ff5f94b4ee7e90cb3f5b249ac69580e5"} Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.275692 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe1ef695256ceeb1bb0ed94ba1c35484ff5f94b4ee7e90cb3f5b249ac69580e5" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.275751 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-bdhw6" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.285885 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b62eda75-27dc-4a93-9ef5-a729483eb7a0","Type":"ContainerStarted","Data":"692f24c910457e9a89a0fb2f8ef6c94883a979330e90c6cbd75277df199ae233"} Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.562363 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-66cdf565f-6lzqw"] Nov 25 09:44:35 crc kubenswrapper[4776]: E1125 09:44:35.562703 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ff2a086-8bce-46e2-b294-3b3f31e39b4e" containerName="barbican-db-sync" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.562718 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ff2a086-8bce-46e2-b294-3b3f31e39b4e" containerName="barbican-db-sync" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.562883 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ff2a086-8bce-46e2-b294-3b3f31e39b4e" containerName="barbican-db-sync" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.563713 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-66cdf565f-6lzqw" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.570535 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.570703 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-c8v8c" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.570812 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.601617 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-66cdf565f-6lzqw"] Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.621405 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-56988fbb4-nlxs9"] Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.622907 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-56988fbb4-nlxs9" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.636726 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.691136 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-config-data\") pod \"barbican-worker-66cdf565f-6lzqw\" (UID: \"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f\") " pod="openstack/barbican-worker-66cdf565f-6lzqw" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.691512 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-config-data-custom\") pod \"barbican-worker-66cdf565f-6lzqw\" (UID: \"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f\") " pod="openstack/barbican-worker-66cdf565f-6lzqw" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.691613 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zm22g\" (UniqueName: \"kubernetes.io/projected/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-kube-api-access-zm22g\") pod \"barbican-keystone-listener-56988fbb4-nlxs9\" (UID: \"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45\") " pod="openstack/barbican-keystone-listener-56988fbb4-nlxs9" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.691647 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6b2d\" (UniqueName: \"kubernetes.io/projected/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-kube-api-access-w6b2d\") pod \"barbican-worker-66cdf565f-6lzqw\" (UID: \"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f\") " pod="openstack/barbican-worker-66cdf565f-6lzqw" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.691676 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-logs\") pod \"barbican-worker-66cdf565f-6lzqw\" (UID: \"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f\") " pod="openstack/barbican-worker-66cdf565f-6lzqw" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.691700 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-logs\") pod \"barbican-keystone-listener-56988fbb4-nlxs9\" (UID: \"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45\") " pod="openstack/barbican-keystone-listener-56988fbb4-nlxs9" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.691779 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-combined-ca-bundle\") pod \"barbican-worker-66cdf565f-6lzqw\" (UID: \"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f\") " pod="openstack/barbican-worker-66cdf565f-6lzqw" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.691812 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-combined-ca-bundle\") pod \"barbican-keystone-listener-56988fbb4-nlxs9\" (UID: \"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45\") " pod="openstack/barbican-keystone-listener-56988fbb4-nlxs9" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.691872 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-config-data\") pod \"barbican-keystone-listener-56988fbb4-nlxs9\" (UID: \"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45\") " pod="openstack/barbican-keystone-listener-56988fbb4-nlxs9" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.691959 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-config-data-custom\") pod \"barbican-keystone-listener-56988fbb4-nlxs9\" (UID: \"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45\") " pod="openstack/barbican-keystone-listener-56988fbb4-nlxs9" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.716373 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-56988fbb4-nlxs9"] Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.812011 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-combined-ca-bundle\") pod \"barbican-worker-66cdf565f-6lzqw\" (UID: \"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f\") " pod="openstack/barbican-worker-66cdf565f-6lzqw" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.812056 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-combined-ca-bundle\") pod \"barbican-keystone-listener-56988fbb4-nlxs9\" (UID: \"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45\") " pod="openstack/barbican-keystone-listener-56988fbb4-nlxs9" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.812093 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-config-data\") pod \"barbican-keystone-listener-56988fbb4-nlxs9\" (UID: \"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45\") " pod="openstack/barbican-keystone-listener-56988fbb4-nlxs9" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.812205 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-config-data-custom\") pod \"barbican-keystone-listener-56988fbb4-nlxs9\" (UID: \"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45\") " pod="openstack/barbican-keystone-listener-56988fbb4-nlxs9" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.812252 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-config-data\") pod \"barbican-worker-66cdf565f-6lzqw\" (UID: \"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f\") " pod="openstack/barbican-worker-66cdf565f-6lzqw" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.812288 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-config-data-custom\") pod \"barbican-worker-66cdf565f-6lzqw\" (UID: \"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f\") " pod="openstack/barbican-worker-66cdf565f-6lzqw" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.812371 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zm22g\" (UniqueName: \"kubernetes.io/projected/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-kube-api-access-zm22g\") pod \"barbican-keystone-listener-56988fbb4-nlxs9\" (UID: \"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45\") " pod="openstack/barbican-keystone-listener-56988fbb4-nlxs9" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.812396 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6b2d\" (UniqueName: \"kubernetes.io/projected/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-kube-api-access-w6b2d\") pod \"barbican-worker-66cdf565f-6lzqw\" (UID: \"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f\") " pod="openstack/barbican-worker-66cdf565f-6lzqw" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.812416 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-logs\") pod \"barbican-worker-66cdf565f-6lzqw\" (UID: \"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f\") " pod="openstack/barbican-worker-66cdf565f-6lzqw" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.812431 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-logs\") pod \"barbican-keystone-listener-56988fbb4-nlxs9\" (UID: \"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45\") " pod="openstack/barbican-keystone-listener-56988fbb4-nlxs9" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.812882 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-logs\") pod \"barbican-keystone-listener-56988fbb4-nlxs9\" (UID: \"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45\") " pod="openstack/barbican-keystone-listener-56988fbb4-nlxs9" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.826693 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-logs\") pod \"barbican-worker-66cdf565f-6lzqw\" (UID: \"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f\") " pod="openstack/barbican-worker-66cdf565f-6lzqw" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.863809 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6b2d\" (UniqueName: \"kubernetes.io/projected/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-kube-api-access-w6b2d\") pod \"barbican-worker-66cdf565f-6lzqw\" (UID: \"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f\") " pod="openstack/barbican-worker-66cdf565f-6lzqw" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.882745 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-config-data\") pod \"barbican-worker-66cdf565f-6lzqw\" (UID: \"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f\") " pod="openstack/barbican-worker-66cdf565f-6lzqw" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.883527 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-combined-ca-bundle\") pod \"barbican-keystone-listener-56988fbb4-nlxs9\" (UID: \"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45\") " pod="openstack/barbican-keystone-listener-56988fbb4-nlxs9" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.883653 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-config-data-custom\") pod \"barbican-worker-66cdf565f-6lzqw\" (UID: \"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f\") " pod="openstack/barbican-worker-66cdf565f-6lzqw" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.899504 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c654c9745-whq24"] Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.899941 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6c654c9745-whq24" podUID="3d2378da-fcb1-4ca6-8906-ae48da346e95" containerName="dnsmasq-dns" containerID="cri-o://fdf727eac872d98e72351108b80191b34fb3b7d5ebcee63350880bfe42d78f37" gracePeriod=10 Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.901719 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-config-data-custom\") pod \"barbican-keystone-listener-56988fbb4-nlxs9\" (UID: \"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45\") " pod="openstack/barbican-keystone-listener-56988fbb4-nlxs9" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.904049 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-config-data\") pod \"barbican-keystone-listener-56988fbb4-nlxs9\" (UID: \"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45\") " pod="openstack/barbican-keystone-listener-56988fbb4-nlxs9" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.912308 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6c654c9745-whq24" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.913786 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zm22g\" (UniqueName: \"kubernetes.io/projected/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-kube-api-access-zm22g\") pod \"barbican-keystone-listener-56988fbb4-nlxs9\" (UID: \"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45\") " pod="openstack/barbican-keystone-listener-56988fbb4-nlxs9" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.914339 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-combined-ca-bundle\") pod \"barbican-worker-66cdf565f-6lzqw\" (UID: \"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f\") " pod="openstack/barbican-worker-66cdf565f-6lzqw" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.946136 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5cc67f459c-7m6ns"] Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.952726 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" Nov 25 09:44:35 crc kubenswrapper[4776]: I1125 09:44:35.980896 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cc67f459c-7m6ns"] Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.026540 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-66cdf565f-6lzqw" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.027040 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-ovsdbserver-sb\") pod \"dnsmasq-dns-5cc67f459c-7m6ns\" (UID: \"616a7df2-c899-43d8-9217-a1cf49ad8add\") " pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.027195 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-ovsdbserver-nb\") pod \"dnsmasq-dns-5cc67f459c-7m6ns\" (UID: \"616a7df2-c899-43d8-9217-a1cf49ad8add\") " pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.027262 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-dns-swift-storage-0\") pod \"dnsmasq-dns-5cc67f459c-7m6ns\" (UID: \"616a7df2-c899-43d8-9217-a1cf49ad8add\") " pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.027302 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5k6l4\" (UniqueName: \"kubernetes.io/projected/616a7df2-c899-43d8-9217-a1cf49ad8add-kube-api-access-5k6l4\") pod \"dnsmasq-dns-5cc67f459c-7m6ns\" (UID: \"616a7df2-c899-43d8-9217-a1cf49ad8add\") " pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.027351 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-dns-svc\") pod \"dnsmasq-dns-5cc67f459c-7m6ns\" (UID: \"616a7df2-c899-43d8-9217-a1cf49ad8add\") " pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.027424 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-config\") pod \"dnsmasq-dns-5cc67f459c-7m6ns\" (UID: \"616a7df2-c899-43d8-9217-a1cf49ad8add\") " pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.018055 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6b8747677b-bmrqm"] Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.030949 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6b8747677b-bmrqm" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.054764 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6b8747677b-bmrqm"] Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.061000 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.061392 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-56988fbb4-nlxs9" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.137453 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-config\") pod \"dnsmasq-dns-5cc67f459c-7m6ns\" (UID: \"616a7df2-c899-43d8-9217-a1cf49ad8add\") " pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.137976 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3dd1ad2-32fd-4c42-bbaf-e42269987973-logs\") pod \"barbican-api-6b8747677b-bmrqm\" (UID: \"e3dd1ad2-32fd-4c42-bbaf-e42269987973\") " pod="openstack/barbican-api-6b8747677b-bmrqm" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.138021 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3dd1ad2-32fd-4c42-bbaf-e42269987973-combined-ca-bundle\") pod \"barbican-api-6b8747677b-bmrqm\" (UID: \"e3dd1ad2-32fd-4c42-bbaf-e42269987973\") " pod="openstack/barbican-api-6b8747677b-bmrqm" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.138048 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72ktx\" (UniqueName: \"kubernetes.io/projected/e3dd1ad2-32fd-4c42-bbaf-e42269987973-kube-api-access-72ktx\") pod \"barbican-api-6b8747677b-bmrqm\" (UID: \"e3dd1ad2-32fd-4c42-bbaf-e42269987973\") " pod="openstack/barbican-api-6b8747677b-bmrqm" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.138175 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-ovsdbserver-sb\") pod \"dnsmasq-dns-5cc67f459c-7m6ns\" (UID: \"616a7df2-c899-43d8-9217-a1cf49ad8add\") " pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.138298 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3dd1ad2-32fd-4c42-bbaf-e42269987973-config-data\") pod \"barbican-api-6b8747677b-bmrqm\" (UID: \"e3dd1ad2-32fd-4c42-bbaf-e42269987973\") " pod="openstack/barbican-api-6b8747677b-bmrqm" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.138364 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-ovsdbserver-nb\") pod \"dnsmasq-dns-5cc67f459c-7m6ns\" (UID: \"616a7df2-c899-43d8-9217-a1cf49ad8add\") " pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.138408 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-dns-swift-storage-0\") pod \"dnsmasq-dns-5cc67f459c-7m6ns\" (UID: \"616a7df2-c899-43d8-9217-a1cf49ad8add\") " pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.138452 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5k6l4\" (UniqueName: \"kubernetes.io/projected/616a7df2-c899-43d8-9217-a1cf49ad8add-kube-api-access-5k6l4\") pod \"dnsmasq-dns-5cc67f459c-7m6ns\" (UID: \"616a7df2-c899-43d8-9217-a1cf49ad8add\") " pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.138478 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-dns-svc\") pod \"dnsmasq-dns-5cc67f459c-7m6ns\" (UID: \"616a7df2-c899-43d8-9217-a1cf49ad8add\") " pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.138539 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3dd1ad2-32fd-4c42-bbaf-e42269987973-config-data-custom\") pod \"barbican-api-6b8747677b-bmrqm\" (UID: \"e3dd1ad2-32fd-4c42-bbaf-e42269987973\") " pod="openstack/barbican-api-6b8747677b-bmrqm" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.138556 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-config\") pod \"dnsmasq-dns-5cc67f459c-7m6ns\" (UID: \"616a7df2-c899-43d8-9217-a1cf49ad8add\") " pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.139144 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-dns-swift-storage-0\") pod \"dnsmasq-dns-5cc67f459c-7m6ns\" (UID: \"616a7df2-c899-43d8-9217-a1cf49ad8add\") " pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.139597 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-ovsdbserver-nb\") pod \"dnsmasq-dns-5cc67f459c-7m6ns\" (UID: \"616a7df2-c899-43d8-9217-a1cf49ad8add\") " pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.154048 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-dns-svc\") pod \"dnsmasq-dns-5cc67f459c-7m6ns\" (UID: \"616a7df2-c899-43d8-9217-a1cf49ad8add\") " pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.154447 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-ovsdbserver-sb\") pod \"dnsmasq-dns-5cc67f459c-7m6ns\" (UID: \"616a7df2-c899-43d8-9217-a1cf49ad8add\") " pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.165864 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5k6l4\" (UniqueName: \"kubernetes.io/projected/616a7df2-c899-43d8-9217-a1cf49ad8add-kube-api-access-5k6l4\") pod \"dnsmasq-dns-5cc67f459c-7m6ns\" (UID: \"616a7df2-c899-43d8-9217-a1cf49ad8add\") " pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.240664 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3dd1ad2-32fd-4c42-bbaf-e42269987973-logs\") pod \"barbican-api-6b8747677b-bmrqm\" (UID: \"e3dd1ad2-32fd-4c42-bbaf-e42269987973\") " pod="openstack/barbican-api-6b8747677b-bmrqm" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.240902 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3dd1ad2-32fd-4c42-bbaf-e42269987973-combined-ca-bundle\") pod \"barbican-api-6b8747677b-bmrqm\" (UID: \"e3dd1ad2-32fd-4c42-bbaf-e42269987973\") " pod="openstack/barbican-api-6b8747677b-bmrqm" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.240927 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72ktx\" (UniqueName: \"kubernetes.io/projected/e3dd1ad2-32fd-4c42-bbaf-e42269987973-kube-api-access-72ktx\") pod \"barbican-api-6b8747677b-bmrqm\" (UID: \"e3dd1ad2-32fd-4c42-bbaf-e42269987973\") " pod="openstack/barbican-api-6b8747677b-bmrqm" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.241031 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3dd1ad2-32fd-4c42-bbaf-e42269987973-config-data\") pod \"barbican-api-6b8747677b-bmrqm\" (UID: \"e3dd1ad2-32fd-4c42-bbaf-e42269987973\") " pod="openstack/barbican-api-6b8747677b-bmrqm" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.241125 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3dd1ad2-32fd-4c42-bbaf-e42269987973-config-data-custom\") pod \"barbican-api-6b8747677b-bmrqm\" (UID: \"e3dd1ad2-32fd-4c42-bbaf-e42269987973\") " pod="openstack/barbican-api-6b8747677b-bmrqm" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.299036 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3dd1ad2-32fd-4c42-bbaf-e42269987973-logs\") pod \"barbican-api-6b8747677b-bmrqm\" (UID: \"e3dd1ad2-32fd-4c42-bbaf-e42269987973\") " pod="openstack/barbican-api-6b8747677b-bmrqm" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.301593 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3dd1ad2-32fd-4c42-bbaf-e42269987973-config-data-custom\") pod \"barbican-api-6b8747677b-bmrqm\" (UID: \"e3dd1ad2-32fd-4c42-bbaf-e42269987973\") " pod="openstack/barbican-api-6b8747677b-bmrqm" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.301612 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72ktx\" (UniqueName: \"kubernetes.io/projected/e3dd1ad2-32fd-4c42-bbaf-e42269987973-kube-api-access-72ktx\") pod \"barbican-api-6b8747677b-bmrqm\" (UID: \"e3dd1ad2-32fd-4c42-bbaf-e42269987973\") " pod="openstack/barbican-api-6b8747677b-bmrqm" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.304285 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3dd1ad2-32fd-4c42-bbaf-e42269987973-config-data\") pod \"barbican-api-6b8747677b-bmrqm\" (UID: \"e3dd1ad2-32fd-4c42-bbaf-e42269987973\") " pod="openstack/barbican-api-6b8747677b-bmrqm" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.313474 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3dd1ad2-32fd-4c42-bbaf-e42269987973-combined-ca-bundle\") pod \"barbican-api-6b8747677b-bmrqm\" (UID: \"e3dd1ad2-32fd-4c42-bbaf-e42269987973\") " pod="openstack/barbican-api-6b8747677b-bmrqm" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.315690 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.418519 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6b8747677b-bmrqm" Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.632159 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-66cdf565f-6lzqw"] Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.712152 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cc67f459c-7m6ns"] Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.762171 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-56988fbb4-nlxs9"] Nov 25 09:44:36 crc kubenswrapper[4776]: I1125 09:44:36.825552 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6b8747677b-bmrqm"] Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.132055 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c654c9745-whq24" Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.276577 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-ovsdbserver-sb\") pod \"3d2378da-fcb1-4ca6-8906-ae48da346e95\" (UID: \"3d2378da-fcb1-4ca6-8906-ae48da346e95\") " Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.276781 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-dns-svc\") pod \"3d2378da-fcb1-4ca6-8906-ae48da346e95\" (UID: \"3d2378da-fcb1-4ca6-8906-ae48da346e95\") " Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.276818 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-ovsdbserver-nb\") pod \"3d2378da-fcb1-4ca6-8906-ae48da346e95\" (UID: \"3d2378da-fcb1-4ca6-8906-ae48da346e95\") " Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.276850 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-dns-swift-storage-0\") pod \"3d2378da-fcb1-4ca6-8906-ae48da346e95\" (UID: \"3d2378da-fcb1-4ca6-8906-ae48da346e95\") " Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.276881 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-config\") pod \"3d2378da-fcb1-4ca6-8906-ae48da346e95\" (UID: \"3d2378da-fcb1-4ca6-8906-ae48da346e95\") " Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.276950 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ls52k\" (UniqueName: \"kubernetes.io/projected/3d2378da-fcb1-4ca6-8906-ae48da346e95-kube-api-access-ls52k\") pod \"3d2378da-fcb1-4ca6-8906-ae48da346e95\" (UID: \"3d2378da-fcb1-4ca6-8906-ae48da346e95\") " Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.282315 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d2378da-fcb1-4ca6-8906-ae48da346e95-kube-api-access-ls52k" (OuterVolumeSpecName: "kube-api-access-ls52k") pod "3d2378da-fcb1-4ca6-8906-ae48da346e95" (UID: "3d2378da-fcb1-4ca6-8906-ae48da346e95"). InnerVolumeSpecName "kube-api-access-ls52k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.322413 4776 generic.go:334] "Generic (PLEG): container finished" podID="616a7df2-c899-43d8-9217-a1cf49ad8add" containerID="7d85333bbe9017116a2c05c0c6b8ef47fd7bc62e15ff3825cc0cec533f34ab74" exitCode=0 Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.322537 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" event={"ID":"616a7df2-c899-43d8-9217-a1cf49ad8add","Type":"ContainerDied","Data":"7d85333bbe9017116a2c05c0c6b8ef47fd7bc62e15ff3825cc0cec533f34ab74"} Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.322572 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" event={"ID":"616a7df2-c899-43d8-9217-a1cf49ad8add","Type":"ContainerStarted","Data":"bdb568b81ceb846fbe30935ec55b9d437a6593da8864783d4feddc24ea29731e"} Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.325894 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-66cdf565f-6lzqw" event={"ID":"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f","Type":"ContainerStarted","Data":"b690b96a14b9b4d3e66ad1db4ce9b8289a92134e04d4486aa5aa883c1148be9b"} Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.333180 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-56988fbb4-nlxs9" event={"ID":"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45","Type":"ContainerStarted","Data":"ee48efd287e21fe11c445356b7f2fed53754f03152b8e09e0cec9fc4b8e05b3c"} Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.337183 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b8747677b-bmrqm" event={"ID":"e3dd1ad2-32fd-4c42-bbaf-e42269987973","Type":"ContainerStarted","Data":"b4ff1ba01bcbbe2c5b08eb7d71d3b56219363949b9c1aab72bd432ffca17d4c1"} Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.337230 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b8747677b-bmrqm" event={"ID":"e3dd1ad2-32fd-4c42-bbaf-e42269987973","Type":"ContainerStarted","Data":"1e8b9f104b52295070ebd96b4f5451896dce3430f9c1303d99e15022d1fbf0b8"} Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.341658 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-config" (OuterVolumeSpecName: "config") pod "3d2378da-fcb1-4ca6-8906-ae48da346e95" (UID: "3d2378da-fcb1-4ca6-8906-ae48da346e95"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.345402 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3d2378da-fcb1-4ca6-8906-ae48da346e95" (UID: "3d2378da-fcb1-4ca6-8906-ae48da346e95"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.346834 4776 generic.go:334] "Generic (PLEG): container finished" podID="3d2378da-fcb1-4ca6-8906-ae48da346e95" containerID="fdf727eac872d98e72351108b80191b34fb3b7d5ebcee63350880bfe42d78f37" exitCode=0 Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.346888 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c654c9745-whq24" event={"ID":"3d2378da-fcb1-4ca6-8906-ae48da346e95","Type":"ContainerDied","Data":"fdf727eac872d98e72351108b80191b34fb3b7d5ebcee63350880bfe42d78f37"} Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.346913 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c654c9745-whq24" event={"ID":"3d2378da-fcb1-4ca6-8906-ae48da346e95","Type":"ContainerDied","Data":"c4756ea9775ac58ef3226a2d6d0d3612568a801c8743e71c78aa176436223403"} Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.346928 4776 scope.go:117] "RemoveContainer" containerID="fdf727eac872d98e72351108b80191b34fb3b7d5ebcee63350880bfe42d78f37" Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.347035 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c654c9745-whq24" Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.354149 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b62eda75-27dc-4a93-9ef5-a729483eb7a0","Type":"ContainerStarted","Data":"a4e627a7fd3e589e8546a2863dea9146033d4da7accb56f9b48af3afc1697807"} Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.354785 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3d2378da-fcb1-4ca6-8906-ae48da346e95" (UID: "3d2378da-fcb1-4ca6-8906-ae48da346e95"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.355340 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3d2378da-fcb1-4ca6-8906-ae48da346e95" (UID: "3d2378da-fcb1-4ca6-8906-ae48da346e95"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.358139 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3d2378da-fcb1-4ca6-8906-ae48da346e95" (UID: "3d2378da-fcb1-4ca6-8906-ae48da346e95"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.378607 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ls52k\" (UniqueName: \"kubernetes.io/projected/3d2378da-fcb1-4ca6-8906-ae48da346e95-kube-api-access-ls52k\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.378677 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.378692 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.378705 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.378718 4776 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.378763 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d2378da-fcb1-4ca6-8906-ae48da346e95-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.385428 4776 scope.go:117] "RemoveContainer" containerID="98b75f64d71dad3aa5ba54b90468297f0304b42e7e1c65945da0c8d495925174" Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.409604 4776 scope.go:117] "RemoveContainer" containerID="fdf727eac872d98e72351108b80191b34fb3b7d5ebcee63350880bfe42d78f37" Nov 25 09:44:37 crc kubenswrapper[4776]: E1125 09:44:37.410321 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdf727eac872d98e72351108b80191b34fb3b7d5ebcee63350880bfe42d78f37\": container with ID starting with fdf727eac872d98e72351108b80191b34fb3b7d5ebcee63350880bfe42d78f37 not found: ID does not exist" containerID="fdf727eac872d98e72351108b80191b34fb3b7d5ebcee63350880bfe42d78f37" Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.410385 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdf727eac872d98e72351108b80191b34fb3b7d5ebcee63350880bfe42d78f37"} err="failed to get container status \"fdf727eac872d98e72351108b80191b34fb3b7d5ebcee63350880bfe42d78f37\": rpc error: code = NotFound desc = could not find container \"fdf727eac872d98e72351108b80191b34fb3b7d5ebcee63350880bfe42d78f37\": container with ID starting with fdf727eac872d98e72351108b80191b34fb3b7d5ebcee63350880bfe42d78f37 not found: ID does not exist" Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.410416 4776 scope.go:117] "RemoveContainer" containerID="98b75f64d71dad3aa5ba54b90468297f0304b42e7e1c65945da0c8d495925174" Nov 25 09:44:37 crc kubenswrapper[4776]: E1125 09:44:37.410765 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98b75f64d71dad3aa5ba54b90468297f0304b42e7e1c65945da0c8d495925174\": container with ID starting with 98b75f64d71dad3aa5ba54b90468297f0304b42e7e1c65945da0c8d495925174 not found: ID does not exist" containerID="98b75f64d71dad3aa5ba54b90468297f0304b42e7e1c65945da0c8d495925174" Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.410806 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98b75f64d71dad3aa5ba54b90468297f0304b42e7e1c65945da0c8d495925174"} err="failed to get container status \"98b75f64d71dad3aa5ba54b90468297f0304b42e7e1c65945da0c8d495925174\": rpc error: code = NotFound desc = could not find container \"98b75f64d71dad3aa5ba54b90468297f0304b42e7e1c65945da0c8d495925174\": container with ID starting with 98b75f64d71dad3aa5ba54b90468297f0304b42e7e1c65945da0c8d495925174 not found: ID does not exist" Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.711854 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c654c9745-whq24"] Nov 25 09:44:37 crc kubenswrapper[4776]: I1125 09:44:37.734096 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c654c9745-whq24"] Nov 25 09:44:38 crc kubenswrapper[4776]: I1125 09:44:38.391313 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b62eda75-27dc-4a93-9ef5-a729483eb7a0","Type":"ContainerStarted","Data":"0201f2a461319eda4465429bd9c908e2a8642d9da6bc14ed48d25627d89f3b29"} Nov 25 09:44:38 crc kubenswrapper[4776]: I1125 09:44:38.392728 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 09:44:38 crc kubenswrapper[4776]: I1125 09:44:38.396604 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" event={"ID":"616a7df2-c899-43d8-9217-a1cf49ad8add","Type":"ContainerStarted","Data":"83c6af89df781455182ffdd3eb5e6c0222b267c62aacd11088a1acfa23cf26d6"} Nov 25 09:44:38 crc kubenswrapper[4776]: I1125 09:44:38.397319 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" Nov 25 09:44:38 crc kubenswrapper[4776]: I1125 09:44:38.400423 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b8747677b-bmrqm" event={"ID":"e3dd1ad2-32fd-4c42-bbaf-e42269987973","Type":"ContainerStarted","Data":"40c228ae0f5503fa2aa9af33eaa8e948a2b937a4476a096a469b2ce4731b8ac6"} Nov 25 09:44:38 crc kubenswrapper[4776]: I1125 09:44:38.400924 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6b8747677b-bmrqm" Nov 25 09:44:38 crc kubenswrapper[4776]: I1125 09:44:38.400950 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6b8747677b-bmrqm" Nov 25 09:44:38 crc kubenswrapper[4776]: I1125 09:44:38.413569 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.769260347 podStartE2EDuration="7.413551535s" podCreationTimestamp="2025-11-25 09:44:31 +0000 UTC" firstStartedPulling="2025-11-25 09:44:32.268301963 +0000 UTC m=+1217.309361516" lastFinishedPulling="2025-11-25 09:44:37.912593151 +0000 UTC m=+1222.953652704" observedRunningTime="2025-11-25 09:44:38.41335058 +0000 UTC m=+1223.454410133" watchObservedRunningTime="2025-11-25 09:44:38.413551535 +0000 UTC m=+1223.454611088" Nov 25 09:44:38 crc kubenswrapper[4776]: I1125 09:44:38.432899 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6b8747677b-bmrqm" podStartSLOduration=3.4328773200000002 podStartE2EDuration="3.43287732s" podCreationTimestamp="2025-11-25 09:44:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:44:38.429539786 +0000 UTC m=+1223.470599349" watchObservedRunningTime="2025-11-25 09:44:38.43287732 +0000 UTC m=+1223.473936873" Nov 25 09:44:38 crc kubenswrapper[4776]: I1125 09:44:38.458933 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" podStartSLOduration=3.458908553 podStartE2EDuration="3.458908553s" podCreationTimestamp="2025-11-25 09:44:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:44:38.455632601 +0000 UTC m=+1223.496692174" watchObservedRunningTime="2025-11-25 09:44:38.458908553 +0000 UTC m=+1223.499968106" Nov 25 09:44:39 crc kubenswrapper[4776]: I1125 09:44:39.411310 4776 generic.go:334] "Generic (PLEG): container finished" podID="e522431f-0956-49f7-9c46-b0202824bd56" containerID="5cb968840c90f1277c3bd3ba5789c34ed207afcd5657d28a1ee8b953a59bcbe3" exitCode=0 Nov 25 09:44:39 crc kubenswrapper[4776]: I1125 09:44:39.411663 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-djwk2" event={"ID":"e522431f-0956-49f7-9c46-b0202824bd56","Type":"ContainerDied","Data":"5cb968840c90f1277c3bd3ba5789c34ed207afcd5657d28a1ee8b953a59bcbe3"} Nov 25 09:44:39 crc kubenswrapper[4776]: I1125 09:44:39.673129 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d2378da-fcb1-4ca6-8906-ae48da346e95" path="/var/lib/kubelet/pods/3d2378da-fcb1-4ca6-8906-ae48da346e95/volumes" Nov 25 09:44:39 crc kubenswrapper[4776]: I1125 09:44:39.990902 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-769ddf488d-46ssj"] Nov 25 09:44:39 crc kubenswrapper[4776]: E1125 09:44:39.991300 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d2378da-fcb1-4ca6-8906-ae48da346e95" containerName="dnsmasq-dns" Nov 25 09:44:39 crc kubenswrapper[4776]: I1125 09:44:39.991317 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d2378da-fcb1-4ca6-8906-ae48da346e95" containerName="dnsmasq-dns" Nov 25 09:44:39 crc kubenswrapper[4776]: E1125 09:44:39.991336 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d2378da-fcb1-4ca6-8906-ae48da346e95" containerName="init" Nov 25 09:44:39 crc kubenswrapper[4776]: I1125 09:44:39.991342 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d2378da-fcb1-4ca6-8906-ae48da346e95" containerName="init" Nov 25 09:44:39 crc kubenswrapper[4776]: I1125 09:44:39.991526 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d2378da-fcb1-4ca6-8906-ae48da346e95" containerName="dnsmasq-dns" Nov 25 09:44:39 crc kubenswrapper[4776]: I1125 09:44:39.992489 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:44:39 crc kubenswrapper[4776]: I1125 09:44:39.996905 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 25 09:44:39 crc kubenswrapper[4776]: I1125 09:44:39.997731 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 25 09:44:40 crc kubenswrapper[4776]: I1125 09:44:40.004804 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-769ddf488d-46ssj"] Nov 25 09:44:40 crc kubenswrapper[4776]: I1125 09:44:40.135377 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-internal-tls-certs\") pod \"barbican-api-769ddf488d-46ssj\" (UID: \"4702d881-d701-41a4-b36f-2d063a8ae246\") " pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:44:40 crc kubenswrapper[4776]: I1125 09:44:40.135439 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4702d881-d701-41a4-b36f-2d063a8ae246-logs\") pod \"barbican-api-769ddf488d-46ssj\" (UID: \"4702d881-d701-41a4-b36f-2d063a8ae246\") " pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:44:40 crc kubenswrapper[4776]: I1125 09:44:40.135461 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldgn5\" (UniqueName: \"kubernetes.io/projected/4702d881-d701-41a4-b36f-2d063a8ae246-kube-api-access-ldgn5\") pod \"barbican-api-769ddf488d-46ssj\" (UID: \"4702d881-d701-41a4-b36f-2d063a8ae246\") " pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:44:40 crc kubenswrapper[4776]: I1125 09:44:40.135484 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-config-data\") pod \"barbican-api-769ddf488d-46ssj\" (UID: \"4702d881-d701-41a4-b36f-2d063a8ae246\") " pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:44:40 crc kubenswrapper[4776]: I1125 09:44:40.135523 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-config-data-custom\") pod \"barbican-api-769ddf488d-46ssj\" (UID: \"4702d881-d701-41a4-b36f-2d063a8ae246\") " pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:44:40 crc kubenswrapper[4776]: I1125 09:44:40.135568 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-public-tls-certs\") pod \"barbican-api-769ddf488d-46ssj\" (UID: \"4702d881-d701-41a4-b36f-2d063a8ae246\") " pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:44:40 crc kubenswrapper[4776]: I1125 09:44:40.135631 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-combined-ca-bundle\") pod \"barbican-api-769ddf488d-46ssj\" (UID: \"4702d881-d701-41a4-b36f-2d063a8ae246\") " pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:44:40 crc kubenswrapper[4776]: I1125 09:44:40.237271 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-combined-ca-bundle\") pod \"barbican-api-769ddf488d-46ssj\" (UID: \"4702d881-d701-41a4-b36f-2d063a8ae246\") " pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:44:40 crc kubenswrapper[4776]: I1125 09:44:40.237571 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-internal-tls-certs\") pod \"barbican-api-769ddf488d-46ssj\" (UID: \"4702d881-d701-41a4-b36f-2d063a8ae246\") " pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:44:40 crc kubenswrapper[4776]: I1125 09:44:40.237675 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4702d881-d701-41a4-b36f-2d063a8ae246-logs\") pod \"barbican-api-769ddf488d-46ssj\" (UID: \"4702d881-d701-41a4-b36f-2d063a8ae246\") " pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:44:40 crc kubenswrapper[4776]: I1125 09:44:40.237753 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldgn5\" (UniqueName: \"kubernetes.io/projected/4702d881-d701-41a4-b36f-2d063a8ae246-kube-api-access-ldgn5\") pod \"barbican-api-769ddf488d-46ssj\" (UID: \"4702d881-d701-41a4-b36f-2d063a8ae246\") " pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:44:40 crc kubenswrapper[4776]: I1125 09:44:40.237872 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-config-data\") pod \"barbican-api-769ddf488d-46ssj\" (UID: \"4702d881-d701-41a4-b36f-2d063a8ae246\") " pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:44:40 crc kubenswrapper[4776]: I1125 09:44:40.238492 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-config-data-custom\") pod \"barbican-api-769ddf488d-46ssj\" (UID: \"4702d881-d701-41a4-b36f-2d063a8ae246\") " pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:44:40 crc kubenswrapper[4776]: I1125 09:44:40.238647 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-public-tls-certs\") pod \"barbican-api-769ddf488d-46ssj\" (UID: \"4702d881-d701-41a4-b36f-2d063a8ae246\") " pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:44:40 crc kubenswrapper[4776]: I1125 09:44:40.238119 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4702d881-d701-41a4-b36f-2d063a8ae246-logs\") pod \"barbican-api-769ddf488d-46ssj\" (UID: \"4702d881-d701-41a4-b36f-2d063a8ae246\") " pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:44:40 crc kubenswrapper[4776]: I1125 09:44:40.243012 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-internal-tls-certs\") pod \"barbican-api-769ddf488d-46ssj\" (UID: \"4702d881-d701-41a4-b36f-2d063a8ae246\") " pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:44:40 crc kubenswrapper[4776]: I1125 09:44:40.243409 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-config-data-custom\") pod \"barbican-api-769ddf488d-46ssj\" (UID: \"4702d881-d701-41a4-b36f-2d063a8ae246\") " pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:44:40 crc kubenswrapper[4776]: I1125 09:44:40.245230 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-config-data\") pod \"barbican-api-769ddf488d-46ssj\" (UID: \"4702d881-d701-41a4-b36f-2d063a8ae246\") " pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:44:40 crc kubenswrapper[4776]: I1125 09:44:40.245675 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-combined-ca-bundle\") pod \"barbican-api-769ddf488d-46ssj\" (UID: \"4702d881-d701-41a4-b36f-2d063a8ae246\") " pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:44:40 crc kubenswrapper[4776]: I1125 09:44:40.249028 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-public-tls-certs\") pod \"barbican-api-769ddf488d-46ssj\" (UID: \"4702d881-d701-41a4-b36f-2d063a8ae246\") " pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:44:40 crc kubenswrapper[4776]: I1125 09:44:40.259840 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldgn5\" (UniqueName: \"kubernetes.io/projected/4702d881-d701-41a4-b36f-2d063a8ae246-kube-api-access-ldgn5\") pod \"barbican-api-769ddf488d-46ssj\" (UID: \"4702d881-d701-41a4-b36f-2d063a8ae246\") " pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:44:40 crc kubenswrapper[4776]: I1125 09:44:40.309829 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.367812 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-djwk2" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.467547 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e522431f-0956-49f7-9c46-b0202824bd56-db-sync-config-data\") pod \"e522431f-0956-49f7-9c46-b0202824bd56\" (UID: \"e522431f-0956-49f7-9c46-b0202824bd56\") " Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.467820 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e522431f-0956-49f7-9c46-b0202824bd56-combined-ca-bundle\") pod \"e522431f-0956-49f7-9c46-b0202824bd56\" (UID: \"e522431f-0956-49f7-9c46-b0202824bd56\") " Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.467992 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e522431f-0956-49f7-9c46-b0202824bd56-config-data\") pod \"e522431f-0956-49f7-9c46-b0202824bd56\" (UID: \"e522431f-0956-49f7-9c46-b0202824bd56\") " Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.468162 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e522431f-0956-49f7-9c46-b0202824bd56-scripts\") pod \"e522431f-0956-49f7-9c46-b0202824bd56\" (UID: \"e522431f-0956-49f7-9c46-b0202824bd56\") " Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.468350 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rntzg\" (UniqueName: \"kubernetes.io/projected/e522431f-0956-49f7-9c46-b0202824bd56-kube-api-access-rntzg\") pod \"e522431f-0956-49f7-9c46-b0202824bd56\" (UID: \"e522431f-0956-49f7-9c46-b0202824bd56\") " Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.468468 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e522431f-0956-49f7-9c46-b0202824bd56-etc-machine-id\") pod \"e522431f-0956-49f7-9c46-b0202824bd56\" (UID: \"e522431f-0956-49f7-9c46-b0202824bd56\") " Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.469097 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e522431f-0956-49f7-9c46-b0202824bd56-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e522431f-0956-49f7-9c46-b0202824bd56" (UID: "e522431f-0956-49f7-9c46-b0202824bd56"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.489435 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e522431f-0956-49f7-9c46-b0202824bd56-scripts" (OuterVolumeSpecName: "scripts") pod "e522431f-0956-49f7-9c46-b0202824bd56" (UID: "e522431f-0956-49f7-9c46-b0202824bd56"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.489854 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e522431f-0956-49f7-9c46-b0202824bd56-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "e522431f-0956-49f7-9c46-b0202824bd56" (UID: "e522431f-0956-49f7-9c46-b0202824bd56"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.505458 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e522431f-0956-49f7-9c46-b0202824bd56-kube-api-access-rntzg" (OuterVolumeSpecName: "kube-api-access-rntzg") pod "e522431f-0956-49f7-9c46-b0202824bd56" (UID: "e522431f-0956-49f7-9c46-b0202824bd56"). InnerVolumeSpecName "kube-api-access-rntzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.507134 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.532558 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.571933 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rntzg\" (UniqueName: \"kubernetes.io/projected/e522431f-0956-49f7-9c46-b0202824bd56-kube-api-access-rntzg\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.571967 4776 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e522431f-0956-49f7-9c46-b0202824bd56-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.571981 4776 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e522431f-0956-49f7-9c46-b0202824bd56-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.571993 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e522431f-0956-49f7-9c46-b0202824bd56-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.574076 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e522431f-0956-49f7-9c46-b0202824bd56-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e522431f-0956-49f7-9c46-b0202824bd56" (UID: "e522431f-0956-49f7-9c46-b0202824bd56"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.574259 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-djwk2" event={"ID":"e522431f-0956-49f7-9c46-b0202824bd56","Type":"ContainerDied","Data":"622a6e97957a89d6ee65cf73b5cdcd0bbdad37b21925e1450350bdaafaf94d2d"} Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.574308 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="622a6e97957a89d6ee65cf73b5cdcd0bbdad37b21925e1450350bdaafaf94d2d" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.574346 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-djwk2" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.585259 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e522431f-0956-49f7-9c46-b0202824bd56-config-data" (OuterVolumeSpecName: "config-data") pod "e522431f-0956-49f7-9c46-b0202824bd56" (UID: "e522431f-0956-49f7-9c46-b0202824bd56"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.674370 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e522431f-0956-49f7-9c46-b0202824bd56-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.674867 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e522431f-0956-49f7-9c46-b0202824bd56-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.942172 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 09:44:41 crc kubenswrapper[4776]: E1125 09:44:41.942614 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e522431f-0956-49f7-9c46-b0202824bd56" containerName="cinder-db-sync" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.942633 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="e522431f-0956-49f7-9c46-b0202824bd56" containerName="cinder-db-sync" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.942787 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="e522431f-0956-49f7-9c46-b0202824bd56" containerName="cinder-db-sync" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.943681 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.955519 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.962534 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.962759 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-5lrgb" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.963029 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.991194 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\") " pod="openstack/cinder-scheduler-0" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.991232 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\") " pod="openstack/cinder-scheduler-0" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.991264 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\") " pod="openstack/cinder-scheduler-0" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.991292 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-scripts\") pod \"cinder-scheduler-0\" (UID: \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\") " pod="openstack/cinder-scheduler-0" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.991389 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-config-data\") pod \"cinder-scheduler-0\" (UID: \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\") " pod="openstack/cinder-scheduler-0" Nov 25 09:44:41 crc kubenswrapper[4776]: I1125 09:44:41.991474 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qrzw\" (UniqueName: \"kubernetes.io/projected/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-kube-api-access-2qrzw\") pod \"cinder-scheduler-0\" (UID: \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\") " pod="openstack/cinder-scheduler-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.009727 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.089160 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cc67f459c-7m6ns"] Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.089404 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" podUID="616a7df2-c899-43d8-9217-a1cf49ad8add" containerName="dnsmasq-dns" containerID="cri-o://83c6af89df781455182ffdd3eb5e6c0222b267c62aacd11088a1acfa23cf26d6" gracePeriod=10 Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.093802 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qrzw\" (UniqueName: \"kubernetes.io/projected/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-kube-api-access-2qrzw\") pod \"cinder-scheduler-0\" (UID: \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\") " pod="openstack/cinder-scheduler-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.093896 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\") " pod="openstack/cinder-scheduler-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.093920 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\") " pod="openstack/cinder-scheduler-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.093950 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\") " pod="openstack/cinder-scheduler-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.093979 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-scripts\") pod \"cinder-scheduler-0\" (UID: \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\") " pod="openstack/cinder-scheduler-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.094053 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-config-data\") pod \"cinder-scheduler-0\" (UID: \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\") " pod="openstack/cinder-scheduler-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.095026 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.096960 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\") " pod="openstack/cinder-scheduler-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.100618 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\") " pod="openstack/cinder-scheduler-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.104226 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\") " pod="openstack/cinder-scheduler-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.105358 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-scripts\") pod \"cinder-scheduler-0\" (UID: \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\") " pod="openstack/cinder-scheduler-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.118084 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-config-data\") pod \"cinder-scheduler-0\" (UID: \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\") " pod="openstack/cinder-scheduler-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.132660 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qrzw\" (UniqueName: \"kubernetes.io/projected/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-kube-api-access-2qrzw\") pod \"cinder-scheduler-0\" (UID: \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\") " pod="openstack/cinder-scheduler-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.137607 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-797bbc649-ncxwv"] Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.153282 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-797bbc649-ncxwv" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.196933 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-ovsdbserver-sb\") pod \"dnsmasq-dns-797bbc649-ncxwv\" (UID: \"ecc97771-b405-44c7-93e1-d24c86e705ed\") " pod="openstack/dnsmasq-dns-797bbc649-ncxwv" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.196995 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-dns-swift-storage-0\") pod \"dnsmasq-dns-797bbc649-ncxwv\" (UID: \"ecc97771-b405-44c7-93e1-d24c86e705ed\") " pod="openstack/dnsmasq-dns-797bbc649-ncxwv" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.197021 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vplwb\" (UniqueName: \"kubernetes.io/projected/ecc97771-b405-44c7-93e1-d24c86e705ed-kube-api-access-vplwb\") pod \"dnsmasq-dns-797bbc649-ncxwv\" (UID: \"ecc97771-b405-44c7-93e1-d24c86e705ed\") " pod="openstack/dnsmasq-dns-797bbc649-ncxwv" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.197079 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-config\") pod \"dnsmasq-dns-797bbc649-ncxwv\" (UID: \"ecc97771-b405-44c7-93e1-d24c86e705ed\") " pod="openstack/dnsmasq-dns-797bbc649-ncxwv" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.197100 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-dns-svc\") pod \"dnsmasq-dns-797bbc649-ncxwv\" (UID: \"ecc97771-b405-44c7-93e1-d24c86e705ed\") " pod="openstack/dnsmasq-dns-797bbc649-ncxwv" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.197288 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-ovsdbserver-nb\") pod \"dnsmasq-dns-797bbc649-ncxwv\" (UID: \"ecc97771-b405-44c7-93e1-d24c86e705ed\") " pod="openstack/dnsmasq-dns-797bbc649-ncxwv" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.199483 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-797bbc649-ncxwv"] Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.234884 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.247192 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.255809 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.266536 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.266607 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-769ddf488d-46ssj"] Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.338399 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.358590 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ce61c9c-20b2-462c-b187-519d56e75f56-config-data-custom\") pod \"cinder-api-0\" (UID: \"8ce61c9c-20b2-462c-b187-519d56e75f56\") " pod="openstack/cinder-api-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.358642 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8ce61c9c-20b2-462c-b187-519d56e75f56-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8ce61c9c-20b2-462c-b187-519d56e75f56\") " pod="openstack/cinder-api-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.358737 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-ovsdbserver-sb\") pod \"dnsmasq-dns-797bbc649-ncxwv\" (UID: \"ecc97771-b405-44c7-93e1-d24c86e705ed\") " pod="openstack/dnsmasq-dns-797bbc649-ncxwv" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.358789 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-dns-swift-storage-0\") pod \"dnsmasq-dns-797bbc649-ncxwv\" (UID: \"ecc97771-b405-44c7-93e1-d24c86e705ed\") " pod="openstack/dnsmasq-dns-797bbc649-ncxwv" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.358819 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h47bg\" (UniqueName: \"kubernetes.io/projected/8ce61c9c-20b2-462c-b187-519d56e75f56-kube-api-access-h47bg\") pod \"cinder-api-0\" (UID: \"8ce61c9c-20b2-462c-b187-519d56e75f56\") " pod="openstack/cinder-api-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.358844 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vplwb\" (UniqueName: \"kubernetes.io/projected/ecc97771-b405-44c7-93e1-d24c86e705ed-kube-api-access-vplwb\") pod \"dnsmasq-dns-797bbc649-ncxwv\" (UID: \"ecc97771-b405-44c7-93e1-d24c86e705ed\") " pod="openstack/dnsmasq-dns-797bbc649-ncxwv" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.358965 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-config\") pod \"dnsmasq-dns-797bbc649-ncxwv\" (UID: \"ecc97771-b405-44c7-93e1-d24c86e705ed\") " pod="openstack/dnsmasq-dns-797bbc649-ncxwv" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.358995 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ce61c9c-20b2-462c-b187-519d56e75f56-config-data\") pod \"cinder-api-0\" (UID: \"8ce61c9c-20b2-462c-b187-519d56e75f56\") " pod="openstack/cinder-api-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.359023 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-dns-svc\") pod \"dnsmasq-dns-797bbc649-ncxwv\" (UID: \"ecc97771-b405-44c7-93e1-d24c86e705ed\") " pod="openstack/dnsmasq-dns-797bbc649-ncxwv" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.359132 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ce61c9c-20b2-462c-b187-519d56e75f56-logs\") pod \"cinder-api-0\" (UID: \"8ce61c9c-20b2-462c-b187-519d56e75f56\") " pod="openstack/cinder-api-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.359172 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ce61c9c-20b2-462c-b187-519d56e75f56-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8ce61c9c-20b2-462c-b187-519d56e75f56\") " pod="openstack/cinder-api-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.359197 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-ovsdbserver-nb\") pod \"dnsmasq-dns-797bbc649-ncxwv\" (UID: \"ecc97771-b405-44c7-93e1-d24c86e705ed\") " pod="openstack/dnsmasq-dns-797bbc649-ncxwv" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.359224 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ce61c9c-20b2-462c-b187-519d56e75f56-scripts\") pod \"cinder-api-0\" (UID: \"8ce61c9c-20b2-462c-b187-519d56e75f56\") " pod="openstack/cinder-api-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.360466 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-ovsdbserver-sb\") pod \"dnsmasq-dns-797bbc649-ncxwv\" (UID: \"ecc97771-b405-44c7-93e1-d24c86e705ed\") " pod="openstack/dnsmasq-dns-797bbc649-ncxwv" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.361455 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-dns-swift-storage-0\") pod \"dnsmasq-dns-797bbc649-ncxwv\" (UID: \"ecc97771-b405-44c7-93e1-d24c86e705ed\") " pod="openstack/dnsmasq-dns-797bbc649-ncxwv" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.376843 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-dns-svc\") pod \"dnsmasq-dns-797bbc649-ncxwv\" (UID: \"ecc97771-b405-44c7-93e1-d24c86e705ed\") " pod="openstack/dnsmasq-dns-797bbc649-ncxwv" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.388306 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-config\") pod \"dnsmasq-dns-797bbc649-ncxwv\" (UID: \"ecc97771-b405-44c7-93e1-d24c86e705ed\") " pod="openstack/dnsmasq-dns-797bbc649-ncxwv" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.388721 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vplwb\" (UniqueName: \"kubernetes.io/projected/ecc97771-b405-44c7-93e1-d24c86e705ed-kube-api-access-vplwb\") pod \"dnsmasq-dns-797bbc649-ncxwv\" (UID: \"ecc97771-b405-44c7-93e1-d24c86e705ed\") " pod="openstack/dnsmasq-dns-797bbc649-ncxwv" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.390020 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-ovsdbserver-nb\") pod \"dnsmasq-dns-797bbc649-ncxwv\" (UID: \"ecc97771-b405-44c7-93e1-d24c86e705ed\") " pod="openstack/dnsmasq-dns-797bbc649-ncxwv" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.462324 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8ce61c9c-20b2-462c-b187-519d56e75f56-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8ce61c9c-20b2-462c-b187-519d56e75f56\") " pod="openstack/cinder-api-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.462367 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ce61c9c-20b2-462c-b187-519d56e75f56-config-data-custom\") pod \"cinder-api-0\" (UID: \"8ce61c9c-20b2-462c-b187-519d56e75f56\") " pod="openstack/cinder-api-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.462431 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h47bg\" (UniqueName: \"kubernetes.io/projected/8ce61c9c-20b2-462c-b187-519d56e75f56-kube-api-access-h47bg\") pod \"cinder-api-0\" (UID: \"8ce61c9c-20b2-462c-b187-519d56e75f56\") " pod="openstack/cinder-api-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.462486 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ce61c9c-20b2-462c-b187-519d56e75f56-config-data\") pod \"cinder-api-0\" (UID: \"8ce61c9c-20b2-462c-b187-519d56e75f56\") " pod="openstack/cinder-api-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.462530 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ce61c9c-20b2-462c-b187-519d56e75f56-logs\") pod \"cinder-api-0\" (UID: \"8ce61c9c-20b2-462c-b187-519d56e75f56\") " pod="openstack/cinder-api-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.462553 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ce61c9c-20b2-462c-b187-519d56e75f56-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8ce61c9c-20b2-462c-b187-519d56e75f56\") " pod="openstack/cinder-api-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.462571 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ce61c9c-20b2-462c-b187-519d56e75f56-scripts\") pod \"cinder-api-0\" (UID: \"8ce61c9c-20b2-462c-b187-519d56e75f56\") " pod="openstack/cinder-api-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.466732 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8ce61c9c-20b2-462c-b187-519d56e75f56-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8ce61c9c-20b2-462c-b187-519d56e75f56\") " pod="openstack/cinder-api-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.470896 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ce61c9c-20b2-462c-b187-519d56e75f56-scripts\") pod \"cinder-api-0\" (UID: \"8ce61c9c-20b2-462c-b187-519d56e75f56\") " pod="openstack/cinder-api-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.471195 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ce61c9c-20b2-462c-b187-519d56e75f56-logs\") pod \"cinder-api-0\" (UID: \"8ce61c9c-20b2-462c-b187-519d56e75f56\") " pod="openstack/cinder-api-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.474553 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ce61c9c-20b2-462c-b187-519d56e75f56-config-data-custom\") pod \"cinder-api-0\" (UID: \"8ce61c9c-20b2-462c-b187-519d56e75f56\") " pod="openstack/cinder-api-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.474859 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-797bbc649-ncxwv" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.477052 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ce61c9c-20b2-462c-b187-519d56e75f56-config-data\") pod \"cinder-api-0\" (UID: \"8ce61c9c-20b2-462c-b187-519d56e75f56\") " pod="openstack/cinder-api-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.502431 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ce61c9c-20b2-462c-b187-519d56e75f56-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8ce61c9c-20b2-462c-b187-519d56e75f56\") " pod="openstack/cinder-api-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.547861 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h47bg\" (UniqueName: \"kubernetes.io/projected/8ce61c9c-20b2-462c-b187-519d56e75f56-kube-api-access-h47bg\") pod \"cinder-api-0\" (UID: \"8ce61c9c-20b2-462c-b187-519d56e75f56\") " pod="openstack/cinder-api-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.611565 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-56988fbb4-nlxs9" event={"ID":"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45","Type":"ContainerStarted","Data":"a1a93104c41ae0baf6a1781b8d29302ccef9fef9da152e6274b2e92e08df9a8c"} Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.613046 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-769ddf488d-46ssj" event={"ID":"4702d881-d701-41a4-b36f-2d063a8ae246","Type":"ContainerStarted","Data":"df59e2ce88d3eada192f146759b343022c729c3eedcbc0cb1bf5166b45a8f96f"} Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.614694 4776 generic.go:334] "Generic (PLEG): container finished" podID="616a7df2-c899-43d8-9217-a1cf49ad8add" containerID="83c6af89df781455182ffdd3eb5e6c0222b267c62aacd11088a1acfa23cf26d6" exitCode=0 Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.614732 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" event={"ID":"616a7df2-c899-43d8-9217-a1cf49ad8add","Type":"ContainerDied","Data":"83c6af89df781455182ffdd3eb5e6c0222b267c62aacd11088a1acfa23cf26d6"} Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.617630 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-66cdf565f-6lzqw" event={"ID":"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f","Type":"ContainerStarted","Data":"87f8cbf578187a46d6e2d2c1082d1ed105b039c2788db71db6340e4c1b8457d4"} Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.754554 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 09:44:42 crc kubenswrapper[4776]: I1125 09:44:42.974933 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.156263 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-797bbc649-ncxwv"] Nov 25 09:44:43 crc kubenswrapper[4776]: W1125 09:44:43.199098 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecc97771_b405_44c7_93e1_d24c86e705ed.slice/crio-39476d42ea0743b82f53bd110fd3296e26eda4cb6adc970cfd1f9baf61dd1eb7 WatchSource:0}: Error finding container 39476d42ea0743b82f53bd110fd3296e26eda4cb6adc970cfd1f9baf61dd1eb7: Status 404 returned error can't find the container with id 39476d42ea0743b82f53bd110fd3296e26eda4cb6adc970cfd1f9baf61dd1eb7 Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.511471 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 09:44:43 crc kubenswrapper[4776]: W1125 09:44:43.545027 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ce61c9c_20b2_462c_b187_519d56e75f56.slice/crio-3ec68c3c80aa2f18db7444b6f5e50a4edc05713e5dbd3a6b2050341e4382f21a WatchSource:0}: Error finding container 3ec68c3c80aa2f18db7444b6f5e50a4edc05713e5dbd3a6b2050341e4382f21a: Status 404 returned error can't find the container with id 3ec68c3c80aa2f18db7444b6f5e50a4edc05713e5dbd3a6b2050341e4382f21a Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.571650 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.665248 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797bbc649-ncxwv" event={"ID":"ecc97771-b405-44c7-93e1-d24c86e705ed","Type":"ContainerStarted","Data":"39476d42ea0743b82f53bd110fd3296e26eda4cb6adc970cfd1f9baf61dd1eb7"} Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.701693 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-dns-svc\") pod \"616a7df2-c899-43d8-9217-a1cf49ad8add\" (UID: \"616a7df2-c899-43d8-9217-a1cf49ad8add\") " Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.701773 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-config\") pod \"616a7df2-c899-43d8-9217-a1cf49ad8add\" (UID: \"616a7df2-c899-43d8-9217-a1cf49ad8add\") " Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.702002 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5k6l4\" (UniqueName: \"kubernetes.io/projected/616a7df2-c899-43d8-9217-a1cf49ad8add-kube-api-access-5k6l4\") pod \"616a7df2-c899-43d8-9217-a1cf49ad8add\" (UID: \"616a7df2-c899-43d8-9217-a1cf49ad8add\") " Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.702050 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-ovsdbserver-nb\") pod \"616a7df2-c899-43d8-9217-a1cf49ad8add\" (UID: \"616a7df2-c899-43d8-9217-a1cf49ad8add\") " Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.702082 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-ovsdbserver-sb\") pod \"616a7df2-c899-43d8-9217-a1cf49ad8add\" (UID: \"616a7df2-c899-43d8-9217-a1cf49ad8add\") " Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.702178 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-dns-swift-storage-0\") pod \"616a7df2-c899-43d8-9217-a1cf49ad8add\" (UID: \"616a7df2-c899-43d8-9217-a1cf49ad8add\") " Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.726494 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-769ddf488d-46ssj" event={"ID":"4702d881-d701-41a4-b36f-2d063a8ae246","Type":"ContainerStarted","Data":"d45871f326a3dbb0f683b3d995f92f15682fd356f7dd9ec10b42a715df612c77"} Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.733637 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" event={"ID":"616a7df2-c899-43d8-9217-a1cf49ad8add","Type":"ContainerDied","Data":"bdb568b81ceb846fbe30935ec55b9d437a6593da8864783d4feddc24ea29731e"} Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.733702 4776 scope.go:117] "RemoveContainer" containerID="83c6af89df781455182ffdd3eb5e6c0222b267c62aacd11088a1acfa23cf26d6" Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.733872 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cc67f459c-7m6ns" Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.743550 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/616a7df2-c899-43d8-9217-a1cf49ad8add-kube-api-access-5k6l4" (OuterVolumeSpecName: "kube-api-access-5k6l4") pod "616a7df2-c899-43d8-9217-a1cf49ad8add" (UID: "616a7df2-c899-43d8-9217-a1cf49ad8add"). InnerVolumeSpecName "kube-api-access-5k6l4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.768420 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-66cdf565f-6lzqw" event={"ID":"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f","Type":"ContainerStarted","Data":"9779df52088b267add1650137fc3fc3f92ad6d3ea2e75a34659668d2ef3813bb"} Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.781457 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1b5f86ea-71cb-4e40-a7f5-cf77663097b2","Type":"ContainerStarted","Data":"39d7a267e133109c3eb942eeb89edf7b75238d7e3499584caf39ac794980898c"} Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.804396 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5k6l4\" (UniqueName: \"kubernetes.io/projected/616a7df2-c899-43d8-9217-a1cf49ad8add-kube-api-access-5k6l4\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.810794 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "616a7df2-c899-43d8-9217-a1cf49ad8add" (UID: "616a7df2-c899-43d8-9217-a1cf49ad8add"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.820038 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8ce61c9c-20b2-462c-b187-519d56e75f56","Type":"ContainerStarted","Data":"3ec68c3c80aa2f18db7444b6f5e50a4edc05713e5dbd3a6b2050341e4382f21a"} Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.837641 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-66cdf565f-6lzqw" podStartSLOduration=4.181024942 podStartE2EDuration="8.83762115s" podCreationTimestamp="2025-11-25 09:44:35 +0000 UTC" firstStartedPulling="2025-11-25 09:44:36.729002096 +0000 UTC m=+1221.770061649" lastFinishedPulling="2025-11-25 09:44:41.385598304 +0000 UTC m=+1226.426657857" observedRunningTime="2025-11-25 09:44:43.809820363 +0000 UTC m=+1228.850879916" watchObservedRunningTime="2025-11-25 09:44:43.83762115 +0000 UTC m=+1228.878680693" Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.838930 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "616a7df2-c899-43d8-9217-a1cf49ad8add" (UID: "616a7df2-c899-43d8-9217-a1cf49ad8add"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.841627 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "616a7df2-c899-43d8-9217-a1cf49ad8add" (UID: "616a7df2-c899-43d8-9217-a1cf49ad8add"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.846237 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-56988fbb4-nlxs9" event={"ID":"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45","Type":"ContainerStarted","Data":"63cb3f8363fb0824de5241eef2c1f027163930b39c0a67a62ab2eadf418c85d7"} Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.857979 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-config" (OuterVolumeSpecName: "config") pod "616a7df2-c899-43d8-9217-a1cf49ad8add" (UID: "616a7df2-c899-43d8-9217-a1cf49ad8add"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.859594 4776 scope.go:117] "RemoveContainer" containerID="7d85333bbe9017116a2c05c0c6b8ef47fd7bc62e15ff3825cc0cec533f34ab74" Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.866511 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "616a7df2-c899-43d8-9217-a1cf49ad8add" (UID: "616a7df2-c899-43d8-9217-a1cf49ad8add"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.888149 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-56988fbb4-nlxs9" podStartSLOduration=4.253054069 podStartE2EDuration="8.888132357s" podCreationTimestamp="2025-11-25 09:44:35 +0000 UTC" firstStartedPulling="2025-11-25 09:44:36.766783284 +0000 UTC m=+1221.807842837" lastFinishedPulling="2025-11-25 09:44:41.401861572 +0000 UTC m=+1226.442921125" observedRunningTime="2025-11-25 09:44:43.876821423 +0000 UTC m=+1228.917880976" watchObservedRunningTime="2025-11-25 09:44:43.888132357 +0000 UTC m=+1228.929191910" Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.909502 4776 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.909535 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.909545 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.909553 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:43 crc kubenswrapper[4776]: I1125 09:44:43.909563 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/616a7df2-c899-43d8-9217-a1cf49ad8add-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:44 crc kubenswrapper[4776]: I1125 09:44:44.105998 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cc67f459c-7m6ns"] Nov 25 09:44:44 crc kubenswrapper[4776]: I1125 09:44:44.118133 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5cc67f459c-7m6ns"] Nov 25 09:44:44 crc kubenswrapper[4776]: I1125 09:44:44.865581 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-769ddf488d-46ssj" event={"ID":"4702d881-d701-41a4-b36f-2d063a8ae246","Type":"ContainerStarted","Data":"4e9d85a2660614a7e89218283e8f24a0ddfcb9cad63c01ab8119453aceaad099"} Nov 25 09:44:44 crc kubenswrapper[4776]: I1125 09:44:44.865912 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:44:44 crc kubenswrapper[4776]: I1125 09:44:44.865929 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:44:44 crc kubenswrapper[4776]: I1125 09:44:44.869912 4776 generic.go:334] "Generic (PLEG): container finished" podID="ecc97771-b405-44c7-93e1-d24c86e705ed" containerID="7e1922940c86bbc7756196c6757fc4aec95313f98843e7b454724c9580f1a220" exitCode=0 Nov 25 09:44:44 crc kubenswrapper[4776]: I1125 09:44:44.870052 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797bbc649-ncxwv" event={"ID":"ecc97771-b405-44c7-93e1-d24c86e705ed","Type":"ContainerDied","Data":"7e1922940c86bbc7756196c6757fc4aec95313f98843e7b454724c9580f1a220"} Nov 25 09:44:44 crc kubenswrapper[4776]: I1125 09:44:44.909342 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-769ddf488d-46ssj" podStartSLOduration=5.909327459 podStartE2EDuration="5.909327459s" podCreationTimestamp="2025-11-25 09:44:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:44:44.908407465 +0000 UTC m=+1229.949467018" watchObservedRunningTime="2025-11-25 09:44:44.909327459 +0000 UTC m=+1229.950387012" Nov 25 09:44:45 crc kubenswrapper[4776]: I1125 09:44:45.680603 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="616a7df2-c899-43d8-9217-a1cf49ad8add" path="/var/lib/kubelet/pods/616a7df2-c899-43d8-9217-a1cf49ad8add/volumes" Nov 25 09:44:45 crc kubenswrapper[4776]: I1125 09:44:45.879884 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797bbc649-ncxwv" event={"ID":"ecc97771-b405-44c7-93e1-d24c86e705ed","Type":"ContainerStarted","Data":"ddeb90174916915ba1e5c7a4e9f66a3160fb3d00c4ccd556f36880eae5c2083b"} Nov 25 09:44:45 crc kubenswrapper[4776]: I1125 09:44:45.881781 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8ce61c9c-20b2-462c-b187-519d56e75f56","Type":"ContainerStarted","Data":"ed6f27472e2324b01d1d07ab6b85981456aa7805810f98f575417846cd62ede4"} Nov 25 09:44:46 crc kubenswrapper[4776]: I1125 09:44:46.006161 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 25 09:44:46 crc kubenswrapper[4776]: I1125 09:44:46.774550 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-667769d556-lqtlp" Nov 25 09:44:46 crc kubenswrapper[4776]: I1125 09:44:46.892326 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-797bbc649-ncxwv" Nov 25 09:44:46 crc kubenswrapper[4776]: I1125 09:44:46.914792 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-667769d556-lqtlp" Nov 25 09:44:46 crc kubenswrapper[4776]: I1125 09:44:46.936289 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-797bbc649-ncxwv" podStartSLOduration=4.9362700440000005 podStartE2EDuration="4.936270044s" podCreationTimestamp="2025-11-25 09:44:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:44:46.929616568 +0000 UTC m=+1231.970676141" watchObservedRunningTime="2025-11-25 09:44:46.936270044 +0000 UTC m=+1231.977329597" Nov 25 09:44:47 crc kubenswrapper[4776]: I1125 09:44:47.392038 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:44:47 crc kubenswrapper[4776]: I1125 09:44:47.818262 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:44:47 crc kubenswrapper[4776]: I1125 09:44:47.818332 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:44:47 crc kubenswrapper[4776]: I1125 09:44:47.923115 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8ce61c9c-20b2-462c-b187-519d56e75f56","Type":"ContainerStarted","Data":"27dbc31408be0c7000f15db205a2fa4f8f0563ea0cdf47359535299432ac1569"} Nov 25 09:44:47 crc kubenswrapper[4776]: I1125 09:44:47.924267 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="8ce61c9c-20b2-462c-b187-519d56e75f56" containerName="cinder-api-log" containerID="cri-o://ed6f27472e2324b01d1d07ab6b85981456aa7805810f98f575417846cd62ede4" gracePeriod=30 Nov 25 09:44:47 crc kubenswrapper[4776]: I1125 09:44:47.924282 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="8ce61c9c-20b2-462c-b187-519d56e75f56" containerName="cinder-api" containerID="cri-o://27dbc31408be0c7000f15db205a2fa4f8f0563ea0cdf47359535299432ac1569" gracePeriod=30 Nov 25 09:44:47 crc kubenswrapper[4776]: I1125 09:44:47.960567 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.960535854 podStartE2EDuration="5.960535854s" podCreationTimestamp="2025-11-25 09:44:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:44:47.951667321 +0000 UTC m=+1232.992726874" watchObservedRunningTime="2025-11-25 09:44:47.960535854 +0000 UTC m=+1233.001595447" Nov 25 09:44:48 crc kubenswrapper[4776]: I1125 09:44:48.577564 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6b8747677b-bmrqm" Nov 25 09:44:48 crc kubenswrapper[4776]: I1125 09:44:48.931122 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6b8747677b-bmrqm" Nov 25 09:44:48 crc kubenswrapper[4776]: I1125 09:44:48.949310 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1b5f86ea-71cb-4e40-a7f5-cf77663097b2","Type":"ContainerStarted","Data":"66ab8a7afff10625abbd356ae5d00854f89dbc0981db05c07f3a3b986fea6221"} Nov 25 09:44:48 crc kubenswrapper[4776]: I1125 09:44:48.949387 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1b5f86ea-71cb-4e40-a7f5-cf77663097b2","Type":"ContainerStarted","Data":"e5971edb9cbfda783a246c9d315408bc030ae25de88ac8fc101bbe339f9f7b07"} Nov 25 09:44:48 crc kubenswrapper[4776]: I1125 09:44:48.954466 4776 generic.go:334] "Generic (PLEG): container finished" podID="8ce61c9c-20b2-462c-b187-519d56e75f56" containerID="ed6f27472e2324b01d1d07ab6b85981456aa7805810f98f575417846cd62ede4" exitCode=143 Nov 25 09:44:48 crc kubenswrapper[4776]: I1125 09:44:48.955237 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8ce61c9c-20b2-462c-b187-519d56e75f56","Type":"ContainerDied","Data":"ed6f27472e2324b01d1d07ab6b85981456aa7805810f98f575417846cd62ede4"} Nov 25 09:44:48 crc kubenswrapper[4776]: I1125 09:44:48.977314 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.703829716 podStartE2EDuration="7.977293854s" podCreationTimestamp="2025-11-25 09:44:41 +0000 UTC" firstStartedPulling="2025-11-25 09:44:42.99588036 +0000 UTC m=+1228.036939913" lastFinishedPulling="2025-11-25 09:44:47.269344498 +0000 UTC m=+1232.310404051" observedRunningTime="2025-11-25 09:44:48.970721459 +0000 UTC m=+1234.011781002" watchObservedRunningTime="2025-11-25 09:44:48.977293854 +0000 UTC m=+1234.018353407" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.086515 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 25 09:44:51 crc kubenswrapper[4776]: E1125 09:44:51.087968 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="616a7df2-c899-43d8-9217-a1cf49ad8add" containerName="init" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.088047 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="616a7df2-c899-43d8-9217-a1cf49ad8add" containerName="init" Nov 25 09:44:51 crc kubenswrapper[4776]: E1125 09:44:51.088137 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="616a7df2-c899-43d8-9217-a1cf49ad8add" containerName="dnsmasq-dns" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.088194 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="616a7df2-c899-43d8-9217-a1cf49ad8add" containerName="dnsmasq-dns" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.088412 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="616a7df2-c899-43d8-9217-a1cf49ad8add" containerName="dnsmasq-dns" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.089069 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.094518 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.094821 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-vlpqr" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.095152 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.144115 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.197011 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/aa774bae-d2c2-47eb-92d8-2d65c3201acc-openstack-config-secret\") pod \"openstackclient\" (UID: \"aa774bae-d2c2-47eb-92d8-2d65c3201acc\") " pod="openstack/openstackclient" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.197093 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/aa774bae-d2c2-47eb-92d8-2d65c3201acc-openstack-config\") pod \"openstackclient\" (UID: \"aa774bae-d2c2-47eb-92d8-2d65c3201acc\") " pod="openstack/openstackclient" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.197119 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp24n\" (UniqueName: \"kubernetes.io/projected/aa774bae-d2c2-47eb-92d8-2d65c3201acc-kube-api-access-dp24n\") pod \"openstackclient\" (UID: \"aa774bae-d2c2-47eb-92d8-2d65c3201acc\") " pod="openstack/openstackclient" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.197690 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa774bae-d2c2-47eb-92d8-2d65c3201acc-combined-ca-bundle\") pod \"openstackclient\" (UID: \"aa774bae-d2c2-47eb-92d8-2d65c3201acc\") " pod="openstack/openstackclient" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.299421 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa774bae-d2c2-47eb-92d8-2d65c3201acc-combined-ca-bundle\") pod \"openstackclient\" (UID: \"aa774bae-d2c2-47eb-92d8-2d65c3201acc\") " pod="openstack/openstackclient" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.299813 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/aa774bae-d2c2-47eb-92d8-2d65c3201acc-openstack-config-secret\") pod \"openstackclient\" (UID: \"aa774bae-d2c2-47eb-92d8-2d65c3201acc\") " pod="openstack/openstackclient" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.299870 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/aa774bae-d2c2-47eb-92d8-2d65c3201acc-openstack-config\") pod \"openstackclient\" (UID: \"aa774bae-d2c2-47eb-92d8-2d65c3201acc\") " pod="openstack/openstackclient" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.299900 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp24n\" (UniqueName: \"kubernetes.io/projected/aa774bae-d2c2-47eb-92d8-2d65c3201acc-kube-api-access-dp24n\") pod \"openstackclient\" (UID: \"aa774bae-d2c2-47eb-92d8-2d65c3201acc\") " pod="openstack/openstackclient" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.300782 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/aa774bae-d2c2-47eb-92d8-2d65c3201acc-openstack-config\") pod \"openstackclient\" (UID: \"aa774bae-d2c2-47eb-92d8-2d65c3201acc\") " pod="openstack/openstackclient" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.305568 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/aa774bae-d2c2-47eb-92d8-2d65c3201acc-openstack-config-secret\") pod \"openstackclient\" (UID: \"aa774bae-d2c2-47eb-92d8-2d65c3201acc\") " pod="openstack/openstackclient" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.308814 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa774bae-d2c2-47eb-92d8-2d65c3201acc-combined-ca-bundle\") pod \"openstackclient\" (UID: \"aa774bae-d2c2-47eb-92d8-2d65c3201acc\") " pod="openstack/openstackclient" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.317475 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 25 09:44:51 crc kubenswrapper[4776]: E1125 09:44:51.318177 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-dp24n], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="aa774bae-d2c2-47eb-92d8-2d65c3201acc" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.319315 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp24n\" (UniqueName: \"kubernetes.io/projected/aa774bae-d2c2-47eb-92d8-2d65c3201acc-kube-api-access-dp24n\") pod \"openstackclient\" (UID: \"aa774bae-d2c2-47eb-92d8-2d65c3201acc\") " pod="openstack/openstackclient" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.334463 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.384687 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.386594 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.406326 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.502995 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk879\" (UniqueName: \"kubernetes.io/projected/3fc11cf2-c6c7-4250-b7f4-9c68d78152eb-kube-api-access-xk879\") pod \"openstackclient\" (UID: \"3fc11cf2-c6c7-4250-b7f4-9c68d78152eb\") " pod="openstack/openstackclient" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.503045 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3fc11cf2-c6c7-4250-b7f4-9c68d78152eb-openstack-config-secret\") pod \"openstackclient\" (UID: \"3fc11cf2-c6c7-4250-b7f4-9c68d78152eb\") " pod="openstack/openstackclient" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.503104 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fc11cf2-c6c7-4250-b7f4-9c68d78152eb-combined-ca-bundle\") pod \"openstackclient\" (UID: \"3fc11cf2-c6c7-4250-b7f4-9c68d78152eb\") " pod="openstack/openstackclient" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.503164 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3fc11cf2-c6c7-4250-b7f4-9c68d78152eb-openstack-config\") pod \"openstackclient\" (UID: \"3fc11cf2-c6c7-4250-b7f4-9c68d78152eb\") " pod="openstack/openstackclient" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.604688 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3fc11cf2-c6c7-4250-b7f4-9c68d78152eb-openstack-config\") pod \"openstackclient\" (UID: \"3fc11cf2-c6c7-4250-b7f4-9c68d78152eb\") " pod="openstack/openstackclient" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.605124 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xk879\" (UniqueName: \"kubernetes.io/projected/3fc11cf2-c6c7-4250-b7f4-9c68d78152eb-kube-api-access-xk879\") pod \"openstackclient\" (UID: \"3fc11cf2-c6c7-4250-b7f4-9c68d78152eb\") " pod="openstack/openstackclient" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.605165 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3fc11cf2-c6c7-4250-b7f4-9c68d78152eb-openstack-config-secret\") pod \"openstackclient\" (UID: \"3fc11cf2-c6c7-4250-b7f4-9c68d78152eb\") " pod="openstack/openstackclient" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.605256 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fc11cf2-c6c7-4250-b7f4-9c68d78152eb-combined-ca-bundle\") pod \"openstackclient\" (UID: \"3fc11cf2-c6c7-4250-b7f4-9c68d78152eb\") " pod="openstack/openstackclient" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.605670 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3fc11cf2-c6c7-4250-b7f4-9c68d78152eb-openstack-config\") pod \"openstackclient\" (UID: \"3fc11cf2-c6c7-4250-b7f4-9c68d78152eb\") " pod="openstack/openstackclient" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.610399 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3fc11cf2-c6c7-4250-b7f4-9c68d78152eb-openstack-config-secret\") pod \"openstackclient\" (UID: \"3fc11cf2-c6c7-4250-b7f4-9c68d78152eb\") " pod="openstack/openstackclient" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.611890 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fc11cf2-c6c7-4250-b7f4-9c68d78152eb-combined-ca-bundle\") pod \"openstackclient\" (UID: \"3fc11cf2-c6c7-4250-b7f4-9c68d78152eb\") " pod="openstack/openstackclient" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.627468 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk879\" (UniqueName: \"kubernetes.io/projected/3fc11cf2-c6c7-4250-b7f4-9c68d78152eb-kube-api-access-xk879\") pod \"openstackclient\" (UID: \"3fc11cf2-c6c7-4250-b7f4-9c68d78152eb\") " pod="openstack/openstackclient" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.717974 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 09:44:51 crc kubenswrapper[4776]: I1125 09:44:51.987644 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 09:44:52 crc kubenswrapper[4776]: I1125 09:44:52.008851 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 09:44:52 crc kubenswrapper[4776]: I1125 09:44:52.012413 4776 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="aa774bae-d2c2-47eb-92d8-2d65c3201acc" podUID="3fc11cf2-c6c7-4250-b7f4-9c68d78152eb" Nov 25 09:44:52 crc kubenswrapper[4776]: I1125 09:44:52.117703 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa774bae-d2c2-47eb-92d8-2d65c3201acc-combined-ca-bundle\") pod \"aa774bae-d2c2-47eb-92d8-2d65c3201acc\" (UID: \"aa774bae-d2c2-47eb-92d8-2d65c3201acc\") " Nov 25 09:44:52 crc kubenswrapper[4776]: I1125 09:44:52.117810 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dp24n\" (UniqueName: \"kubernetes.io/projected/aa774bae-d2c2-47eb-92d8-2d65c3201acc-kube-api-access-dp24n\") pod \"aa774bae-d2c2-47eb-92d8-2d65c3201acc\" (UID: \"aa774bae-d2c2-47eb-92d8-2d65c3201acc\") " Nov 25 09:44:52 crc kubenswrapper[4776]: I1125 09:44:52.117892 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/aa774bae-d2c2-47eb-92d8-2d65c3201acc-openstack-config-secret\") pod \"aa774bae-d2c2-47eb-92d8-2d65c3201acc\" (UID: \"aa774bae-d2c2-47eb-92d8-2d65c3201acc\") " Nov 25 09:44:52 crc kubenswrapper[4776]: I1125 09:44:52.118003 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/aa774bae-d2c2-47eb-92d8-2d65c3201acc-openstack-config\") pod \"aa774bae-d2c2-47eb-92d8-2d65c3201acc\" (UID: \"aa774bae-d2c2-47eb-92d8-2d65c3201acc\") " Nov 25 09:44:52 crc kubenswrapper[4776]: I1125 09:44:52.118588 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa774bae-d2c2-47eb-92d8-2d65c3201acc-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "aa774bae-d2c2-47eb-92d8-2d65c3201acc" (UID: "aa774bae-d2c2-47eb-92d8-2d65c3201acc"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:44:52 crc kubenswrapper[4776]: I1125 09:44:52.122446 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa774bae-d2c2-47eb-92d8-2d65c3201acc-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "aa774bae-d2c2-47eb-92d8-2d65c3201acc" (UID: "aa774bae-d2c2-47eb-92d8-2d65c3201acc"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:52 crc kubenswrapper[4776]: I1125 09:44:52.124261 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa774bae-d2c2-47eb-92d8-2d65c3201acc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aa774bae-d2c2-47eb-92d8-2d65c3201acc" (UID: "aa774bae-d2c2-47eb-92d8-2d65c3201acc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:52 crc kubenswrapper[4776]: I1125 09:44:52.126195 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa774bae-d2c2-47eb-92d8-2d65c3201acc-kube-api-access-dp24n" (OuterVolumeSpecName: "kube-api-access-dp24n") pod "aa774bae-d2c2-47eb-92d8-2d65c3201acc" (UID: "aa774bae-d2c2-47eb-92d8-2d65c3201acc"). InnerVolumeSpecName "kube-api-access-dp24n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:44:52 crc kubenswrapper[4776]: I1125 09:44:52.221054 4776 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/aa774bae-d2c2-47eb-92d8-2d65c3201acc-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:52 crc kubenswrapper[4776]: I1125 09:44:52.221182 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa774bae-d2c2-47eb-92d8-2d65c3201acc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:52 crc kubenswrapper[4776]: I1125 09:44:52.221196 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dp24n\" (UniqueName: \"kubernetes.io/projected/aa774bae-d2c2-47eb-92d8-2d65c3201acc-kube-api-access-dp24n\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:52 crc kubenswrapper[4776]: I1125 09:44:52.221210 4776 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/aa774bae-d2c2-47eb-92d8-2d65c3201acc-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:52 crc kubenswrapper[4776]: W1125 09:44:52.265613 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3fc11cf2_c6c7_4250_b7f4_9c68d78152eb.slice/crio-34023a10b4dbea1967b560435b4de781ef6c5d0c9f5d8b63082342a5974732d4 WatchSource:0}: Error finding container 34023a10b4dbea1967b560435b4de781ef6c5d0c9f5d8b63082342a5974732d4: Status 404 returned error can't find the container with id 34023a10b4dbea1967b560435b4de781ef6c5d0c9f5d8b63082342a5974732d4 Nov 25 09:44:52 crc kubenswrapper[4776]: I1125 09:44:52.265912 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 25 09:44:52 crc kubenswrapper[4776]: I1125 09:44:52.267693 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 25 09:44:52 crc kubenswrapper[4776]: I1125 09:44:52.323912 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:44:52 crc kubenswrapper[4776]: I1125 09:44:52.379331 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:44:52 crc kubenswrapper[4776]: I1125 09:44:52.456331 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6b8747677b-bmrqm"] Nov 25 09:44:52 crc kubenswrapper[4776]: I1125 09:44:52.456683 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6b8747677b-bmrqm" podUID="e3dd1ad2-32fd-4c42-bbaf-e42269987973" containerName="barbican-api-log" containerID="cri-o://b4ff1ba01bcbbe2c5b08eb7d71d3b56219363949b9c1aab72bd432ffca17d4c1" gracePeriod=30 Nov 25 09:44:52 crc kubenswrapper[4776]: I1125 09:44:52.456815 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6b8747677b-bmrqm" podUID="e3dd1ad2-32fd-4c42-bbaf-e42269987973" containerName="barbican-api" containerID="cri-o://40c228ae0f5503fa2aa9af33eaa8e948a2b937a4476a096a469b2ce4731b8ac6" gracePeriod=30 Nov 25 09:44:52 crc kubenswrapper[4776]: I1125 09:44:52.476282 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-797bbc649-ncxwv" Nov 25 09:44:52 crc kubenswrapper[4776]: I1125 09:44:52.545716 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76c58b6d97-jbkjb"] Nov 25 09:44:52 crc kubenswrapper[4776]: I1125 09:44:52.545958 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" podUID="624c8057-3576-4249-9a63-5bb0dee2da0a" containerName="dnsmasq-dns" containerID="cri-o://da36d0549e285577f46082c15a99da5ebd9c6b191beddf067a2807915ebcfe1c" gracePeriod=10 Nov 25 09:44:52 crc kubenswrapper[4776]: I1125 09:44:52.757444 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 25 09:44:53 crc kubenswrapper[4776]: I1125 09:44:53.000643 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"3fc11cf2-c6c7-4250-b7f4-9c68d78152eb","Type":"ContainerStarted","Data":"34023a10b4dbea1967b560435b4de781ef6c5d0c9f5d8b63082342a5974732d4"} Nov 25 09:44:53 crc kubenswrapper[4776]: I1125 09:44:53.026285 4776 generic.go:334] "Generic (PLEG): container finished" podID="624c8057-3576-4249-9a63-5bb0dee2da0a" containerID="da36d0549e285577f46082c15a99da5ebd9c6b191beddf067a2807915ebcfe1c" exitCode=0 Nov 25 09:44:53 crc kubenswrapper[4776]: I1125 09:44:53.026370 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" event={"ID":"624c8057-3576-4249-9a63-5bb0dee2da0a","Type":"ContainerDied","Data":"da36d0549e285577f46082c15a99da5ebd9c6b191beddf067a2807915ebcfe1c"} Nov 25 09:44:53 crc kubenswrapper[4776]: I1125 09:44:53.038425 4776 generic.go:334] "Generic (PLEG): container finished" podID="e3dd1ad2-32fd-4c42-bbaf-e42269987973" containerID="b4ff1ba01bcbbe2c5b08eb7d71d3b56219363949b9c1aab72bd432ffca17d4c1" exitCode=143 Nov 25 09:44:53 crc kubenswrapper[4776]: I1125 09:44:53.039444 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b8747677b-bmrqm" event={"ID":"e3dd1ad2-32fd-4c42-bbaf-e42269987973","Type":"ContainerDied","Data":"b4ff1ba01bcbbe2c5b08eb7d71d3b56219363949b9c1aab72bd432ffca17d4c1"} Nov 25 09:44:53 crc kubenswrapper[4776]: I1125 09:44:53.039483 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 09:44:53 crc kubenswrapper[4776]: I1125 09:44:53.052830 4776 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="aa774bae-d2c2-47eb-92d8-2d65c3201acc" podUID="3fc11cf2-c6c7-4250-b7f4-9c68d78152eb" Nov 25 09:44:53 crc kubenswrapper[4776]: E1125 09:44:53.160976 4776 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa774bae_d2c2_47eb_92d8_2d65c3201acc.slice\": RecentStats: unable to find data in memory cache]" Nov 25 09:44:53 crc kubenswrapper[4776]: I1125 09:44:53.239121 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" Nov 25 09:44:53 crc kubenswrapper[4776]: I1125 09:44:53.352442 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-dns-svc\") pod \"624c8057-3576-4249-9a63-5bb0dee2da0a\" (UID: \"624c8057-3576-4249-9a63-5bb0dee2da0a\") " Nov 25 09:44:53 crc kubenswrapper[4776]: I1125 09:44:53.352576 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-config\") pod \"624c8057-3576-4249-9a63-5bb0dee2da0a\" (UID: \"624c8057-3576-4249-9a63-5bb0dee2da0a\") " Nov 25 09:44:53 crc kubenswrapper[4776]: I1125 09:44:53.352618 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8c67\" (UniqueName: \"kubernetes.io/projected/624c8057-3576-4249-9a63-5bb0dee2da0a-kube-api-access-x8c67\") pod \"624c8057-3576-4249-9a63-5bb0dee2da0a\" (UID: \"624c8057-3576-4249-9a63-5bb0dee2da0a\") " Nov 25 09:44:53 crc kubenswrapper[4776]: I1125 09:44:53.352643 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-ovsdbserver-sb\") pod \"624c8057-3576-4249-9a63-5bb0dee2da0a\" (UID: \"624c8057-3576-4249-9a63-5bb0dee2da0a\") " Nov 25 09:44:53 crc kubenswrapper[4776]: I1125 09:44:53.352695 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-ovsdbserver-nb\") pod \"624c8057-3576-4249-9a63-5bb0dee2da0a\" (UID: \"624c8057-3576-4249-9a63-5bb0dee2da0a\") " Nov 25 09:44:53 crc kubenswrapper[4776]: I1125 09:44:53.352787 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-dns-swift-storage-0\") pod \"624c8057-3576-4249-9a63-5bb0dee2da0a\" (UID: \"624c8057-3576-4249-9a63-5bb0dee2da0a\") " Nov 25 09:44:53 crc kubenswrapper[4776]: I1125 09:44:53.387288 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/624c8057-3576-4249-9a63-5bb0dee2da0a-kube-api-access-x8c67" (OuterVolumeSpecName: "kube-api-access-x8c67") pod "624c8057-3576-4249-9a63-5bb0dee2da0a" (UID: "624c8057-3576-4249-9a63-5bb0dee2da0a"). InnerVolumeSpecName "kube-api-access-x8c67". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:44:53 crc kubenswrapper[4776]: I1125 09:44:53.463894 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8c67\" (UniqueName: \"kubernetes.io/projected/624c8057-3576-4249-9a63-5bb0dee2da0a-kube-api-access-x8c67\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:53 crc kubenswrapper[4776]: I1125 09:44:53.478950 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "624c8057-3576-4249-9a63-5bb0dee2da0a" (UID: "624c8057-3576-4249-9a63-5bb0dee2da0a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:44:53 crc kubenswrapper[4776]: I1125 09:44:53.480388 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "624c8057-3576-4249-9a63-5bb0dee2da0a" (UID: "624c8057-3576-4249-9a63-5bb0dee2da0a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:44:53 crc kubenswrapper[4776]: I1125 09:44:53.483407 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-config" (OuterVolumeSpecName: "config") pod "624c8057-3576-4249-9a63-5bb0dee2da0a" (UID: "624c8057-3576-4249-9a63-5bb0dee2da0a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:44:53 crc kubenswrapper[4776]: I1125 09:44:53.531532 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "624c8057-3576-4249-9a63-5bb0dee2da0a" (UID: "624c8057-3576-4249-9a63-5bb0dee2da0a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:44:53 crc kubenswrapper[4776]: I1125 09:44:53.538282 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "624c8057-3576-4249-9a63-5bb0dee2da0a" (UID: "624c8057-3576-4249-9a63-5bb0dee2da0a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:44:53 crc kubenswrapper[4776]: I1125 09:44:53.567265 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:53 crc kubenswrapper[4776]: I1125 09:44:53.567303 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:53 crc kubenswrapper[4776]: I1125 09:44:53.567316 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:53 crc kubenswrapper[4776]: I1125 09:44:53.567329 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:53 crc kubenswrapper[4776]: I1125 09:44:53.567340 4776 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/624c8057-3576-4249-9a63-5bb0dee2da0a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:53 crc kubenswrapper[4776]: I1125 09:44:53.691662 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa774bae-d2c2-47eb-92d8-2d65c3201acc" path="/var/lib/kubelet/pods/aa774bae-d2c2-47eb-92d8-2d65c3201acc/volumes" Nov 25 09:44:54 crc kubenswrapper[4776]: I1125 09:44:54.048963 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" event={"ID":"624c8057-3576-4249-9a63-5bb0dee2da0a","Type":"ContainerDied","Data":"81aa9892ab356d20fd260d0a5c2590947e25c2167e81aeecf818a69d278ffc27"} Nov 25 09:44:54 crc kubenswrapper[4776]: I1125 09:44:54.049024 4776 scope.go:117] "RemoveContainer" containerID="da36d0549e285577f46082c15a99da5ebd9c6b191beddf067a2807915ebcfe1c" Nov 25 09:44:54 crc kubenswrapper[4776]: I1125 09:44:54.049311 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76c58b6d97-jbkjb" Nov 25 09:44:54 crc kubenswrapper[4776]: I1125 09:44:54.104731 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76c58b6d97-jbkjb"] Nov 25 09:44:54 crc kubenswrapper[4776]: I1125 09:44:54.122589 4776 scope.go:117] "RemoveContainer" containerID="fd389d7b041963d9f2233f5abfaa28dabd63ef7d03b7319bb290986791d92571" Nov 25 09:44:54 crc kubenswrapper[4776]: I1125 09:44:54.122722 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-76c58b6d97-jbkjb"] Nov 25 09:44:55 crc kubenswrapper[4776]: I1125 09:44:55.113373 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 25 09:44:55 crc kubenswrapper[4776]: I1125 09:44:55.672996 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="624c8057-3576-4249-9a63-5bb0dee2da0a" path="/var/lib/kubelet/pods/624c8057-3576-4249-9a63-5bb0dee2da0a/volumes" Nov 25 09:44:56 crc kubenswrapper[4776]: I1125 09:44:56.095289 4776 generic.go:334] "Generic (PLEG): container finished" podID="e3dd1ad2-32fd-4c42-bbaf-e42269987973" containerID="40c228ae0f5503fa2aa9af33eaa8e948a2b937a4476a096a469b2ce4731b8ac6" exitCode=0 Nov 25 09:44:56 crc kubenswrapper[4776]: I1125 09:44:56.095341 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b8747677b-bmrqm" event={"ID":"e3dd1ad2-32fd-4c42-bbaf-e42269987973","Type":"ContainerDied","Data":"40c228ae0f5503fa2aa9af33eaa8e948a2b937a4476a096a469b2ce4731b8ac6"} Nov 25 09:44:56 crc kubenswrapper[4776]: I1125 09:44:56.279842 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6b8747677b-bmrqm" Nov 25 09:44:56 crc kubenswrapper[4776]: I1125 09:44:56.433415 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3dd1ad2-32fd-4c42-bbaf-e42269987973-config-data-custom\") pod \"e3dd1ad2-32fd-4c42-bbaf-e42269987973\" (UID: \"e3dd1ad2-32fd-4c42-bbaf-e42269987973\") " Nov 25 09:44:56 crc kubenswrapper[4776]: I1125 09:44:56.433457 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72ktx\" (UniqueName: \"kubernetes.io/projected/e3dd1ad2-32fd-4c42-bbaf-e42269987973-kube-api-access-72ktx\") pod \"e3dd1ad2-32fd-4c42-bbaf-e42269987973\" (UID: \"e3dd1ad2-32fd-4c42-bbaf-e42269987973\") " Nov 25 09:44:56 crc kubenswrapper[4776]: I1125 09:44:56.433606 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3dd1ad2-32fd-4c42-bbaf-e42269987973-logs\") pod \"e3dd1ad2-32fd-4c42-bbaf-e42269987973\" (UID: \"e3dd1ad2-32fd-4c42-bbaf-e42269987973\") " Nov 25 09:44:56 crc kubenswrapper[4776]: I1125 09:44:56.433632 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3dd1ad2-32fd-4c42-bbaf-e42269987973-config-data\") pod \"e3dd1ad2-32fd-4c42-bbaf-e42269987973\" (UID: \"e3dd1ad2-32fd-4c42-bbaf-e42269987973\") " Nov 25 09:44:56 crc kubenswrapper[4776]: I1125 09:44:56.433683 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3dd1ad2-32fd-4c42-bbaf-e42269987973-combined-ca-bundle\") pod \"e3dd1ad2-32fd-4c42-bbaf-e42269987973\" (UID: \"e3dd1ad2-32fd-4c42-bbaf-e42269987973\") " Nov 25 09:44:56 crc kubenswrapper[4776]: I1125 09:44:56.435552 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3dd1ad2-32fd-4c42-bbaf-e42269987973-logs" (OuterVolumeSpecName: "logs") pod "e3dd1ad2-32fd-4c42-bbaf-e42269987973" (UID: "e3dd1ad2-32fd-4c42-bbaf-e42269987973"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:44:56 crc kubenswrapper[4776]: I1125 09:44:56.441058 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3dd1ad2-32fd-4c42-bbaf-e42269987973-kube-api-access-72ktx" (OuterVolumeSpecName: "kube-api-access-72ktx") pod "e3dd1ad2-32fd-4c42-bbaf-e42269987973" (UID: "e3dd1ad2-32fd-4c42-bbaf-e42269987973"). InnerVolumeSpecName "kube-api-access-72ktx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:44:56 crc kubenswrapper[4776]: I1125 09:44:56.452888 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3dd1ad2-32fd-4c42-bbaf-e42269987973-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e3dd1ad2-32fd-4c42-bbaf-e42269987973" (UID: "e3dd1ad2-32fd-4c42-bbaf-e42269987973"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:56 crc kubenswrapper[4776]: I1125 09:44:56.490268 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3dd1ad2-32fd-4c42-bbaf-e42269987973-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e3dd1ad2-32fd-4c42-bbaf-e42269987973" (UID: "e3dd1ad2-32fd-4c42-bbaf-e42269987973"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:56 crc kubenswrapper[4776]: I1125 09:44:56.523257 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3dd1ad2-32fd-4c42-bbaf-e42269987973-config-data" (OuterVolumeSpecName: "config-data") pod "e3dd1ad2-32fd-4c42-bbaf-e42269987973" (UID: "e3dd1ad2-32fd-4c42-bbaf-e42269987973"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:44:56 crc kubenswrapper[4776]: I1125 09:44:56.539319 4776 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e3dd1ad2-32fd-4c42-bbaf-e42269987973-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:56 crc kubenswrapper[4776]: I1125 09:44:56.539366 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72ktx\" (UniqueName: \"kubernetes.io/projected/e3dd1ad2-32fd-4c42-bbaf-e42269987973-kube-api-access-72ktx\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:56 crc kubenswrapper[4776]: I1125 09:44:56.539382 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3dd1ad2-32fd-4c42-bbaf-e42269987973-logs\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:56 crc kubenswrapper[4776]: I1125 09:44:56.539395 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3dd1ad2-32fd-4c42-bbaf-e42269987973-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:56 crc kubenswrapper[4776]: I1125 09:44:56.539407 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3dd1ad2-32fd-4c42-bbaf-e42269987973-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:44:57 crc kubenswrapper[4776]: I1125 09:44:57.105862 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b8747677b-bmrqm" event={"ID":"e3dd1ad2-32fd-4c42-bbaf-e42269987973","Type":"ContainerDied","Data":"1e8b9f104b52295070ebd96b4f5451896dce3430f9c1303d99e15022d1fbf0b8"} Nov 25 09:44:57 crc kubenswrapper[4776]: I1125 09:44:57.105918 4776 scope.go:117] "RemoveContainer" containerID="40c228ae0f5503fa2aa9af33eaa8e948a2b937a4476a096a469b2ce4731b8ac6" Nov 25 09:44:57 crc kubenswrapper[4776]: I1125 09:44:57.105935 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6b8747677b-bmrqm" Nov 25 09:44:57 crc kubenswrapper[4776]: I1125 09:44:57.133168 4776 scope.go:117] "RemoveContainer" containerID="b4ff1ba01bcbbe2c5b08eb7d71d3b56219363949b9c1aab72bd432ffca17d4c1" Nov 25 09:44:57 crc kubenswrapper[4776]: I1125 09:44:57.145543 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6b8747677b-bmrqm"] Nov 25 09:44:57 crc kubenswrapper[4776]: I1125 09:44:57.152655 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6b8747677b-bmrqm"] Nov 25 09:44:57 crc kubenswrapper[4776]: I1125 09:44:57.422688 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5f4c844d98-nt7dn" Nov 25 09:44:57 crc kubenswrapper[4776]: I1125 09:44:57.678653 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3dd1ad2-32fd-4c42-bbaf-e42269987973" path="/var/lib/kubelet/pods/e3dd1ad2-32fd-4c42-bbaf-e42269987973/volumes" Nov 25 09:44:57 crc kubenswrapper[4776]: I1125 09:44:57.873641 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 25 09:44:57 crc kubenswrapper[4776]: I1125 09:44:57.942402 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.117181 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="1b5f86ea-71cb-4e40-a7f5-cf77663097b2" containerName="cinder-scheduler" containerID="cri-o://e5971edb9cbfda783a246c9d315408bc030ae25de88ac8fc101bbe339f9f7b07" gracePeriod=30 Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.117287 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="1b5f86ea-71cb-4e40-a7f5-cf77663097b2" containerName="probe" containerID="cri-o://66ab8a7afff10625abbd356ae5d00854f89dbc0981db05c07f3a3b986fea6221" gracePeriod=30 Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.319610 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-5864b7d7fc-g9z2r"] Nov 25 09:44:58 crc kubenswrapper[4776]: E1125 09:44:58.319985 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3dd1ad2-32fd-4c42-bbaf-e42269987973" containerName="barbican-api-log" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.320003 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3dd1ad2-32fd-4c42-bbaf-e42269987973" containerName="barbican-api-log" Nov 25 09:44:58 crc kubenswrapper[4776]: E1125 09:44:58.320028 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3dd1ad2-32fd-4c42-bbaf-e42269987973" containerName="barbican-api" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.320037 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3dd1ad2-32fd-4c42-bbaf-e42269987973" containerName="barbican-api" Nov 25 09:44:58 crc kubenswrapper[4776]: E1125 09:44:58.320048 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="624c8057-3576-4249-9a63-5bb0dee2da0a" containerName="init" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.320054 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="624c8057-3576-4249-9a63-5bb0dee2da0a" containerName="init" Nov 25 09:44:58 crc kubenswrapper[4776]: E1125 09:44:58.320084 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="624c8057-3576-4249-9a63-5bb0dee2da0a" containerName="dnsmasq-dns" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.320090 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="624c8057-3576-4249-9a63-5bb0dee2da0a" containerName="dnsmasq-dns" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.324286 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="624c8057-3576-4249-9a63-5bb0dee2da0a" containerName="dnsmasq-dns" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.324331 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3dd1ad2-32fd-4c42-bbaf-e42269987973" containerName="barbican-api-log" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.324344 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3dd1ad2-32fd-4c42-bbaf-e42269987973" containerName="barbican-api" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.325342 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.329612 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.329817 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.329988 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.342317 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5864b7d7fc-g9z2r"] Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.487178 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-combined-ca-bundle\") pod \"swift-proxy-5864b7d7fc-g9z2r\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.487238 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-internal-tls-certs\") pod \"swift-proxy-5864b7d7fc-g9z2r\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.487276 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-config-data\") pod \"swift-proxy-5864b7d7fc-g9z2r\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.487311 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-run-httpd\") pod \"swift-proxy-5864b7d7fc-g9z2r\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.487344 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-public-tls-certs\") pod \"swift-proxy-5864b7d7fc-g9z2r\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.487372 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-log-httpd\") pod \"swift-proxy-5864b7d7fc-g9z2r\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.487391 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zj7k\" (UniqueName: \"kubernetes.io/projected/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-kube-api-access-7zj7k\") pod \"swift-proxy-5864b7d7fc-g9z2r\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.487409 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-etc-swift\") pod \"swift-proxy-5864b7d7fc-g9z2r\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.589036 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-internal-tls-certs\") pod \"swift-proxy-5864b7d7fc-g9z2r\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.589122 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-config-data\") pod \"swift-proxy-5864b7d7fc-g9z2r\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.589158 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-run-httpd\") pod \"swift-proxy-5864b7d7fc-g9z2r\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.589192 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-public-tls-certs\") pod \"swift-proxy-5864b7d7fc-g9z2r\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.589220 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-log-httpd\") pod \"swift-proxy-5864b7d7fc-g9z2r\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.589242 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zj7k\" (UniqueName: \"kubernetes.io/projected/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-kube-api-access-7zj7k\") pod \"swift-proxy-5864b7d7fc-g9z2r\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.589263 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-etc-swift\") pod \"swift-proxy-5864b7d7fc-g9z2r\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.589340 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-combined-ca-bundle\") pod \"swift-proxy-5864b7d7fc-g9z2r\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.589829 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-run-httpd\") pod \"swift-proxy-5864b7d7fc-g9z2r\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.589868 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-log-httpd\") pod \"swift-proxy-5864b7d7fc-g9z2r\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.597846 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-public-tls-certs\") pod \"swift-proxy-5864b7d7fc-g9z2r\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.598049 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-config-data\") pod \"swift-proxy-5864b7d7fc-g9z2r\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.601972 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-etc-swift\") pod \"swift-proxy-5864b7d7fc-g9z2r\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.602971 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-combined-ca-bundle\") pod \"swift-proxy-5864b7d7fc-g9z2r\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.609002 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-internal-tls-certs\") pod \"swift-proxy-5864b7d7fc-g9z2r\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.612905 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zj7k\" (UniqueName: \"kubernetes.io/projected/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-kube-api-access-7zj7k\") pod \"swift-proxy-5864b7d7fc-g9z2r\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:44:58 crc kubenswrapper[4776]: I1125 09:44:58.645918 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:44:59 crc kubenswrapper[4776]: I1125 09:44:59.128029 4776 generic.go:334] "Generic (PLEG): container finished" podID="1b5f86ea-71cb-4e40-a7f5-cf77663097b2" containerID="66ab8a7afff10625abbd356ae5d00854f89dbc0981db05c07f3a3b986fea6221" exitCode=0 Nov 25 09:44:59 crc kubenswrapper[4776]: I1125 09:44:59.128084 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1b5f86ea-71cb-4e40-a7f5-cf77663097b2","Type":"ContainerDied","Data":"66ab8a7afff10625abbd356ae5d00854f89dbc0981db05c07f3a3b986fea6221"} Nov 25 09:44:59 crc kubenswrapper[4776]: I1125 09:44:59.921316 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-54dbffbb57-kmqfb" Nov 25 09:45:00 crc kubenswrapper[4776]: I1125 09:45:00.008772 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5f4c844d98-nt7dn"] Nov 25 09:45:00 crc kubenswrapper[4776]: I1125 09:45:00.009092 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5f4c844d98-nt7dn" podUID="27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0" containerName="neutron-httpd" containerID="cri-o://d8c7c2e271eae150bda7090f4dde04fe427c80ff86bd65b1e36e0f137aac0a52" gracePeriod=30 Nov 25 09:45:00 crc kubenswrapper[4776]: I1125 09:45:00.009056 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5f4c844d98-nt7dn" podUID="27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0" containerName="neutron-api" containerID="cri-o://780469763d9074be88958dd18def678a1f9122fe313a2bc79714b4d658d275e9" gracePeriod=30 Nov 25 09:45:00 crc kubenswrapper[4776]: I1125 09:45:00.149037 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401065-tf5xr"] Nov 25 09:45:00 crc kubenswrapper[4776]: I1125 09:45:00.150429 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-tf5xr" Nov 25 09:45:00 crc kubenswrapper[4776]: I1125 09:45:00.157393 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 09:45:00 crc kubenswrapper[4776]: I1125 09:45:00.159209 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 09:45:00 crc kubenswrapper[4776]: I1125 09:45:00.167813 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401065-tf5xr"] Nov 25 09:45:00 crc kubenswrapper[4776]: I1125 09:45:00.180688 4776 generic.go:334] "Generic (PLEG): container finished" podID="1b5f86ea-71cb-4e40-a7f5-cf77663097b2" containerID="e5971edb9cbfda783a246c9d315408bc030ae25de88ac8fc101bbe339f9f7b07" exitCode=0 Nov 25 09:45:00 crc kubenswrapper[4776]: I1125 09:45:00.180730 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1b5f86ea-71cb-4e40-a7f5-cf77663097b2","Type":"ContainerDied","Data":"e5971edb9cbfda783a246c9d315408bc030ae25de88ac8fc101bbe339f9f7b07"} Nov 25 09:45:00 crc kubenswrapper[4776]: I1125 09:45:00.326003 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swr8x\" (UniqueName: \"kubernetes.io/projected/ffedd17d-0d6c-4f5c-88b5-a11e9931d1af-kube-api-access-swr8x\") pod \"collect-profiles-29401065-tf5xr\" (UID: \"ffedd17d-0d6c-4f5c-88b5-a11e9931d1af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-tf5xr" Nov 25 09:45:00 crc kubenswrapper[4776]: I1125 09:45:00.326269 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ffedd17d-0d6c-4f5c-88b5-a11e9931d1af-secret-volume\") pod \"collect-profiles-29401065-tf5xr\" (UID: \"ffedd17d-0d6c-4f5c-88b5-a11e9931d1af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-tf5xr" Nov 25 09:45:00 crc kubenswrapper[4776]: I1125 09:45:00.326338 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ffedd17d-0d6c-4f5c-88b5-a11e9931d1af-config-volume\") pod \"collect-profiles-29401065-tf5xr\" (UID: \"ffedd17d-0d6c-4f5c-88b5-a11e9931d1af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-tf5xr" Nov 25 09:45:00 crc kubenswrapper[4776]: I1125 09:45:00.428143 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ffedd17d-0d6c-4f5c-88b5-a11e9931d1af-secret-volume\") pod \"collect-profiles-29401065-tf5xr\" (UID: \"ffedd17d-0d6c-4f5c-88b5-a11e9931d1af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-tf5xr" Nov 25 09:45:00 crc kubenswrapper[4776]: I1125 09:45:00.428217 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ffedd17d-0d6c-4f5c-88b5-a11e9931d1af-config-volume\") pod \"collect-profiles-29401065-tf5xr\" (UID: \"ffedd17d-0d6c-4f5c-88b5-a11e9931d1af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-tf5xr" Nov 25 09:45:00 crc kubenswrapper[4776]: I1125 09:45:00.428337 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swr8x\" (UniqueName: \"kubernetes.io/projected/ffedd17d-0d6c-4f5c-88b5-a11e9931d1af-kube-api-access-swr8x\") pod \"collect-profiles-29401065-tf5xr\" (UID: \"ffedd17d-0d6c-4f5c-88b5-a11e9931d1af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-tf5xr" Nov 25 09:45:00 crc kubenswrapper[4776]: I1125 09:45:00.431804 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ffedd17d-0d6c-4f5c-88b5-a11e9931d1af-config-volume\") pod \"collect-profiles-29401065-tf5xr\" (UID: \"ffedd17d-0d6c-4f5c-88b5-a11e9931d1af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-tf5xr" Nov 25 09:45:00 crc kubenswrapper[4776]: I1125 09:45:00.436722 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ffedd17d-0d6c-4f5c-88b5-a11e9931d1af-secret-volume\") pod \"collect-profiles-29401065-tf5xr\" (UID: \"ffedd17d-0d6c-4f5c-88b5-a11e9931d1af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-tf5xr" Nov 25 09:45:00 crc kubenswrapper[4776]: I1125 09:45:00.449786 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swr8x\" (UniqueName: \"kubernetes.io/projected/ffedd17d-0d6c-4f5c-88b5-a11e9931d1af-kube-api-access-swr8x\") pod \"collect-profiles-29401065-tf5xr\" (UID: \"ffedd17d-0d6c-4f5c-88b5-a11e9931d1af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-tf5xr" Nov 25 09:45:00 crc kubenswrapper[4776]: I1125 09:45:00.529444 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-tf5xr" Nov 25 09:45:01 crc kubenswrapper[4776]: I1125 09:45:01.203875 4776 generic.go:334] "Generic (PLEG): container finished" podID="27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0" containerID="d8c7c2e271eae150bda7090f4dde04fe427c80ff86bd65b1e36e0f137aac0a52" exitCode=0 Nov 25 09:45:01 crc kubenswrapper[4776]: I1125 09:45:01.203940 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5f4c844d98-nt7dn" event={"ID":"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0","Type":"ContainerDied","Data":"d8c7c2e271eae150bda7090f4dde04fe427c80ff86bd65b1e36e0f137aac0a52"} Nov 25 09:45:01 crc kubenswrapper[4776]: I1125 09:45:01.678379 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 25 09:45:02 crc kubenswrapper[4776]: I1125 09:45:02.255548 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:45:02 crc kubenswrapper[4776]: I1125 09:45:02.255962 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b62eda75-27dc-4a93-9ef5-a729483eb7a0" containerName="proxy-httpd" containerID="cri-o://0201f2a461319eda4465429bd9c908e2a8642d9da6bc14ed48d25627d89f3b29" gracePeriod=30 Nov 25 09:45:02 crc kubenswrapper[4776]: I1125 09:45:02.256002 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b62eda75-27dc-4a93-9ef5-a729483eb7a0" containerName="sg-core" containerID="cri-o://a4e627a7fd3e589e8546a2863dea9146033d4da7accb56f9b48af3afc1697807" gracePeriod=30 Nov 25 09:45:02 crc kubenswrapper[4776]: I1125 09:45:02.256058 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b62eda75-27dc-4a93-9ef5-a729483eb7a0" containerName="ceilometer-notification-agent" containerID="cri-o://692f24c910457e9a89a0fb2f8ef6c94883a979330e90c6cbd75277df199ae233" gracePeriod=30 Nov 25 09:45:02 crc kubenswrapper[4776]: I1125 09:45:02.256312 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b62eda75-27dc-4a93-9ef5-a729483eb7a0" containerName="ceilometer-central-agent" containerID="cri-o://26c5b8e121542f7c1cf527f237c648db13a0b1275674c0685082bcf0225acb5e" gracePeriod=30 Nov 25 09:45:03 crc kubenswrapper[4776]: I1125 09:45:03.230827 4776 generic.go:334] "Generic (PLEG): container finished" podID="b62eda75-27dc-4a93-9ef5-a729483eb7a0" containerID="0201f2a461319eda4465429bd9c908e2a8642d9da6bc14ed48d25627d89f3b29" exitCode=0 Nov 25 09:45:03 crc kubenswrapper[4776]: I1125 09:45:03.230865 4776 generic.go:334] "Generic (PLEG): container finished" podID="b62eda75-27dc-4a93-9ef5-a729483eb7a0" containerID="a4e627a7fd3e589e8546a2863dea9146033d4da7accb56f9b48af3afc1697807" exitCode=2 Nov 25 09:45:03 crc kubenswrapper[4776]: I1125 09:45:03.230877 4776 generic.go:334] "Generic (PLEG): container finished" podID="b62eda75-27dc-4a93-9ef5-a729483eb7a0" containerID="26c5b8e121542f7c1cf527f237c648db13a0b1275674c0685082bcf0225acb5e" exitCode=0 Nov 25 09:45:03 crc kubenswrapper[4776]: I1125 09:45:03.230903 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b62eda75-27dc-4a93-9ef5-a729483eb7a0","Type":"ContainerDied","Data":"0201f2a461319eda4465429bd9c908e2a8642d9da6bc14ed48d25627d89f3b29"} Nov 25 09:45:03 crc kubenswrapper[4776]: I1125 09:45:03.230951 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b62eda75-27dc-4a93-9ef5-a729483eb7a0","Type":"ContainerDied","Data":"a4e627a7fd3e589e8546a2863dea9146033d4da7accb56f9b48af3afc1697807"} Nov 25 09:45:03 crc kubenswrapper[4776]: I1125 09:45:03.230962 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b62eda75-27dc-4a93-9ef5-a729483eb7a0","Type":"ContainerDied","Data":"26c5b8e121542f7c1cf527f237c648db13a0b1275674c0685082bcf0225acb5e"} Nov 25 09:45:04 crc kubenswrapper[4776]: I1125 09:45:04.864803 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 09:45:04 crc kubenswrapper[4776]: I1125 09:45:04.865641 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="c595957f-32bd-455a-b212-8f2536102d7f" containerName="glance-log" containerID="cri-o://7e9a37b45f237cc3a95404fab964097152209924df283c3b016813d90db66b84" gracePeriod=30 Nov 25 09:45:04 crc kubenswrapper[4776]: I1125 09:45:04.865706 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="c595957f-32bd-455a-b212-8f2536102d7f" containerName="glance-httpd" containerID="cri-o://3ca0b83197510418dd82fff89f9a0b35f7b3833f3df2f065cda0aa3374493db9" gracePeriod=30 Nov 25 09:45:05 crc kubenswrapper[4776]: I1125 09:45:05.250438 4776 generic.go:334] "Generic (PLEG): container finished" podID="c595957f-32bd-455a-b212-8f2536102d7f" containerID="7e9a37b45f237cc3a95404fab964097152209924df283c3b016813d90db66b84" exitCode=143 Nov 25 09:45:05 crc kubenswrapper[4776]: I1125 09:45:05.250484 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c595957f-32bd-455a-b212-8f2536102d7f","Type":"ContainerDied","Data":"7e9a37b45f237cc3a95404fab964097152209924df283c3b016813d90db66b84"} Nov 25 09:45:06 crc kubenswrapper[4776]: I1125 09:45:06.281084 4776 generic.go:334] "Generic (PLEG): container finished" podID="b62eda75-27dc-4a93-9ef5-a729483eb7a0" containerID="692f24c910457e9a89a0fb2f8ef6c94883a979330e90c6cbd75277df199ae233" exitCode=0 Nov 25 09:45:06 crc kubenswrapper[4776]: I1125 09:45:06.281423 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b62eda75-27dc-4a93-9ef5-a729483eb7a0","Type":"ContainerDied","Data":"692f24c910457e9a89a0fb2f8ef6c94883a979330e90c6cbd75277df199ae233"} Nov 25 09:45:06 crc kubenswrapper[4776]: I1125 09:45:06.284870 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401065-tf5xr"] Nov 25 09:45:06 crc kubenswrapper[4776]: I1125 09:45:06.624112 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5864b7d7fc-g9z2r"] Nov 25 09:45:06 crc kubenswrapper[4776]: W1125 09:45:06.761581 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podffedd17d_0d6c_4f5c_88b5_a11e9931d1af.slice/crio-9ee4dcf540289cdb4c9e237c8fbf68d4c6bb2da4661b83e0176038a26ba6f0ad WatchSource:0}: Error finding container 9ee4dcf540289cdb4c9e237c8fbf68d4c6bb2da4661b83e0176038a26ba6f0ad: Status 404 returned error can't find the container with id 9ee4dcf540289cdb4c9e237c8fbf68d4c6bb2da4661b83e0176038a26ba6f0ad Nov 25 09:45:06 crc kubenswrapper[4776]: W1125 09:45:06.765750 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf89d44cf_1ca0_4581_aa8c_e86e0f3c61d9.slice/crio-54bc9c507715dcff53a21f00cb2fd667946240a7b4842b710dfb49b81a04e354 WatchSource:0}: Error finding container 54bc9c507715dcff53a21f00cb2fd667946240a7b4842b710dfb49b81a04e354: Status 404 returned error can't find the container with id 54bc9c507715dcff53a21f00cb2fd667946240a7b4842b710dfb49b81a04e354 Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.177907 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.205923 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.281373 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-config-data-custom\") pod \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\" (UID: \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\") " Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.282431 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-combined-ca-bundle\") pod \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\" (UID: \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\") " Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.282615 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-etc-machine-id\") pod \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\" (UID: \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\") " Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.282780 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-config-data\") pod \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\" (UID: \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\") " Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.282907 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qrzw\" (UniqueName: \"kubernetes.io/projected/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-kube-api-access-2qrzw\") pod \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\" (UID: \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\") " Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.283019 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-scripts\") pod \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\" (UID: \"1b5f86ea-71cb-4e40-a7f5-cf77663097b2\") " Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.286869 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1b5f86ea-71cb-4e40-a7f5-cf77663097b2" (UID: "1b5f86ea-71cb-4e40-a7f5-cf77663097b2"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.289455 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-scripts" (OuterVolumeSpecName: "scripts") pod "1b5f86ea-71cb-4e40-a7f5-cf77663097b2" (UID: "1b5f86ea-71cb-4e40-a7f5-cf77663097b2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.289514 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1b5f86ea-71cb-4e40-a7f5-cf77663097b2" (UID: "1b5f86ea-71cb-4e40-a7f5-cf77663097b2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.297296 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-kube-api-access-2qrzw" (OuterVolumeSpecName: "kube-api-access-2qrzw") pod "1b5f86ea-71cb-4e40-a7f5-cf77663097b2" (UID: "1b5f86ea-71cb-4e40-a7f5-cf77663097b2"). InnerVolumeSpecName "kube-api-access-2qrzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.304336 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-tf5xr" event={"ID":"ffedd17d-0d6c-4f5c-88b5-a11e9931d1af","Type":"ContainerStarted","Data":"9ee4dcf540289cdb4c9e237c8fbf68d4c6bb2da4661b83e0176038a26ba6f0ad"} Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.324379 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b62eda75-27dc-4a93-9ef5-a729483eb7a0","Type":"ContainerDied","Data":"291ad566091d112b62fe0169f01df0accd1db826a1ec99183d4c8a28556aa0c4"} Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.324437 4776 scope.go:117] "RemoveContainer" containerID="0201f2a461319eda4465429bd9c908e2a8642d9da6bc14ed48d25627d89f3b29" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.324598 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.330807 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5864b7d7fc-g9z2r" event={"ID":"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9","Type":"ContainerStarted","Data":"54bc9c507715dcff53a21f00cb2fd667946240a7b4842b710dfb49b81a04e354"} Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.333644 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1b5f86ea-71cb-4e40-a7f5-cf77663097b2","Type":"ContainerDied","Data":"39d7a267e133109c3eb942eeb89edf7b75238d7e3499584caf39ac794980898c"} Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.333753 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.362000 4776 scope.go:117] "RemoveContainer" containerID="a4e627a7fd3e589e8546a2863dea9146033d4da7accb56f9b48af3afc1697807" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.384433 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5492f\" (UniqueName: \"kubernetes.io/projected/b62eda75-27dc-4a93-9ef5-a729483eb7a0-kube-api-access-5492f\") pod \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\" (UID: \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\") " Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.384517 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b62eda75-27dc-4a93-9ef5-a729483eb7a0-sg-core-conf-yaml\") pod \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\" (UID: \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\") " Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.384566 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b62eda75-27dc-4a93-9ef5-a729483eb7a0-run-httpd\") pod \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\" (UID: \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\") " Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.384948 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b62eda75-27dc-4a93-9ef5-a729483eb7a0-log-httpd\") pod \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\" (UID: \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\") " Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.384991 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b62eda75-27dc-4a93-9ef5-a729483eb7a0-combined-ca-bundle\") pod \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\" (UID: \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\") " Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.385044 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b62eda75-27dc-4a93-9ef5-a729483eb7a0-config-data\") pod \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\" (UID: \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\") " Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.385110 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b62eda75-27dc-4a93-9ef5-a729483eb7a0-scripts\") pod \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\" (UID: \"b62eda75-27dc-4a93-9ef5-a729483eb7a0\") " Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.385488 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b62eda75-27dc-4a93-9ef5-a729483eb7a0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b62eda75-27dc-4a93-9ef5-a729483eb7a0" (UID: "b62eda75-27dc-4a93-9ef5-a729483eb7a0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.385883 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b62eda75-27dc-4a93-9ef5-a729483eb7a0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b62eda75-27dc-4a93-9ef5-a729483eb7a0" (UID: "b62eda75-27dc-4a93-9ef5-a729483eb7a0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.385937 4776 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.385960 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qrzw\" (UniqueName: \"kubernetes.io/projected/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-kube-api-access-2qrzw\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.385975 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.385987 4776 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b62eda75-27dc-4a93-9ef5-a729483eb7a0-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.385998 4776 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.393043 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b62eda75-27dc-4a93-9ef5-a729483eb7a0-scripts" (OuterVolumeSpecName: "scripts") pod "b62eda75-27dc-4a93-9ef5-a729483eb7a0" (UID: "b62eda75-27dc-4a93-9ef5-a729483eb7a0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.396813 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b62eda75-27dc-4a93-9ef5-a729483eb7a0-kube-api-access-5492f" (OuterVolumeSpecName: "kube-api-access-5492f") pod "b62eda75-27dc-4a93-9ef5-a729483eb7a0" (UID: "b62eda75-27dc-4a93-9ef5-a729483eb7a0"). InnerVolumeSpecName "kube-api-access-5492f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.404245 4776 scope.go:117] "RemoveContainer" containerID="692f24c910457e9a89a0fb2f8ef6c94883a979330e90c6cbd75277df199ae233" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.426200 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1b5f86ea-71cb-4e40-a7f5-cf77663097b2" (UID: "1b5f86ea-71cb-4e40-a7f5-cf77663097b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.435958 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b62eda75-27dc-4a93-9ef5-a729483eb7a0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b62eda75-27dc-4a93-9ef5-a729483eb7a0" (UID: "b62eda75-27dc-4a93-9ef5-a729483eb7a0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.444977 4776 scope.go:117] "RemoveContainer" containerID="26c5b8e121542f7c1cf527f237c648db13a0b1275674c0685082bcf0225acb5e" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.468534 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-config-data" (OuterVolumeSpecName: "config-data") pod "1b5f86ea-71cb-4e40-a7f5-cf77663097b2" (UID: "1b5f86ea-71cb-4e40-a7f5-cf77663097b2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.468662 4776 scope.go:117] "RemoveContainer" containerID="66ab8a7afff10625abbd356ae5d00854f89dbc0981db05c07f3a3b986fea6221" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.487381 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.487408 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b62eda75-27dc-4a93-9ef5-a729483eb7a0-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.487418 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b5f86ea-71cb-4e40-a7f5-cf77663097b2-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.487427 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5492f\" (UniqueName: \"kubernetes.io/projected/b62eda75-27dc-4a93-9ef5-a729483eb7a0-kube-api-access-5492f\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.487437 4776 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b62eda75-27dc-4a93-9ef5-a729483eb7a0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.487445 4776 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b62eda75-27dc-4a93-9ef5-a729483eb7a0-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.492290 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b62eda75-27dc-4a93-9ef5-a729483eb7a0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b62eda75-27dc-4a93-9ef5-a729483eb7a0" (UID: "b62eda75-27dc-4a93-9ef5-a729483eb7a0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.492313 4776 scope.go:117] "RemoveContainer" containerID="e5971edb9cbfda783a246c9d315408bc030ae25de88ac8fc101bbe339f9f7b07" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.527182 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b62eda75-27dc-4a93-9ef5-a729483eb7a0-config-data" (OuterVolumeSpecName: "config-data") pod "b62eda75-27dc-4a93-9ef5-a729483eb7a0" (UID: "b62eda75-27dc-4a93-9ef5-a729483eb7a0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.591671 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b62eda75-27dc-4a93-9ef5-a729483eb7a0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.591729 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b62eda75-27dc-4a93-9ef5-a729483eb7a0-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.701719 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.730132 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.762828 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.803373 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.841935 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:45:07 crc kubenswrapper[4776]: E1125 09:45:07.842333 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b5f86ea-71cb-4e40-a7f5-cf77663097b2" containerName="probe" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.842347 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b5f86ea-71cb-4e40-a7f5-cf77663097b2" containerName="probe" Nov 25 09:45:07 crc kubenswrapper[4776]: E1125 09:45:07.842364 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b62eda75-27dc-4a93-9ef5-a729483eb7a0" containerName="sg-core" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.842371 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="b62eda75-27dc-4a93-9ef5-a729483eb7a0" containerName="sg-core" Nov 25 09:45:07 crc kubenswrapper[4776]: E1125 09:45:07.842388 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b62eda75-27dc-4a93-9ef5-a729483eb7a0" containerName="ceilometer-central-agent" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.842395 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="b62eda75-27dc-4a93-9ef5-a729483eb7a0" containerName="ceilometer-central-agent" Nov 25 09:45:07 crc kubenswrapper[4776]: E1125 09:45:07.842408 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b62eda75-27dc-4a93-9ef5-a729483eb7a0" containerName="ceilometer-notification-agent" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.842413 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="b62eda75-27dc-4a93-9ef5-a729483eb7a0" containerName="ceilometer-notification-agent" Nov 25 09:45:07 crc kubenswrapper[4776]: E1125 09:45:07.842421 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b62eda75-27dc-4a93-9ef5-a729483eb7a0" containerName="proxy-httpd" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.842427 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="b62eda75-27dc-4a93-9ef5-a729483eb7a0" containerName="proxy-httpd" Nov 25 09:45:07 crc kubenswrapper[4776]: E1125 09:45:07.842436 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b5f86ea-71cb-4e40-a7f5-cf77663097b2" containerName="cinder-scheduler" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.842442 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b5f86ea-71cb-4e40-a7f5-cf77663097b2" containerName="cinder-scheduler" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.842614 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b5f86ea-71cb-4e40-a7f5-cf77663097b2" containerName="cinder-scheduler" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.842627 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="b62eda75-27dc-4a93-9ef5-a729483eb7a0" containerName="sg-core" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.842634 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b5f86ea-71cb-4e40-a7f5-cf77663097b2" containerName="probe" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.842644 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="b62eda75-27dc-4a93-9ef5-a729483eb7a0" containerName="ceilometer-central-agent" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.842652 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="b62eda75-27dc-4a93-9ef5-a729483eb7a0" containerName="proxy-httpd" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.842667 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="b62eda75-27dc-4a93-9ef5-a729483eb7a0" containerName="ceilometer-notification-agent" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.844253 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.850368 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.850671 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.878893 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.938194 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.939669 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.956514 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 25 09:45:07 crc kubenswrapper[4776]: I1125 09:45:07.973426 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.007336 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc920d71-4308-4d90-a891-8b760651ccc9-scripts\") pod \"cinder-scheduler-0\" (UID: \"bc920d71-4308-4d90-a891-8b760651ccc9\") " pod="openstack/cinder-scheduler-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.007385 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9686x\" (UniqueName: \"kubernetes.io/projected/bc920d71-4308-4d90-a891-8b760651ccc9-kube-api-access-9686x\") pod \"cinder-scheduler-0\" (UID: \"bc920d71-4308-4d90-a891-8b760651ccc9\") " pod="openstack/cinder-scheduler-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.007444 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc920d71-4308-4d90-a891-8b760651ccc9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bc920d71-4308-4d90-a891-8b760651ccc9\") " pod="openstack/cinder-scheduler-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.007488 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/768b7f4e-d631-4cc7-ad91-7e1099c8b354-scripts\") pod \"ceilometer-0\" (UID: \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\") " pod="openstack/ceilometer-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.007512 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bc920d71-4308-4d90-a891-8b760651ccc9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bc920d71-4308-4d90-a891-8b760651ccc9\") " pod="openstack/cinder-scheduler-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.007589 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/768b7f4e-d631-4cc7-ad91-7e1099c8b354-config-data\") pod \"ceilometer-0\" (UID: \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\") " pod="openstack/ceilometer-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.007641 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/768b7f4e-d631-4cc7-ad91-7e1099c8b354-run-httpd\") pod \"ceilometer-0\" (UID: \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\") " pod="openstack/ceilometer-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.007700 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bc920d71-4308-4d90-a891-8b760651ccc9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bc920d71-4308-4d90-a891-8b760651ccc9\") " pod="openstack/cinder-scheduler-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.007730 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/768b7f4e-d631-4cc7-ad91-7e1099c8b354-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\") " pod="openstack/ceilometer-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.007774 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/768b7f4e-d631-4cc7-ad91-7e1099c8b354-log-httpd\") pod \"ceilometer-0\" (UID: \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\") " pod="openstack/ceilometer-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.007857 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md7lz\" (UniqueName: \"kubernetes.io/projected/768b7f4e-d631-4cc7-ad91-7e1099c8b354-kube-api-access-md7lz\") pod \"ceilometer-0\" (UID: \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\") " pod="openstack/ceilometer-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.007914 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc920d71-4308-4d90-a891-8b760651ccc9-config-data\") pod \"cinder-scheduler-0\" (UID: \"bc920d71-4308-4d90-a891-8b760651ccc9\") " pod="openstack/cinder-scheduler-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.007948 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/768b7f4e-d631-4cc7-ad91-7e1099c8b354-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\") " pod="openstack/ceilometer-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.109785 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc920d71-4308-4d90-a891-8b760651ccc9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bc920d71-4308-4d90-a891-8b760651ccc9\") " pod="openstack/cinder-scheduler-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.110099 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/768b7f4e-d631-4cc7-ad91-7e1099c8b354-scripts\") pod \"ceilometer-0\" (UID: \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\") " pod="openstack/ceilometer-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.110238 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bc920d71-4308-4d90-a891-8b760651ccc9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bc920d71-4308-4d90-a891-8b760651ccc9\") " pod="openstack/cinder-scheduler-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.110385 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/768b7f4e-d631-4cc7-ad91-7e1099c8b354-config-data\") pod \"ceilometer-0\" (UID: \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\") " pod="openstack/ceilometer-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.110547 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/768b7f4e-d631-4cc7-ad91-7e1099c8b354-run-httpd\") pod \"ceilometer-0\" (UID: \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\") " pod="openstack/ceilometer-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.110691 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bc920d71-4308-4d90-a891-8b760651ccc9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bc920d71-4308-4d90-a891-8b760651ccc9\") " pod="openstack/cinder-scheduler-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.110811 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/768b7f4e-d631-4cc7-ad91-7e1099c8b354-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\") " pod="openstack/ceilometer-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.110927 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/768b7f4e-d631-4cc7-ad91-7e1099c8b354-log-httpd\") pod \"ceilometer-0\" (UID: \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\") " pod="openstack/ceilometer-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.111096 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md7lz\" (UniqueName: \"kubernetes.io/projected/768b7f4e-d631-4cc7-ad91-7e1099c8b354-kube-api-access-md7lz\") pod \"ceilometer-0\" (UID: \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\") " pod="openstack/ceilometer-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.110817 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bc920d71-4308-4d90-a891-8b760651ccc9-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bc920d71-4308-4d90-a891-8b760651ccc9\") " pod="openstack/cinder-scheduler-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.110956 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/768b7f4e-d631-4cc7-ad91-7e1099c8b354-run-httpd\") pod \"ceilometer-0\" (UID: \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\") " pod="openstack/ceilometer-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.111288 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/768b7f4e-d631-4cc7-ad91-7e1099c8b354-log-httpd\") pod \"ceilometer-0\" (UID: \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\") " pod="openstack/ceilometer-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.111398 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc920d71-4308-4d90-a891-8b760651ccc9-config-data\") pod \"cinder-scheduler-0\" (UID: \"bc920d71-4308-4d90-a891-8b760651ccc9\") " pod="openstack/cinder-scheduler-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.111516 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/768b7f4e-d631-4cc7-ad91-7e1099c8b354-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\") " pod="openstack/ceilometer-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.111671 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc920d71-4308-4d90-a891-8b760651ccc9-scripts\") pod \"cinder-scheduler-0\" (UID: \"bc920d71-4308-4d90-a891-8b760651ccc9\") " pod="openstack/cinder-scheduler-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.111820 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9686x\" (UniqueName: \"kubernetes.io/projected/bc920d71-4308-4d90-a891-8b760651ccc9-kube-api-access-9686x\") pod \"cinder-scheduler-0\" (UID: \"bc920d71-4308-4d90-a891-8b760651ccc9\") " pod="openstack/cinder-scheduler-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.116315 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/768b7f4e-d631-4cc7-ad91-7e1099c8b354-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\") " pod="openstack/ceilometer-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.116882 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/768b7f4e-d631-4cc7-ad91-7e1099c8b354-scripts\") pod \"ceilometer-0\" (UID: \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\") " pod="openstack/ceilometer-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.117672 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc920d71-4308-4d90-a891-8b760651ccc9-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bc920d71-4308-4d90-a891-8b760651ccc9\") " pod="openstack/cinder-scheduler-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.119218 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bc920d71-4308-4d90-a891-8b760651ccc9-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bc920d71-4308-4d90-a891-8b760651ccc9\") " pod="openstack/cinder-scheduler-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.119584 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc920d71-4308-4d90-a891-8b760651ccc9-scripts\") pod \"cinder-scheduler-0\" (UID: \"bc920d71-4308-4d90-a891-8b760651ccc9\") " pod="openstack/cinder-scheduler-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.128532 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc920d71-4308-4d90-a891-8b760651ccc9-config-data\") pod \"cinder-scheduler-0\" (UID: \"bc920d71-4308-4d90-a891-8b760651ccc9\") " pod="openstack/cinder-scheduler-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.129514 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/768b7f4e-d631-4cc7-ad91-7e1099c8b354-config-data\") pod \"ceilometer-0\" (UID: \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\") " pod="openstack/ceilometer-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.130138 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/768b7f4e-d631-4cc7-ad91-7e1099c8b354-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\") " pod="openstack/ceilometer-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.131644 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9686x\" (UniqueName: \"kubernetes.io/projected/bc920d71-4308-4d90-a891-8b760651ccc9-kube-api-access-9686x\") pod \"cinder-scheduler-0\" (UID: \"bc920d71-4308-4d90-a891-8b760651ccc9\") " pod="openstack/cinder-scheduler-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.135586 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md7lz\" (UniqueName: \"kubernetes.io/projected/768b7f4e-d631-4cc7-ad91-7e1099c8b354-kube-api-access-md7lz\") pod \"ceilometer-0\" (UID: \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\") " pod="openstack/ceilometer-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.216747 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.315896 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.355715 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"3fc11cf2-c6c7-4250-b7f4-9c68d78152eb","Type":"ContainerStarted","Data":"5b70abe095c3919e69bde8512f0bf19cca9ec9011ee75d534001611515145f6c"} Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.371371 4776 generic.go:334] "Generic (PLEG): container finished" podID="ffedd17d-0d6c-4f5c-88b5-a11e9931d1af" containerID="190f0504e8bf2e09fc36c3fc20e9878eb5565206fdf350f5381773a2a111c697" exitCode=0 Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.371477 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-tf5xr" event={"ID":"ffedd17d-0d6c-4f5c-88b5-a11e9931d1af","Type":"ContainerDied","Data":"190f0504e8bf2e09fc36c3fc20e9878eb5565206fdf350f5381773a2a111c697"} Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.397802 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.781285651 podStartE2EDuration="17.39777788s" podCreationTimestamp="2025-11-25 09:44:51 +0000 UTC" firstStartedPulling="2025-11-25 09:44:52.268626609 +0000 UTC m=+1237.309686162" lastFinishedPulling="2025-11-25 09:45:06.885118838 +0000 UTC m=+1251.926178391" observedRunningTime="2025-11-25 09:45:08.378338372 +0000 UTC m=+1253.419397925" watchObservedRunningTime="2025-11-25 09:45:08.39777788 +0000 UTC m=+1253.438837443" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.413793 4776 generic.go:334] "Generic (PLEG): container finished" podID="c595957f-32bd-455a-b212-8f2536102d7f" containerID="3ca0b83197510418dd82fff89f9a0b35f7b3833f3df2f065cda0aa3374493db9" exitCode=0 Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.414123 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c595957f-32bd-455a-b212-8f2536102d7f","Type":"ContainerDied","Data":"3ca0b83197510418dd82fff89f9a0b35f7b3833f3df2f065cda0aa3374493db9"} Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.449410 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5864b7d7fc-g9z2r" event={"ID":"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9","Type":"ContainerStarted","Data":"94b41979aa9c10d3bd3737ef33206925359824fb82b712cb4748abcf10f1e055"} Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.449515 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5864b7d7fc-g9z2r" event={"ID":"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9","Type":"ContainerStarted","Data":"8c4015a72b16a36f9de326c38c3b6c1efa071b368ac1bacb90a14ea00d371022"} Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.449579 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.449616 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.507824 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-5864b7d7fc-g9z2r" podStartSLOduration=10.507793548 podStartE2EDuration="10.507793548s" podCreationTimestamp="2025-11-25 09:44:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:45:08.483916039 +0000 UTC m=+1253.524975602" watchObservedRunningTime="2025-11-25 09:45:08.507793548 +0000 UTC m=+1253.548853101" Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.752118 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:45:08 crc kubenswrapper[4776]: I1125 09:45:08.887491 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.026237 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.047225 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c595957f-32bd-455a-b212-8f2536102d7f-logs\") pod \"c595957f-32bd-455a-b212-8f2536102d7f\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.047271 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"c595957f-32bd-455a-b212-8f2536102d7f\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.047311 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c595957f-32bd-455a-b212-8f2536102d7f-httpd-run\") pod \"c595957f-32bd-455a-b212-8f2536102d7f\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.047338 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c595957f-32bd-455a-b212-8f2536102d7f-combined-ca-bundle\") pod \"c595957f-32bd-455a-b212-8f2536102d7f\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.047371 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c595957f-32bd-455a-b212-8f2536102d7f-internal-tls-certs\") pod \"c595957f-32bd-455a-b212-8f2536102d7f\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.047763 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c595957f-32bd-455a-b212-8f2536102d7f-config-data\") pod \"c595957f-32bd-455a-b212-8f2536102d7f\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.047814 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbx2d\" (UniqueName: \"kubernetes.io/projected/c595957f-32bd-455a-b212-8f2536102d7f-kube-api-access-sbx2d\") pod \"c595957f-32bd-455a-b212-8f2536102d7f\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.047919 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c595957f-32bd-455a-b212-8f2536102d7f-scripts\") pod \"c595957f-32bd-455a-b212-8f2536102d7f\" (UID: \"c595957f-32bd-455a-b212-8f2536102d7f\") " Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.049228 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c595957f-32bd-455a-b212-8f2536102d7f-logs" (OuterVolumeSpecName: "logs") pod "c595957f-32bd-455a-b212-8f2536102d7f" (UID: "c595957f-32bd-455a-b212-8f2536102d7f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.049889 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c595957f-32bd-455a-b212-8f2536102d7f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c595957f-32bd-455a-b212-8f2536102d7f" (UID: "c595957f-32bd-455a-b212-8f2536102d7f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.050490 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "c595957f-32bd-455a-b212-8f2536102d7f" (UID: "c595957f-32bd-455a-b212-8f2536102d7f"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.052977 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c595957f-32bd-455a-b212-8f2536102d7f-scripts" (OuterVolumeSpecName: "scripts") pod "c595957f-32bd-455a-b212-8f2536102d7f" (UID: "c595957f-32bd-455a-b212-8f2536102d7f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.060310 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c595957f-32bd-455a-b212-8f2536102d7f-kube-api-access-sbx2d" (OuterVolumeSpecName: "kube-api-access-sbx2d") pod "c595957f-32bd-455a-b212-8f2536102d7f" (UID: "c595957f-32bd-455a-b212-8f2536102d7f"). InnerVolumeSpecName "kube-api-access-sbx2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.107431 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c595957f-32bd-455a-b212-8f2536102d7f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c595957f-32bd-455a-b212-8f2536102d7f" (UID: "c595957f-32bd-455a-b212-8f2536102d7f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.151348 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c595957f-32bd-455a-b212-8f2536102d7f-logs\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.152546 4776 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.158159 4776 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c595957f-32bd-455a-b212-8f2536102d7f-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.158201 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c595957f-32bd-455a-b212-8f2536102d7f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.158220 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbx2d\" (UniqueName: \"kubernetes.io/projected/c595957f-32bd-455a-b212-8f2536102d7f-kube-api-access-sbx2d\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.158233 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c595957f-32bd-455a-b212-8f2536102d7f-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.168601 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c595957f-32bd-455a-b212-8f2536102d7f-config-data" (OuterVolumeSpecName: "config-data") pod "c595957f-32bd-455a-b212-8f2536102d7f" (UID: "c595957f-32bd-455a-b212-8f2536102d7f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.181199 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c595957f-32bd-455a-b212-8f2536102d7f-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c595957f-32bd-455a-b212-8f2536102d7f" (UID: "c595957f-32bd-455a-b212-8f2536102d7f"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.186387 4776 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.260858 4776 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.260908 4776 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c595957f-32bd-455a-b212-8f2536102d7f-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.260923 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c595957f-32bd-455a-b212-8f2536102d7f-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.479614 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"768b7f4e-d631-4cc7-ad91-7e1099c8b354","Type":"ContainerStarted","Data":"4c55ce75f985912e742dc4721c3d41f8d89b500b82fa5f2748e397b897f3fea0"} Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.493835 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bc920d71-4308-4d90-a891-8b760651ccc9","Type":"ContainerStarted","Data":"a26a73284ec7ca88f3d4569feec4777b80aa3ca2f76bd5f8814d4c904b763b0d"} Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.513493 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c595957f-32bd-455a-b212-8f2536102d7f","Type":"ContainerDied","Data":"65ba093605052686fb8026a5ab79fac757efd95fc530300ccdbf7a8e1ef95bd9"} Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.513571 4776 scope.go:117] "RemoveContainer" containerID="3ca0b83197510418dd82fff89f9a0b35f7b3833f3df2f065cda0aa3374493db9" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.513732 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.600139 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.619529 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.660726 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 09:45:09 crc kubenswrapper[4776]: E1125 09:45:09.661330 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c595957f-32bd-455a-b212-8f2536102d7f" containerName="glance-httpd" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.661355 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="c595957f-32bd-455a-b212-8f2536102d7f" containerName="glance-httpd" Nov 25 09:45:09 crc kubenswrapper[4776]: E1125 09:45:09.661377 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c595957f-32bd-455a-b212-8f2536102d7f" containerName="glance-log" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.661385 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="c595957f-32bd-455a-b212-8f2536102d7f" containerName="glance-log" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.661615 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="c595957f-32bd-455a-b212-8f2536102d7f" containerName="glance-log" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.661632 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="c595957f-32bd-455a-b212-8f2536102d7f" containerName="glance-httpd" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.662784 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.669134 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.682620 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.686319 4776 scope.go:117] "RemoveContainer" containerID="7e9a37b45f237cc3a95404fab964097152209924df283c3b016813d90db66b84" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.811137 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b708b100-166c-49b1-a47b-76d47ba2c6e5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.811207 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.811238 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b708b100-166c-49b1-a47b-76d47ba2c6e5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.811279 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkxgs\" (UniqueName: \"kubernetes.io/projected/b708b100-166c-49b1-a47b-76d47ba2c6e5-kube-api-access-dkxgs\") pod \"glance-default-internal-api-0\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.811328 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b708b100-166c-49b1-a47b-76d47ba2c6e5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.811346 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b708b100-166c-49b1-a47b-76d47ba2c6e5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.811364 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b708b100-166c-49b1-a47b-76d47ba2c6e5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.811407 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b708b100-166c-49b1-a47b-76d47ba2c6e5-logs\") pod \"glance-default-internal-api-0\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.843359 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b5f86ea-71cb-4e40-a7f5-cf77663097b2" path="/var/lib/kubelet/pods/1b5f86ea-71cb-4e40-a7f5-cf77663097b2/volumes" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.844667 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b62eda75-27dc-4a93-9ef5-a729483eb7a0" path="/var/lib/kubelet/pods/b62eda75-27dc-4a93-9ef5-a729483eb7a0/volumes" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.845908 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c595957f-32bd-455a-b212-8f2536102d7f" path="/var/lib/kubelet/pods/c595957f-32bd-455a-b212-8f2536102d7f/volumes" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.846633 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.917509 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.918510 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b708b100-166c-49b1-a47b-76d47ba2c6e5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.918381 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-internal-api-0" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.919394 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkxgs\" (UniqueName: \"kubernetes.io/projected/b708b100-166c-49b1-a47b-76d47ba2c6e5-kube-api-access-dkxgs\") pod \"glance-default-internal-api-0\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.919514 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b708b100-166c-49b1-a47b-76d47ba2c6e5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.919566 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b708b100-166c-49b1-a47b-76d47ba2c6e5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.919594 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b708b100-166c-49b1-a47b-76d47ba2c6e5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.919637 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b708b100-166c-49b1-a47b-76d47ba2c6e5-logs\") pod \"glance-default-internal-api-0\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.919765 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b708b100-166c-49b1-a47b-76d47ba2c6e5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.920778 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b708b100-166c-49b1-a47b-76d47ba2c6e5-logs\") pod \"glance-default-internal-api-0\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.921349 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b708b100-166c-49b1-a47b-76d47ba2c6e5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.927974 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b708b100-166c-49b1-a47b-76d47ba2c6e5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.928500 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b708b100-166c-49b1-a47b-76d47ba2c6e5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.929165 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b708b100-166c-49b1-a47b-76d47ba2c6e5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.931811 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b708b100-166c-49b1-a47b-76d47ba2c6e5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.942056 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkxgs\" (UniqueName: \"kubernetes.io/projected/b708b100-166c-49b1-a47b-76d47ba2c6e5-kube-api-access-dkxgs\") pod \"glance-default-internal-api-0\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:45:09 crc kubenswrapper[4776]: I1125 09:45:09.965584 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-internal-api-0\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " pod="openstack/glance-default-internal-api-0" Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.046641 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.282206 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-tf5xr" Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.431046 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ffedd17d-0d6c-4f5c-88b5-a11e9931d1af-config-volume\") pod \"ffedd17d-0d6c-4f5c-88b5-a11e9931d1af\" (UID: \"ffedd17d-0d6c-4f5c-88b5-a11e9931d1af\") " Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.431666 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swr8x\" (UniqueName: \"kubernetes.io/projected/ffedd17d-0d6c-4f5c-88b5-a11e9931d1af-kube-api-access-swr8x\") pod \"ffedd17d-0d6c-4f5c-88b5-a11e9931d1af\" (UID: \"ffedd17d-0d6c-4f5c-88b5-a11e9931d1af\") " Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.431740 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ffedd17d-0d6c-4f5c-88b5-a11e9931d1af-secret-volume\") pod \"ffedd17d-0d6c-4f5c-88b5-a11e9931d1af\" (UID: \"ffedd17d-0d6c-4f5c-88b5-a11e9931d1af\") " Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.435465 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffedd17d-0d6c-4f5c-88b5-a11e9931d1af-config-volume" (OuterVolumeSpecName: "config-volume") pod "ffedd17d-0d6c-4f5c-88b5-a11e9931d1af" (UID: "ffedd17d-0d6c-4f5c-88b5-a11e9931d1af"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.444290 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffedd17d-0d6c-4f5c-88b5-a11e9931d1af-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ffedd17d-0d6c-4f5c-88b5-a11e9931d1af" (UID: "ffedd17d-0d6c-4f5c-88b5-a11e9931d1af"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.454351 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffedd17d-0d6c-4f5c-88b5-a11e9931d1af-kube-api-access-swr8x" (OuterVolumeSpecName: "kube-api-access-swr8x") pod "ffedd17d-0d6c-4f5c-88b5-a11e9931d1af" (UID: "ffedd17d-0d6c-4f5c-88b5-a11e9931d1af"). InnerVolumeSpecName "kube-api-access-swr8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.536551 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5f4c844d98-nt7dn" Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.537961 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swr8x\" (UniqueName: \"kubernetes.io/projected/ffedd17d-0d6c-4f5c-88b5-a11e9931d1af-kube-api-access-swr8x\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.537988 4776 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ffedd17d-0d6c-4f5c-88b5-a11e9931d1af-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.538002 4776 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ffedd17d-0d6c-4f5c-88b5-a11e9931d1af-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.558650 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bc920d71-4308-4d90-a891-8b760651ccc9","Type":"ContainerStarted","Data":"6bffff98e55f5e2fe0110f4a9d127c1dff4b4c552606101a667c0fb53d5e9b16"} Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.600194 4776 generic.go:334] "Generic (PLEG): container finished" podID="27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0" containerID="780469763d9074be88958dd18def678a1f9122fe313a2bc79714b4d658d275e9" exitCode=0 Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.600311 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5f4c844d98-nt7dn" Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.600317 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5f4c844d98-nt7dn" event={"ID":"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0","Type":"ContainerDied","Data":"780469763d9074be88958dd18def678a1f9122fe313a2bc79714b4d658d275e9"} Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.600385 4776 scope.go:117] "RemoveContainer" containerID="d8c7c2e271eae150bda7090f4dde04fe427c80ff86bd65b1e36e0f137aac0a52" Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.604495 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"768b7f4e-d631-4cc7-ad91-7e1099c8b354","Type":"ContainerStarted","Data":"6515b87d6beb7f876edae119be02a45958d18ae4c801f6d415a59604678f66d8"} Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.607823 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-tf5xr" event={"ID":"ffedd17d-0d6c-4f5c-88b5-a11e9931d1af","Type":"ContainerDied","Data":"9ee4dcf540289cdb4c9e237c8fbf68d4c6bb2da4661b83e0176038a26ba6f0ad"} Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.607856 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ee4dcf540289cdb4c9e237c8fbf68d4c6bb2da4661b83e0176038a26ba6f0ad" Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.607907 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-tf5xr" Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.639453 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-combined-ca-bundle\") pod \"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0\" (UID: \"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0\") " Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.639589 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xts9\" (UniqueName: \"kubernetes.io/projected/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-kube-api-access-7xts9\") pod \"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0\" (UID: \"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0\") " Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.639644 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-httpd-config\") pod \"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0\" (UID: \"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0\") " Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.639733 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-ovndb-tls-certs\") pod \"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0\" (UID: \"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0\") " Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.639771 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-config\") pod \"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0\" (UID: \"27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0\") " Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.671031 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0" (UID: "27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.680644 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-kube-api-access-7xts9" (OuterVolumeSpecName: "kube-api-access-7xts9") pod "27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0" (UID: "27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0"). InnerVolumeSpecName "kube-api-access-7xts9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.682129 4776 scope.go:117] "RemoveContainer" containerID="780469763d9074be88958dd18def678a1f9122fe313a2bc79714b4d658d275e9" Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.692210 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.692445 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="e06b81f6-6161-46ae-9830-cd735c64d2f0" containerName="kube-state-metrics" containerID="cri-o://51ee7ef170e152f40514f8d16bd62c383594344be3ed28e799c200c3721328e7" gracePeriod=30 Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.743862 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xts9\" (UniqueName: \"kubernetes.io/projected/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-kube-api-access-7xts9\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.744441 4776 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.774334 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-config" (OuterVolumeSpecName: "config") pod "27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0" (UID: "27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.783390 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0" (UID: "27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.807022 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0" (UID: "27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.847516 4776 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.847557 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.847571 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.931988 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.957001 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5f4c844d98-nt7dn"] Nov 25 09:45:10 crc kubenswrapper[4776]: I1125 09:45:10.967553 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5f4c844d98-nt7dn"] Nov 25 09:45:10 crc kubenswrapper[4776]: W1125 09:45:10.973827 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb708b100_166c_49b1_a47b_76d47ba2c6e5.slice/crio-928800eb5246288399c4e346cf14e1b06defe30c42280c8cbd4ccdd157393af7 WatchSource:0}: Error finding container 928800eb5246288399c4e346cf14e1b06defe30c42280c8cbd4ccdd157393af7: Status 404 returned error can't find the container with id 928800eb5246288399c4e346cf14e1b06defe30c42280c8cbd4ccdd157393af7 Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.418345 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.563530 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g46km\" (UniqueName: \"kubernetes.io/projected/e06b81f6-6161-46ae-9830-cd735c64d2f0-kube-api-access-g46km\") pod \"e06b81f6-6161-46ae-9830-cd735c64d2f0\" (UID: \"e06b81f6-6161-46ae-9830-cd735c64d2f0\") " Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.574606 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e06b81f6-6161-46ae-9830-cd735c64d2f0-kube-api-access-g46km" (OuterVolumeSpecName: "kube-api-access-g46km") pod "e06b81f6-6161-46ae-9830-cd735c64d2f0" (UID: "e06b81f6-6161-46ae-9830-cd735c64d2f0"). InnerVolumeSpecName "kube-api-access-g46km". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.631795 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bc920d71-4308-4d90-a891-8b760651ccc9","Type":"ContainerStarted","Data":"c9f1b69c74818d42cf402965ae75f33a682f61bb14e9874a7f7904a0e8298e5e"} Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.646882 4776 generic.go:334] "Generic (PLEG): container finished" podID="e06b81f6-6161-46ae-9830-cd735c64d2f0" containerID="51ee7ef170e152f40514f8d16bd62c383594344be3ed28e799c200c3721328e7" exitCode=2 Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.646988 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e06b81f6-6161-46ae-9830-cd735c64d2f0","Type":"ContainerDied","Data":"51ee7ef170e152f40514f8d16bd62c383594344be3ed28e799c200c3721328e7"} Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.647022 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e06b81f6-6161-46ae-9830-cd735c64d2f0","Type":"ContainerDied","Data":"781f22edcf0938a2b04712393256587ccd5b53a5cb9b50329ea8a3dc55ebe642"} Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.647042 4776 scope.go:117] "RemoveContainer" containerID="51ee7ef170e152f40514f8d16bd62c383594344be3ed28e799c200c3721328e7" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.647208 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.661056 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.66103058 podStartE2EDuration="4.66103058s" podCreationTimestamp="2025-11-25 09:45:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:45:11.658055446 +0000 UTC m=+1256.699114999" watchObservedRunningTime="2025-11-25 09:45:11.66103058 +0000 UTC m=+1256.702090143" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.663423 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b708b100-166c-49b1-a47b-76d47ba2c6e5","Type":"ContainerStarted","Data":"928800eb5246288399c4e346cf14e1b06defe30c42280c8cbd4ccdd157393af7"} Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.667651 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g46km\" (UniqueName: \"kubernetes.io/projected/e06b81f6-6161-46ae-9830-cd735c64d2f0-kube-api-access-g46km\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.740829 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0" path="/var/lib/kubelet/pods/27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0/volumes" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.741692 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.744204 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.755950 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 09:45:11 crc kubenswrapper[4776]: E1125 09:45:11.756402 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e06b81f6-6161-46ae-9830-cd735c64d2f0" containerName="kube-state-metrics" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.756421 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="e06b81f6-6161-46ae-9830-cd735c64d2f0" containerName="kube-state-metrics" Nov 25 09:45:11 crc kubenswrapper[4776]: E1125 09:45:11.756433 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0" containerName="neutron-api" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.756440 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0" containerName="neutron-api" Nov 25 09:45:11 crc kubenswrapper[4776]: E1125 09:45:11.756453 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0" containerName="neutron-httpd" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.756459 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0" containerName="neutron-httpd" Nov 25 09:45:11 crc kubenswrapper[4776]: E1125 09:45:11.756489 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffedd17d-0d6c-4f5c-88b5-a11e9931d1af" containerName="collect-profiles" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.756494 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffedd17d-0d6c-4f5c-88b5-a11e9931d1af" containerName="collect-profiles" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.756663 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0" containerName="neutron-httpd" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.756680 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="27cb0fd2-6e6d-4e2a-b5d8-a1940c1ef8b0" containerName="neutron-api" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.756689 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffedd17d-0d6c-4f5c-88b5-a11e9931d1af" containerName="collect-profiles" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.756702 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="e06b81f6-6161-46ae-9830-cd735c64d2f0" containerName="kube-state-metrics" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.757299 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.772165 4776 scope.go:117] "RemoveContainer" containerID="51ee7ef170e152f40514f8d16bd62c383594344be3ed28e799c200c3721328e7" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.776121 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.776301 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 25 09:45:11 crc kubenswrapper[4776]: E1125 09:45:11.776813 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51ee7ef170e152f40514f8d16bd62c383594344be3ed28e799c200c3721328e7\": container with ID starting with 51ee7ef170e152f40514f8d16bd62c383594344be3ed28e799c200c3721328e7 not found: ID does not exist" containerID="51ee7ef170e152f40514f8d16bd62c383594344be3ed28e799c200c3721328e7" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.776848 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51ee7ef170e152f40514f8d16bd62c383594344be3ed28e799c200c3721328e7"} err="failed to get container status \"51ee7ef170e152f40514f8d16bd62c383594344be3ed28e799c200c3721328e7\": rpc error: code = NotFound desc = could not find container \"51ee7ef170e152f40514f8d16bd62c383594344be3ed28e799c200c3721328e7\": container with ID starting with 51ee7ef170e152f40514f8d16bd62c383594344be3ed28e799c200c3721328e7 not found: ID does not exist" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.782412 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.872249 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09f0463c-f091-495e-9773-c07fe902c200-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"09f0463c-f091-495e-9773-c07fe902c200\") " pod="openstack/kube-state-metrics-0" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.872355 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfg9g\" (UniqueName: \"kubernetes.io/projected/09f0463c-f091-495e-9773-c07fe902c200-kube-api-access-tfg9g\") pod \"kube-state-metrics-0\" (UID: \"09f0463c-f091-495e-9773-c07fe902c200\") " pod="openstack/kube-state-metrics-0" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.872557 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/09f0463c-f091-495e-9773-c07fe902c200-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"09f0463c-f091-495e-9773-c07fe902c200\") " pod="openstack/kube-state-metrics-0" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.872592 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/09f0463c-f091-495e-9773-c07fe902c200-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"09f0463c-f091-495e-9773-c07fe902c200\") " pod="openstack/kube-state-metrics-0" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.973746 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/09f0463c-f091-495e-9773-c07fe902c200-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"09f0463c-f091-495e-9773-c07fe902c200\") " pod="openstack/kube-state-metrics-0" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.973799 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/09f0463c-f091-495e-9773-c07fe902c200-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"09f0463c-f091-495e-9773-c07fe902c200\") " pod="openstack/kube-state-metrics-0" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.973864 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09f0463c-f091-495e-9773-c07fe902c200-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"09f0463c-f091-495e-9773-c07fe902c200\") " pod="openstack/kube-state-metrics-0" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.973926 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfg9g\" (UniqueName: \"kubernetes.io/projected/09f0463c-f091-495e-9773-c07fe902c200-kube-api-access-tfg9g\") pod \"kube-state-metrics-0\" (UID: \"09f0463c-f091-495e-9773-c07fe902c200\") " pod="openstack/kube-state-metrics-0" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.979666 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09f0463c-f091-495e-9773-c07fe902c200-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"09f0463c-f091-495e-9773-c07fe902c200\") " pod="openstack/kube-state-metrics-0" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.983694 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/09f0463c-f091-495e-9773-c07fe902c200-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"09f0463c-f091-495e-9773-c07fe902c200\") " pod="openstack/kube-state-metrics-0" Nov 25 09:45:11 crc kubenswrapper[4776]: I1125 09:45:11.984441 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/09f0463c-f091-495e-9773-c07fe902c200-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"09f0463c-f091-495e-9773-c07fe902c200\") " pod="openstack/kube-state-metrics-0" Nov 25 09:45:12 crc kubenswrapper[4776]: I1125 09:45:11.999962 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfg9g\" (UniqueName: \"kubernetes.io/projected/09f0463c-f091-495e-9773-c07fe902c200-kube-api-access-tfg9g\") pod \"kube-state-metrics-0\" (UID: \"09f0463c-f091-495e-9773-c07fe902c200\") " pod="openstack/kube-state-metrics-0" Nov 25 09:45:12 crc kubenswrapper[4776]: I1125 09:45:12.238461 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 09:45:12 crc kubenswrapper[4776]: I1125 09:45:12.676424 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b708b100-166c-49b1-a47b-76d47ba2c6e5","Type":"ContainerStarted","Data":"7f99367ae1299a4af7ca7bdb87418a9b017add504f1ee9377bbcccc241edd6fb"} Nov 25 09:45:12 crc kubenswrapper[4776]: I1125 09:45:12.680422 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"768b7f4e-d631-4cc7-ad91-7e1099c8b354","Type":"ContainerStarted","Data":"01030c402508d2fdc1e43c8737b1f1f54cd200360facb57d6a5733ecae97896a"} Nov 25 09:45:12 crc kubenswrapper[4776]: I1125 09:45:12.812043 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 09:45:12 crc kubenswrapper[4776]: W1125 09:45:12.823100 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09f0463c_f091_495e_9773_c07fe902c200.slice/crio-d1ffc8968dfe50f1ebb42b0a99c53bc306bdc7bc1023dd5d4b8c82859dbf7e39 WatchSource:0}: Error finding container d1ffc8968dfe50f1ebb42b0a99c53bc306bdc7bc1023dd5d4b8c82859dbf7e39: Status 404 returned error can't find the container with id d1ffc8968dfe50f1ebb42b0a99c53bc306bdc7bc1023dd5d4b8c82859dbf7e39 Nov 25 09:45:13 crc kubenswrapper[4776]: I1125 09:45:13.316024 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 25 09:45:13 crc kubenswrapper[4776]: I1125 09:45:13.409660 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:45:13 crc kubenswrapper[4776]: I1125 09:45:13.661410 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:45:13 crc kubenswrapper[4776]: I1125 09:45:13.663782 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:45:13 crc kubenswrapper[4776]: I1125 09:45:13.677516 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e06b81f6-6161-46ae-9830-cd735c64d2f0" path="/var/lib/kubelet/pods/e06b81f6-6161-46ae-9830-cd735c64d2f0/volumes" Nov 25 09:45:13 crc kubenswrapper[4776]: I1125 09:45:13.716317 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b708b100-166c-49b1-a47b-76d47ba2c6e5","Type":"ContainerStarted","Data":"5919b100098159f61f76d181a1ae66bfafcb2916a56d7fe4fe1cee93ad3c32d0"} Nov 25 09:45:13 crc kubenswrapper[4776]: I1125 09:45:13.719154 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"09f0463c-f091-495e-9773-c07fe902c200","Type":"ContainerStarted","Data":"d1ffc8968dfe50f1ebb42b0a99c53bc306bdc7bc1023dd5d4b8c82859dbf7e39"} Nov 25 09:45:13 crc kubenswrapper[4776]: I1125 09:45:13.809277 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.809252333 podStartE2EDuration="4.809252333s" podCreationTimestamp="2025-11-25 09:45:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:45:13.788533953 +0000 UTC m=+1258.829593506" watchObservedRunningTime="2025-11-25 09:45:13.809252333 +0000 UTC m=+1258.850311886" Nov 25 09:45:15 crc kubenswrapper[4776]: I1125 09:45:15.655683 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 09:45:15 crc kubenswrapper[4776]: I1125 09:45:15.656577 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="85e1e1f5-a606-427e-a997-55dfad758e2e" containerName="glance-log" containerID="cri-o://f4038d3a7a822862f0092282cd6f4b6eaa945eeff59e6d0ac36071314e20b11d" gracePeriod=30 Nov 25 09:45:15 crc kubenswrapper[4776]: I1125 09:45:15.657115 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="85e1e1f5-a606-427e-a997-55dfad758e2e" containerName="glance-httpd" containerID="cri-o://776925a5c003ac3e724c8cfd6aff1758a1de35d749a1353e22ea01e705238a2e" gracePeriod=30 Nov 25 09:45:15 crc kubenswrapper[4776]: I1125 09:45:15.747433 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"09f0463c-f091-495e-9773-c07fe902c200","Type":"ContainerStarted","Data":"94684aa85a999df01587ff8336fcf45d1847b413f000fe304eae44dda233eae6"} Nov 25 09:45:15 crc kubenswrapper[4776]: I1125 09:45:15.748731 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 25 09:45:15 crc kubenswrapper[4776]: I1125 09:45:15.752050 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"768b7f4e-d631-4cc7-ad91-7e1099c8b354","Type":"ContainerStarted","Data":"5f04e4d9c26603eeb50c21d39c52c1a5e24f749d1eff7c8516f0fc36f3f2ef84"} Nov 25 09:45:16 crc kubenswrapper[4776]: I1125 09:45:16.762950 4776 generic.go:334] "Generic (PLEG): container finished" podID="85e1e1f5-a606-427e-a997-55dfad758e2e" containerID="f4038d3a7a822862f0092282cd6f4b6eaa945eeff59e6d0ac36071314e20b11d" exitCode=143 Nov 25 09:45:16 crc kubenswrapper[4776]: I1125 09:45:16.763014 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"85e1e1f5-a606-427e-a997-55dfad758e2e","Type":"ContainerDied","Data":"f4038d3a7a822862f0092282cd6f4b6eaa945eeff59e6d0ac36071314e20b11d"} Nov 25 09:45:16 crc kubenswrapper[4776]: I1125 09:45:16.768092 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="768b7f4e-d631-4cc7-ad91-7e1099c8b354" containerName="ceilometer-central-agent" containerID="cri-o://6515b87d6beb7f876edae119be02a45958d18ae4c801f6d415a59604678f66d8" gracePeriod=30 Nov 25 09:45:16 crc kubenswrapper[4776]: I1125 09:45:16.768373 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"768b7f4e-d631-4cc7-ad91-7e1099c8b354","Type":"ContainerStarted","Data":"c4b874dd8103aee5096d8a55ec80822635fd055cddb05d21f01b9482504a4e70"} Nov 25 09:45:16 crc kubenswrapper[4776]: I1125 09:45:16.768415 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 09:45:16 crc kubenswrapper[4776]: I1125 09:45:16.769086 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="768b7f4e-d631-4cc7-ad91-7e1099c8b354" containerName="proxy-httpd" containerID="cri-o://c4b874dd8103aee5096d8a55ec80822635fd055cddb05d21f01b9482504a4e70" gracePeriod=30 Nov 25 09:45:16 crc kubenswrapper[4776]: I1125 09:45:16.769158 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="768b7f4e-d631-4cc7-ad91-7e1099c8b354" containerName="sg-core" containerID="cri-o://5f04e4d9c26603eeb50c21d39c52c1a5e24f749d1eff7c8516f0fc36f3f2ef84" gracePeriod=30 Nov 25 09:45:16 crc kubenswrapper[4776]: I1125 09:45:16.769203 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="768b7f4e-d631-4cc7-ad91-7e1099c8b354" containerName="ceilometer-notification-agent" containerID="cri-o://01030c402508d2fdc1e43c8737b1f1f54cd200360facb57d6a5733ecae97896a" gracePeriod=30 Nov 25 09:45:16 crc kubenswrapper[4776]: I1125 09:45:16.804765 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.342893118 podStartE2EDuration="5.804736927s" podCreationTimestamp="2025-11-25 09:45:11 +0000 UTC" firstStartedPulling="2025-11-25 09:45:12.825912138 +0000 UTC m=+1257.866971691" lastFinishedPulling="2025-11-25 09:45:15.287755947 +0000 UTC m=+1260.328815500" observedRunningTime="2025-11-25 09:45:15.767578612 +0000 UTC m=+1260.808638165" watchObservedRunningTime="2025-11-25 09:45:16.804736927 +0000 UTC m=+1261.845796480" Nov 25 09:45:16 crc kubenswrapper[4776]: I1125 09:45:16.814303 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.113556813 podStartE2EDuration="9.813903757s" podCreationTimestamp="2025-11-25 09:45:07 +0000 UTC" firstStartedPulling="2025-11-25 09:45:08.758940888 +0000 UTC m=+1253.800000441" lastFinishedPulling="2025-11-25 09:45:16.459287842 +0000 UTC m=+1261.500347385" observedRunningTime="2025-11-25 09:45:16.78852403 +0000 UTC m=+1261.829583593" watchObservedRunningTime="2025-11-25 09:45:16.813903757 +0000 UTC m=+1261.854963330" Nov 25 09:45:17 crc kubenswrapper[4776]: I1125 09:45:17.779121 4776 generic.go:334] "Generic (PLEG): container finished" podID="768b7f4e-d631-4cc7-ad91-7e1099c8b354" containerID="5f04e4d9c26603eeb50c21d39c52c1a5e24f749d1eff7c8516f0fc36f3f2ef84" exitCode=2 Nov 25 09:45:17 crc kubenswrapper[4776]: I1125 09:45:17.779158 4776 generic.go:334] "Generic (PLEG): container finished" podID="768b7f4e-d631-4cc7-ad91-7e1099c8b354" containerID="01030c402508d2fdc1e43c8737b1f1f54cd200360facb57d6a5733ecae97896a" exitCode=0 Nov 25 09:45:17 crc kubenswrapper[4776]: I1125 09:45:17.779169 4776 generic.go:334] "Generic (PLEG): container finished" podID="768b7f4e-d631-4cc7-ad91-7e1099c8b354" containerID="6515b87d6beb7f876edae119be02a45958d18ae4c801f6d415a59604678f66d8" exitCode=0 Nov 25 09:45:17 crc kubenswrapper[4776]: I1125 09:45:17.779188 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"768b7f4e-d631-4cc7-ad91-7e1099c8b354","Type":"ContainerDied","Data":"5f04e4d9c26603eeb50c21d39c52c1a5e24f749d1eff7c8516f0fc36f3f2ef84"} Nov 25 09:45:17 crc kubenswrapper[4776]: I1125 09:45:17.779223 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"768b7f4e-d631-4cc7-ad91-7e1099c8b354","Type":"ContainerDied","Data":"01030c402508d2fdc1e43c8737b1f1f54cd200360facb57d6a5733ecae97896a"} Nov 25 09:45:17 crc kubenswrapper[4776]: I1125 09:45:17.779235 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"768b7f4e-d631-4cc7-ad91-7e1099c8b354","Type":"ContainerDied","Data":"6515b87d6beb7f876edae119be02a45958d18ae4c801f6d415a59604678f66d8"} Nov 25 09:45:17 crc kubenswrapper[4776]: I1125 09:45:17.818282 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:45:17 crc kubenswrapper[4776]: I1125 09:45:17.818354 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:45:17 crc kubenswrapper[4776]: I1125 09:45:17.818405 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 09:45:17 crc kubenswrapper[4776]: I1125 09:45:17.819220 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bbf70cb3f31b21dd9b2d356f05e12d08a7b4f65263789554ac47b2758bc077be"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 09:45:17 crc kubenswrapper[4776]: I1125 09:45:17.819286 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://bbf70cb3f31b21dd9b2d356f05e12d08a7b4f65263789554ac47b2758bc077be" gracePeriod=600 Nov 25 09:45:18 crc kubenswrapper[4776]: I1125 09:45:18.549793 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 25 09:45:18 crc kubenswrapper[4776]: I1125 09:45:18.792127 4776 generic.go:334] "Generic (PLEG): container finished" podID="8ce61c9c-20b2-462c-b187-519d56e75f56" containerID="27dbc31408be0c7000f15db205a2fa4f8f0563ea0cdf47359535299432ac1569" exitCode=137 Nov 25 09:45:18 crc kubenswrapper[4776]: I1125 09:45:18.792171 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8ce61c9c-20b2-462c-b187-519d56e75f56","Type":"ContainerDied","Data":"27dbc31408be0c7000f15db205a2fa4f8f0563ea0cdf47359535299432ac1569"} Nov 25 09:45:18 crc kubenswrapper[4776]: I1125 09:45:18.795640 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="bbf70cb3f31b21dd9b2d356f05e12d08a7b4f65263789554ac47b2758bc077be" exitCode=0 Nov 25 09:45:18 crc kubenswrapper[4776]: I1125 09:45:18.795707 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"bbf70cb3f31b21dd9b2d356f05e12d08a7b4f65263789554ac47b2758bc077be"} Nov 25 09:45:18 crc kubenswrapper[4776]: I1125 09:45:18.795750 4776 scope.go:117] "RemoveContainer" containerID="6132d3cd67d80b93c28c890778e005f6af8e158a1aed030b0062e3edb3f984f8" Nov 25 09:45:19 crc kubenswrapper[4776]: I1125 09:45:19.807922 4776 generic.go:334] "Generic (PLEG): container finished" podID="85e1e1f5-a606-427e-a997-55dfad758e2e" containerID="776925a5c003ac3e724c8cfd6aff1758a1de35d749a1353e22ea01e705238a2e" exitCode=0 Nov 25 09:45:19 crc kubenswrapper[4776]: I1125 09:45:19.808414 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"85e1e1f5-a606-427e-a997-55dfad758e2e","Type":"ContainerDied","Data":"776925a5c003ac3e724c8cfd6aff1758a1de35d749a1353e22ea01e705238a2e"} Nov 25 09:45:19 crc kubenswrapper[4776]: I1125 09:45:19.812012 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"41bf411c5d6ef6679dd0fb3b86628f86f08b8ccef2b894d89b25148005eb3847"} Nov 25 09:45:19 crc kubenswrapper[4776]: I1125 09:45:19.984588 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.047300 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.047410 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.057056 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ce61c9c-20b2-462c-b187-519d56e75f56-scripts\") pod \"8ce61c9c-20b2-462c-b187-519d56e75f56\" (UID: \"8ce61c9c-20b2-462c-b187-519d56e75f56\") " Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.057110 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ce61c9c-20b2-462c-b187-519d56e75f56-config-data-custom\") pod \"8ce61c9c-20b2-462c-b187-519d56e75f56\" (UID: \"8ce61c9c-20b2-462c-b187-519d56e75f56\") " Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.057173 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ce61c9c-20b2-462c-b187-519d56e75f56-config-data\") pod \"8ce61c9c-20b2-462c-b187-519d56e75f56\" (UID: \"8ce61c9c-20b2-462c-b187-519d56e75f56\") " Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.057288 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h47bg\" (UniqueName: \"kubernetes.io/projected/8ce61c9c-20b2-462c-b187-519d56e75f56-kube-api-access-h47bg\") pod \"8ce61c9c-20b2-462c-b187-519d56e75f56\" (UID: \"8ce61c9c-20b2-462c-b187-519d56e75f56\") " Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.057338 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8ce61c9c-20b2-462c-b187-519d56e75f56-etc-machine-id\") pod \"8ce61c9c-20b2-462c-b187-519d56e75f56\" (UID: \"8ce61c9c-20b2-462c-b187-519d56e75f56\") " Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.057360 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ce61c9c-20b2-462c-b187-519d56e75f56-logs\") pod \"8ce61c9c-20b2-462c-b187-519d56e75f56\" (UID: \"8ce61c9c-20b2-462c-b187-519d56e75f56\") " Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.057388 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ce61c9c-20b2-462c-b187-519d56e75f56-combined-ca-bundle\") pod \"8ce61c9c-20b2-462c-b187-519d56e75f56\" (UID: \"8ce61c9c-20b2-462c-b187-519d56e75f56\") " Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.058833 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8ce61c9c-20b2-462c-b187-519d56e75f56-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "8ce61c9c-20b2-462c-b187-519d56e75f56" (UID: "8ce61c9c-20b2-462c-b187-519d56e75f56"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.059176 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ce61c9c-20b2-462c-b187-519d56e75f56-logs" (OuterVolumeSpecName: "logs") pod "8ce61c9c-20b2-462c-b187-519d56e75f56" (UID: "8ce61c9c-20b2-462c-b187-519d56e75f56"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.067139 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ce61c9c-20b2-462c-b187-519d56e75f56-kube-api-access-h47bg" (OuterVolumeSpecName: "kube-api-access-h47bg") pod "8ce61c9c-20b2-462c-b187-519d56e75f56" (UID: "8ce61c9c-20b2-462c-b187-519d56e75f56"). InnerVolumeSpecName "kube-api-access-h47bg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.068329 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ce61c9c-20b2-462c-b187-519d56e75f56-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8ce61c9c-20b2-462c-b187-519d56e75f56" (UID: "8ce61c9c-20b2-462c-b187-519d56e75f56"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.085281 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ce61c9c-20b2-462c-b187-519d56e75f56-scripts" (OuterVolumeSpecName: "scripts") pod "8ce61c9c-20b2-462c-b187-519d56e75f56" (UID: "8ce61c9c-20b2-462c-b187-519d56e75f56"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.104687 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.110587 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.161556 4776 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8ce61c9c-20b2-462c-b187-519d56e75f56-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.161588 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ce61c9c-20b2-462c-b187-519d56e75f56-logs\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.161600 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ce61c9c-20b2-462c-b187-519d56e75f56-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.161611 4776 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ce61c9c-20b2-462c-b187-519d56e75f56-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.161622 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h47bg\" (UniqueName: \"kubernetes.io/projected/8ce61c9c-20b2-462c-b187-519d56e75f56-kube-api-access-h47bg\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.164414 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ce61c9c-20b2-462c-b187-519d56e75f56-config-data" (OuterVolumeSpecName: "config-data") pod "8ce61c9c-20b2-462c-b187-519d56e75f56" (UID: "8ce61c9c-20b2-462c-b187-519d56e75f56"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.167235 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ce61c9c-20b2-462c-b187-519d56e75f56-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8ce61c9c-20b2-462c-b187-519d56e75f56" (UID: "8ce61c9c-20b2-462c-b187-519d56e75f56"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.229721 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.264863 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ce61c9c-20b2-462c-b187-519d56e75f56-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.264908 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ce61c9c-20b2-462c-b187-519d56e75f56-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.366111 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/85e1e1f5-a606-427e-a997-55dfad758e2e-httpd-run\") pod \"85e1e1f5-a606-427e-a997-55dfad758e2e\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.366193 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85e1e1f5-a606-427e-a997-55dfad758e2e-public-tls-certs\") pod \"85e1e1f5-a606-427e-a997-55dfad758e2e\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.366236 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85e1e1f5-a606-427e-a997-55dfad758e2e-scripts\") pod \"85e1e1f5-a606-427e-a997-55dfad758e2e\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.366295 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"85e1e1f5-a606-427e-a997-55dfad758e2e\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.366354 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85e1e1f5-a606-427e-a997-55dfad758e2e-config-data\") pod \"85e1e1f5-a606-427e-a997-55dfad758e2e\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.366375 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85e1e1f5-a606-427e-a997-55dfad758e2e-combined-ca-bundle\") pod \"85e1e1f5-a606-427e-a997-55dfad758e2e\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.366398 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85e1e1f5-a606-427e-a997-55dfad758e2e-logs\") pod \"85e1e1f5-a606-427e-a997-55dfad758e2e\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.366438 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8q55\" (UniqueName: \"kubernetes.io/projected/85e1e1f5-a606-427e-a997-55dfad758e2e-kube-api-access-k8q55\") pod \"85e1e1f5-a606-427e-a997-55dfad758e2e\" (UID: \"85e1e1f5-a606-427e-a997-55dfad758e2e\") " Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.367242 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85e1e1f5-a606-427e-a997-55dfad758e2e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "85e1e1f5-a606-427e-a997-55dfad758e2e" (UID: "85e1e1f5-a606-427e-a997-55dfad758e2e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.367618 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85e1e1f5-a606-427e-a997-55dfad758e2e-logs" (OuterVolumeSpecName: "logs") pod "85e1e1f5-a606-427e-a997-55dfad758e2e" (UID: "85e1e1f5-a606-427e-a997-55dfad758e2e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.370449 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85e1e1f5-a606-427e-a997-55dfad758e2e-scripts" (OuterVolumeSpecName: "scripts") pod "85e1e1f5-a606-427e-a997-55dfad758e2e" (UID: "85e1e1f5-a606-427e-a997-55dfad758e2e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.371221 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85e1e1f5-a606-427e-a997-55dfad758e2e-kube-api-access-k8q55" (OuterVolumeSpecName: "kube-api-access-k8q55") pod "85e1e1f5-a606-427e-a997-55dfad758e2e" (UID: "85e1e1f5-a606-427e-a997-55dfad758e2e"). InnerVolumeSpecName "kube-api-access-k8q55". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.372030 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "85e1e1f5-a606-427e-a997-55dfad758e2e" (UID: "85e1e1f5-a606-427e-a997-55dfad758e2e"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.393241 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85e1e1f5-a606-427e-a997-55dfad758e2e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85e1e1f5-a606-427e-a997-55dfad758e2e" (UID: "85e1e1f5-a606-427e-a997-55dfad758e2e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.422503 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85e1e1f5-a606-427e-a997-55dfad758e2e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "85e1e1f5-a606-427e-a997-55dfad758e2e" (UID: "85e1e1f5-a606-427e-a997-55dfad758e2e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.429761 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85e1e1f5-a606-427e-a997-55dfad758e2e-config-data" (OuterVolumeSpecName: "config-data") pod "85e1e1f5-a606-427e-a997-55dfad758e2e" (UID: "85e1e1f5-a606-427e-a997-55dfad758e2e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.468622 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85e1e1f5-a606-427e-a997-55dfad758e2e-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.468700 4776 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.468761 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85e1e1f5-a606-427e-a997-55dfad758e2e-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.468784 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85e1e1f5-a606-427e-a997-55dfad758e2e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.468804 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85e1e1f5-a606-427e-a997-55dfad758e2e-logs\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.468819 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8q55\" (UniqueName: \"kubernetes.io/projected/85e1e1f5-a606-427e-a997-55dfad758e2e-kube-api-access-k8q55\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.468835 4776 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/85e1e1f5-a606-427e-a997-55dfad758e2e-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.468851 4776 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85e1e1f5-a606-427e-a997-55dfad758e2e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.493001 4776 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.570498 4776 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.824418 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"85e1e1f5-a606-427e-a997-55dfad758e2e","Type":"ContainerDied","Data":"4b6a02d7412a1011e58348b8a6e71bada513cbdf4f20fcaa4f6f8b852e55909a"} Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.824485 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.824499 4776 scope.go:117] "RemoveContainer" containerID="776925a5c003ac3e724c8cfd6aff1758a1de35d749a1353e22ea01e705238a2e" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.831197 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.831161 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8ce61c9c-20b2-462c-b187-519d56e75f56","Type":"ContainerDied","Data":"3ec68c3c80aa2f18db7444b6f5e50a4edc05713e5dbd3a6b2050341e4382f21a"} Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.832721 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.832752 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.882782 4776 scope.go:117] "RemoveContainer" containerID="f4038d3a7a822862f0092282cd6f4b6eaa945eeff59e6d0ac36071314e20b11d" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.912938 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.934473 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.958956 4776 scope.go:117] "RemoveContainer" containerID="27dbc31408be0c7000f15db205a2fa4f8f0563ea0cdf47359535299432ac1569" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.959124 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 09:45:20 crc kubenswrapper[4776]: E1125 09:45:20.959533 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85e1e1f5-a606-427e-a997-55dfad758e2e" containerName="glance-log" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.959545 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="85e1e1f5-a606-427e-a997-55dfad758e2e" containerName="glance-log" Nov 25 09:45:20 crc kubenswrapper[4776]: E1125 09:45:20.959558 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85e1e1f5-a606-427e-a997-55dfad758e2e" containerName="glance-httpd" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.959566 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="85e1e1f5-a606-427e-a997-55dfad758e2e" containerName="glance-httpd" Nov 25 09:45:20 crc kubenswrapper[4776]: E1125 09:45:20.959588 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ce61c9c-20b2-462c-b187-519d56e75f56" containerName="cinder-api-log" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.959595 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ce61c9c-20b2-462c-b187-519d56e75f56" containerName="cinder-api-log" Nov 25 09:45:20 crc kubenswrapper[4776]: E1125 09:45:20.959607 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ce61c9c-20b2-462c-b187-519d56e75f56" containerName="cinder-api" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.959612 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ce61c9c-20b2-462c-b187-519d56e75f56" containerName="cinder-api" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.959817 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="85e1e1f5-a606-427e-a997-55dfad758e2e" containerName="glance-httpd" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.959828 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="85e1e1f5-a606-427e-a997-55dfad758e2e" containerName="glance-log" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.959842 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ce61c9c-20b2-462c-b187-519d56e75f56" containerName="cinder-api" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.959852 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ce61c9c-20b2-462c-b187-519d56e75f56" containerName="cinder-api-log" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.960860 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.964327 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.964568 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 25 09:45:20 crc kubenswrapper[4776]: I1125 09:45:20.982541 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.006594 4776 scope.go:117] "RemoveContainer" containerID="ed6f27472e2324b01d1d07ab6b85981456aa7805810f98f575417846cd62ede4" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.013127 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.033325 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.046588 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.048392 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.054149 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.054326 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.054454 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.057928 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.089885 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-config-data\") pod \"glance-default-external-api-0\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " pod="openstack/glance-default-external-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.089950 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " pod="openstack/glance-default-external-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.090005 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " pod="openstack/glance-default-external-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.090192 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " pod="openstack/glance-default-external-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.090222 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-logs\") pod \"glance-default-external-api-0\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " pod="openstack/glance-default-external-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.090251 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jznnd\" (UniqueName: \"kubernetes.io/projected/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-kube-api-access-jznnd\") pod \"glance-default-external-api-0\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " pod="openstack/glance-default-external-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.090303 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " pod="openstack/glance-default-external-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.090346 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-scripts\") pod \"glance-default-external-api-0\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " pod="openstack/glance-default-external-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.192250 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-scripts\") pod \"glance-default-external-api-0\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " pod="openstack/glance-default-external-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.192574 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.192605 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-config-data\") pod \"glance-default-external-api-0\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " pod="openstack/glance-default-external-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.192629 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " pod="openstack/glance-default-external-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.192652 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86e320e0-80ee-40df-b3a1-a48cb810a435-logs\") pod \"cinder-api-0\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.192679 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " pod="openstack/glance-default-external-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.192709 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-public-tls-certs\") pod \"cinder-api-0\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.192730 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-config-data\") pod \"cinder-api-0\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.192747 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.192766 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-scripts\") pod \"cinder-api-0\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.192804 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k277j\" (UniqueName: \"kubernetes.io/projected/86e320e0-80ee-40df-b3a1-a48cb810a435-kube-api-access-k277j\") pod \"cinder-api-0\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.192830 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " pod="openstack/glance-default-external-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.192848 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-logs\") pod \"glance-default-external-api-0\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " pod="openstack/glance-default-external-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.192864 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jznnd\" (UniqueName: \"kubernetes.io/projected/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-kube-api-access-jznnd\") pod \"glance-default-external-api-0\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " pod="openstack/glance-default-external-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.192888 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " pod="openstack/glance-default-external-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.192917 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-config-data-custom\") pod \"cinder-api-0\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.192937 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/86e320e0-80ee-40df-b3a1-a48cb810a435-etc-machine-id\") pod \"cinder-api-0\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.193847 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " pod="openstack/glance-default-external-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.193955 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.194785 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-logs\") pod \"glance-default-external-api-0\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " pod="openstack/glance-default-external-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.204159 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " pod="openstack/glance-default-external-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.204712 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-scripts\") pod \"glance-default-external-api-0\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " pod="openstack/glance-default-external-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.206252 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-config-data\") pod \"glance-default-external-api-0\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " pod="openstack/glance-default-external-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.206750 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " pod="openstack/glance-default-external-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.212245 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jznnd\" (UniqueName: \"kubernetes.io/projected/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-kube-api-access-jznnd\") pod \"glance-default-external-api-0\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " pod="openstack/glance-default-external-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.223755 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " pod="openstack/glance-default-external-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.283523 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.294133 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-public-tls-certs\") pod \"cinder-api-0\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.294189 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-config-data\") pod \"cinder-api-0\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.294221 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.294253 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-scripts\") pod \"cinder-api-0\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.294298 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k277j\" (UniqueName: \"kubernetes.io/projected/86e320e0-80ee-40df-b3a1-a48cb810a435-kube-api-access-k277j\") pod \"cinder-api-0\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.294362 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-config-data-custom\") pod \"cinder-api-0\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.294390 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/86e320e0-80ee-40df-b3a1-a48cb810a435-etc-machine-id\") pod \"cinder-api-0\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.294439 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.294487 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86e320e0-80ee-40df-b3a1-a48cb810a435-logs\") pod \"cinder-api-0\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.294546 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/86e320e0-80ee-40df-b3a1-a48cb810a435-etc-machine-id\") pod \"cinder-api-0\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.294916 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86e320e0-80ee-40df-b3a1-a48cb810a435-logs\") pod \"cinder-api-0\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.297806 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.300679 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-scripts\") pod \"cinder-api-0\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.304014 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-config-data\") pod \"cinder-api-0\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.310542 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.311018 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-config-data-custom\") pod \"cinder-api-0\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.313411 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-public-tls-certs\") pod \"cinder-api-0\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.317103 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k277j\" (UniqueName: \"kubernetes.io/projected/86e320e0-80ee-40df-b3a1-a48cb810a435-kube-api-access-k277j\") pod \"cinder-api-0\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.366311 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.680850 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85e1e1f5-a606-427e-a997-55dfad758e2e" path="/var/lib/kubelet/pods/85e1e1f5-a606-427e-a997-55dfad758e2e/volumes" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.681896 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ce61c9c-20b2-462c-b187-519d56e75f56" path="/var/lib/kubelet/pods/8ce61c9c-20b2-462c-b187-519d56e75f56/volumes" Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.909863 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 09:45:21 crc kubenswrapper[4776]: W1125 09:45:21.919407 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86e320e0_80ee_40df_b3a1_a48cb810a435.slice/crio-e533e247825031f6a8cc6b6a24f05322c2eb862b637377dc7167d71281aadbac WatchSource:0}: Error finding container e533e247825031f6a8cc6b6a24f05322c2eb862b637377dc7167d71281aadbac: Status 404 returned error can't find the container with id e533e247825031f6a8cc6b6a24f05322c2eb862b637377dc7167d71281aadbac Nov 25 09:45:21 crc kubenswrapper[4776]: W1125 09:45:21.991538 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2cda6de_ef12_42f4_a6e5_824c2db3cd01.slice/crio-42441d71e06e7f040dbf095b2f3e3d0ef18caf0085fa5e84a1f401d5ef9342a1 WatchSource:0}: Error finding container 42441d71e06e7f040dbf095b2f3e3d0ef18caf0085fa5e84a1f401d5ef9342a1: Status 404 returned error can't find the container with id 42441d71e06e7f040dbf095b2f3e3d0ef18caf0085fa5e84a1f401d5ef9342a1 Nov 25 09:45:21 crc kubenswrapper[4776]: I1125 09:45:21.992056 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 09:45:22 crc kubenswrapper[4776]: I1125 09:45:22.405646 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 25 09:45:22 crc kubenswrapper[4776]: I1125 09:45:22.851848 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d2cda6de-ef12-42f4-a6e5-824c2db3cd01","Type":"ContainerStarted","Data":"0bd7925c5eadaefd62b355bbd7aebe47ce493b779918483df9c72dc9a141944b"} Nov 25 09:45:22 crc kubenswrapper[4776]: I1125 09:45:22.852224 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d2cda6de-ef12-42f4-a6e5-824c2db3cd01","Type":"ContainerStarted","Data":"42441d71e06e7f040dbf095b2f3e3d0ef18caf0085fa5e84a1f401d5ef9342a1"} Nov 25 09:45:22 crc kubenswrapper[4776]: I1125 09:45:22.853640 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"86e320e0-80ee-40df-b3a1-a48cb810a435","Type":"ContainerStarted","Data":"643bd96dcd87a61d61b437895cd783519ab3ff279196abf17121bbe3e4d485dd"} Nov 25 09:45:22 crc kubenswrapper[4776]: I1125 09:45:22.853689 4776 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 09:45:22 crc kubenswrapper[4776]: I1125 09:45:22.853701 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"86e320e0-80ee-40df-b3a1-a48cb810a435","Type":"ContainerStarted","Data":"e533e247825031f6a8cc6b6a24f05322c2eb862b637377dc7167d71281aadbac"} Nov 25 09:45:22 crc kubenswrapper[4776]: I1125 09:45:22.853706 4776 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 09:45:23 crc kubenswrapper[4776]: I1125 09:45:23.330792 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 25 09:45:23 crc kubenswrapper[4776]: I1125 09:45:23.509268 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 25 09:45:24 crc kubenswrapper[4776]: I1125 09:45:24.871923 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"86e320e0-80ee-40df-b3a1-a48cb810a435","Type":"ContainerStarted","Data":"968312305c48cd2a75a04c3e2a43d37d9ed640ee14d080081ad79c323f2cf0f5"} Nov 25 09:45:24 crc kubenswrapper[4776]: I1125 09:45:24.872489 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 25 09:45:24 crc kubenswrapper[4776]: I1125 09:45:24.874089 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d2cda6de-ef12-42f4-a6e5-824c2db3cd01","Type":"ContainerStarted","Data":"9fddb69738d8e8fd199d1f7a677b07c2d8ceff1639dbe9784b2c6fd247da4ae1"} Nov 25 09:45:24 crc kubenswrapper[4776]: I1125 09:45:24.889702 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.889681688 podStartE2EDuration="4.889681688s" podCreationTimestamp="2025-11-25 09:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:45:24.889152265 +0000 UTC m=+1269.930211818" watchObservedRunningTime="2025-11-25 09:45:24.889681688 +0000 UTC m=+1269.930741241" Nov 25 09:45:24 crc kubenswrapper[4776]: I1125 09:45:24.934500 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.934472861 podStartE2EDuration="4.934472861s" podCreationTimestamp="2025-11-25 09:45:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:45:24.909786022 +0000 UTC m=+1269.950845575" watchObservedRunningTime="2025-11-25 09:45:24.934472861 +0000 UTC m=+1269.975532414" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.378713 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-t5rsg"] Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.380712 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-t5rsg" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.403922 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-t5rsg"] Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.469464 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-96l8q"] Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.470764 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-96l8q" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.492946 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b686729-83aa-48c6-9cce-cd28b26dd4b4-operator-scripts\") pod \"nova-api-db-create-t5rsg\" (UID: \"2b686729-83aa-48c6-9cce-cd28b26dd4b4\") " pod="openstack/nova-api-db-create-t5rsg" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.493059 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgrrf\" (UniqueName: \"kubernetes.io/projected/2b686729-83aa-48c6-9cce-cd28b26dd4b4-kube-api-access-mgrrf\") pod \"nova-api-db-create-t5rsg\" (UID: \"2b686729-83aa-48c6-9cce-cd28b26dd4b4\") " pod="openstack/nova-api-db-create-t5rsg" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.504148 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-3b46-account-create-kg8tk"] Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.505824 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3b46-account-create-kg8tk" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.507943 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.524645 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-96l8q"] Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.533162 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-3b46-account-create-kg8tk"] Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.594913 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-474qp\" (UniqueName: \"kubernetes.io/projected/15bbd95b-2071-4988-83c7-8411c8a974fb-kube-api-access-474qp\") pod \"nova-cell0-db-create-96l8q\" (UID: \"15bbd95b-2071-4988-83c7-8411c8a974fb\") " pod="openstack/nova-cell0-db-create-96l8q" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.595058 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgrrf\" (UniqueName: \"kubernetes.io/projected/2b686729-83aa-48c6-9cce-cd28b26dd4b4-kube-api-access-mgrrf\") pod \"nova-api-db-create-t5rsg\" (UID: \"2b686729-83aa-48c6-9cce-cd28b26dd4b4\") " pod="openstack/nova-api-db-create-t5rsg" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.595131 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdk9c\" (UniqueName: \"kubernetes.io/projected/eadc8c7e-32f4-4bbf-8ac3-51ebbe471718-kube-api-access-wdk9c\") pod \"nova-api-3b46-account-create-kg8tk\" (UID: \"eadc8c7e-32f4-4bbf-8ac3-51ebbe471718\") " pod="openstack/nova-api-3b46-account-create-kg8tk" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.595163 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eadc8c7e-32f4-4bbf-8ac3-51ebbe471718-operator-scripts\") pod \"nova-api-3b46-account-create-kg8tk\" (UID: \"eadc8c7e-32f4-4bbf-8ac3-51ebbe471718\") " pod="openstack/nova-api-3b46-account-create-kg8tk" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.595185 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15bbd95b-2071-4988-83c7-8411c8a974fb-operator-scripts\") pod \"nova-cell0-db-create-96l8q\" (UID: \"15bbd95b-2071-4988-83c7-8411c8a974fb\") " pod="openstack/nova-cell0-db-create-96l8q" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.595358 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b686729-83aa-48c6-9cce-cd28b26dd4b4-operator-scripts\") pod \"nova-api-db-create-t5rsg\" (UID: \"2b686729-83aa-48c6-9cce-cd28b26dd4b4\") " pod="openstack/nova-api-db-create-t5rsg" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.596103 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b686729-83aa-48c6-9cce-cd28b26dd4b4-operator-scripts\") pod \"nova-api-db-create-t5rsg\" (UID: \"2b686729-83aa-48c6-9cce-cd28b26dd4b4\") " pod="openstack/nova-api-db-create-t5rsg" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.612581 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgrrf\" (UniqueName: \"kubernetes.io/projected/2b686729-83aa-48c6-9cce-cd28b26dd4b4-kube-api-access-mgrrf\") pod \"nova-api-db-create-t5rsg\" (UID: \"2b686729-83aa-48c6-9cce-cd28b26dd4b4\") " pod="openstack/nova-api-db-create-t5rsg" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.672384 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-hc2rc"] Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.673560 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-hc2rc" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.680964 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-9dc4-account-create-cj5jv"] Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.682353 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9dc4-account-create-cj5jv" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.684241 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.694826 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-hc2rc"] Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.697118 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15bbd95b-2071-4988-83c7-8411c8a974fb-operator-scripts\") pod \"nova-cell0-db-create-96l8q\" (UID: \"15bbd95b-2071-4988-83c7-8411c8a974fb\") " pod="openstack/nova-cell0-db-create-96l8q" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.697248 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-474qp\" (UniqueName: \"kubernetes.io/projected/15bbd95b-2071-4988-83c7-8411c8a974fb-kube-api-access-474qp\") pod \"nova-cell0-db-create-96l8q\" (UID: \"15bbd95b-2071-4988-83c7-8411c8a974fb\") " pod="openstack/nova-cell0-db-create-96l8q" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.697537 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdk9c\" (UniqueName: \"kubernetes.io/projected/eadc8c7e-32f4-4bbf-8ac3-51ebbe471718-kube-api-access-wdk9c\") pod \"nova-api-3b46-account-create-kg8tk\" (UID: \"eadc8c7e-32f4-4bbf-8ac3-51ebbe471718\") " pod="openstack/nova-api-3b46-account-create-kg8tk" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.697573 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eadc8c7e-32f4-4bbf-8ac3-51ebbe471718-operator-scripts\") pod \"nova-api-3b46-account-create-kg8tk\" (UID: \"eadc8c7e-32f4-4bbf-8ac3-51ebbe471718\") " pod="openstack/nova-api-3b46-account-create-kg8tk" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.698371 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15bbd95b-2071-4988-83c7-8411c8a974fb-operator-scripts\") pod \"nova-cell0-db-create-96l8q\" (UID: \"15bbd95b-2071-4988-83c7-8411c8a974fb\") " pod="openstack/nova-cell0-db-create-96l8q" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.698421 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eadc8c7e-32f4-4bbf-8ac3-51ebbe471718-operator-scripts\") pod \"nova-api-3b46-account-create-kg8tk\" (UID: \"eadc8c7e-32f4-4bbf-8ac3-51ebbe471718\") " pod="openstack/nova-api-3b46-account-create-kg8tk" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.703590 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-t5rsg" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.710051 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-9dc4-account-create-cj5jv"] Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.716603 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdk9c\" (UniqueName: \"kubernetes.io/projected/eadc8c7e-32f4-4bbf-8ac3-51ebbe471718-kube-api-access-wdk9c\") pod \"nova-api-3b46-account-create-kg8tk\" (UID: \"eadc8c7e-32f4-4bbf-8ac3-51ebbe471718\") " pod="openstack/nova-api-3b46-account-create-kg8tk" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.717031 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-474qp\" (UniqueName: \"kubernetes.io/projected/15bbd95b-2071-4988-83c7-8411c8a974fb-kube-api-access-474qp\") pod \"nova-cell0-db-create-96l8q\" (UID: \"15bbd95b-2071-4988-83c7-8411c8a974fb\") " pod="openstack/nova-cell0-db-create-96l8q" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.790359 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-96l8q" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.801481 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knlzn\" (UniqueName: \"kubernetes.io/projected/610a6a76-1c03-49a0-850f-ee5131308f7b-kube-api-access-knlzn\") pod \"nova-cell1-db-create-hc2rc\" (UID: \"610a6a76-1c03-49a0-850f-ee5131308f7b\") " pod="openstack/nova-cell1-db-create-hc2rc" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.801645 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7hx5\" (UniqueName: \"kubernetes.io/projected/789daf57-0bbd-4e06-84d4-37e9ca59e390-kube-api-access-b7hx5\") pod \"nova-cell0-9dc4-account-create-cj5jv\" (UID: \"789daf57-0bbd-4e06-84d4-37e9ca59e390\") " pod="openstack/nova-cell0-9dc4-account-create-cj5jv" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.801804 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/610a6a76-1c03-49a0-850f-ee5131308f7b-operator-scripts\") pod \"nova-cell1-db-create-hc2rc\" (UID: \"610a6a76-1c03-49a0-850f-ee5131308f7b\") " pod="openstack/nova-cell1-db-create-hc2rc" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.801844 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/789daf57-0bbd-4e06-84d4-37e9ca59e390-operator-scripts\") pod \"nova-cell0-9dc4-account-create-cj5jv\" (UID: \"789daf57-0bbd-4e06-84d4-37e9ca59e390\") " pod="openstack/nova-cell0-9dc4-account-create-cj5jv" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.829026 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3b46-account-create-kg8tk" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.877686 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-a02b-account-create-dpj2t"] Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.879145 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-a02b-account-create-dpj2t" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.882176 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.895691 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-a02b-account-create-dpj2t"] Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.903000 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/610a6a76-1c03-49a0-850f-ee5131308f7b-operator-scripts\") pod \"nova-cell1-db-create-hc2rc\" (UID: \"610a6a76-1c03-49a0-850f-ee5131308f7b\") " pod="openstack/nova-cell1-db-create-hc2rc" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.903042 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/789daf57-0bbd-4e06-84d4-37e9ca59e390-operator-scripts\") pod \"nova-cell0-9dc4-account-create-cj5jv\" (UID: \"789daf57-0bbd-4e06-84d4-37e9ca59e390\") " pod="openstack/nova-cell0-9dc4-account-create-cj5jv" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.903152 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knlzn\" (UniqueName: \"kubernetes.io/projected/610a6a76-1c03-49a0-850f-ee5131308f7b-kube-api-access-knlzn\") pod \"nova-cell1-db-create-hc2rc\" (UID: \"610a6a76-1c03-49a0-850f-ee5131308f7b\") " pod="openstack/nova-cell1-db-create-hc2rc" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.903221 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7hx5\" (UniqueName: \"kubernetes.io/projected/789daf57-0bbd-4e06-84d4-37e9ca59e390-kube-api-access-b7hx5\") pod \"nova-cell0-9dc4-account-create-cj5jv\" (UID: \"789daf57-0bbd-4e06-84d4-37e9ca59e390\") " pod="openstack/nova-cell0-9dc4-account-create-cj5jv" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.903731 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/610a6a76-1c03-49a0-850f-ee5131308f7b-operator-scripts\") pod \"nova-cell1-db-create-hc2rc\" (UID: \"610a6a76-1c03-49a0-850f-ee5131308f7b\") " pod="openstack/nova-cell1-db-create-hc2rc" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.904172 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/789daf57-0bbd-4e06-84d4-37e9ca59e390-operator-scripts\") pod \"nova-cell0-9dc4-account-create-cj5jv\" (UID: \"789daf57-0bbd-4e06-84d4-37e9ca59e390\") " pod="openstack/nova-cell0-9dc4-account-create-cj5jv" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.921397 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knlzn\" (UniqueName: \"kubernetes.io/projected/610a6a76-1c03-49a0-850f-ee5131308f7b-kube-api-access-knlzn\") pod \"nova-cell1-db-create-hc2rc\" (UID: \"610a6a76-1c03-49a0-850f-ee5131308f7b\") " pod="openstack/nova-cell1-db-create-hc2rc" Nov 25 09:45:26 crc kubenswrapper[4776]: I1125 09:45:26.921875 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7hx5\" (UniqueName: \"kubernetes.io/projected/789daf57-0bbd-4e06-84d4-37e9ca59e390-kube-api-access-b7hx5\") pod \"nova-cell0-9dc4-account-create-cj5jv\" (UID: \"789daf57-0bbd-4e06-84d4-37e9ca59e390\") " pod="openstack/nova-cell0-9dc4-account-create-cj5jv" Nov 25 09:45:27 crc kubenswrapper[4776]: I1125 09:45:27.000588 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-hc2rc" Nov 25 09:45:27 crc kubenswrapper[4776]: I1125 09:45:27.005130 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cae7ff0c-7548-455d-85c8-3e68554658ce-operator-scripts\") pod \"nova-cell1-a02b-account-create-dpj2t\" (UID: \"cae7ff0c-7548-455d-85c8-3e68554658ce\") " pod="openstack/nova-cell1-a02b-account-create-dpj2t" Nov 25 09:45:27 crc kubenswrapper[4776]: I1125 09:45:27.005385 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rg4j\" (UniqueName: \"kubernetes.io/projected/cae7ff0c-7548-455d-85c8-3e68554658ce-kube-api-access-2rg4j\") pod \"nova-cell1-a02b-account-create-dpj2t\" (UID: \"cae7ff0c-7548-455d-85c8-3e68554658ce\") " pod="openstack/nova-cell1-a02b-account-create-dpj2t" Nov 25 09:45:27 crc kubenswrapper[4776]: I1125 09:45:27.007210 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9dc4-account-create-cj5jv" Nov 25 09:45:27 crc kubenswrapper[4776]: I1125 09:45:27.107940 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rg4j\" (UniqueName: \"kubernetes.io/projected/cae7ff0c-7548-455d-85c8-3e68554658ce-kube-api-access-2rg4j\") pod \"nova-cell1-a02b-account-create-dpj2t\" (UID: \"cae7ff0c-7548-455d-85c8-3e68554658ce\") " pod="openstack/nova-cell1-a02b-account-create-dpj2t" Nov 25 09:45:27 crc kubenswrapper[4776]: I1125 09:45:27.108437 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cae7ff0c-7548-455d-85c8-3e68554658ce-operator-scripts\") pod \"nova-cell1-a02b-account-create-dpj2t\" (UID: \"cae7ff0c-7548-455d-85c8-3e68554658ce\") " pod="openstack/nova-cell1-a02b-account-create-dpj2t" Nov 25 09:45:27 crc kubenswrapper[4776]: I1125 09:45:27.109452 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cae7ff0c-7548-455d-85c8-3e68554658ce-operator-scripts\") pod \"nova-cell1-a02b-account-create-dpj2t\" (UID: \"cae7ff0c-7548-455d-85c8-3e68554658ce\") " pod="openstack/nova-cell1-a02b-account-create-dpj2t" Nov 25 09:45:27 crc kubenswrapper[4776]: I1125 09:45:27.129733 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rg4j\" (UniqueName: \"kubernetes.io/projected/cae7ff0c-7548-455d-85c8-3e68554658ce-kube-api-access-2rg4j\") pod \"nova-cell1-a02b-account-create-dpj2t\" (UID: \"cae7ff0c-7548-455d-85c8-3e68554658ce\") " pod="openstack/nova-cell1-a02b-account-create-dpj2t" Nov 25 09:45:27 crc kubenswrapper[4776]: I1125 09:45:27.226652 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-a02b-account-create-dpj2t" Nov 25 09:45:27 crc kubenswrapper[4776]: I1125 09:45:27.376762 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-3b46-account-create-kg8tk"] Nov 25 09:45:27 crc kubenswrapper[4776]: W1125 09:45:27.383376 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeadc8c7e_32f4_4bbf_8ac3_51ebbe471718.slice/crio-48bfa5d0c85a086f4312bea3bef7801423ee5d46443ee4da35210c0657df7f7e WatchSource:0}: Error finding container 48bfa5d0c85a086f4312bea3bef7801423ee5d46443ee4da35210c0657df7f7e: Status 404 returned error can't find the container with id 48bfa5d0c85a086f4312bea3bef7801423ee5d46443ee4da35210c0657df7f7e Nov 25 09:45:27 crc kubenswrapper[4776]: W1125 09:45:27.385419 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15bbd95b_2071_4988_83c7_8411c8a974fb.slice/crio-335b76d86b0e231a7d9b2565c084a457e86fd6a7274093a39474afa721443227 WatchSource:0}: Error finding container 335b76d86b0e231a7d9b2565c084a457e86fd6a7274093a39474afa721443227: Status 404 returned error can't find the container with id 335b76d86b0e231a7d9b2565c084a457e86fd6a7274093a39474afa721443227 Nov 25 09:45:27 crc kubenswrapper[4776]: I1125 09:45:27.388355 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-96l8q"] Nov 25 09:45:27 crc kubenswrapper[4776]: I1125 09:45:27.454580 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-t5rsg"] Nov 25 09:45:27 crc kubenswrapper[4776]: W1125 09:45:27.455349 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b686729_83aa_48c6_9cce_cd28b26dd4b4.slice/crio-c67e21c5097fc690c2580ce63600e95b5d99c9913d595b5c652e9b3b4492d167 WatchSource:0}: Error finding container c67e21c5097fc690c2580ce63600e95b5d99c9913d595b5c652e9b3b4492d167: Status 404 returned error can't find the container with id c67e21c5097fc690c2580ce63600e95b5d99c9913d595b5c652e9b3b4492d167 Nov 25 09:45:27 crc kubenswrapper[4776]: I1125 09:45:27.523348 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-hc2rc"] Nov 25 09:45:27 crc kubenswrapper[4776]: I1125 09:45:27.536103 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-9dc4-account-create-cj5jv"] Nov 25 09:45:27 crc kubenswrapper[4776]: W1125 09:45:27.538890 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod610a6a76_1c03_49a0_850f_ee5131308f7b.slice/crio-bcc5e54787dd7b2e8af725670bbbd4e7cfc0dcb3ebe8adc71a05eb0abdd09c20 WatchSource:0}: Error finding container bcc5e54787dd7b2e8af725670bbbd4e7cfc0dcb3ebe8adc71a05eb0abdd09c20: Status 404 returned error can't find the container with id bcc5e54787dd7b2e8af725670bbbd4e7cfc0dcb3ebe8adc71a05eb0abdd09c20 Nov 25 09:45:27 crc kubenswrapper[4776]: W1125 09:45:27.540686 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod789daf57_0bbd_4e06_84d4_37e9ca59e390.slice/crio-2c222bdd5794026a23d247a04539e80acdd0aaf3500ec9ec855460bb24dd66e4 WatchSource:0}: Error finding container 2c222bdd5794026a23d247a04539e80acdd0aaf3500ec9ec855460bb24dd66e4: Status 404 returned error can't find the container with id 2c222bdd5794026a23d247a04539e80acdd0aaf3500ec9ec855460bb24dd66e4 Nov 25 09:45:27 crc kubenswrapper[4776]: I1125 09:45:27.681456 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-a02b-account-create-dpj2t"] Nov 25 09:45:27 crc kubenswrapper[4776]: W1125 09:45:27.694458 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcae7ff0c_7548_455d_85c8_3e68554658ce.slice/crio-78f82d72b9460f3ae620b01a37b15fa9dcd68e694ad55687d67e18c6a2053e3e WatchSource:0}: Error finding container 78f82d72b9460f3ae620b01a37b15fa9dcd68e694ad55687d67e18c6a2053e3e: Status 404 returned error can't find the container with id 78f82d72b9460f3ae620b01a37b15fa9dcd68e694ad55687d67e18c6a2053e3e Nov 25 09:45:27 crc kubenswrapper[4776]: I1125 09:45:27.919380 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-hc2rc" event={"ID":"610a6a76-1c03-49a0-850f-ee5131308f7b","Type":"ContainerStarted","Data":"bcc5e54787dd7b2e8af725670bbbd4e7cfc0dcb3ebe8adc71a05eb0abdd09c20"} Nov 25 09:45:27 crc kubenswrapper[4776]: I1125 09:45:27.921599 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3b46-account-create-kg8tk" event={"ID":"eadc8c7e-32f4-4bbf-8ac3-51ebbe471718","Type":"ContainerStarted","Data":"02c4c64a3ee9a1d43c07e566240a85097d6186935e3986969648ee476f3e015f"} Nov 25 09:45:27 crc kubenswrapper[4776]: I1125 09:45:27.921633 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3b46-account-create-kg8tk" event={"ID":"eadc8c7e-32f4-4bbf-8ac3-51ebbe471718","Type":"ContainerStarted","Data":"48bfa5d0c85a086f4312bea3bef7801423ee5d46443ee4da35210c0657df7f7e"} Nov 25 09:45:27 crc kubenswrapper[4776]: I1125 09:45:27.923989 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-t5rsg" event={"ID":"2b686729-83aa-48c6-9cce-cd28b26dd4b4","Type":"ContainerStarted","Data":"53eb409078616814ade62b3d6be13d09644f9faf11f7f4738e36db9a3a38eefb"} Nov 25 09:45:27 crc kubenswrapper[4776]: I1125 09:45:27.924020 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-t5rsg" event={"ID":"2b686729-83aa-48c6-9cce-cd28b26dd4b4","Type":"ContainerStarted","Data":"c67e21c5097fc690c2580ce63600e95b5d99c9913d595b5c652e9b3b4492d167"} Nov 25 09:45:27 crc kubenswrapper[4776]: I1125 09:45:27.925295 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-96l8q" event={"ID":"15bbd95b-2071-4988-83c7-8411c8a974fb","Type":"ContainerStarted","Data":"ea0e61b5432e05c0fedaf2f966d0d05c1d4707abe42eef391fb793278d77f5c7"} Nov 25 09:45:27 crc kubenswrapper[4776]: I1125 09:45:27.925321 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-96l8q" event={"ID":"15bbd95b-2071-4988-83c7-8411c8a974fb","Type":"ContainerStarted","Data":"335b76d86b0e231a7d9b2565c084a457e86fd6a7274093a39474afa721443227"} Nov 25 09:45:27 crc kubenswrapper[4776]: I1125 09:45:27.929167 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9dc4-account-create-cj5jv" event={"ID":"789daf57-0bbd-4e06-84d4-37e9ca59e390","Type":"ContainerStarted","Data":"2c222bdd5794026a23d247a04539e80acdd0aaf3500ec9ec855460bb24dd66e4"} Nov 25 09:45:27 crc kubenswrapper[4776]: I1125 09:45:27.931344 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-a02b-account-create-dpj2t" event={"ID":"cae7ff0c-7548-455d-85c8-3e68554658ce","Type":"ContainerStarted","Data":"78f82d72b9460f3ae620b01a37b15fa9dcd68e694ad55687d67e18c6a2053e3e"} Nov 25 09:45:27 crc kubenswrapper[4776]: I1125 09:45:27.940121 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-3b46-account-create-kg8tk" podStartSLOduration=1.94009841 podStartE2EDuration="1.94009841s" podCreationTimestamp="2025-11-25 09:45:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:45:27.935735151 +0000 UTC m=+1272.976794714" watchObservedRunningTime="2025-11-25 09:45:27.94009841 +0000 UTC m=+1272.981157963" Nov 25 09:45:27 crc kubenswrapper[4776]: I1125 09:45:27.959614 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-96l8q" podStartSLOduration=1.9595937989999999 podStartE2EDuration="1.959593799s" podCreationTimestamp="2025-11-25 09:45:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:45:27.951945947 +0000 UTC m=+1272.993005510" watchObservedRunningTime="2025-11-25 09:45:27.959593799 +0000 UTC m=+1273.000653352" Nov 25 09:45:28 crc kubenswrapper[4776]: I1125 09:45:28.947421 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9dc4-account-create-cj5jv" event={"ID":"789daf57-0bbd-4e06-84d4-37e9ca59e390","Type":"ContainerStarted","Data":"ef86e26a3c64bb3fa983b0e4cb1ba97b5d01472982d3395d5f9382b30981a3d2"} Nov 25 09:45:28 crc kubenswrapper[4776]: I1125 09:45:28.952465 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-a02b-account-create-dpj2t" event={"ID":"cae7ff0c-7548-455d-85c8-3e68554658ce","Type":"ContainerStarted","Data":"1505fc4e03c873c1557f6ddb2bb810c54cb98dd95eb6841cce7981791552b151"} Nov 25 09:45:28 crc kubenswrapper[4776]: I1125 09:45:28.957801 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-hc2rc" event={"ID":"610a6a76-1c03-49a0-850f-ee5131308f7b","Type":"ContainerStarted","Data":"c2a85ebe5755f4af658dd25fd030364579d92bdc461690f7c2025585b4779bd8"} Nov 25 09:45:28 crc kubenswrapper[4776]: I1125 09:45:28.987776 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-9dc4-account-create-cj5jv" podStartSLOduration=2.987756298 podStartE2EDuration="2.987756298s" podCreationTimestamp="2025-11-25 09:45:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:45:28.965595432 +0000 UTC m=+1274.006654985" watchObservedRunningTime="2025-11-25 09:45:28.987756298 +0000 UTC m=+1274.028815861" Nov 25 09:45:28 crc kubenswrapper[4776]: I1125 09:45:28.990051 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-hc2rc" podStartSLOduration=2.990041195 podStartE2EDuration="2.990041195s" podCreationTimestamp="2025-11-25 09:45:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:45:28.983920112 +0000 UTC m=+1274.024979665" watchObservedRunningTime="2025-11-25 09:45:28.990041195 +0000 UTC m=+1274.031100748" Nov 25 09:45:28 crc kubenswrapper[4776]: I1125 09:45:28.999207 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-a02b-account-create-dpj2t" podStartSLOduration=2.9991924450000003 podStartE2EDuration="2.999192445s" podCreationTimestamp="2025-11-25 09:45:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:45:28.997984185 +0000 UTC m=+1274.039043748" watchObservedRunningTime="2025-11-25 09:45:28.999192445 +0000 UTC m=+1274.040251998" Nov 25 09:45:29 crc kubenswrapper[4776]: I1125 09:45:29.021369 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-t5rsg" podStartSLOduration=3.021350131 podStartE2EDuration="3.021350131s" podCreationTimestamp="2025-11-25 09:45:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:45:29.010931909 +0000 UTC m=+1274.051991462" watchObservedRunningTime="2025-11-25 09:45:29.021350131 +0000 UTC m=+1274.062409684" Nov 25 09:45:29 crc kubenswrapper[4776]: I1125 09:45:29.972659 4776 generic.go:334] "Generic (PLEG): container finished" podID="2b686729-83aa-48c6-9cce-cd28b26dd4b4" containerID="53eb409078616814ade62b3d6be13d09644f9faf11f7f4738e36db9a3a38eefb" exitCode=0 Nov 25 09:45:29 crc kubenswrapper[4776]: I1125 09:45:29.972727 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-t5rsg" event={"ID":"2b686729-83aa-48c6-9cce-cd28b26dd4b4","Type":"ContainerDied","Data":"53eb409078616814ade62b3d6be13d09644f9faf11f7f4738e36db9a3a38eefb"} Nov 25 09:45:29 crc kubenswrapper[4776]: I1125 09:45:29.975096 4776 generic.go:334] "Generic (PLEG): container finished" podID="15bbd95b-2071-4988-83c7-8411c8a974fb" containerID="ea0e61b5432e05c0fedaf2f966d0d05c1d4707abe42eef391fb793278d77f5c7" exitCode=0 Nov 25 09:45:29 crc kubenswrapper[4776]: I1125 09:45:29.975146 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-96l8q" event={"ID":"15bbd95b-2071-4988-83c7-8411c8a974fb","Type":"ContainerDied","Data":"ea0e61b5432e05c0fedaf2f966d0d05c1d4707abe42eef391fb793278d77f5c7"} Nov 25 09:45:29 crc kubenswrapper[4776]: I1125 09:45:29.977502 4776 generic.go:334] "Generic (PLEG): container finished" podID="789daf57-0bbd-4e06-84d4-37e9ca59e390" containerID="ef86e26a3c64bb3fa983b0e4cb1ba97b5d01472982d3395d5f9382b30981a3d2" exitCode=0 Nov 25 09:45:29 crc kubenswrapper[4776]: I1125 09:45:29.977564 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9dc4-account-create-cj5jv" event={"ID":"789daf57-0bbd-4e06-84d4-37e9ca59e390","Type":"ContainerDied","Data":"ef86e26a3c64bb3fa983b0e4cb1ba97b5d01472982d3395d5f9382b30981a3d2"} Nov 25 09:45:29 crc kubenswrapper[4776]: I1125 09:45:29.978942 4776 generic.go:334] "Generic (PLEG): container finished" podID="cae7ff0c-7548-455d-85c8-3e68554658ce" containerID="1505fc4e03c873c1557f6ddb2bb810c54cb98dd95eb6841cce7981791552b151" exitCode=0 Nov 25 09:45:29 crc kubenswrapper[4776]: I1125 09:45:29.978997 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-a02b-account-create-dpj2t" event={"ID":"cae7ff0c-7548-455d-85c8-3e68554658ce","Type":"ContainerDied","Data":"1505fc4e03c873c1557f6ddb2bb810c54cb98dd95eb6841cce7981791552b151"} Nov 25 09:45:29 crc kubenswrapper[4776]: I1125 09:45:29.980233 4776 generic.go:334] "Generic (PLEG): container finished" podID="610a6a76-1c03-49a0-850f-ee5131308f7b" containerID="c2a85ebe5755f4af658dd25fd030364579d92bdc461690f7c2025585b4779bd8" exitCode=0 Nov 25 09:45:29 crc kubenswrapper[4776]: I1125 09:45:29.980283 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-hc2rc" event={"ID":"610a6a76-1c03-49a0-850f-ee5131308f7b","Type":"ContainerDied","Data":"c2a85ebe5755f4af658dd25fd030364579d92bdc461690f7c2025585b4779bd8"} Nov 25 09:45:29 crc kubenswrapper[4776]: I1125 09:45:29.981782 4776 generic.go:334] "Generic (PLEG): container finished" podID="eadc8c7e-32f4-4bbf-8ac3-51ebbe471718" containerID="02c4c64a3ee9a1d43c07e566240a85097d6186935e3986969648ee476f3e015f" exitCode=0 Nov 25 09:45:29 crc kubenswrapper[4776]: I1125 09:45:29.981850 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3b46-account-create-kg8tk" event={"ID":"eadc8c7e-32f4-4bbf-8ac3-51ebbe471718","Type":"ContainerDied","Data":"02c4c64a3ee9a1d43c07e566240a85097d6186935e3986969648ee476f3e015f"} Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.285281 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.285724 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.342690 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.356369 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.467922 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-96l8q" Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.602783 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15bbd95b-2071-4988-83c7-8411c8a974fb-operator-scripts\") pod \"15bbd95b-2071-4988-83c7-8411c8a974fb\" (UID: \"15bbd95b-2071-4988-83c7-8411c8a974fb\") " Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.603308 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-474qp\" (UniqueName: \"kubernetes.io/projected/15bbd95b-2071-4988-83c7-8411c8a974fb-kube-api-access-474qp\") pod \"15bbd95b-2071-4988-83c7-8411c8a974fb\" (UID: \"15bbd95b-2071-4988-83c7-8411c8a974fb\") " Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.603737 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15bbd95b-2071-4988-83c7-8411c8a974fb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "15bbd95b-2071-4988-83c7-8411c8a974fb" (UID: "15bbd95b-2071-4988-83c7-8411c8a974fb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.612836 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15bbd95b-2071-4988-83c7-8411c8a974fb-kube-api-access-474qp" (OuterVolumeSpecName: "kube-api-access-474qp") pod "15bbd95b-2071-4988-83c7-8411c8a974fb" (UID: "15bbd95b-2071-4988-83c7-8411c8a974fb"). InnerVolumeSpecName "kube-api-access-474qp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.705043 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-474qp\" (UniqueName: \"kubernetes.io/projected/15bbd95b-2071-4988-83c7-8411c8a974fb-kube-api-access-474qp\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.705087 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15bbd95b-2071-4988-83c7-8411c8a974fb-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.746759 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-a02b-account-create-dpj2t" Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.760772 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-hc2rc" Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.770084 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-t5rsg" Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.780780 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9dc4-account-create-cj5jv" Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.796628 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3b46-account-create-kg8tk" Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.909017 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7hx5\" (UniqueName: \"kubernetes.io/projected/789daf57-0bbd-4e06-84d4-37e9ca59e390-kube-api-access-b7hx5\") pod \"789daf57-0bbd-4e06-84d4-37e9ca59e390\" (UID: \"789daf57-0bbd-4e06-84d4-37e9ca59e390\") " Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.909133 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/789daf57-0bbd-4e06-84d4-37e9ca59e390-operator-scripts\") pod \"789daf57-0bbd-4e06-84d4-37e9ca59e390\" (UID: \"789daf57-0bbd-4e06-84d4-37e9ca59e390\") " Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.909161 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/610a6a76-1c03-49a0-850f-ee5131308f7b-operator-scripts\") pod \"610a6a76-1c03-49a0-850f-ee5131308f7b\" (UID: \"610a6a76-1c03-49a0-850f-ee5131308f7b\") " Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.909200 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rg4j\" (UniqueName: \"kubernetes.io/projected/cae7ff0c-7548-455d-85c8-3e68554658ce-kube-api-access-2rg4j\") pod \"cae7ff0c-7548-455d-85c8-3e68554658ce\" (UID: \"cae7ff0c-7548-455d-85c8-3e68554658ce\") " Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.909577 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eadc8c7e-32f4-4bbf-8ac3-51ebbe471718-operator-scripts\") pod \"eadc8c7e-32f4-4bbf-8ac3-51ebbe471718\" (UID: \"eadc8c7e-32f4-4bbf-8ac3-51ebbe471718\") " Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.909647 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knlzn\" (UniqueName: \"kubernetes.io/projected/610a6a76-1c03-49a0-850f-ee5131308f7b-kube-api-access-knlzn\") pod \"610a6a76-1c03-49a0-850f-ee5131308f7b\" (UID: \"610a6a76-1c03-49a0-850f-ee5131308f7b\") " Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.909675 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdk9c\" (UniqueName: \"kubernetes.io/projected/eadc8c7e-32f4-4bbf-8ac3-51ebbe471718-kube-api-access-wdk9c\") pod \"eadc8c7e-32f4-4bbf-8ac3-51ebbe471718\" (UID: \"eadc8c7e-32f4-4bbf-8ac3-51ebbe471718\") " Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.909700 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cae7ff0c-7548-455d-85c8-3e68554658ce-operator-scripts\") pod \"cae7ff0c-7548-455d-85c8-3e68554658ce\" (UID: \"cae7ff0c-7548-455d-85c8-3e68554658ce\") " Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.909734 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgrrf\" (UniqueName: \"kubernetes.io/projected/2b686729-83aa-48c6-9cce-cd28b26dd4b4-kube-api-access-mgrrf\") pod \"2b686729-83aa-48c6-9cce-cd28b26dd4b4\" (UID: \"2b686729-83aa-48c6-9cce-cd28b26dd4b4\") " Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.909786 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b686729-83aa-48c6-9cce-cd28b26dd4b4-operator-scripts\") pod \"2b686729-83aa-48c6-9cce-cd28b26dd4b4\" (UID: \"2b686729-83aa-48c6-9cce-cd28b26dd4b4\") " Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.909856 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/789daf57-0bbd-4e06-84d4-37e9ca59e390-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "789daf57-0bbd-4e06-84d4-37e9ca59e390" (UID: "789daf57-0bbd-4e06-84d4-37e9ca59e390"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.910262 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/789daf57-0bbd-4e06-84d4-37e9ca59e390-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.910727 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b686729-83aa-48c6-9cce-cd28b26dd4b4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2b686729-83aa-48c6-9cce-cd28b26dd4b4" (UID: "2b686729-83aa-48c6-9cce-cd28b26dd4b4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.913024 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eadc8c7e-32f4-4bbf-8ac3-51ebbe471718-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "eadc8c7e-32f4-4bbf-8ac3-51ebbe471718" (UID: "eadc8c7e-32f4-4bbf-8ac3-51ebbe471718"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.913101 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cae7ff0c-7548-455d-85c8-3e68554658ce-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cae7ff0c-7548-455d-85c8-3e68554658ce" (UID: "cae7ff0c-7548-455d-85c8-3e68554658ce"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.913144 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/610a6a76-1c03-49a0-850f-ee5131308f7b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "610a6a76-1c03-49a0-850f-ee5131308f7b" (UID: "610a6a76-1c03-49a0-850f-ee5131308f7b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.913501 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/610a6a76-1c03-49a0-850f-ee5131308f7b-kube-api-access-knlzn" (OuterVolumeSpecName: "kube-api-access-knlzn") pod "610a6a76-1c03-49a0-850f-ee5131308f7b" (UID: "610a6a76-1c03-49a0-850f-ee5131308f7b"). InnerVolumeSpecName "kube-api-access-knlzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.914261 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/789daf57-0bbd-4e06-84d4-37e9ca59e390-kube-api-access-b7hx5" (OuterVolumeSpecName: "kube-api-access-b7hx5") pod "789daf57-0bbd-4e06-84d4-37e9ca59e390" (UID: "789daf57-0bbd-4e06-84d4-37e9ca59e390"). InnerVolumeSpecName "kube-api-access-b7hx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.914319 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cae7ff0c-7548-455d-85c8-3e68554658ce-kube-api-access-2rg4j" (OuterVolumeSpecName: "kube-api-access-2rg4j") pod "cae7ff0c-7548-455d-85c8-3e68554658ce" (UID: "cae7ff0c-7548-455d-85c8-3e68554658ce"). InnerVolumeSpecName "kube-api-access-2rg4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.916243 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eadc8c7e-32f4-4bbf-8ac3-51ebbe471718-kube-api-access-wdk9c" (OuterVolumeSpecName: "kube-api-access-wdk9c") pod "eadc8c7e-32f4-4bbf-8ac3-51ebbe471718" (UID: "eadc8c7e-32f4-4bbf-8ac3-51ebbe471718"). InnerVolumeSpecName "kube-api-access-wdk9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:45:31 crc kubenswrapper[4776]: I1125 09:45:31.916288 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b686729-83aa-48c6-9cce-cd28b26dd4b4-kube-api-access-mgrrf" (OuterVolumeSpecName: "kube-api-access-mgrrf") pod "2b686729-83aa-48c6-9cce-cd28b26dd4b4" (UID: "2b686729-83aa-48c6-9cce-cd28b26dd4b4"). InnerVolumeSpecName "kube-api-access-mgrrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.000932 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3b46-account-create-kg8tk" event={"ID":"eadc8c7e-32f4-4bbf-8ac3-51ebbe471718","Type":"ContainerDied","Data":"48bfa5d0c85a086f4312bea3bef7801423ee5d46443ee4da35210c0657df7f7e"} Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.000971 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48bfa5d0c85a086f4312bea3bef7801423ee5d46443ee4da35210c0657df7f7e" Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.001020 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3b46-account-create-kg8tk" Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.004508 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-t5rsg" event={"ID":"2b686729-83aa-48c6-9cce-cd28b26dd4b4","Type":"ContainerDied","Data":"c67e21c5097fc690c2580ce63600e95b5d99c9913d595b5c652e9b3b4492d167"} Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.004593 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c67e21c5097fc690c2580ce63600e95b5d99c9913d595b5c652e9b3b4492d167" Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.004527 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-t5rsg" Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.006214 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-96l8q" event={"ID":"15bbd95b-2071-4988-83c7-8411c8a974fb","Type":"ContainerDied","Data":"335b76d86b0e231a7d9b2565c084a457e86fd6a7274093a39474afa721443227"} Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.006257 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="335b76d86b0e231a7d9b2565c084a457e86fd6a7274093a39474afa721443227" Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.006348 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-96l8q" Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.008511 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9dc4-account-create-cj5jv" event={"ID":"789daf57-0bbd-4e06-84d4-37e9ca59e390","Type":"ContainerDied","Data":"2c222bdd5794026a23d247a04539e80acdd0aaf3500ec9ec855460bb24dd66e4"} Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.008535 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c222bdd5794026a23d247a04539e80acdd0aaf3500ec9ec855460bb24dd66e4" Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.008576 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9dc4-account-create-cj5jv" Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.013217 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgrrf\" (UniqueName: \"kubernetes.io/projected/2b686729-83aa-48c6-9cce-cd28b26dd4b4-kube-api-access-mgrrf\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.013249 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b686729-83aa-48c6-9cce-cd28b26dd4b4-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.013262 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7hx5\" (UniqueName: \"kubernetes.io/projected/789daf57-0bbd-4e06-84d4-37e9ca59e390-kube-api-access-b7hx5\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.013272 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/610a6a76-1c03-49a0-850f-ee5131308f7b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.013282 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rg4j\" (UniqueName: \"kubernetes.io/projected/cae7ff0c-7548-455d-85c8-3e68554658ce-kube-api-access-2rg4j\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.013292 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eadc8c7e-32f4-4bbf-8ac3-51ebbe471718-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.013301 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knlzn\" (UniqueName: \"kubernetes.io/projected/610a6a76-1c03-49a0-850f-ee5131308f7b-kube-api-access-knlzn\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.013313 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdk9c\" (UniqueName: \"kubernetes.io/projected/eadc8c7e-32f4-4bbf-8ac3-51ebbe471718-kube-api-access-wdk9c\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.013322 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cae7ff0c-7548-455d-85c8-3e68554658ce-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.015580 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-a02b-account-create-dpj2t" Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.015597 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-a02b-account-create-dpj2t" event={"ID":"cae7ff0c-7548-455d-85c8-3e68554658ce","Type":"ContainerDied","Data":"78f82d72b9460f3ae620b01a37b15fa9dcd68e694ad55687d67e18c6a2053e3e"} Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.015634 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78f82d72b9460f3ae620b01a37b15fa9dcd68e694ad55687d67e18c6a2053e3e" Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.017537 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-hc2rc" Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.017577 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-hc2rc" event={"ID":"610a6a76-1c03-49a0-850f-ee5131308f7b","Type":"ContainerDied","Data":"bcc5e54787dd7b2e8af725670bbbd4e7cfc0dcb3ebe8adc71a05eb0abdd09c20"} Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.017599 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bcc5e54787dd7b2e8af725670bbbd4e7cfc0dcb3ebe8adc71a05eb0abdd09c20" Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.017857 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 25 09:45:32 crc kubenswrapper[4776]: I1125 09:45:32.017896 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 25 09:45:33 crc kubenswrapper[4776]: I1125 09:45:33.517531 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 25 09:45:34 crc kubenswrapper[4776]: I1125 09:45:34.359144 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 25 09:45:34 crc kubenswrapper[4776]: I1125 09:45:34.359720 4776 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 09:45:34 crc kubenswrapper[4776]: I1125 09:45:34.363506 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.076323 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jfmbd"] Nov 25 09:45:37 crc kubenswrapper[4776]: E1125 09:45:37.077125 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b686729-83aa-48c6-9cce-cd28b26dd4b4" containerName="mariadb-database-create" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.077138 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b686729-83aa-48c6-9cce-cd28b26dd4b4" containerName="mariadb-database-create" Nov 25 09:45:37 crc kubenswrapper[4776]: E1125 09:45:37.077158 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cae7ff0c-7548-455d-85c8-3e68554658ce" containerName="mariadb-account-create" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.077164 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="cae7ff0c-7548-455d-85c8-3e68554658ce" containerName="mariadb-account-create" Nov 25 09:45:37 crc kubenswrapper[4776]: E1125 09:45:37.077174 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="789daf57-0bbd-4e06-84d4-37e9ca59e390" containerName="mariadb-account-create" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.077180 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="789daf57-0bbd-4e06-84d4-37e9ca59e390" containerName="mariadb-account-create" Nov 25 09:45:37 crc kubenswrapper[4776]: E1125 09:45:37.077187 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eadc8c7e-32f4-4bbf-8ac3-51ebbe471718" containerName="mariadb-account-create" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.077192 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="eadc8c7e-32f4-4bbf-8ac3-51ebbe471718" containerName="mariadb-account-create" Nov 25 09:45:37 crc kubenswrapper[4776]: E1125 09:45:37.077206 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="610a6a76-1c03-49a0-850f-ee5131308f7b" containerName="mariadb-database-create" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.077211 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="610a6a76-1c03-49a0-850f-ee5131308f7b" containerName="mariadb-database-create" Nov 25 09:45:37 crc kubenswrapper[4776]: E1125 09:45:37.077227 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15bbd95b-2071-4988-83c7-8411c8a974fb" containerName="mariadb-database-create" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.077233 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="15bbd95b-2071-4988-83c7-8411c8a974fb" containerName="mariadb-database-create" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.077394 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="15bbd95b-2071-4988-83c7-8411c8a974fb" containerName="mariadb-database-create" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.077408 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="cae7ff0c-7548-455d-85c8-3e68554658ce" containerName="mariadb-account-create" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.077417 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b686729-83aa-48c6-9cce-cd28b26dd4b4" containerName="mariadb-database-create" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.077430 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="789daf57-0bbd-4e06-84d4-37e9ca59e390" containerName="mariadb-account-create" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.077444 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="eadc8c7e-32f4-4bbf-8ac3-51ebbe471718" containerName="mariadb-account-create" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.077452 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="610a6a76-1c03-49a0-850f-ee5131308f7b" containerName="mariadb-database-create" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.078018 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jfmbd" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.081466 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.081469 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-8tfjb" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.081605 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.090682 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jfmbd"] Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.220935 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wg6n7\" (UniqueName: \"kubernetes.io/projected/19f20b56-a02b-4831-932c-0a58f91544da-kube-api-access-wg6n7\") pod \"nova-cell0-conductor-db-sync-jfmbd\" (UID: \"19f20b56-a02b-4831-932c-0a58f91544da\") " pod="openstack/nova-cell0-conductor-db-sync-jfmbd" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.221044 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19f20b56-a02b-4831-932c-0a58f91544da-scripts\") pod \"nova-cell0-conductor-db-sync-jfmbd\" (UID: \"19f20b56-a02b-4831-932c-0a58f91544da\") " pod="openstack/nova-cell0-conductor-db-sync-jfmbd" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.221107 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19f20b56-a02b-4831-932c-0a58f91544da-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-jfmbd\" (UID: \"19f20b56-a02b-4831-932c-0a58f91544da\") " pod="openstack/nova-cell0-conductor-db-sync-jfmbd" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.221187 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19f20b56-a02b-4831-932c-0a58f91544da-config-data\") pod \"nova-cell0-conductor-db-sync-jfmbd\" (UID: \"19f20b56-a02b-4831-932c-0a58f91544da\") " pod="openstack/nova-cell0-conductor-db-sync-jfmbd" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.322732 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg6n7\" (UniqueName: \"kubernetes.io/projected/19f20b56-a02b-4831-932c-0a58f91544da-kube-api-access-wg6n7\") pod \"nova-cell0-conductor-db-sync-jfmbd\" (UID: \"19f20b56-a02b-4831-932c-0a58f91544da\") " pod="openstack/nova-cell0-conductor-db-sync-jfmbd" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.322928 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19f20b56-a02b-4831-932c-0a58f91544da-scripts\") pod \"nova-cell0-conductor-db-sync-jfmbd\" (UID: \"19f20b56-a02b-4831-932c-0a58f91544da\") " pod="openstack/nova-cell0-conductor-db-sync-jfmbd" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.323008 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19f20b56-a02b-4831-932c-0a58f91544da-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-jfmbd\" (UID: \"19f20b56-a02b-4831-932c-0a58f91544da\") " pod="openstack/nova-cell0-conductor-db-sync-jfmbd" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.323148 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19f20b56-a02b-4831-932c-0a58f91544da-config-data\") pod \"nova-cell0-conductor-db-sync-jfmbd\" (UID: \"19f20b56-a02b-4831-932c-0a58f91544da\") " pod="openstack/nova-cell0-conductor-db-sync-jfmbd" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.328936 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19f20b56-a02b-4831-932c-0a58f91544da-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-jfmbd\" (UID: \"19f20b56-a02b-4831-932c-0a58f91544da\") " pod="openstack/nova-cell0-conductor-db-sync-jfmbd" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.329245 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19f20b56-a02b-4831-932c-0a58f91544da-scripts\") pod \"nova-cell0-conductor-db-sync-jfmbd\" (UID: \"19f20b56-a02b-4831-932c-0a58f91544da\") " pod="openstack/nova-cell0-conductor-db-sync-jfmbd" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.336933 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19f20b56-a02b-4831-932c-0a58f91544da-config-data\") pod \"nova-cell0-conductor-db-sync-jfmbd\" (UID: \"19f20b56-a02b-4831-932c-0a58f91544da\") " pod="openstack/nova-cell0-conductor-db-sync-jfmbd" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.338705 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg6n7\" (UniqueName: \"kubernetes.io/projected/19f20b56-a02b-4831-932c-0a58f91544da-kube-api-access-wg6n7\") pod \"nova-cell0-conductor-db-sync-jfmbd\" (UID: \"19f20b56-a02b-4831-932c-0a58f91544da\") " pod="openstack/nova-cell0-conductor-db-sync-jfmbd" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.395178 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jfmbd" Nov 25 09:45:37 crc kubenswrapper[4776]: I1125 09:45:37.880133 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jfmbd"] Nov 25 09:45:37 crc kubenswrapper[4776]: W1125 09:45:37.886906 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19f20b56_a02b_4831_932c_0a58f91544da.slice/crio-615f6e9755f15fe770538509e9a53dfc23fe8cf94241ee15d18504f4af4b8c40 WatchSource:0}: Error finding container 615f6e9755f15fe770538509e9a53dfc23fe8cf94241ee15d18504f4af4b8c40: Status 404 returned error can't find the container with id 615f6e9755f15fe770538509e9a53dfc23fe8cf94241ee15d18504f4af4b8c40 Nov 25 09:45:38 crc kubenswrapper[4776]: I1125 09:45:38.088687 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jfmbd" event={"ID":"19f20b56-a02b-4831-932c-0a58f91544da","Type":"ContainerStarted","Data":"615f6e9755f15fe770538509e9a53dfc23fe8cf94241ee15d18504f4af4b8c40"} Nov 25 09:45:38 crc kubenswrapper[4776]: I1125 09:45:38.221477 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="768b7f4e-d631-4cc7-ad91-7e1099c8b354" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 25 09:45:45 crc kubenswrapper[4776]: I1125 09:45:45.145212 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jfmbd" event={"ID":"19f20b56-a02b-4831-932c-0a58f91544da","Type":"ContainerStarted","Data":"de673d55efb03d21b4bc57b9296a8b59f87945fbe34c20d80c4980cad983590b"} Nov 25 09:45:45 crc kubenswrapper[4776]: I1125 09:45:45.170392 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-jfmbd" podStartSLOduration=1.199683576 podStartE2EDuration="8.170370639s" podCreationTimestamp="2025-11-25 09:45:37 +0000 UTC" firstStartedPulling="2025-11-25 09:45:37.889319091 +0000 UTC m=+1282.930378634" lastFinishedPulling="2025-11-25 09:45:44.860006144 +0000 UTC m=+1289.901065697" observedRunningTime="2025-11-25 09:45:45.159802054 +0000 UTC m=+1290.200861617" watchObservedRunningTime="2025-11-25 09:45:45.170370639 +0000 UTC m=+1290.211430192" Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.175425 4776 generic.go:334] "Generic (PLEG): container finished" podID="768b7f4e-d631-4cc7-ad91-7e1099c8b354" containerID="c4b874dd8103aee5096d8a55ec80822635fd055cddb05d21f01b9482504a4e70" exitCode=137 Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.175455 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.175491 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"768b7f4e-d631-4cc7-ad91-7e1099c8b354","Type":"ContainerDied","Data":"c4b874dd8103aee5096d8a55ec80822635fd055cddb05d21f01b9482504a4e70"} Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.176369 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"768b7f4e-d631-4cc7-ad91-7e1099c8b354","Type":"ContainerDied","Data":"4c55ce75f985912e742dc4721c3d41f8d89b500b82fa5f2748e397b897f3fea0"} Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.176437 4776 scope.go:117] "RemoveContainer" containerID="c4b874dd8103aee5096d8a55ec80822635fd055cddb05d21f01b9482504a4e70" Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.202435 4776 scope.go:117] "RemoveContainer" containerID="5f04e4d9c26603eeb50c21d39c52c1a5e24f749d1eff7c8516f0fc36f3f2ef84" Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.222896 4776 scope.go:117] "RemoveContainer" containerID="01030c402508d2fdc1e43c8737b1f1f54cd200360facb57d6a5733ecae97896a" Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.227155 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-md7lz\" (UniqueName: \"kubernetes.io/projected/768b7f4e-d631-4cc7-ad91-7e1099c8b354-kube-api-access-md7lz\") pod \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\" (UID: \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\") " Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.227229 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/768b7f4e-d631-4cc7-ad91-7e1099c8b354-sg-core-conf-yaml\") pod \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\" (UID: \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\") " Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.227292 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/768b7f4e-d631-4cc7-ad91-7e1099c8b354-config-data\") pod \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\" (UID: \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\") " Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.227327 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/768b7f4e-d631-4cc7-ad91-7e1099c8b354-scripts\") pod \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\" (UID: \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\") " Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.227569 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/768b7f4e-d631-4cc7-ad91-7e1099c8b354-combined-ca-bundle\") pod \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\" (UID: \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\") " Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.227620 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/768b7f4e-d631-4cc7-ad91-7e1099c8b354-log-httpd\") pod \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\" (UID: \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\") " Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.227688 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/768b7f4e-d631-4cc7-ad91-7e1099c8b354-run-httpd\") pod \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\" (UID: \"768b7f4e-d631-4cc7-ad91-7e1099c8b354\") " Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.228763 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/768b7f4e-d631-4cc7-ad91-7e1099c8b354-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "768b7f4e-d631-4cc7-ad91-7e1099c8b354" (UID: "768b7f4e-d631-4cc7-ad91-7e1099c8b354"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.229062 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/768b7f4e-d631-4cc7-ad91-7e1099c8b354-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "768b7f4e-d631-4cc7-ad91-7e1099c8b354" (UID: "768b7f4e-d631-4cc7-ad91-7e1099c8b354"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.229122 4776 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/768b7f4e-d631-4cc7-ad91-7e1099c8b354-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.233922 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/768b7f4e-d631-4cc7-ad91-7e1099c8b354-scripts" (OuterVolumeSpecName: "scripts") pod "768b7f4e-d631-4cc7-ad91-7e1099c8b354" (UID: "768b7f4e-d631-4cc7-ad91-7e1099c8b354"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.234609 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/768b7f4e-d631-4cc7-ad91-7e1099c8b354-kube-api-access-md7lz" (OuterVolumeSpecName: "kube-api-access-md7lz") pod "768b7f4e-d631-4cc7-ad91-7e1099c8b354" (UID: "768b7f4e-d631-4cc7-ad91-7e1099c8b354"). InnerVolumeSpecName "kube-api-access-md7lz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.243030 4776 scope.go:117] "RemoveContainer" containerID="6515b87d6beb7f876edae119be02a45958d18ae4c801f6d415a59604678f66d8" Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.255171 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/768b7f4e-d631-4cc7-ad91-7e1099c8b354-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "768b7f4e-d631-4cc7-ad91-7e1099c8b354" (UID: "768b7f4e-d631-4cc7-ad91-7e1099c8b354"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.302595 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/768b7f4e-d631-4cc7-ad91-7e1099c8b354-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "768b7f4e-d631-4cc7-ad91-7e1099c8b354" (UID: "768b7f4e-d631-4cc7-ad91-7e1099c8b354"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.318410 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/768b7f4e-d631-4cc7-ad91-7e1099c8b354-config-data" (OuterVolumeSpecName: "config-data") pod "768b7f4e-d631-4cc7-ad91-7e1099c8b354" (UID: "768b7f4e-d631-4cc7-ad91-7e1099c8b354"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.331314 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/768b7f4e-d631-4cc7-ad91-7e1099c8b354-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.331344 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/768b7f4e-d631-4cc7-ad91-7e1099c8b354-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.331353 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/768b7f4e-d631-4cc7-ad91-7e1099c8b354-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.331366 4776 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/768b7f4e-d631-4cc7-ad91-7e1099c8b354-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.331374 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-md7lz\" (UniqueName: \"kubernetes.io/projected/768b7f4e-d631-4cc7-ad91-7e1099c8b354-kube-api-access-md7lz\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.331385 4776 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/768b7f4e-d631-4cc7-ad91-7e1099c8b354-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.335511 4776 scope.go:117] "RemoveContainer" containerID="c4b874dd8103aee5096d8a55ec80822635fd055cddb05d21f01b9482504a4e70" Nov 25 09:45:47 crc kubenswrapper[4776]: E1125 09:45:47.336045 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4b874dd8103aee5096d8a55ec80822635fd055cddb05d21f01b9482504a4e70\": container with ID starting with c4b874dd8103aee5096d8a55ec80822635fd055cddb05d21f01b9482504a4e70 not found: ID does not exist" containerID="c4b874dd8103aee5096d8a55ec80822635fd055cddb05d21f01b9482504a4e70" Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.336097 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4b874dd8103aee5096d8a55ec80822635fd055cddb05d21f01b9482504a4e70"} err="failed to get container status \"c4b874dd8103aee5096d8a55ec80822635fd055cddb05d21f01b9482504a4e70\": rpc error: code = NotFound desc = could not find container \"c4b874dd8103aee5096d8a55ec80822635fd055cddb05d21f01b9482504a4e70\": container with ID starting with c4b874dd8103aee5096d8a55ec80822635fd055cddb05d21f01b9482504a4e70 not found: ID does not exist" Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.336122 4776 scope.go:117] "RemoveContainer" containerID="5f04e4d9c26603eeb50c21d39c52c1a5e24f749d1eff7c8516f0fc36f3f2ef84" Nov 25 09:45:47 crc kubenswrapper[4776]: E1125 09:45:47.336420 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f04e4d9c26603eeb50c21d39c52c1a5e24f749d1eff7c8516f0fc36f3f2ef84\": container with ID starting with 5f04e4d9c26603eeb50c21d39c52c1a5e24f749d1eff7c8516f0fc36f3f2ef84 not found: ID does not exist" containerID="5f04e4d9c26603eeb50c21d39c52c1a5e24f749d1eff7c8516f0fc36f3f2ef84" Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.336466 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f04e4d9c26603eeb50c21d39c52c1a5e24f749d1eff7c8516f0fc36f3f2ef84"} err="failed to get container status \"5f04e4d9c26603eeb50c21d39c52c1a5e24f749d1eff7c8516f0fc36f3f2ef84\": rpc error: code = NotFound desc = could not find container \"5f04e4d9c26603eeb50c21d39c52c1a5e24f749d1eff7c8516f0fc36f3f2ef84\": container with ID starting with 5f04e4d9c26603eeb50c21d39c52c1a5e24f749d1eff7c8516f0fc36f3f2ef84 not found: ID does not exist" Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.336491 4776 scope.go:117] "RemoveContainer" containerID="01030c402508d2fdc1e43c8737b1f1f54cd200360facb57d6a5733ecae97896a" Nov 25 09:45:47 crc kubenswrapper[4776]: E1125 09:45:47.336767 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01030c402508d2fdc1e43c8737b1f1f54cd200360facb57d6a5733ecae97896a\": container with ID starting with 01030c402508d2fdc1e43c8737b1f1f54cd200360facb57d6a5733ecae97896a not found: ID does not exist" containerID="01030c402508d2fdc1e43c8737b1f1f54cd200360facb57d6a5733ecae97896a" Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.336797 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01030c402508d2fdc1e43c8737b1f1f54cd200360facb57d6a5733ecae97896a"} err="failed to get container status \"01030c402508d2fdc1e43c8737b1f1f54cd200360facb57d6a5733ecae97896a\": rpc error: code = NotFound desc = could not find container \"01030c402508d2fdc1e43c8737b1f1f54cd200360facb57d6a5733ecae97896a\": container with ID starting with 01030c402508d2fdc1e43c8737b1f1f54cd200360facb57d6a5733ecae97896a not found: ID does not exist" Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.336813 4776 scope.go:117] "RemoveContainer" containerID="6515b87d6beb7f876edae119be02a45958d18ae4c801f6d415a59604678f66d8" Nov 25 09:45:47 crc kubenswrapper[4776]: E1125 09:45:47.337001 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6515b87d6beb7f876edae119be02a45958d18ae4c801f6d415a59604678f66d8\": container with ID starting with 6515b87d6beb7f876edae119be02a45958d18ae4c801f6d415a59604678f66d8 not found: ID does not exist" containerID="6515b87d6beb7f876edae119be02a45958d18ae4c801f6d415a59604678f66d8" Nov 25 09:45:47 crc kubenswrapper[4776]: I1125 09:45:47.337023 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6515b87d6beb7f876edae119be02a45958d18ae4c801f6d415a59604678f66d8"} err="failed to get container status \"6515b87d6beb7f876edae119be02a45958d18ae4c801f6d415a59604678f66d8\": rpc error: code = NotFound desc = could not find container \"6515b87d6beb7f876edae119be02a45958d18ae4c801f6d415a59604678f66d8\": container with ID starting with 6515b87d6beb7f876edae119be02a45958d18ae4c801f6d415a59604678f66d8 not found: ID does not exist" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.201831 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.233907 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.245992 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.260629 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:45:48 crc kubenswrapper[4776]: E1125 09:45:48.261171 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="768b7f4e-d631-4cc7-ad91-7e1099c8b354" containerName="ceilometer-notification-agent" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.261205 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="768b7f4e-d631-4cc7-ad91-7e1099c8b354" containerName="ceilometer-notification-agent" Nov 25 09:45:48 crc kubenswrapper[4776]: E1125 09:45:48.261283 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="768b7f4e-d631-4cc7-ad91-7e1099c8b354" containerName="ceilometer-central-agent" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.261297 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="768b7f4e-d631-4cc7-ad91-7e1099c8b354" containerName="ceilometer-central-agent" Nov 25 09:45:48 crc kubenswrapper[4776]: E1125 09:45:48.261352 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="768b7f4e-d631-4cc7-ad91-7e1099c8b354" containerName="sg-core" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.261365 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="768b7f4e-d631-4cc7-ad91-7e1099c8b354" containerName="sg-core" Nov 25 09:45:48 crc kubenswrapper[4776]: E1125 09:45:48.261393 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="768b7f4e-d631-4cc7-ad91-7e1099c8b354" containerName="proxy-httpd" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.261402 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="768b7f4e-d631-4cc7-ad91-7e1099c8b354" containerName="proxy-httpd" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.261710 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="768b7f4e-d631-4cc7-ad91-7e1099c8b354" containerName="ceilometer-notification-agent" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.261755 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="768b7f4e-d631-4cc7-ad91-7e1099c8b354" containerName="sg-core" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.261777 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="768b7f4e-d631-4cc7-ad91-7e1099c8b354" containerName="ceilometer-central-agent" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.261802 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="768b7f4e-d631-4cc7-ad91-7e1099c8b354" containerName="proxy-httpd" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.268593 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.273144 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.273666 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.275504 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.275815 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.349346 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-config-data\") pod \"ceilometer-0\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " pod="openstack/ceilometer-0" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.349405 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-scripts\") pod \"ceilometer-0\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " pod="openstack/ceilometer-0" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.349427 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " pod="openstack/ceilometer-0" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.349462 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzwnc\" (UniqueName: \"kubernetes.io/projected/f1e0bba8-06f9-481c-a780-bfc17ec5222c-kube-api-access-pzwnc\") pod \"ceilometer-0\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " pod="openstack/ceilometer-0" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.349543 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " pod="openstack/ceilometer-0" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.349581 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f1e0bba8-06f9-481c-a780-bfc17ec5222c-log-httpd\") pod \"ceilometer-0\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " pod="openstack/ceilometer-0" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.349637 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " pod="openstack/ceilometer-0" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.349740 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f1e0bba8-06f9-481c-a780-bfc17ec5222c-run-httpd\") pod \"ceilometer-0\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " pod="openstack/ceilometer-0" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.451130 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f1e0bba8-06f9-481c-a780-bfc17ec5222c-run-httpd\") pod \"ceilometer-0\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " pod="openstack/ceilometer-0" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.451260 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-config-data\") pod \"ceilometer-0\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " pod="openstack/ceilometer-0" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.451316 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-scripts\") pod \"ceilometer-0\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " pod="openstack/ceilometer-0" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.451337 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " pod="openstack/ceilometer-0" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.451375 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzwnc\" (UniqueName: \"kubernetes.io/projected/f1e0bba8-06f9-481c-a780-bfc17ec5222c-kube-api-access-pzwnc\") pod \"ceilometer-0\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " pod="openstack/ceilometer-0" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.451422 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " pod="openstack/ceilometer-0" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.451441 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f1e0bba8-06f9-481c-a780-bfc17ec5222c-log-httpd\") pod \"ceilometer-0\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " pod="openstack/ceilometer-0" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.451489 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " pod="openstack/ceilometer-0" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.451818 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f1e0bba8-06f9-481c-a780-bfc17ec5222c-run-httpd\") pod \"ceilometer-0\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " pod="openstack/ceilometer-0" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.452225 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f1e0bba8-06f9-481c-a780-bfc17ec5222c-log-httpd\") pod \"ceilometer-0\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " pod="openstack/ceilometer-0" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.457412 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " pod="openstack/ceilometer-0" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.457594 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " pod="openstack/ceilometer-0" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.458784 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " pod="openstack/ceilometer-0" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.468672 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-config-data\") pod \"ceilometer-0\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " pod="openstack/ceilometer-0" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.472055 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-scripts\") pod \"ceilometer-0\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " pod="openstack/ceilometer-0" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.472938 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzwnc\" (UniqueName: \"kubernetes.io/projected/f1e0bba8-06f9-481c-a780-bfc17ec5222c-kube-api-access-pzwnc\") pod \"ceilometer-0\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " pod="openstack/ceilometer-0" Nov 25 09:45:48 crc kubenswrapper[4776]: I1125 09:45:48.590654 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 09:45:49 crc kubenswrapper[4776]: I1125 09:45:49.022006 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:45:49 crc kubenswrapper[4776]: I1125 09:45:49.215102 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f1e0bba8-06f9-481c-a780-bfc17ec5222c","Type":"ContainerStarted","Data":"0c10415e998a38a6825c9564b5c9eb05f9577d6610db0f9e10b2b4a9fa196edd"} Nov 25 09:45:49 crc kubenswrapper[4776]: I1125 09:45:49.676571 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="768b7f4e-d631-4cc7-ad91-7e1099c8b354" path="/var/lib/kubelet/pods/768b7f4e-d631-4cc7-ad91-7e1099c8b354/volumes" Nov 25 09:45:50 crc kubenswrapper[4776]: I1125 09:45:50.225455 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f1e0bba8-06f9-481c-a780-bfc17ec5222c","Type":"ContainerStarted","Data":"a61faed17a75f0fadba92e5ee69f5a2f1d3c2259dffec60af1df198f74f2f188"} Nov 25 09:45:51 crc kubenswrapper[4776]: I1125 09:45:51.234687 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f1e0bba8-06f9-481c-a780-bfc17ec5222c","Type":"ContainerStarted","Data":"bf1ec9e8754cbd3f26c19b9836d8b9f855007d555bfd31191664e08b436499e8"} Nov 25 09:45:52 crc kubenswrapper[4776]: I1125 09:45:52.247299 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f1e0bba8-06f9-481c-a780-bfc17ec5222c","Type":"ContainerStarted","Data":"b6d6e1b8d62ded83fbd5439c8721571564758b292c4ad38aae991861620d9cfb"} Nov 25 09:45:54 crc kubenswrapper[4776]: I1125 09:45:54.270736 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f1e0bba8-06f9-481c-a780-bfc17ec5222c","Type":"ContainerStarted","Data":"30c6ca260ebb54bae727c7b09725297576062adf42235ac89f7f0b9f8fe54531"} Nov 25 09:45:54 crc kubenswrapper[4776]: I1125 09:45:54.271388 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 09:45:54 crc kubenswrapper[4776]: I1125 09:45:54.300217 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.643970478 podStartE2EDuration="6.300195479s" podCreationTimestamp="2025-11-25 09:45:48 +0000 UTC" firstStartedPulling="2025-11-25 09:45:49.026213923 +0000 UTC m=+1294.067273476" lastFinishedPulling="2025-11-25 09:45:53.682438924 +0000 UTC m=+1298.723498477" observedRunningTime="2025-11-25 09:45:54.294244769 +0000 UTC m=+1299.335304322" watchObservedRunningTime="2025-11-25 09:45:54.300195479 +0000 UTC m=+1299.341255042" Nov 25 09:45:58 crc kubenswrapper[4776]: I1125 09:45:58.305361 4776 generic.go:334] "Generic (PLEG): container finished" podID="19f20b56-a02b-4831-932c-0a58f91544da" containerID="de673d55efb03d21b4bc57b9296a8b59f87945fbe34c20d80c4980cad983590b" exitCode=0 Nov 25 09:45:58 crc kubenswrapper[4776]: I1125 09:45:58.305461 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jfmbd" event={"ID":"19f20b56-a02b-4831-932c-0a58f91544da","Type":"ContainerDied","Data":"de673d55efb03d21b4bc57b9296a8b59f87945fbe34c20d80c4980cad983590b"} Nov 25 09:45:59 crc kubenswrapper[4776]: I1125 09:45:59.599045 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jfmbd" Nov 25 09:45:59 crc kubenswrapper[4776]: I1125 09:45:59.689593 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19f20b56-a02b-4831-932c-0a58f91544da-config-data\") pod \"19f20b56-a02b-4831-932c-0a58f91544da\" (UID: \"19f20b56-a02b-4831-932c-0a58f91544da\") " Nov 25 09:45:59 crc kubenswrapper[4776]: I1125 09:45:59.689871 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19f20b56-a02b-4831-932c-0a58f91544da-combined-ca-bundle\") pod \"19f20b56-a02b-4831-932c-0a58f91544da\" (UID: \"19f20b56-a02b-4831-932c-0a58f91544da\") " Nov 25 09:45:59 crc kubenswrapper[4776]: I1125 09:45:59.689903 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wg6n7\" (UniqueName: \"kubernetes.io/projected/19f20b56-a02b-4831-932c-0a58f91544da-kube-api-access-wg6n7\") pod \"19f20b56-a02b-4831-932c-0a58f91544da\" (UID: \"19f20b56-a02b-4831-932c-0a58f91544da\") " Nov 25 09:45:59 crc kubenswrapper[4776]: I1125 09:45:59.690685 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19f20b56-a02b-4831-932c-0a58f91544da-scripts\") pod \"19f20b56-a02b-4831-932c-0a58f91544da\" (UID: \"19f20b56-a02b-4831-932c-0a58f91544da\") " Nov 25 09:45:59 crc kubenswrapper[4776]: I1125 09:45:59.695080 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19f20b56-a02b-4831-932c-0a58f91544da-kube-api-access-wg6n7" (OuterVolumeSpecName: "kube-api-access-wg6n7") pod "19f20b56-a02b-4831-932c-0a58f91544da" (UID: "19f20b56-a02b-4831-932c-0a58f91544da"). InnerVolumeSpecName "kube-api-access-wg6n7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:45:59 crc kubenswrapper[4776]: I1125 09:45:59.695223 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19f20b56-a02b-4831-932c-0a58f91544da-scripts" (OuterVolumeSpecName: "scripts") pod "19f20b56-a02b-4831-932c-0a58f91544da" (UID: "19f20b56-a02b-4831-932c-0a58f91544da"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:59 crc kubenswrapper[4776]: I1125 09:45:59.716328 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19f20b56-a02b-4831-932c-0a58f91544da-config-data" (OuterVolumeSpecName: "config-data") pod "19f20b56-a02b-4831-932c-0a58f91544da" (UID: "19f20b56-a02b-4831-932c-0a58f91544da"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:59 crc kubenswrapper[4776]: I1125 09:45:59.722736 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19f20b56-a02b-4831-932c-0a58f91544da-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "19f20b56-a02b-4831-932c-0a58f91544da" (UID: "19f20b56-a02b-4831-932c-0a58f91544da"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:45:59 crc kubenswrapper[4776]: I1125 09:45:59.793424 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19f20b56-a02b-4831-932c-0a58f91544da-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:59 crc kubenswrapper[4776]: I1125 09:45:59.793454 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wg6n7\" (UniqueName: \"kubernetes.io/projected/19f20b56-a02b-4831-932c-0a58f91544da-kube-api-access-wg6n7\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:59 crc kubenswrapper[4776]: I1125 09:45:59.793468 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19f20b56-a02b-4831-932c-0a58f91544da-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:45:59 crc kubenswrapper[4776]: I1125 09:45:59.793483 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19f20b56-a02b-4831-932c-0a58f91544da-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:00 crc kubenswrapper[4776]: I1125 09:46:00.324296 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jfmbd" event={"ID":"19f20b56-a02b-4831-932c-0a58f91544da","Type":"ContainerDied","Data":"615f6e9755f15fe770538509e9a53dfc23fe8cf94241ee15d18504f4af4b8c40"} Nov 25 09:46:00 crc kubenswrapper[4776]: I1125 09:46:00.324726 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="615f6e9755f15fe770538509e9a53dfc23fe8cf94241ee15d18504f4af4b8c40" Nov 25 09:46:00 crc kubenswrapper[4776]: I1125 09:46:00.324343 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jfmbd" Nov 25 09:46:00 crc kubenswrapper[4776]: I1125 09:46:00.403564 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 09:46:00 crc kubenswrapper[4776]: E1125 09:46:00.403916 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19f20b56-a02b-4831-932c-0a58f91544da" containerName="nova-cell0-conductor-db-sync" Nov 25 09:46:00 crc kubenswrapper[4776]: I1125 09:46:00.403934 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="19f20b56-a02b-4831-932c-0a58f91544da" containerName="nova-cell0-conductor-db-sync" Nov 25 09:46:00 crc kubenswrapper[4776]: I1125 09:46:00.404156 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="19f20b56-a02b-4831-932c-0a58f91544da" containerName="nova-cell0-conductor-db-sync" Nov 25 09:46:00 crc kubenswrapper[4776]: I1125 09:46:00.404777 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 25 09:46:00 crc kubenswrapper[4776]: I1125 09:46:00.406529 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-8tfjb" Nov 25 09:46:00 crc kubenswrapper[4776]: I1125 09:46:00.406881 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 25 09:46:00 crc kubenswrapper[4776]: I1125 09:46:00.420551 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 09:46:00 crc kubenswrapper[4776]: I1125 09:46:00.506227 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4\") " pod="openstack/nova-cell0-conductor-0" Nov 25 09:46:00 crc kubenswrapper[4776]: I1125 09:46:00.506533 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4\") " pod="openstack/nova-cell0-conductor-0" Nov 25 09:46:00 crc kubenswrapper[4776]: I1125 09:46:00.506661 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf7l8\" (UniqueName: \"kubernetes.io/projected/f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4-kube-api-access-bf7l8\") pod \"nova-cell0-conductor-0\" (UID: \"f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4\") " pod="openstack/nova-cell0-conductor-0" Nov 25 09:46:00 crc kubenswrapper[4776]: I1125 09:46:00.607943 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4\") " pod="openstack/nova-cell0-conductor-0" Nov 25 09:46:00 crc kubenswrapper[4776]: I1125 09:46:00.608002 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4\") " pod="openstack/nova-cell0-conductor-0" Nov 25 09:46:00 crc kubenswrapper[4776]: I1125 09:46:00.608052 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf7l8\" (UniqueName: \"kubernetes.io/projected/f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4-kube-api-access-bf7l8\") pod \"nova-cell0-conductor-0\" (UID: \"f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4\") " pod="openstack/nova-cell0-conductor-0" Nov 25 09:46:00 crc kubenswrapper[4776]: I1125 09:46:00.613044 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4\") " pod="openstack/nova-cell0-conductor-0" Nov 25 09:46:00 crc kubenswrapper[4776]: I1125 09:46:00.613563 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4\") " pod="openstack/nova-cell0-conductor-0" Nov 25 09:46:00 crc kubenswrapper[4776]: I1125 09:46:00.624391 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf7l8\" (UniqueName: \"kubernetes.io/projected/f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4-kube-api-access-bf7l8\") pod \"nova-cell0-conductor-0\" (UID: \"f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4\") " pod="openstack/nova-cell0-conductor-0" Nov 25 09:46:00 crc kubenswrapper[4776]: I1125 09:46:00.764925 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 25 09:46:01 crc kubenswrapper[4776]: W1125 09:46:01.175289 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9cc962d_25ca_4dcb_9fb3_06eaa655c3b4.slice/crio-a7cce054b7bef15b7a4a6a15d9d6f8782b9044f4308d061712276043ffdd055d WatchSource:0}: Error finding container a7cce054b7bef15b7a4a6a15d9d6f8782b9044f4308d061712276043ffdd055d: Status 404 returned error can't find the container with id a7cce054b7bef15b7a4a6a15d9d6f8782b9044f4308d061712276043ffdd055d Nov 25 09:46:01 crc kubenswrapper[4776]: I1125 09:46:01.175687 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 09:46:01 crc kubenswrapper[4776]: I1125 09:46:01.336685 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4","Type":"ContainerStarted","Data":"a7cce054b7bef15b7a4a6a15d9d6f8782b9044f4308d061712276043ffdd055d"} Nov 25 09:46:02 crc kubenswrapper[4776]: I1125 09:46:02.349783 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4","Type":"ContainerStarted","Data":"51468cb0acece8ce1a386bdf49793e6322b08cddac5cf23fc46abef1f5330e22"} Nov 25 09:46:02 crc kubenswrapper[4776]: I1125 09:46:02.350156 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 25 09:46:02 crc kubenswrapper[4776]: I1125 09:46:02.364841 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.364818749 podStartE2EDuration="2.364818749s" podCreationTimestamp="2025-11-25 09:46:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:46:02.362687446 +0000 UTC m=+1307.403746999" watchObservedRunningTime="2025-11-25 09:46:02.364818749 +0000 UTC m=+1307.405878302" Nov 25 09:46:10 crc kubenswrapper[4776]: I1125 09:46:10.801575 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.301767 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-69dpf"] Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.303230 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-69dpf" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.306124 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.306971 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.314567 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-69dpf"] Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.384930 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22bb0550-ee44-46e6-b45e-1158524bfa3e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-69dpf\" (UID: \"22bb0550-ee44-46e6-b45e-1158524bfa3e\") " pod="openstack/nova-cell0-cell-mapping-69dpf" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.384994 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22bb0550-ee44-46e6-b45e-1158524bfa3e-scripts\") pod \"nova-cell0-cell-mapping-69dpf\" (UID: \"22bb0550-ee44-46e6-b45e-1158524bfa3e\") " pod="openstack/nova-cell0-cell-mapping-69dpf" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.385085 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22bb0550-ee44-46e6-b45e-1158524bfa3e-config-data\") pod \"nova-cell0-cell-mapping-69dpf\" (UID: \"22bb0550-ee44-46e6-b45e-1158524bfa3e\") " pod="openstack/nova-cell0-cell-mapping-69dpf" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.385120 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7gsr\" (UniqueName: \"kubernetes.io/projected/22bb0550-ee44-46e6-b45e-1158524bfa3e-kube-api-access-z7gsr\") pod \"nova-cell0-cell-mapping-69dpf\" (UID: \"22bb0550-ee44-46e6-b45e-1158524bfa3e\") " pod="openstack/nova-cell0-cell-mapping-69dpf" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.486669 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22bb0550-ee44-46e6-b45e-1158524bfa3e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-69dpf\" (UID: \"22bb0550-ee44-46e6-b45e-1158524bfa3e\") " pod="openstack/nova-cell0-cell-mapping-69dpf" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.486720 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22bb0550-ee44-46e6-b45e-1158524bfa3e-scripts\") pod \"nova-cell0-cell-mapping-69dpf\" (UID: \"22bb0550-ee44-46e6-b45e-1158524bfa3e\") " pod="openstack/nova-cell0-cell-mapping-69dpf" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.486777 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22bb0550-ee44-46e6-b45e-1158524bfa3e-config-data\") pod \"nova-cell0-cell-mapping-69dpf\" (UID: \"22bb0550-ee44-46e6-b45e-1158524bfa3e\") " pod="openstack/nova-cell0-cell-mapping-69dpf" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.486808 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7gsr\" (UniqueName: \"kubernetes.io/projected/22bb0550-ee44-46e6-b45e-1158524bfa3e-kube-api-access-z7gsr\") pod \"nova-cell0-cell-mapping-69dpf\" (UID: \"22bb0550-ee44-46e6-b45e-1158524bfa3e\") " pod="openstack/nova-cell0-cell-mapping-69dpf" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.495034 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22bb0550-ee44-46e6-b45e-1158524bfa3e-scripts\") pod \"nova-cell0-cell-mapping-69dpf\" (UID: \"22bb0550-ee44-46e6-b45e-1158524bfa3e\") " pod="openstack/nova-cell0-cell-mapping-69dpf" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.495124 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22bb0550-ee44-46e6-b45e-1158524bfa3e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-69dpf\" (UID: \"22bb0550-ee44-46e6-b45e-1158524bfa3e\") " pod="openstack/nova-cell0-cell-mapping-69dpf" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.498121 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22bb0550-ee44-46e6-b45e-1158524bfa3e-config-data\") pod \"nova-cell0-cell-mapping-69dpf\" (UID: \"22bb0550-ee44-46e6-b45e-1158524bfa3e\") " pod="openstack/nova-cell0-cell-mapping-69dpf" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.520739 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7gsr\" (UniqueName: \"kubernetes.io/projected/22bb0550-ee44-46e6-b45e-1158524bfa3e-kube-api-access-z7gsr\") pod \"nova-cell0-cell-mapping-69dpf\" (UID: \"22bb0550-ee44-46e6-b45e-1158524bfa3e\") " pod="openstack/nova-cell0-cell-mapping-69dpf" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.559344 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.560880 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.574368 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.577640 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.579121 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.583683 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.587936 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c4q4\" (UniqueName: \"kubernetes.io/projected/d7225aed-20ef-47e7-94d3-ae900a0e55db-kube-api-access-8c4q4\") pod \"nova-api-0\" (UID: \"d7225aed-20ef-47e7-94d3-ae900a0e55db\") " pod="openstack/nova-api-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.588000 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/603ba880-f1f5-4893-9e81-bd2031b42192-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"603ba880-f1f5-4893-9e81-bd2031b42192\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.588027 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7225aed-20ef-47e7-94d3-ae900a0e55db-config-data\") pod \"nova-api-0\" (UID: \"d7225aed-20ef-47e7-94d3-ae900a0e55db\") " pod="openstack/nova-api-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.588085 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7225aed-20ef-47e7-94d3-ae900a0e55db-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d7225aed-20ef-47e7-94d3-ae900a0e55db\") " pod="openstack/nova-api-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.588106 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7225aed-20ef-47e7-94d3-ae900a0e55db-logs\") pod \"nova-api-0\" (UID: \"d7225aed-20ef-47e7-94d3-ae900a0e55db\") " pod="openstack/nova-api-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.588123 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnxhv\" (UniqueName: \"kubernetes.io/projected/603ba880-f1f5-4893-9e81-bd2031b42192-kube-api-access-mnxhv\") pod \"nova-cell1-novncproxy-0\" (UID: \"603ba880-f1f5-4893-9e81-bd2031b42192\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.588174 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/603ba880-f1f5-4893-9e81-bd2031b42192-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"603ba880-f1f5-4893-9e81-bd2031b42192\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.593641 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.608610 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.635628 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-69dpf" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.694652 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.696105 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.698036 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/603ba880-f1f5-4893-9e81-bd2031b42192-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"603ba880-f1f5-4893-9e81-bd2031b42192\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.698107 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c4q4\" (UniqueName: \"kubernetes.io/projected/d7225aed-20ef-47e7-94d3-ae900a0e55db-kube-api-access-8c4q4\") pod \"nova-api-0\" (UID: \"d7225aed-20ef-47e7-94d3-ae900a0e55db\") " pod="openstack/nova-api-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.698152 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/603ba880-f1f5-4893-9e81-bd2031b42192-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"603ba880-f1f5-4893-9e81-bd2031b42192\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.698177 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7225aed-20ef-47e7-94d3-ae900a0e55db-config-data\") pod \"nova-api-0\" (UID: \"d7225aed-20ef-47e7-94d3-ae900a0e55db\") " pod="openstack/nova-api-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.698221 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7225aed-20ef-47e7-94d3-ae900a0e55db-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d7225aed-20ef-47e7-94d3-ae900a0e55db\") " pod="openstack/nova-api-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.698239 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7225aed-20ef-47e7-94d3-ae900a0e55db-logs\") pod \"nova-api-0\" (UID: \"d7225aed-20ef-47e7-94d3-ae900a0e55db\") " pod="openstack/nova-api-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.698258 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnxhv\" (UniqueName: \"kubernetes.io/projected/603ba880-f1f5-4893-9e81-bd2031b42192-kube-api-access-mnxhv\") pod \"nova-cell1-novncproxy-0\" (UID: \"603ba880-f1f5-4893-9e81-bd2031b42192\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.703716 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/603ba880-f1f5-4893-9e81-bd2031b42192-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"603ba880-f1f5-4893-9e81-bd2031b42192\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.705683 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7225aed-20ef-47e7-94d3-ae900a0e55db-logs\") pod \"nova-api-0\" (UID: \"d7225aed-20ef-47e7-94d3-ae900a0e55db\") " pod="openstack/nova-api-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.706793 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7225aed-20ef-47e7-94d3-ae900a0e55db-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d7225aed-20ef-47e7-94d3-ae900a0e55db\") " pod="openstack/nova-api-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.709870 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7225aed-20ef-47e7-94d3-ae900a0e55db-config-data\") pod \"nova-api-0\" (UID: \"d7225aed-20ef-47e7-94d3-ae900a0e55db\") " pod="openstack/nova-api-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.721136 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.722608 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.724436 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/603ba880-f1f5-4893-9e81-bd2031b42192-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"603ba880-f1f5-4893-9e81-bd2031b42192\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.777608 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.782170 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.786090 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.799869 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d317bcd-8815-4e1c-8821-630f492c0be6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4d317bcd-8815-4e1c-8821-630f492c0be6\") " pod="openstack/nova-scheduler-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.799937 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzjgk\" (UniqueName: \"kubernetes.io/projected/4d317bcd-8815-4e1c-8821-630f492c0be6-kube-api-access-vzjgk\") pod \"nova-scheduler-0\" (UID: \"4d317bcd-8815-4e1c-8821-630f492c0be6\") " pod="openstack/nova-scheduler-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.799964 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2qgk\" (UniqueName: \"kubernetes.io/projected/1d5147b1-a584-4e96-895e-12064817d745-kube-api-access-w2qgk\") pod \"nova-metadata-0\" (UID: \"1d5147b1-a584-4e96-895e-12064817d745\") " pod="openstack/nova-metadata-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.800003 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d317bcd-8815-4e1c-8821-630f492c0be6-config-data\") pod \"nova-scheduler-0\" (UID: \"4d317bcd-8815-4e1c-8821-630f492c0be6\") " pod="openstack/nova-scheduler-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.800080 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d5147b1-a584-4e96-895e-12064817d745-logs\") pod \"nova-metadata-0\" (UID: \"1d5147b1-a584-4e96-895e-12064817d745\") " pod="openstack/nova-metadata-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.800105 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d5147b1-a584-4e96-895e-12064817d745-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1d5147b1-a584-4e96-895e-12064817d745\") " pod="openstack/nova-metadata-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.800142 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d5147b1-a584-4e96-895e-12064817d745-config-data\") pod \"nova-metadata-0\" (UID: \"1d5147b1-a584-4e96-895e-12064817d745\") " pod="openstack/nova-metadata-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.836761 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnxhv\" (UniqueName: \"kubernetes.io/projected/603ba880-f1f5-4893-9e81-bd2031b42192-kube-api-access-mnxhv\") pod \"nova-cell1-novncproxy-0\" (UID: \"603ba880-f1f5-4893-9e81-bd2031b42192\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.856592 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.882683 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c4q4\" (UniqueName: \"kubernetes.io/projected/d7225aed-20ef-47e7-94d3-ae900a0e55db-kube-api-access-8c4q4\") pod \"nova-api-0\" (UID: \"d7225aed-20ef-47e7-94d3-ae900a0e55db\") " pod="openstack/nova-api-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.902838 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzjgk\" (UniqueName: \"kubernetes.io/projected/4d317bcd-8815-4e1c-8821-630f492c0be6-kube-api-access-vzjgk\") pod \"nova-scheduler-0\" (UID: \"4d317bcd-8815-4e1c-8821-630f492c0be6\") " pod="openstack/nova-scheduler-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.902927 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2qgk\" (UniqueName: \"kubernetes.io/projected/1d5147b1-a584-4e96-895e-12064817d745-kube-api-access-w2qgk\") pod \"nova-metadata-0\" (UID: \"1d5147b1-a584-4e96-895e-12064817d745\") " pod="openstack/nova-metadata-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.902996 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d317bcd-8815-4e1c-8821-630f492c0be6-config-data\") pod \"nova-scheduler-0\" (UID: \"4d317bcd-8815-4e1c-8821-630f492c0be6\") " pod="openstack/nova-scheduler-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.903133 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d5147b1-a584-4e96-895e-12064817d745-logs\") pod \"nova-metadata-0\" (UID: \"1d5147b1-a584-4e96-895e-12064817d745\") " pod="openstack/nova-metadata-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.903174 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d5147b1-a584-4e96-895e-12064817d745-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1d5147b1-a584-4e96-895e-12064817d745\") " pod="openstack/nova-metadata-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.903233 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d5147b1-a584-4e96-895e-12064817d745-config-data\") pod \"nova-metadata-0\" (UID: \"1d5147b1-a584-4e96-895e-12064817d745\") " pod="openstack/nova-metadata-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.903320 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d317bcd-8815-4e1c-8821-630f492c0be6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4d317bcd-8815-4e1c-8821-630f492c0be6\") " pod="openstack/nova-scheduler-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.907615 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d5147b1-a584-4e96-895e-12064817d745-logs\") pod \"nova-metadata-0\" (UID: \"1d5147b1-a584-4e96-895e-12064817d745\") " pod="openstack/nova-metadata-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.910161 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d317bcd-8815-4e1c-8821-630f492c0be6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4d317bcd-8815-4e1c-8821-630f492c0be6\") " pod="openstack/nova-scheduler-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.920561 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.921446 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d317bcd-8815-4e1c-8821-630f492c0be6-config-data\") pod \"nova-scheduler-0\" (UID: \"4d317bcd-8815-4e1c-8821-630f492c0be6\") " pod="openstack/nova-scheduler-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.921814 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d5147b1-a584-4e96-895e-12064817d745-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1d5147b1-a584-4e96-895e-12064817d745\") " pod="openstack/nova-metadata-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.921844 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d5147b1-a584-4e96-895e-12064817d745-config-data\") pod \"nova-metadata-0\" (UID: \"1d5147b1-a584-4e96-895e-12064817d745\") " pod="openstack/nova-metadata-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.935561 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:11 crc kubenswrapper[4776]: I1125 09:46:11.940590 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzjgk\" (UniqueName: \"kubernetes.io/projected/4d317bcd-8815-4e1c-8821-630f492c0be6-kube-api-access-vzjgk\") pod \"nova-scheduler-0\" (UID: \"4d317bcd-8815-4e1c-8821-630f492c0be6\") " pod="openstack/nova-scheduler-0" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.020761 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2qgk\" (UniqueName: \"kubernetes.io/projected/1d5147b1-a584-4e96-895e-12064817d745-kube-api-access-w2qgk\") pod \"nova-metadata-0\" (UID: \"1d5147b1-a584-4e96-895e-12064817d745\") " pod="openstack/nova-metadata-0" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.063133 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5dd7c4987f-4kphp"] Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.064748 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.089024 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5dd7c4987f-4kphp"] Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.121465 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-ovsdbserver-sb\") pod \"dnsmasq-dns-5dd7c4987f-4kphp\" (UID: \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\") " pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.121522 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fqn5\" (UniqueName: \"kubernetes.io/projected/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-kube-api-access-6fqn5\") pod \"dnsmasq-dns-5dd7c4987f-4kphp\" (UID: \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\") " pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.121555 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-ovsdbserver-nb\") pod \"dnsmasq-dns-5dd7c4987f-4kphp\" (UID: \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\") " pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.121607 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-dns-svc\") pod \"dnsmasq-dns-5dd7c4987f-4kphp\" (UID: \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\") " pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.121664 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-dns-swift-storage-0\") pod \"dnsmasq-dns-5dd7c4987f-4kphp\" (UID: \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\") " pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.121696 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-config\") pod \"dnsmasq-dns-5dd7c4987f-4kphp\" (UID: \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\") " pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.122125 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.134402 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.225455 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-dns-svc\") pod \"dnsmasq-dns-5dd7c4987f-4kphp\" (UID: \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\") " pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.225755 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-dns-swift-storage-0\") pod \"dnsmasq-dns-5dd7c4987f-4kphp\" (UID: \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\") " pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.225783 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-config\") pod \"dnsmasq-dns-5dd7c4987f-4kphp\" (UID: \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\") " pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.227212 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-dns-svc\") pod \"dnsmasq-dns-5dd7c4987f-4kphp\" (UID: \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\") " pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.239618 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-ovsdbserver-sb\") pod \"dnsmasq-dns-5dd7c4987f-4kphp\" (UID: \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\") " pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.239670 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fqn5\" (UniqueName: \"kubernetes.io/projected/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-kube-api-access-6fqn5\") pod \"dnsmasq-dns-5dd7c4987f-4kphp\" (UID: \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\") " pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.239746 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-ovsdbserver-nb\") pod \"dnsmasq-dns-5dd7c4987f-4kphp\" (UID: \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\") " pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.264223 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-config\") pod \"dnsmasq-dns-5dd7c4987f-4kphp\" (UID: \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\") " pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.264494 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-ovsdbserver-nb\") pod \"dnsmasq-dns-5dd7c4987f-4kphp\" (UID: \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\") " pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.265314 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-dns-swift-storage-0\") pod \"dnsmasq-dns-5dd7c4987f-4kphp\" (UID: \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\") " pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.266036 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-ovsdbserver-sb\") pod \"dnsmasq-dns-5dd7c4987f-4kphp\" (UID: \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\") " pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.293312 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fqn5\" (UniqueName: \"kubernetes.io/projected/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-kube-api-access-6fqn5\") pod \"dnsmasq-dns-5dd7c4987f-4kphp\" (UID: \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\") " pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.405040 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.766491 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-69dpf"] Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.844096 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7nsh6"] Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.846995 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7nsh6" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.854685 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7nsh6"] Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.862153 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.862498 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 25 09:46:12 crc kubenswrapper[4776]: W1125 09:46:12.923917 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod603ba880_f1f5_4893_9e81_bd2031b42192.slice/crio-077710e85b301d3997b57e5d2152740d26f14feda8819d0fece8eda2069c9d6d WatchSource:0}: Error finding container 077710e85b301d3997b57e5d2152740d26f14feda8819d0fece8eda2069c9d6d: Status 404 returned error can't find the container with id 077710e85b301d3997b57e5d2152740d26f14feda8819d0fece8eda2069c9d6d Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.924938 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.936901 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.964166 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d32cd325-5a9d-4a3b-873f-fccab96da880-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-7nsh6\" (UID: \"d32cd325-5a9d-4a3b-873f-fccab96da880\") " pod="openstack/nova-cell1-conductor-db-sync-7nsh6" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.964246 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d32cd325-5a9d-4a3b-873f-fccab96da880-scripts\") pod \"nova-cell1-conductor-db-sync-7nsh6\" (UID: \"d32cd325-5a9d-4a3b-873f-fccab96da880\") " pod="openstack/nova-cell1-conductor-db-sync-7nsh6" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.964294 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzjpq\" (UniqueName: \"kubernetes.io/projected/d32cd325-5a9d-4a3b-873f-fccab96da880-kube-api-access-fzjpq\") pod \"nova-cell1-conductor-db-sync-7nsh6\" (UID: \"d32cd325-5a9d-4a3b-873f-fccab96da880\") " pod="openstack/nova-cell1-conductor-db-sync-7nsh6" Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.964456 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d32cd325-5a9d-4a3b-873f-fccab96da880-config-data\") pod \"nova-cell1-conductor-db-sync-7nsh6\" (UID: \"d32cd325-5a9d-4a3b-873f-fccab96da880\") " pod="openstack/nova-cell1-conductor-db-sync-7nsh6" Nov 25 09:46:12 crc kubenswrapper[4776]: W1125 09:46:12.997456 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7225aed_20ef_47e7_94d3_ae900a0e55db.slice/crio-3e282fc48d98c4f8adda953d5f080ea7e1e230940c269589ab1504539855b444 WatchSource:0}: Error finding container 3e282fc48d98c4f8adda953d5f080ea7e1e230940c269589ab1504539855b444: Status 404 returned error can't find the container with id 3e282fc48d98c4f8adda953d5f080ea7e1e230940c269589ab1504539855b444 Nov 25 09:46:12 crc kubenswrapper[4776]: I1125 09:46:12.997544 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 09:46:13 crc kubenswrapper[4776]: I1125 09:46:13.067216 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d32cd325-5a9d-4a3b-873f-fccab96da880-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-7nsh6\" (UID: \"d32cd325-5a9d-4a3b-873f-fccab96da880\") " pod="openstack/nova-cell1-conductor-db-sync-7nsh6" Nov 25 09:46:13 crc kubenswrapper[4776]: I1125 09:46:13.067270 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d32cd325-5a9d-4a3b-873f-fccab96da880-scripts\") pod \"nova-cell1-conductor-db-sync-7nsh6\" (UID: \"d32cd325-5a9d-4a3b-873f-fccab96da880\") " pod="openstack/nova-cell1-conductor-db-sync-7nsh6" Nov 25 09:46:13 crc kubenswrapper[4776]: I1125 09:46:13.067340 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzjpq\" (UniqueName: \"kubernetes.io/projected/d32cd325-5a9d-4a3b-873f-fccab96da880-kube-api-access-fzjpq\") pod \"nova-cell1-conductor-db-sync-7nsh6\" (UID: \"d32cd325-5a9d-4a3b-873f-fccab96da880\") " pod="openstack/nova-cell1-conductor-db-sync-7nsh6" Nov 25 09:46:13 crc kubenswrapper[4776]: I1125 09:46:13.067495 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d32cd325-5a9d-4a3b-873f-fccab96da880-config-data\") pod \"nova-cell1-conductor-db-sync-7nsh6\" (UID: \"d32cd325-5a9d-4a3b-873f-fccab96da880\") " pod="openstack/nova-cell1-conductor-db-sync-7nsh6" Nov 25 09:46:13 crc kubenswrapper[4776]: I1125 09:46:13.077499 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d32cd325-5a9d-4a3b-873f-fccab96da880-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-7nsh6\" (UID: \"d32cd325-5a9d-4a3b-873f-fccab96da880\") " pod="openstack/nova-cell1-conductor-db-sync-7nsh6" Nov 25 09:46:13 crc kubenswrapper[4776]: I1125 09:46:13.079147 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d32cd325-5a9d-4a3b-873f-fccab96da880-config-data\") pod \"nova-cell1-conductor-db-sync-7nsh6\" (UID: \"d32cd325-5a9d-4a3b-873f-fccab96da880\") " pod="openstack/nova-cell1-conductor-db-sync-7nsh6" Nov 25 09:46:13 crc kubenswrapper[4776]: I1125 09:46:13.087400 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d32cd325-5a9d-4a3b-873f-fccab96da880-scripts\") pod \"nova-cell1-conductor-db-sync-7nsh6\" (UID: \"d32cd325-5a9d-4a3b-873f-fccab96da880\") " pod="openstack/nova-cell1-conductor-db-sync-7nsh6" Nov 25 09:46:13 crc kubenswrapper[4776]: I1125 09:46:13.092725 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzjpq\" (UniqueName: \"kubernetes.io/projected/d32cd325-5a9d-4a3b-873f-fccab96da880-kube-api-access-fzjpq\") pod \"nova-cell1-conductor-db-sync-7nsh6\" (UID: \"d32cd325-5a9d-4a3b-873f-fccab96da880\") " pod="openstack/nova-cell1-conductor-db-sync-7nsh6" Nov 25 09:46:13 crc kubenswrapper[4776]: I1125 09:46:13.107949 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 09:46:13 crc kubenswrapper[4776]: I1125 09:46:13.126978 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5dd7c4987f-4kphp"] Nov 25 09:46:13 crc kubenswrapper[4776]: I1125 09:46:13.325263 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7nsh6" Nov 25 09:46:13 crc kubenswrapper[4776]: I1125 09:46:13.489882 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1d5147b1-a584-4e96-895e-12064817d745","Type":"ContainerStarted","Data":"3747a9104745ebcb18e736565bcb97ef570ff1bca7892dde9a651af5da7386da"} Nov 25 09:46:13 crc kubenswrapper[4776]: I1125 09:46:13.518010 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4d317bcd-8815-4e1c-8821-630f492c0be6","Type":"ContainerStarted","Data":"fa1b3ef6775cf9907debba3e4222e31db218f181b044a1d5dd7e858afa87a615"} Nov 25 09:46:13 crc kubenswrapper[4776]: I1125 09:46:13.526263 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-69dpf" event={"ID":"22bb0550-ee44-46e6-b45e-1158524bfa3e","Type":"ContainerStarted","Data":"5f45f2b73f855ed2de26c3204f7878eea5330ecac2bd8412a8721a9a5a319cfb"} Nov 25 09:46:13 crc kubenswrapper[4776]: I1125 09:46:13.526318 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-69dpf" event={"ID":"22bb0550-ee44-46e6-b45e-1158524bfa3e","Type":"ContainerStarted","Data":"bdbe7d478e3307b01eaaf5d6531526692340621c9fe490e819e19eebdc0ee83b"} Nov 25 09:46:13 crc kubenswrapper[4776]: I1125 09:46:13.533188 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"603ba880-f1f5-4893-9e81-bd2031b42192","Type":"ContainerStarted","Data":"077710e85b301d3997b57e5d2152740d26f14feda8819d0fece8eda2069c9d6d"} Nov 25 09:46:13 crc kubenswrapper[4776]: I1125 09:46:13.539455 4776 generic.go:334] "Generic (PLEG): container finished" podID="d7d021ba-cc92-4ef8-a471-26aa495ea6ab" containerID="e2585e863a2a360dafd403d35d262a6c229963ec204b453fbc68f98d8cde8db2" exitCode=0 Nov 25 09:46:13 crc kubenswrapper[4776]: I1125 09:46:13.539543 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" event={"ID":"d7d021ba-cc92-4ef8-a471-26aa495ea6ab","Type":"ContainerDied","Data":"e2585e863a2a360dafd403d35d262a6c229963ec204b453fbc68f98d8cde8db2"} Nov 25 09:46:13 crc kubenswrapper[4776]: I1125 09:46:13.539575 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" event={"ID":"d7d021ba-cc92-4ef8-a471-26aa495ea6ab","Type":"ContainerStarted","Data":"2c91e3bfc1cfec95c2ce46dfe176d7106a508078b2cf1affec2fe790903f8b1a"} Nov 25 09:46:13 crc kubenswrapper[4776]: I1125 09:46:13.548959 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d7225aed-20ef-47e7-94d3-ae900a0e55db","Type":"ContainerStarted","Data":"3e282fc48d98c4f8adda953d5f080ea7e1e230940c269589ab1504539855b444"} Nov 25 09:46:13 crc kubenswrapper[4776]: I1125 09:46:13.560018 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-69dpf" podStartSLOduration=2.560004555 podStartE2EDuration="2.560004555s" podCreationTimestamp="2025-11-25 09:46:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:46:13.549520332 +0000 UTC m=+1318.590579895" watchObservedRunningTime="2025-11-25 09:46:13.560004555 +0000 UTC m=+1318.601064098" Nov 25 09:46:13 crc kubenswrapper[4776]: I1125 09:46:13.847140 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7nsh6"] Nov 25 09:46:14 crc kubenswrapper[4776]: I1125 09:46:14.565316 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" event={"ID":"d7d021ba-cc92-4ef8-a471-26aa495ea6ab","Type":"ContainerStarted","Data":"f7079fb542df2e880a7d48696479a3e07c7298972160f08d1c04ea9542b544ab"} Nov 25 09:46:14 crc kubenswrapper[4776]: I1125 09:46:14.567024 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" Nov 25 09:46:14 crc kubenswrapper[4776]: I1125 09:46:14.574190 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7nsh6" event={"ID":"d32cd325-5a9d-4a3b-873f-fccab96da880","Type":"ContainerStarted","Data":"d4105354d99c01f581fbadfbf25a6148558ca3890a4fe12519c12245b6fa0449"} Nov 25 09:46:14 crc kubenswrapper[4776]: I1125 09:46:14.574240 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7nsh6" event={"ID":"d32cd325-5a9d-4a3b-873f-fccab96da880","Type":"ContainerStarted","Data":"0cb9c20a6c497e1021d33dbe790bd53ce251d468ea2e069de3bf89328a11968a"} Nov 25 09:46:14 crc kubenswrapper[4776]: I1125 09:46:14.590670 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" podStartSLOduration=3.590646546 podStartE2EDuration="3.590646546s" podCreationTimestamp="2025-11-25 09:46:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:46:14.588757239 +0000 UTC m=+1319.629816792" watchObservedRunningTime="2025-11-25 09:46:14.590646546 +0000 UTC m=+1319.631706099" Nov 25 09:46:14 crc kubenswrapper[4776]: I1125 09:46:14.617601 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-7nsh6" podStartSLOduration=2.617581892 podStartE2EDuration="2.617581892s" podCreationTimestamp="2025-11-25 09:46:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:46:14.614446493 +0000 UTC m=+1319.655506056" watchObservedRunningTime="2025-11-25 09:46:14.617581892 +0000 UTC m=+1319.658641445" Nov 25 09:46:15 crc kubenswrapper[4776]: I1125 09:46:15.868884 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 09:46:15 crc kubenswrapper[4776]: I1125 09:46:15.876732 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 09:46:17 crc kubenswrapper[4776]: I1125 09:46:17.606886 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d7225aed-20ef-47e7-94d3-ae900a0e55db","Type":"ContainerStarted","Data":"ec5d283742b0b75e676274ca46d13d9ee1a7a20face861f085df3eaa3afa6882"} Nov 25 09:46:17 crc kubenswrapper[4776]: I1125 09:46:17.607543 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d7225aed-20ef-47e7-94d3-ae900a0e55db","Type":"ContainerStarted","Data":"d8839360606fe1db88610baea1425be62c1eb5b51bd4704dd86ddc95ba28e782"} Nov 25 09:46:17 crc kubenswrapper[4776]: I1125 09:46:17.609292 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1d5147b1-a584-4e96-895e-12064817d745","Type":"ContainerStarted","Data":"7d92f0bc0fcbfd5817ce91a7f890beccbb99e9a36a9b7343f37580d042ecd9c9"} Nov 25 09:46:17 crc kubenswrapper[4776]: I1125 09:46:17.609330 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1d5147b1-a584-4e96-895e-12064817d745","Type":"ContainerStarted","Data":"8e8aa80ccc220da461dabd3c4860720225360db3c5c6df5ffe0ab81497e76034"} Nov 25 09:46:17 crc kubenswrapper[4776]: I1125 09:46:17.609458 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1d5147b1-a584-4e96-895e-12064817d745" containerName="nova-metadata-log" containerID="cri-o://8e8aa80ccc220da461dabd3c4860720225360db3c5c6df5ffe0ab81497e76034" gracePeriod=30 Nov 25 09:46:17 crc kubenswrapper[4776]: I1125 09:46:17.609756 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1d5147b1-a584-4e96-895e-12064817d745" containerName="nova-metadata-metadata" containerID="cri-o://7d92f0bc0fcbfd5817ce91a7f890beccbb99e9a36a9b7343f37580d042ecd9c9" gracePeriod=30 Nov 25 09:46:17 crc kubenswrapper[4776]: I1125 09:46:17.620345 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4d317bcd-8815-4e1c-8821-630f492c0be6","Type":"ContainerStarted","Data":"9425209167e4619730c55377e996dde9de2b26c83eb3409f8adad3bcd2cd854a"} Nov 25 09:46:17 crc kubenswrapper[4776]: I1125 09:46:17.624935 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"603ba880-f1f5-4893-9e81-bd2031b42192","Type":"ContainerStarted","Data":"0ba1f6da1b94ee190b90d2a6e712f500ba7120cebfced993b8957549e5bc9481"} Nov 25 09:46:17 crc kubenswrapper[4776]: I1125 09:46:17.625124 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="603ba880-f1f5-4893-9e81-bd2031b42192" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://0ba1f6da1b94ee190b90d2a6e712f500ba7120cebfced993b8957549e5bc9481" gracePeriod=30 Nov 25 09:46:17 crc kubenswrapper[4776]: I1125 09:46:17.634649 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.210637015 podStartE2EDuration="6.634629467s" podCreationTimestamp="2025-11-25 09:46:11 +0000 UTC" firstStartedPulling="2025-11-25 09:46:13.000779178 +0000 UTC m=+1318.041838741" lastFinishedPulling="2025-11-25 09:46:16.42477164 +0000 UTC m=+1321.465831193" observedRunningTime="2025-11-25 09:46:17.630101144 +0000 UTC m=+1322.671160707" watchObservedRunningTime="2025-11-25 09:46:17.634629467 +0000 UTC m=+1322.675689020" Nov 25 09:46:17 crc kubenswrapper[4776]: I1125 09:46:17.661345 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.162356314 podStartE2EDuration="6.661322307s" podCreationTimestamp="2025-11-25 09:46:11 +0000 UTC" firstStartedPulling="2025-11-25 09:46:12.926411253 +0000 UTC m=+1317.967470806" lastFinishedPulling="2025-11-25 09:46:16.425377246 +0000 UTC m=+1321.466436799" observedRunningTime="2025-11-25 09:46:17.647264094 +0000 UTC m=+1322.688323647" watchObservedRunningTime="2025-11-25 09:46:17.661322307 +0000 UTC m=+1322.702381860" Nov 25 09:46:17 crc kubenswrapper[4776]: I1125 09:46:17.682315 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.192642523 podStartE2EDuration="6.682296473s" podCreationTimestamp="2025-11-25 09:46:11 +0000 UTC" firstStartedPulling="2025-11-25 09:46:12.938486485 +0000 UTC m=+1317.979546038" lastFinishedPulling="2025-11-25 09:46:16.428140435 +0000 UTC m=+1321.469199988" observedRunningTime="2025-11-25 09:46:17.679568544 +0000 UTC m=+1322.720628107" watchObservedRunningTime="2025-11-25 09:46:17.682296473 +0000 UTC m=+1322.723356026" Nov 25 09:46:17 crc kubenswrapper[4776]: I1125 09:46:17.704771 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.397281587 podStartE2EDuration="6.704747806s" podCreationTimestamp="2025-11-25 09:46:11 +0000 UTC" firstStartedPulling="2025-11-25 09:46:13.123356633 +0000 UTC m=+1318.164416186" lastFinishedPulling="2025-11-25 09:46:16.430822852 +0000 UTC m=+1321.471882405" observedRunningTime="2025-11-25 09:46:17.70251947 +0000 UTC m=+1322.743579033" watchObservedRunningTime="2025-11-25 09:46:17.704747806 +0000 UTC m=+1322.745807359" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.243408 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.350714 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d5147b1-a584-4e96-895e-12064817d745-logs\") pod \"1d5147b1-a584-4e96-895e-12064817d745\" (UID: \"1d5147b1-a584-4e96-895e-12064817d745\") " Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.351001 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d5147b1-a584-4e96-895e-12064817d745-combined-ca-bundle\") pod \"1d5147b1-a584-4e96-895e-12064817d745\" (UID: \"1d5147b1-a584-4e96-895e-12064817d745\") " Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.351037 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2qgk\" (UniqueName: \"kubernetes.io/projected/1d5147b1-a584-4e96-895e-12064817d745-kube-api-access-w2qgk\") pod \"1d5147b1-a584-4e96-895e-12064817d745\" (UID: \"1d5147b1-a584-4e96-895e-12064817d745\") " Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.351103 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d5147b1-a584-4e96-895e-12064817d745-config-data\") pod \"1d5147b1-a584-4e96-895e-12064817d745\" (UID: \"1d5147b1-a584-4e96-895e-12064817d745\") " Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.352925 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d5147b1-a584-4e96-895e-12064817d745-logs" (OuterVolumeSpecName: "logs") pod "1d5147b1-a584-4e96-895e-12064817d745" (UID: "1d5147b1-a584-4e96-895e-12064817d745"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.357623 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d5147b1-a584-4e96-895e-12064817d745-kube-api-access-w2qgk" (OuterVolumeSpecName: "kube-api-access-w2qgk") pod "1d5147b1-a584-4e96-895e-12064817d745" (UID: "1d5147b1-a584-4e96-895e-12064817d745"). InnerVolumeSpecName "kube-api-access-w2qgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.382509 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d5147b1-a584-4e96-895e-12064817d745-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d5147b1-a584-4e96-895e-12064817d745" (UID: "1d5147b1-a584-4e96-895e-12064817d745"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.383440 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d5147b1-a584-4e96-895e-12064817d745-config-data" (OuterVolumeSpecName: "config-data") pod "1d5147b1-a584-4e96-895e-12064817d745" (UID: "1d5147b1-a584-4e96-895e-12064817d745"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.454233 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d5147b1-a584-4e96-895e-12064817d745-logs\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.454288 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d5147b1-a584-4e96-895e-12064817d745-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.454307 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2qgk\" (UniqueName: \"kubernetes.io/projected/1d5147b1-a584-4e96-895e-12064817d745-kube-api-access-w2qgk\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.454319 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d5147b1-a584-4e96-895e-12064817d745-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.641341 4776 generic.go:334] "Generic (PLEG): container finished" podID="1d5147b1-a584-4e96-895e-12064817d745" containerID="7d92f0bc0fcbfd5817ce91a7f890beccbb99e9a36a9b7343f37580d042ecd9c9" exitCode=0 Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.641383 4776 generic.go:334] "Generic (PLEG): container finished" podID="1d5147b1-a584-4e96-895e-12064817d745" containerID="8e8aa80ccc220da461dabd3c4860720225360db3c5c6df5ffe0ab81497e76034" exitCode=143 Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.642629 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.647551 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1d5147b1-a584-4e96-895e-12064817d745","Type":"ContainerDied","Data":"7d92f0bc0fcbfd5817ce91a7f890beccbb99e9a36a9b7343f37580d042ecd9c9"} Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.647708 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.647726 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1d5147b1-a584-4e96-895e-12064817d745","Type":"ContainerDied","Data":"8e8aa80ccc220da461dabd3c4860720225360db3c5c6df5ffe0ab81497e76034"} Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.647739 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1d5147b1-a584-4e96-895e-12064817d745","Type":"ContainerDied","Data":"3747a9104745ebcb18e736565bcb97ef570ff1bca7892dde9a651af5da7386da"} Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.647759 4776 scope.go:117] "RemoveContainer" containerID="7d92f0bc0fcbfd5817ce91a7f890beccbb99e9a36a9b7343f37580d042ecd9c9" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.697272 4776 scope.go:117] "RemoveContainer" containerID="8e8aa80ccc220da461dabd3c4860720225360db3c5c6df5ffe0ab81497e76034" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.740644 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.744881 4776 scope.go:117] "RemoveContainer" containerID="7d92f0bc0fcbfd5817ce91a7f890beccbb99e9a36a9b7343f37580d042ecd9c9" Nov 25 09:46:18 crc kubenswrapper[4776]: E1125 09:46:18.745340 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d92f0bc0fcbfd5817ce91a7f890beccbb99e9a36a9b7343f37580d042ecd9c9\": container with ID starting with 7d92f0bc0fcbfd5817ce91a7f890beccbb99e9a36a9b7343f37580d042ecd9c9 not found: ID does not exist" containerID="7d92f0bc0fcbfd5817ce91a7f890beccbb99e9a36a9b7343f37580d042ecd9c9" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.745379 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d92f0bc0fcbfd5817ce91a7f890beccbb99e9a36a9b7343f37580d042ecd9c9"} err="failed to get container status \"7d92f0bc0fcbfd5817ce91a7f890beccbb99e9a36a9b7343f37580d042ecd9c9\": rpc error: code = NotFound desc = could not find container \"7d92f0bc0fcbfd5817ce91a7f890beccbb99e9a36a9b7343f37580d042ecd9c9\": container with ID starting with 7d92f0bc0fcbfd5817ce91a7f890beccbb99e9a36a9b7343f37580d042ecd9c9 not found: ID does not exist" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.745405 4776 scope.go:117] "RemoveContainer" containerID="8e8aa80ccc220da461dabd3c4860720225360db3c5c6df5ffe0ab81497e76034" Nov 25 09:46:18 crc kubenswrapper[4776]: E1125 09:46:18.745800 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e8aa80ccc220da461dabd3c4860720225360db3c5c6df5ffe0ab81497e76034\": container with ID starting with 8e8aa80ccc220da461dabd3c4860720225360db3c5c6df5ffe0ab81497e76034 not found: ID does not exist" containerID="8e8aa80ccc220da461dabd3c4860720225360db3c5c6df5ffe0ab81497e76034" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.745829 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e8aa80ccc220da461dabd3c4860720225360db3c5c6df5ffe0ab81497e76034"} err="failed to get container status \"8e8aa80ccc220da461dabd3c4860720225360db3c5c6df5ffe0ab81497e76034\": rpc error: code = NotFound desc = could not find container \"8e8aa80ccc220da461dabd3c4860720225360db3c5c6df5ffe0ab81497e76034\": container with ID starting with 8e8aa80ccc220da461dabd3c4860720225360db3c5c6df5ffe0ab81497e76034 not found: ID does not exist" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.745845 4776 scope.go:117] "RemoveContainer" containerID="7d92f0bc0fcbfd5817ce91a7f890beccbb99e9a36a9b7343f37580d042ecd9c9" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.746135 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d92f0bc0fcbfd5817ce91a7f890beccbb99e9a36a9b7343f37580d042ecd9c9"} err="failed to get container status \"7d92f0bc0fcbfd5817ce91a7f890beccbb99e9a36a9b7343f37580d042ecd9c9\": rpc error: code = NotFound desc = could not find container \"7d92f0bc0fcbfd5817ce91a7f890beccbb99e9a36a9b7343f37580d042ecd9c9\": container with ID starting with 7d92f0bc0fcbfd5817ce91a7f890beccbb99e9a36a9b7343f37580d042ecd9c9 not found: ID does not exist" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.746166 4776 scope.go:117] "RemoveContainer" containerID="8e8aa80ccc220da461dabd3c4860720225360db3c5c6df5ffe0ab81497e76034" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.747441 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e8aa80ccc220da461dabd3c4860720225360db3c5c6df5ffe0ab81497e76034"} err="failed to get container status \"8e8aa80ccc220da461dabd3c4860720225360db3c5c6df5ffe0ab81497e76034\": rpc error: code = NotFound desc = could not find container \"8e8aa80ccc220da461dabd3c4860720225360db3c5c6df5ffe0ab81497e76034\": container with ID starting with 8e8aa80ccc220da461dabd3c4860720225360db3c5c6df5ffe0ab81497e76034 not found: ID does not exist" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.751732 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.761889 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 25 09:46:18 crc kubenswrapper[4776]: E1125 09:46:18.762385 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d5147b1-a584-4e96-895e-12064817d745" containerName="nova-metadata-log" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.762410 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d5147b1-a584-4e96-895e-12064817d745" containerName="nova-metadata-log" Nov 25 09:46:18 crc kubenswrapper[4776]: E1125 09:46:18.762456 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d5147b1-a584-4e96-895e-12064817d745" containerName="nova-metadata-metadata" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.762466 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d5147b1-a584-4e96-895e-12064817d745" containerName="nova-metadata-metadata" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.762695 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d5147b1-a584-4e96-895e-12064817d745" containerName="nova-metadata-log" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.762719 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d5147b1-a584-4e96-895e-12064817d745" containerName="nova-metadata-metadata" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.764653 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.774229 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.774472 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.801124 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.866430 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/025c914f-86cb-4a2f-be7e-c8f36b6bff23-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"025c914f-86cb-4a2f-be7e-c8f36b6bff23\") " pod="openstack/nova-metadata-0" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.866879 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/025c914f-86cb-4a2f-be7e-c8f36b6bff23-logs\") pod \"nova-metadata-0\" (UID: \"025c914f-86cb-4a2f-be7e-c8f36b6bff23\") " pod="openstack/nova-metadata-0" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.867079 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j92bf\" (UniqueName: \"kubernetes.io/projected/025c914f-86cb-4a2f-be7e-c8f36b6bff23-kube-api-access-j92bf\") pod \"nova-metadata-0\" (UID: \"025c914f-86cb-4a2f-be7e-c8f36b6bff23\") " pod="openstack/nova-metadata-0" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.867138 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/025c914f-86cb-4a2f-be7e-c8f36b6bff23-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"025c914f-86cb-4a2f-be7e-c8f36b6bff23\") " pod="openstack/nova-metadata-0" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.867182 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/025c914f-86cb-4a2f-be7e-c8f36b6bff23-config-data\") pod \"nova-metadata-0\" (UID: \"025c914f-86cb-4a2f-be7e-c8f36b6bff23\") " pod="openstack/nova-metadata-0" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.970103 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/025c914f-86cb-4a2f-be7e-c8f36b6bff23-logs\") pod \"nova-metadata-0\" (UID: \"025c914f-86cb-4a2f-be7e-c8f36b6bff23\") " pod="openstack/nova-metadata-0" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.970487 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j92bf\" (UniqueName: \"kubernetes.io/projected/025c914f-86cb-4a2f-be7e-c8f36b6bff23-kube-api-access-j92bf\") pod \"nova-metadata-0\" (UID: \"025c914f-86cb-4a2f-be7e-c8f36b6bff23\") " pod="openstack/nova-metadata-0" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.970620 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/025c914f-86cb-4a2f-be7e-c8f36b6bff23-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"025c914f-86cb-4a2f-be7e-c8f36b6bff23\") " pod="openstack/nova-metadata-0" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.970722 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/025c914f-86cb-4a2f-be7e-c8f36b6bff23-config-data\") pod \"nova-metadata-0\" (UID: \"025c914f-86cb-4a2f-be7e-c8f36b6bff23\") " pod="openstack/nova-metadata-0" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.970902 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/025c914f-86cb-4a2f-be7e-c8f36b6bff23-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"025c914f-86cb-4a2f-be7e-c8f36b6bff23\") " pod="openstack/nova-metadata-0" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.971384 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/025c914f-86cb-4a2f-be7e-c8f36b6bff23-logs\") pod \"nova-metadata-0\" (UID: \"025c914f-86cb-4a2f-be7e-c8f36b6bff23\") " pod="openstack/nova-metadata-0" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.976523 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/025c914f-86cb-4a2f-be7e-c8f36b6bff23-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"025c914f-86cb-4a2f-be7e-c8f36b6bff23\") " pod="openstack/nova-metadata-0" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.979794 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/025c914f-86cb-4a2f-be7e-c8f36b6bff23-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"025c914f-86cb-4a2f-be7e-c8f36b6bff23\") " pod="openstack/nova-metadata-0" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.988811 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j92bf\" (UniqueName: \"kubernetes.io/projected/025c914f-86cb-4a2f-be7e-c8f36b6bff23-kube-api-access-j92bf\") pod \"nova-metadata-0\" (UID: \"025c914f-86cb-4a2f-be7e-c8f36b6bff23\") " pod="openstack/nova-metadata-0" Nov 25 09:46:18 crc kubenswrapper[4776]: I1125 09:46:18.993118 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/025c914f-86cb-4a2f-be7e-c8f36b6bff23-config-data\") pod \"nova-metadata-0\" (UID: \"025c914f-86cb-4a2f-be7e-c8f36b6bff23\") " pod="openstack/nova-metadata-0" Nov 25 09:46:19 crc kubenswrapper[4776]: I1125 09:46:19.109714 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 09:46:19 crc kubenswrapper[4776]: I1125 09:46:19.574528 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 09:46:19 crc kubenswrapper[4776]: W1125 09:46:19.578788 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod025c914f_86cb_4a2f_be7e_c8f36b6bff23.slice/crio-5bdd9f35f80b24539ed6de1a4963ab579b279b8ca81e837286ea218a4fb99867 WatchSource:0}: Error finding container 5bdd9f35f80b24539ed6de1a4963ab579b279b8ca81e837286ea218a4fb99867: Status 404 returned error can't find the container with id 5bdd9f35f80b24539ed6de1a4963ab579b279b8ca81e837286ea218a4fb99867 Nov 25 09:46:19 crc kubenswrapper[4776]: I1125 09:46:19.678260 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d5147b1-a584-4e96-895e-12064817d745" path="/var/lib/kubelet/pods/1d5147b1-a584-4e96-895e-12064817d745/volumes" Nov 25 09:46:19 crc kubenswrapper[4776]: I1125 09:46:19.680330 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"025c914f-86cb-4a2f-be7e-c8f36b6bff23","Type":"ContainerStarted","Data":"5bdd9f35f80b24539ed6de1a4963ab579b279b8ca81e837286ea218a4fb99867"} Nov 25 09:46:20 crc kubenswrapper[4776]: I1125 09:46:20.677944 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"025c914f-86cb-4a2f-be7e-c8f36b6bff23","Type":"ContainerStarted","Data":"6ce688668333cce36b7fb19d86784fb224e0be933b4e5a9cae13a961b31e075b"} Nov 25 09:46:20 crc kubenswrapper[4776]: I1125 09:46:20.678249 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"025c914f-86cb-4a2f-be7e-c8f36b6bff23","Type":"ContainerStarted","Data":"bc6e64978e9b27aba5e92e195036d8e71e74329d78ad9f1416273f24afc813ec"} Nov 25 09:46:20 crc kubenswrapper[4776]: I1125 09:46:20.704712 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.704692182 podStartE2EDuration="2.704692182s" podCreationTimestamp="2025-11-25 09:46:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:46:20.703135323 +0000 UTC m=+1325.744194886" watchObservedRunningTime="2025-11-25 09:46:20.704692182 +0000 UTC m=+1325.745751755" Nov 25 09:46:21 crc kubenswrapper[4776]: I1125 09:46:21.921515 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 09:46:21 crc kubenswrapper[4776]: I1125 09:46:21.922834 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 09:46:21 crc kubenswrapper[4776]: I1125 09:46:21.936575 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:22 crc kubenswrapper[4776]: I1125 09:46:22.156473 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 25 09:46:22 crc kubenswrapper[4776]: I1125 09:46:22.156540 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 25 09:46:22 crc kubenswrapper[4776]: I1125 09:46:22.187825 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 25 09:46:22 crc kubenswrapper[4776]: I1125 09:46:22.407498 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" Nov 25 09:46:22 crc kubenswrapper[4776]: I1125 09:46:22.462565 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-797bbc649-ncxwv"] Nov 25 09:46:22 crc kubenswrapper[4776]: I1125 09:46:22.462845 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-797bbc649-ncxwv" podUID="ecc97771-b405-44c7-93e1-d24c86e705ed" containerName="dnsmasq-dns" containerID="cri-o://ddeb90174916915ba1e5c7a4e9f66a3160fb3d00c4ccd556f36880eae5c2083b" gracePeriod=10 Nov 25 09:46:22 crc kubenswrapper[4776]: I1125 09:46:22.703975 4776 generic.go:334] "Generic (PLEG): container finished" podID="ecc97771-b405-44c7-93e1-d24c86e705ed" containerID="ddeb90174916915ba1e5c7a4e9f66a3160fb3d00c4ccd556f36880eae5c2083b" exitCode=0 Nov 25 09:46:22 crc kubenswrapper[4776]: I1125 09:46:22.704075 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797bbc649-ncxwv" event={"ID":"ecc97771-b405-44c7-93e1-d24c86e705ed","Type":"ContainerDied","Data":"ddeb90174916915ba1e5c7a4e9f66a3160fb3d00c4ccd556f36880eae5c2083b"} Nov 25 09:46:22 crc kubenswrapper[4776]: I1125 09:46:22.708308 4776 generic.go:334] "Generic (PLEG): container finished" podID="22bb0550-ee44-46e6-b45e-1158524bfa3e" containerID="5f45f2b73f855ed2de26c3204f7878eea5330ecac2bd8412a8721a9a5a319cfb" exitCode=0 Nov 25 09:46:22 crc kubenswrapper[4776]: I1125 09:46:22.708369 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-69dpf" event={"ID":"22bb0550-ee44-46e6-b45e-1158524bfa3e","Type":"ContainerDied","Data":"5f45f2b73f855ed2de26c3204f7878eea5330ecac2bd8412a8721a9a5a319cfb"} Nov 25 09:46:22 crc kubenswrapper[4776]: I1125 09:46:22.745859 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 25 09:46:23 crc kubenswrapper[4776]: I1125 09:46:23.007785 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d7225aed-20ef-47e7-94d3-ae900a0e55db" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.186:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 09:46:23 crc kubenswrapper[4776]: I1125 09:46:23.009273 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d7225aed-20ef-47e7-94d3-ae900a0e55db" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.186:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 09:46:23 crc kubenswrapper[4776]: I1125 09:46:23.152591 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-797bbc649-ncxwv" Nov 25 09:46:23 crc kubenswrapper[4776]: I1125 09:46:23.262249 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vplwb\" (UniqueName: \"kubernetes.io/projected/ecc97771-b405-44c7-93e1-d24c86e705ed-kube-api-access-vplwb\") pod \"ecc97771-b405-44c7-93e1-d24c86e705ed\" (UID: \"ecc97771-b405-44c7-93e1-d24c86e705ed\") " Nov 25 09:46:23 crc kubenswrapper[4776]: I1125 09:46:23.262335 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-config\") pod \"ecc97771-b405-44c7-93e1-d24c86e705ed\" (UID: \"ecc97771-b405-44c7-93e1-d24c86e705ed\") " Nov 25 09:46:23 crc kubenswrapper[4776]: I1125 09:46:23.262420 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-dns-swift-storage-0\") pod \"ecc97771-b405-44c7-93e1-d24c86e705ed\" (UID: \"ecc97771-b405-44c7-93e1-d24c86e705ed\") " Nov 25 09:46:23 crc kubenswrapper[4776]: I1125 09:46:23.262586 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-dns-svc\") pod \"ecc97771-b405-44c7-93e1-d24c86e705ed\" (UID: \"ecc97771-b405-44c7-93e1-d24c86e705ed\") " Nov 25 09:46:23 crc kubenswrapper[4776]: I1125 09:46:23.262642 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-ovsdbserver-nb\") pod \"ecc97771-b405-44c7-93e1-d24c86e705ed\" (UID: \"ecc97771-b405-44c7-93e1-d24c86e705ed\") " Nov 25 09:46:23 crc kubenswrapper[4776]: I1125 09:46:23.262674 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-ovsdbserver-sb\") pod \"ecc97771-b405-44c7-93e1-d24c86e705ed\" (UID: \"ecc97771-b405-44c7-93e1-d24c86e705ed\") " Nov 25 09:46:23 crc kubenswrapper[4776]: I1125 09:46:23.267938 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecc97771-b405-44c7-93e1-d24c86e705ed-kube-api-access-vplwb" (OuterVolumeSpecName: "kube-api-access-vplwb") pod "ecc97771-b405-44c7-93e1-d24c86e705ed" (UID: "ecc97771-b405-44c7-93e1-d24c86e705ed"). InnerVolumeSpecName "kube-api-access-vplwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:46:23 crc kubenswrapper[4776]: I1125 09:46:23.318616 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-config" (OuterVolumeSpecName: "config") pod "ecc97771-b405-44c7-93e1-d24c86e705ed" (UID: "ecc97771-b405-44c7-93e1-d24c86e705ed"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:46:23 crc kubenswrapper[4776]: I1125 09:46:23.329403 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ecc97771-b405-44c7-93e1-d24c86e705ed" (UID: "ecc97771-b405-44c7-93e1-d24c86e705ed"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:46:23 crc kubenswrapper[4776]: I1125 09:46:23.333735 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ecc97771-b405-44c7-93e1-d24c86e705ed" (UID: "ecc97771-b405-44c7-93e1-d24c86e705ed"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:46:23 crc kubenswrapper[4776]: I1125 09:46:23.337674 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ecc97771-b405-44c7-93e1-d24c86e705ed" (UID: "ecc97771-b405-44c7-93e1-d24c86e705ed"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:46:23 crc kubenswrapper[4776]: I1125 09:46:23.350695 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ecc97771-b405-44c7-93e1-d24c86e705ed" (UID: "ecc97771-b405-44c7-93e1-d24c86e705ed"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:46:23 crc kubenswrapper[4776]: I1125 09:46:23.364663 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vplwb\" (UniqueName: \"kubernetes.io/projected/ecc97771-b405-44c7-93e1-d24c86e705ed-kube-api-access-vplwb\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:23 crc kubenswrapper[4776]: I1125 09:46:23.364717 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:23 crc kubenswrapper[4776]: I1125 09:46:23.364731 4776 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:23 crc kubenswrapper[4776]: I1125 09:46:23.364742 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:23 crc kubenswrapper[4776]: I1125 09:46:23.364754 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:23 crc kubenswrapper[4776]: I1125 09:46:23.364765 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ecc97771-b405-44c7-93e1-d24c86e705ed-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:23 crc kubenswrapper[4776]: I1125 09:46:23.723634 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-797bbc649-ncxwv" event={"ID":"ecc97771-b405-44c7-93e1-d24c86e705ed","Type":"ContainerDied","Data":"39476d42ea0743b82f53bd110fd3296e26eda4cb6adc970cfd1f9baf61dd1eb7"} Nov 25 09:46:23 crc kubenswrapper[4776]: I1125 09:46:23.723724 4776 scope.go:117] "RemoveContainer" containerID="ddeb90174916915ba1e5c7a4e9f66a3160fb3d00c4ccd556f36880eae5c2083b" Nov 25 09:46:23 crc kubenswrapper[4776]: I1125 09:46:23.723810 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-797bbc649-ncxwv" Nov 25 09:46:23 crc kubenswrapper[4776]: I1125 09:46:23.756897 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-797bbc649-ncxwv"] Nov 25 09:46:23 crc kubenswrapper[4776]: I1125 09:46:23.766206 4776 scope.go:117] "RemoveContainer" containerID="7e1922940c86bbc7756196c6757fc4aec95313f98843e7b454724c9580f1a220" Nov 25 09:46:23 crc kubenswrapper[4776]: I1125 09:46:23.768013 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-797bbc649-ncxwv"] Nov 25 09:46:24 crc kubenswrapper[4776]: I1125 09:46:24.111449 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 09:46:24 crc kubenswrapper[4776]: I1125 09:46:24.111533 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 09:46:24 crc kubenswrapper[4776]: I1125 09:46:24.129575 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-69dpf" Nov 25 09:46:24 crc kubenswrapper[4776]: I1125 09:46:24.292469 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22bb0550-ee44-46e6-b45e-1158524bfa3e-combined-ca-bundle\") pod \"22bb0550-ee44-46e6-b45e-1158524bfa3e\" (UID: \"22bb0550-ee44-46e6-b45e-1158524bfa3e\") " Nov 25 09:46:24 crc kubenswrapper[4776]: I1125 09:46:24.292855 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22bb0550-ee44-46e6-b45e-1158524bfa3e-scripts\") pod \"22bb0550-ee44-46e6-b45e-1158524bfa3e\" (UID: \"22bb0550-ee44-46e6-b45e-1158524bfa3e\") " Nov 25 09:46:24 crc kubenswrapper[4776]: I1125 09:46:24.292941 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22bb0550-ee44-46e6-b45e-1158524bfa3e-config-data\") pod \"22bb0550-ee44-46e6-b45e-1158524bfa3e\" (UID: \"22bb0550-ee44-46e6-b45e-1158524bfa3e\") " Nov 25 09:46:24 crc kubenswrapper[4776]: I1125 09:46:24.292970 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7gsr\" (UniqueName: \"kubernetes.io/projected/22bb0550-ee44-46e6-b45e-1158524bfa3e-kube-api-access-z7gsr\") pod \"22bb0550-ee44-46e6-b45e-1158524bfa3e\" (UID: \"22bb0550-ee44-46e6-b45e-1158524bfa3e\") " Nov 25 09:46:24 crc kubenswrapper[4776]: I1125 09:46:24.298409 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22bb0550-ee44-46e6-b45e-1158524bfa3e-kube-api-access-z7gsr" (OuterVolumeSpecName: "kube-api-access-z7gsr") pod "22bb0550-ee44-46e6-b45e-1158524bfa3e" (UID: "22bb0550-ee44-46e6-b45e-1158524bfa3e"). InnerVolumeSpecName "kube-api-access-z7gsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:46:24 crc kubenswrapper[4776]: I1125 09:46:24.298531 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22bb0550-ee44-46e6-b45e-1158524bfa3e-scripts" (OuterVolumeSpecName: "scripts") pod "22bb0550-ee44-46e6-b45e-1158524bfa3e" (UID: "22bb0550-ee44-46e6-b45e-1158524bfa3e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:24 crc kubenswrapper[4776]: I1125 09:46:24.321938 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22bb0550-ee44-46e6-b45e-1158524bfa3e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "22bb0550-ee44-46e6-b45e-1158524bfa3e" (UID: "22bb0550-ee44-46e6-b45e-1158524bfa3e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:24 crc kubenswrapper[4776]: I1125 09:46:24.325739 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22bb0550-ee44-46e6-b45e-1158524bfa3e-config-data" (OuterVolumeSpecName: "config-data") pod "22bb0550-ee44-46e6-b45e-1158524bfa3e" (UID: "22bb0550-ee44-46e6-b45e-1158524bfa3e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:24 crc kubenswrapper[4776]: I1125 09:46:24.395146 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22bb0550-ee44-46e6-b45e-1158524bfa3e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:24 crc kubenswrapper[4776]: I1125 09:46:24.395181 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22bb0550-ee44-46e6-b45e-1158524bfa3e-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:24 crc kubenswrapper[4776]: I1125 09:46:24.395190 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22bb0550-ee44-46e6-b45e-1158524bfa3e-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:24 crc kubenswrapper[4776]: I1125 09:46:24.395199 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7gsr\" (UniqueName: \"kubernetes.io/projected/22bb0550-ee44-46e6-b45e-1158524bfa3e-kube-api-access-z7gsr\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:24 crc kubenswrapper[4776]: I1125 09:46:24.742041 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-69dpf" Nov 25 09:46:24 crc kubenswrapper[4776]: I1125 09:46:24.742109 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-69dpf" event={"ID":"22bb0550-ee44-46e6-b45e-1158524bfa3e","Type":"ContainerDied","Data":"bdbe7d478e3307b01eaaf5d6531526692340621c9fe490e819e19eebdc0ee83b"} Nov 25 09:46:24 crc kubenswrapper[4776]: I1125 09:46:24.742138 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bdbe7d478e3307b01eaaf5d6531526692340621c9fe490e819e19eebdc0ee83b" Nov 25 09:46:24 crc kubenswrapper[4776]: I1125 09:46:24.924020 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 09:46:24 crc kubenswrapper[4776]: I1125 09:46:24.924377 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d7225aed-20ef-47e7-94d3-ae900a0e55db" containerName="nova-api-log" containerID="cri-o://d8839360606fe1db88610baea1425be62c1eb5b51bd4704dd86ddc95ba28e782" gracePeriod=30 Nov 25 09:46:24 crc kubenswrapper[4776]: I1125 09:46:24.924466 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d7225aed-20ef-47e7-94d3-ae900a0e55db" containerName="nova-api-api" containerID="cri-o://ec5d283742b0b75e676274ca46d13d9ee1a7a20face861f085df3eaa3afa6882" gracePeriod=30 Nov 25 09:46:24 crc kubenswrapper[4776]: I1125 09:46:24.937798 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 09:46:24 crc kubenswrapper[4776]: I1125 09:46:24.938001 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="4d317bcd-8815-4e1c-8821-630f492c0be6" containerName="nova-scheduler-scheduler" containerID="cri-o://9425209167e4619730c55377e996dde9de2b26c83eb3409f8adad3bcd2cd854a" gracePeriod=30 Nov 25 09:46:25 crc kubenswrapper[4776]: I1125 09:46:25.000820 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 09:46:25 crc kubenswrapper[4776]: I1125 09:46:25.693619 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecc97771-b405-44c7-93e1-d24c86e705ed" path="/var/lib/kubelet/pods/ecc97771-b405-44c7-93e1-d24c86e705ed/volumes" Nov 25 09:46:25 crc kubenswrapper[4776]: I1125 09:46:25.755383 4776 generic.go:334] "Generic (PLEG): container finished" podID="d7225aed-20ef-47e7-94d3-ae900a0e55db" containerID="d8839360606fe1db88610baea1425be62c1eb5b51bd4704dd86ddc95ba28e782" exitCode=143 Nov 25 09:46:25 crc kubenswrapper[4776]: I1125 09:46:25.755487 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d7225aed-20ef-47e7-94d3-ae900a0e55db","Type":"ContainerDied","Data":"d8839360606fe1db88610baea1425be62c1eb5b51bd4704dd86ddc95ba28e782"} Nov 25 09:46:25 crc kubenswrapper[4776]: I1125 09:46:25.757757 4776 generic.go:334] "Generic (PLEG): container finished" podID="d32cd325-5a9d-4a3b-873f-fccab96da880" containerID="d4105354d99c01f581fbadfbf25a6148558ca3890a4fe12519c12245b6fa0449" exitCode=0 Nov 25 09:46:25 crc kubenswrapper[4776]: I1125 09:46:25.758002 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="025c914f-86cb-4a2f-be7e-c8f36b6bff23" containerName="nova-metadata-log" containerID="cri-o://bc6e64978e9b27aba5e92e195036d8e71e74329d78ad9f1416273f24afc813ec" gracePeriod=30 Nov 25 09:46:25 crc kubenswrapper[4776]: I1125 09:46:25.758348 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7nsh6" event={"ID":"d32cd325-5a9d-4a3b-873f-fccab96da880","Type":"ContainerDied","Data":"d4105354d99c01f581fbadfbf25a6148558ca3890a4fe12519c12245b6fa0449"} Nov 25 09:46:25 crc kubenswrapper[4776]: I1125 09:46:25.760542 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="025c914f-86cb-4a2f-be7e-c8f36b6bff23" containerName="nova-metadata-metadata" containerID="cri-o://6ce688668333cce36b7fb19d86784fb224e0be933b4e5a9cae13a961b31e075b" gracePeriod=30 Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.363603 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.532485 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/025c914f-86cb-4a2f-be7e-c8f36b6bff23-combined-ca-bundle\") pod \"025c914f-86cb-4a2f-be7e-c8f36b6bff23\" (UID: \"025c914f-86cb-4a2f-be7e-c8f36b6bff23\") " Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.532677 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j92bf\" (UniqueName: \"kubernetes.io/projected/025c914f-86cb-4a2f-be7e-c8f36b6bff23-kube-api-access-j92bf\") pod \"025c914f-86cb-4a2f-be7e-c8f36b6bff23\" (UID: \"025c914f-86cb-4a2f-be7e-c8f36b6bff23\") " Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.533154 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/025c914f-86cb-4a2f-be7e-c8f36b6bff23-config-data\") pod \"025c914f-86cb-4a2f-be7e-c8f36b6bff23\" (UID: \"025c914f-86cb-4a2f-be7e-c8f36b6bff23\") " Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.533193 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/025c914f-86cb-4a2f-be7e-c8f36b6bff23-nova-metadata-tls-certs\") pod \"025c914f-86cb-4a2f-be7e-c8f36b6bff23\" (UID: \"025c914f-86cb-4a2f-be7e-c8f36b6bff23\") " Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.533224 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/025c914f-86cb-4a2f-be7e-c8f36b6bff23-logs\") pod \"025c914f-86cb-4a2f-be7e-c8f36b6bff23\" (UID: \"025c914f-86cb-4a2f-be7e-c8f36b6bff23\") " Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.533816 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/025c914f-86cb-4a2f-be7e-c8f36b6bff23-logs" (OuterVolumeSpecName: "logs") pod "025c914f-86cb-4a2f-be7e-c8f36b6bff23" (UID: "025c914f-86cb-4a2f-be7e-c8f36b6bff23"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.533976 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/025c914f-86cb-4a2f-be7e-c8f36b6bff23-logs\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.543368 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/025c914f-86cb-4a2f-be7e-c8f36b6bff23-kube-api-access-j92bf" (OuterVolumeSpecName: "kube-api-access-j92bf") pod "025c914f-86cb-4a2f-be7e-c8f36b6bff23" (UID: "025c914f-86cb-4a2f-be7e-c8f36b6bff23"). InnerVolumeSpecName "kube-api-access-j92bf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.568295 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/025c914f-86cb-4a2f-be7e-c8f36b6bff23-config-data" (OuterVolumeSpecName: "config-data") pod "025c914f-86cb-4a2f-be7e-c8f36b6bff23" (UID: "025c914f-86cb-4a2f-be7e-c8f36b6bff23"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.572975 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/025c914f-86cb-4a2f-be7e-c8f36b6bff23-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "025c914f-86cb-4a2f-be7e-c8f36b6bff23" (UID: "025c914f-86cb-4a2f-be7e-c8f36b6bff23"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.589813 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/025c914f-86cb-4a2f-be7e-c8f36b6bff23-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "025c914f-86cb-4a2f-be7e-c8f36b6bff23" (UID: "025c914f-86cb-4a2f-be7e-c8f36b6bff23"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.637095 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/025c914f-86cb-4a2f-be7e-c8f36b6bff23-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.637179 4776 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/025c914f-86cb-4a2f-be7e-c8f36b6bff23-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.637195 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/025c914f-86cb-4a2f-be7e-c8f36b6bff23-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.637208 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j92bf\" (UniqueName: \"kubernetes.io/projected/025c914f-86cb-4a2f-be7e-c8f36b6bff23-kube-api-access-j92bf\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.773389 4776 generic.go:334] "Generic (PLEG): container finished" podID="025c914f-86cb-4a2f-be7e-c8f36b6bff23" containerID="6ce688668333cce36b7fb19d86784fb224e0be933b4e5a9cae13a961b31e075b" exitCode=0 Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.773456 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"025c914f-86cb-4a2f-be7e-c8f36b6bff23","Type":"ContainerDied","Data":"6ce688668333cce36b7fb19d86784fb224e0be933b4e5a9cae13a961b31e075b"} Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.773518 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"025c914f-86cb-4a2f-be7e-c8f36b6bff23","Type":"ContainerDied","Data":"bc6e64978e9b27aba5e92e195036d8e71e74329d78ad9f1416273f24afc813ec"} Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.773539 4776 scope.go:117] "RemoveContainer" containerID="6ce688668333cce36b7fb19d86784fb224e0be933b4e5a9cae13a961b31e075b" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.773485 4776 generic.go:334] "Generic (PLEG): container finished" podID="025c914f-86cb-4a2f-be7e-c8f36b6bff23" containerID="bc6e64978e9b27aba5e92e195036d8e71e74329d78ad9f1416273f24afc813ec" exitCode=143 Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.773819 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"025c914f-86cb-4a2f-be7e-c8f36b6bff23","Type":"ContainerDied","Data":"5bdd9f35f80b24539ed6de1a4963ab579b279b8ca81e837286ea218a4fb99867"} Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.773413 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.832601 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.846239 4776 scope.go:117] "RemoveContainer" containerID="bc6e64978e9b27aba5e92e195036d8e71e74329d78ad9f1416273f24afc813ec" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.857138 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.871773 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 25 09:46:26 crc kubenswrapper[4776]: E1125 09:46:26.872327 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="025c914f-86cb-4a2f-be7e-c8f36b6bff23" containerName="nova-metadata-log" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.872348 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="025c914f-86cb-4a2f-be7e-c8f36b6bff23" containerName="nova-metadata-log" Nov 25 09:46:26 crc kubenswrapper[4776]: E1125 09:46:26.872373 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecc97771-b405-44c7-93e1-d24c86e705ed" containerName="init" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.872380 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecc97771-b405-44c7-93e1-d24c86e705ed" containerName="init" Nov 25 09:46:26 crc kubenswrapper[4776]: E1125 09:46:26.872402 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22bb0550-ee44-46e6-b45e-1158524bfa3e" containerName="nova-manage" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.872411 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="22bb0550-ee44-46e6-b45e-1158524bfa3e" containerName="nova-manage" Nov 25 09:46:26 crc kubenswrapper[4776]: E1125 09:46:26.872444 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecc97771-b405-44c7-93e1-d24c86e705ed" containerName="dnsmasq-dns" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.872453 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecc97771-b405-44c7-93e1-d24c86e705ed" containerName="dnsmasq-dns" Nov 25 09:46:26 crc kubenswrapper[4776]: E1125 09:46:26.872468 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="025c914f-86cb-4a2f-be7e-c8f36b6bff23" containerName="nova-metadata-metadata" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.872476 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="025c914f-86cb-4a2f-be7e-c8f36b6bff23" containerName="nova-metadata-metadata" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.872746 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="025c914f-86cb-4a2f-be7e-c8f36b6bff23" containerName="nova-metadata-metadata" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.872783 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecc97771-b405-44c7-93e1-d24c86e705ed" containerName="dnsmasq-dns" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.872799 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="025c914f-86cb-4a2f-be7e-c8f36b6bff23" containerName="nova-metadata-log" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.872817 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="22bb0550-ee44-46e6-b45e-1158524bfa3e" containerName="nova-manage" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.874191 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.879787 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.880091 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.892798 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.952564 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0\") " pod="openstack/nova-metadata-0" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.952633 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-config-data\") pod \"nova-metadata-0\" (UID: \"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0\") " pod="openstack/nova-metadata-0" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.952657 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rb2tv\" (UniqueName: \"kubernetes.io/projected/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-kube-api-access-rb2tv\") pod \"nova-metadata-0\" (UID: \"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0\") " pod="openstack/nova-metadata-0" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.952762 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0\") " pod="openstack/nova-metadata-0" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.952818 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-logs\") pod \"nova-metadata-0\" (UID: \"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0\") " pod="openstack/nova-metadata-0" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.957869 4776 scope.go:117] "RemoveContainer" containerID="6ce688668333cce36b7fb19d86784fb224e0be933b4e5a9cae13a961b31e075b" Nov 25 09:46:26 crc kubenswrapper[4776]: E1125 09:46:26.958597 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ce688668333cce36b7fb19d86784fb224e0be933b4e5a9cae13a961b31e075b\": container with ID starting with 6ce688668333cce36b7fb19d86784fb224e0be933b4e5a9cae13a961b31e075b not found: ID does not exist" containerID="6ce688668333cce36b7fb19d86784fb224e0be933b4e5a9cae13a961b31e075b" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.958625 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ce688668333cce36b7fb19d86784fb224e0be933b4e5a9cae13a961b31e075b"} err="failed to get container status \"6ce688668333cce36b7fb19d86784fb224e0be933b4e5a9cae13a961b31e075b\": rpc error: code = NotFound desc = could not find container \"6ce688668333cce36b7fb19d86784fb224e0be933b4e5a9cae13a961b31e075b\": container with ID starting with 6ce688668333cce36b7fb19d86784fb224e0be933b4e5a9cae13a961b31e075b not found: ID does not exist" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.958646 4776 scope.go:117] "RemoveContainer" containerID="bc6e64978e9b27aba5e92e195036d8e71e74329d78ad9f1416273f24afc813ec" Nov 25 09:46:26 crc kubenswrapper[4776]: E1125 09:46:26.958959 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc6e64978e9b27aba5e92e195036d8e71e74329d78ad9f1416273f24afc813ec\": container with ID starting with bc6e64978e9b27aba5e92e195036d8e71e74329d78ad9f1416273f24afc813ec not found: ID does not exist" containerID="bc6e64978e9b27aba5e92e195036d8e71e74329d78ad9f1416273f24afc813ec" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.959018 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc6e64978e9b27aba5e92e195036d8e71e74329d78ad9f1416273f24afc813ec"} err="failed to get container status \"bc6e64978e9b27aba5e92e195036d8e71e74329d78ad9f1416273f24afc813ec\": rpc error: code = NotFound desc = could not find container \"bc6e64978e9b27aba5e92e195036d8e71e74329d78ad9f1416273f24afc813ec\": container with ID starting with bc6e64978e9b27aba5e92e195036d8e71e74329d78ad9f1416273f24afc813ec not found: ID does not exist" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.959047 4776 scope.go:117] "RemoveContainer" containerID="6ce688668333cce36b7fb19d86784fb224e0be933b4e5a9cae13a961b31e075b" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.959342 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ce688668333cce36b7fb19d86784fb224e0be933b4e5a9cae13a961b31e075b"} err="failed to get container status \"6ce688668333cce36b7fb19d86784fb224e0be933b4e5a9cae13a961b31e075b\": rpc error: code = NotFound desc = could not find container \"6ce688668333cce36b7fb19d86784fb224e0be933b4e5a9cae13a961b31e075b\": container with ID starting with 6ce688668333cce36b7fb19d86784fb224e0be933b4e5a9cae13a961b31e075b not found: ID does not exist" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.959365 4776 scope.go:117] "RemoveContainer" containerID="bc6e64978e9b27aba5e92e195036d8e71e74329d78ad9f1416273f24afc813ec" Nov 25 09:46:26 crc kubenswrapper[4776]: I1125 09:46:26.959551 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc6e64978e9b27aba5e92e195036d8e71e74329d78ad9f1416273f24afc813ec"} err="failed to get container status \"bc6e64978e9b27aba5e92e195036d8e71e74329d78ad9f1416273f24afc813ec\": rpc error: code = NotFound desc = could not find container \"bc6e64978e9b27aba5e92e195036d8e71e74329d78ad9f1416273f24afc813ec\": container with ID starting with bc6e64978e9b27aba5e92e195036d8e71e74329d78ad9f1416273f24afc813ec not found: ID does not exist" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.053865 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0\") " pod="openstack/nova-metadata-0" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.053929 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-config-data\") pod \"nova-metadata-0\" (UID: \"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0\") " pod="openstack/nova-metadata-0" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.053949 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rb2tv\" (UniqueName: \"kubernetes.io/projected/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-kube-api-access-rb2tv\") pod \"nova-metadata-0\" (UID: \"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0\") " pod="openstack/nova-metadata-0" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.054035 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0\") " pod="openstack/nova-metadata-0" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.054087 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-logs\") pod \"nova-metadata-0\" (UID: \"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0\") " pod="openstack/nova-metadata-0" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.055213 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-logs\") pod \"nova-metadata-0\" (UID: \"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0\") " pod="openstack/nova-metadata-0" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.059853 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0\") " pod="openstack/nova-metadata-0" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.059932 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0\") " pod="openstack/nova-metadata-0" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.069228 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-config-data\") pod \"nova-metadata-0\" (UID: \"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0\") " pod="openstack/nova-metadata-0" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.072997 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rb2tv\" (UniqueName: \"kubernetes.io/projected/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-kube-api-access-rb2tv\") pod \"nova-metadata-0\" (UID: \"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0\") " pod="openstack/nova-metadata-0" Nov 25 09:46:27 crc kubenswrapper[4776]: E1125 09:46:27.159743 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9425209167e4619730c55377e996dde9de2b26c83eb3409f8adad3bcd2cd854a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 25 09:46:27 crc kubenswrapper[4776]: E1125 09:46:27.162212 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9425209167e4619730c55377e996dde9de2b26c83eb3409f8adad3bcd2cd854a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 25 09:46:27 crc kubenswrapper[4776]: E1125 09:46:27.164210 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9425209167e4619730c55377e996dde9de2b26c83eb3409f8adad3bcd2cd854a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 25 09:46:27 crc kubenswrapper[4776]: E1125 09:46:27.164274 4776 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="4d317bcd-8815-4e1c-8821-630f492c0be6" containerName="nova-scheduler-scheduler" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.164538 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7nsh6" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.251460 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.359351 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d32cd325-5a9d-4a3b-873f-fccab96da880-scripts\") pod \"d32cd325-5a9d-4a3b-873f-fccab96da880\" (UID: \"d32cd325-5a9d-4a3b-873f-fccab96da880\") " Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.359805 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d32cd325-5a9d-4a3b-873f-fccab96da880-combined-ca-bundle\") pod \"d32cd325-5a9d-4a3b-873f-fccab96da880\" (UID: \"d32cd325-5a9d-4a3b-873f-fccab96da880\") " Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.359877 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzjpq\" (UniqueName: \"kubernetes.io/projected/d32cd325-5a9d-4a3b-873f-fccab96da880-kube-api-access-fzjpq\") pod \"d32cd325-5a9d-4a3b-873f-fccab96da880\" (UID: \"d32cd325-5a9d-4a3b-873f-fccab96da880\") " Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.359940 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d32cd325-5a9d-4a3b-873f-fccab96da880-config-data\") pod \"d32cd325-5a9d-4a3b-873f-fccab96da880\" (UID: \"d32cd325-5a9d-4a3b-873f-fccab96da880\") " Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.363809 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d32cd325-5a9d-4a3b-873f-fccab96da880-scripts" (OuterVolumeSpecName: "scripts") pod "d32cd325-5a9d-4a3b-873f-fccab96da880" (UID: "d32cd325-5a9d-4a3b-873f-fccab96da880"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.364427 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d32cd325-5a9d-4a3b-873f-fccab96da880-kube-api-access-fzjpq" (OuterVolumeSpecName: "kube-api-access-fzjpq") pod "d32cd325-5a9d-4a3b-873f-fccab96da880" (UID: "d32cd325-5a9d-4a3b-873f-fccab96da880"). InnerVolumeSpecName "kube-api-access-fzjpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.391319 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d32cd325-5a9d-4a3b-873f-fccab96da880-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d32cd325-5a9d-4a3b-873f-fccab96da880" (UID: "d32cd325-5a9d-4a3b-873f-fccab96da880"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.407234 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d32cd325-5a9d-4a3b-873f-fccab96da880-config-data" (OuterVolumeSpecName: "config-data") pod "d32cd325-5a9d-4a3b-873f-fccab96da880" (UID: "d32cd325-5a9d-4a3b-873f-fccab96da880"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.462596 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d32cd325-5a9d-4a3b-873f-fccab96da880-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.462631 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d32cd325-5a9d-4a3b-873f-fccab96da880-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.462640 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d32cd325-5a9d-4a3b-873f-fccab96da880-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.462653 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzjpq\" (UniqueName: \"kubernetes.io/projected/d32cd325-5a9d-4a3b-873f-fccab96da880-kube-api-access-fzjpq\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.673367 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="025c914f-86cb-4a2f-be7e-c8f36b6bff23" path="/var/lib/kubelet/pods/025c914f-86cb-4a2f-be7e-c8f36b6bff23/volumes" Nov 25 09:46:27 crc kubenswrapper[4776]: W1125 09:46:27.714862 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a35a929_dab0_4c6c_ac0f_2199cb71f4d0.slice/crio-3d239108af315c66806c7a34b54f33898ff8d6979b8f77e45e3c4a31afd64786 WatchSource:0}: Error finding container 3d239108af315c66806c7a34b54f33898ff8d6979b8f77e45e3c4a31afd64786: Status 404 returned error can't find the container with id 3d239108af315c66806c7a34b54f33898ff8d6979b8f77e45e3c4a31afd64786 Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.717401 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.785278 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7nsh6" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.785236 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7nsh6" event={"ID":"d32cd325-5a9d-4a3b-873f-fccab96da880","Type":"ContainerDied","Data":"0cb9c20a6c497e1021d33dbe790bd53ce251d468ea2e069de3bf89328a11968a"} Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.785453 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0cb9c20a6c497e1021d33dbe790bd53ce251d468ea2e069de3bf89328a11968a" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.786904 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0","Type":"ContainerStarted","Data":"3d239108af315c66806c7a34b54f33898ff8d6979b8f77e45e3c4a31afd64786"} Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.851459 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 09:46:27 crc kubenswrapper[4776]: E1125 09:46:27.851920 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d32cd325-5a9d-4a3b-873f-fccab96da880" containerName="nova-cell1-conductor-db-sync" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.851951 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="d32cd325-5a9d-4a3b-873f-fccab96da880" containerName="nova-cell1-conductor-db-sync" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.852271 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="d32cd325-5a9d-4a3b-873f-fccab96da880" containerName="nova-cell1-conductor-db-sync" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.853144 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.857036 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.864214 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.868700 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f182a7e0-ebd8-4258-9269-43a662e39af8-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"f182a7e0-ebd8-4258-9269-43a662e39af8\") " pod="openstack/nova-cell1-conductor-0" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.868850 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f182a7e0-ebd8-4258-9269-43a662e39af8-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"f182a7e0-ebd8-4258-9269-43a662e39af8\") " pod="openstack/nova-cell1-conductor-0" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.868912 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvghd\" (UniqueName: \"kubernetes.io/projected/f182a7e0-ebd8-4258-9269-43a662e39af8-kube-api-access-fvghd\") pod \"nova-cell1-conductor-0\" (UID: \"f182a7e0-ebd8-4258-9269-43a662e39af8\") " pod="openstack/nova-cell1-conductor-0" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.970273 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f182a7e0-ebd8-4258-9269-43a662e39af8-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"f182a7e0-ebd8-4258-9269-43a662e39af8\") " pod="openstack/nova-cell1-conductor-0" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.970335 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvghd\" (UniqueName: \"kubernetes.io/projected/f182a7e0-ebd8-4258-9269-43a662e39af8-kube-api-access-fvghd\") pod \"nova-cell1-conductor-0\" (UID: \"f182a7e0-ebd8-4258-9269-43a662e39af8\") " pod="openstack/nova-cell1-conductor-0" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.970477 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f182a7e0-ebd8-4258-9269-43a662e39af8-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"f182a7e0-ebd8-4258-9269-43a662e39af8\") " pod="openstack/nova-cell1-conductor-0" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.975820 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f182a7e0-ebd8-4258-9269-43a662e39af8-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"f182a7e0-ebd8-4258-9269-43a662e39af8\") " pod="openstack/nova-cell1-conductor-0" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.975849 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f182a7e0-ebd8-4258-9269-43a662e39af8-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"f182a7e0-ebd8-4258-9269-43a662e39af8\") " pod="openstack/nova-cell1-conductor-0" Nov 25 09:46:27 crc kubenswrapper[4776]: I1125 09:46:27.988048 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvghd\" (UniqueName: \"kubernetes.io/projected/f182a7e0-ebd8-4258-9269-43a662e39af8-kube-api-access-fvghd\") pod \"nova-cell1-conductor-0\" (UID: \"f182a7e0-ebd8-4258-9269-43a662e39af8\") " pod="openstack/nova-cell1-conductor-0" Nov 25 09:46:28 crc kubenswrapper[4776]: I1125 09:46:28.170448 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 25 09:46:28 crc kubenswrapper[4776]: I1125 09:46:28.607779 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 09:46:28 crc kubenswrapper[4776]: I1125 09:46:28.794009 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 09:46:28 crc kubenswrapper[4776]: I1125 09:46:28.817609 4776 generic.go:334] "Generic (PLEG): container finished" podID="d7225aed-20ef-47e7-94d3-ae900a0e55db" containerID="ec5d283742b0b75e676274ca46d13d9ee1a7a20face861f085df3eaa3afa6882" exitCode=0 Nov 25 09:46:28 crc kubenswrapper[4776]: I1125 09:46:28.817723 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d7225aed-20ef-47e7-94d3-ae900a0e55db","Type":"ContainerDied","Data":"ec5d283742b0b75e676274ca46d13d9ee1a7a20face861f085df3eaa3afa6882"} Nov 25 09:46:28 crc kubenswrapper[4776]: I1125 09:46:28.817766 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d7225aed-20ef-47e7-94d3-ae900a0e55db","Type":"ContainerDied","Data":"3e282fc48d98c4f8adda953d5f080ea7e1e230940c269589ab1504539855b444"} Nov 25 09:46:28 crc kubenswrapper[4776]: I1125 09:46:28.817786 4776 scope.go:117] "RemoveContainer" containerID="ec5d283742b0b75e676274ca46d13d9ee1a7a20face861f085df3eaa3afa6882" Nov 25 09:46:28 crc kubenswrapper[4776]: I1125 09:46:28.817903 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 09:46:28 crc kubenswrapper[4776]: I1125 09:46:28.823708 4776 generic.go:334] "Generic (PLEG): container finished" podID="4d317bcd-8815-4e1c-8821-630f492c0be6" containerID="9425209167e4619730c55377e996dde9de2b26c83eb3409f8adad3bcd2cd854a" exitCode=0 Nov 25 09:46:28 crc kubenswrapper[4776]: I1125 09:46:28.823836 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4d317bcd-8815-4e1c-8821-630f492c0be6","Type":"ContainerDied","Data":"9425209167e4619730c55377e996dde9de2b26c83eb3409f8adad3bcd2cd854a"} Nov 25 09:46:28 crc kubenswrapper[4776]: I1125 09:46:28.839729 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0","Type":"ContainerStarted","Data":"8746f8f433ab92b6be995d01ccc400704b41df632f434effb9df0790d7335ffb"} Nov 25 09:46:28 crc kubenswrapper[4776]: I1125 09:46:28.839781 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0","Type":"ContainerStarted","Data":"0b13b46852bb397571c1533f5e6b6a73573a684637b6d0c2af1f64451e8b0640"} Nov 25 09:46:28 crc kubenswrapper[4776]: I1125 09:46:28.853710 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"f182a7e0-ebd8-4258-9269-43a662e39af8","Type":"ContainerStarted","Data":"35af05bcb3324eb2b85951a073a1335b4087f998234cc11e4721f613083d3288"} Nov 25 09:46:28 crc kubenswrapper[4776]: I1125 09:46:28.874144 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.874121722 podStartE2EDuration="2.874121722s" podCreationTimestamp="2025-11-25 09:46:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:46:28.865442534 +0000 UTC m=+1333.906502088" watchObservedRunningTime="2025-11-25 09:46:28.874121722 +0000 UTC m=+1333.915181275" Nov 25 09:46:28 crc kubenswrapper[4776]: I1125 09:46:28.876291 4776 scope.go:117] "RemoveContainer" containerID="d8839360606fe1db88610baea1425be62c1eb5b51bd4704dd86ddc95ba28e782" Nov 25 09:46:28 crc kubenswrapper[4776]: I1125 09:46:28.992321 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 09:46:28 crc kubenswrapper[4776]: I1125 09:46:28.994925 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7225aed-20ef-47e7-94d3-ae900a0e55db-config-data\") pod \"d7225aed-20ef-47e7-94d3-ae900a0e55db\" (UID: \"d7225aed-20ef-47e7-94d3-ae900a0e55db\") " Nov 25 09:46:28 crc kubenswrapper[4776]: I1125 09:46:28.995150 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7225aed-20ef-47e7-94d3-ae900a0e55db-combined-ca-bundle\") pod \"d7225aed-20ef-47e7-94d3-ae900a0e55db\" (UID: \"d7225aed-20ef-47e7-94d3-ae900a0e55db\") " Nov 25 09:46:28 crc kubenswrapper[4776]: I1125 09:46:28.995975 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8c4q4\" (UniqueName: \"kubernetes.io/projected/d7225aed-20ef-47e7-94d3-ae900a0e55db-kube-api-access-8c4q4\") pod \"d7225aed-20ef-47e7-94d3-ae900a0e55db\" (UID: \"d7225aed-20ef-47e7-94d3-ae900a0e55db\") " Nov 25 09:46:28 crc kubenswrapper[4776]: I1125 09:46:28.996010 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7225aed-20ef-47e7-94d3-ae900a0e55db-logs\") pod \"d7225aed-20ef-47e7-94d3-ae900a0e55db\" (UID: \"d7225aed-20ef-47e7-94d3-ae900a0e55db\") " Nov 25 09:46:28 crc kubenswrapper[4776]: I1125 09:46:28.996777 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7225aed-20ef-47e7-94d3-ae900a0e55db-logs" (OuterVolumeSpecName: "logs") pod "d7225aed-20ef-47e7-94d3-ae900a0e55db" (UID: "d7225aed-20ef-47e7-94d3-ae900a0e55db"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.002835 4776 scope.go:117] "RemoveContainer" containerID="ec5d283742b0b75e676274ca46d13d9ee1a7a20face861f085df3eaa3afa6882" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.002903 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7225aed-20ef-47e7-94d3-ae900a0e55db-kube-api-access-8c4q4" (OuterVolumeSpecName: "kube-api-access-8c4q4") pod "d7225aed-20ef-47e7-94d3-ae900a0e55db" (UID: "d7225aed-20ef-47e7-94d3-ae900a0e55db"). InnerVolumeSpecName "kube-api-access-8c4q4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:46:29 crc kubenswrapper[4776]: E1125 09:46:29.003727 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec5d283742b0b75e676274ca46d13d9ee1a7a20face861f085df3eaa3afa6882\": container with ID starting with ec5d283742b0b75e676274ca46d13d9ee1a7a20face861f085df3eaa3afa6882 not found: ID does not exist" containerID="ec5d283742b0b75e676274ca46d13d9ee1a7a20face861f085df3eaa3afa6882" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.003763 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec5d283742b0b75e676274ca46d13d9ee1a7a20face861f085df3eaa3afa6882"} err="failed to get container status \"ec5d283742b0b75e676274ca46d13d9ee1a7a20face861f085df3eaa3afa6882\": rpc error: code = NotFound desc = could not find container \"ec5d283742b0b75e676274ca46d13d9ee1a7a20face861f085df3eaa3afa6882\": container with ID starting with ec5d283742b0b75e676274ca46d13d9ee1a7a20face861f085df3eaa3afa6882 not found: ID does not exist" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.003785 4776 scope.go:117] "RemoveContainer" containerID="d8839360606fe1db88610baea1425be62c1eb5b51bd4704dd86ddc95ba28e782" Nov 25 09:46:29 crc kubenswrapper[4776]: E1125 09:46:29.004271 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8839360606fe1db88610baea1425be62c1eb5b51bd4704dd86ddc95ba28e782\": container with ID starting with d8839360606fe1db88610baea1425be62c1eb5b51bd4704dd86ddc95ba28e782 not found: ID does not exist" containerID="d8839360606fe1db88610baea1425be62c1eb5b51bd4704dd86ddc95ba28e782" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.004297 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8839360606fe1db88610baea1425be62c1eb5b51bd4704dd86ddc95ba28e782"} err="failed to get container status \"d8839360606fe1db88610baea1425be62c1eb5b51bd4704dd86ddc95ba28e782\": rpc error: code = NotFound desc = could not find container \"d8839360606fe1db88610baea1425be62c1eb5b51bd4704dd86ddc95ba28e782\": container with ID starting with d8839360606fe1db88610baea1425be62c1eb5b51bd4704dd86ddc95ba28e782 not found: ID does not exist" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.044428 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7225aed-20ef-47e7-94d3-ae900a0e55db-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d7225aed-20ef-47e7-94d3-ae900a0e55db" (UID: "d7225aed-20ef-47e7-94d3-ae900a0e55db"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.047641 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7225aed-20ef-47e7-94d3-ae900a0e55db-config-data" (OuterVolumeSpecName: "config-data") pod "d7225aed-20ef-47e7-94d3-ae900a0e55db" (UID: "d7225aed-20ef-47e7-94d3-ae900a0e55db"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.098192 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d317bcd-8815-4e1c-8821-630f492c0be6-combined-ca-bundle\") pod \"4d317bcd-8815-4e1c-8821-630f492c0be6\" (UID: \"4d317bcd-8815-4e1c-8821-630f492c0be6\") " Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.098770 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzjgk\" (UniqueName: \"kubernetes.io/projected/4d317bcd-8815-4e1c-8821-630f492c0be6-kube-api-access-vzjgk\") pod \"4d317bcd-8815-4e1c-8821-630f492c0be6\" (UID: \"4d317bcd-8815-4e1c-8821-630f492c0be6\") " Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.098891 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d317bcd-8815-4e1c-8821-630f492c0be6-config-data\") pod \"4d317bcd-8815-4e1c-8821-630f492c0be6\" (UID: \"4d317bcd-8815-4e1c-8821-630f492c0be6\") " Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.099417 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7225aed-20ef-47e7-94d3-ae900a0e55db-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.099573 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8c4q4\" (UniqueName: \"kubernetes.io/projected/d7225aed-20ef-47e7-94d3-ae900a0e55db-kube-api-access-8c4q4\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.099698 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7225aed-20ef-47e7-94d3-ae900a0e55db-logs\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.099775 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7225aed-20ef-47e7-94d3-ae900a0e55db-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.102648 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d317bcd-8815-4e1c-8821-630f492c0be6-kube-api-access-vzjgk" (OuterVolumeSpecName: "kube-api-access-vzjgk") pod "4d317bcd-8815-4e1c-8821-630f492c0be6" (UID: "4d317bcd-8815-4e1c-8821-630f492c0be6"). InnerVolumeSpecName "kube-api-access-vzjgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.125074 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d317bcd-8815-4e1c-8821-630f492c0be6-config-data" (OuterVolumeSpecName: "config-data") pod "4d317bcd-8815-4e1c-8821-630f492c0be6" (UID: "4d317bcd-8815-4e1c-8821-630f492c0be6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.129844 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d317bcd-8815-4e1c-8821-630f492c0be6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d317bcd-8815-4e1c-8821-630f492c0be6" (UID: "4d317bcd-8815-4e1c-8821-630f492c0be6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.154942 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.169807 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.184328 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 09:46:29 crc kubenswrapper[4776]: E1125 09:46:29.184796 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7225aed-20ef-47e7-94d3-ae900a0e55db" containerName="nova-api-api" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.184822 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7225aed-20ef-47e7-94d3-ae900a0e55db" containerName="nova-api-api" Nov 25 09:46:29 crc kubenswrapper[4776]: E1125 09:46:29.184866 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d317bcd-8815-4e1c-8821-630f492c0be6" containerName="nova-scheduler-scheduler" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.184874 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d317bcd-8815-4e1c-8821-630f492c0be6" containerName="nova-scheduler-scheduler" Nov 25 09:46:29 crc kubenswrapper[4776]: E1125 09:46:29.184887 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7225aed-20ef-47e7-94d3-ae900a0e55db" containerName="nova-api-log" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.184893 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7225aed-20ef-47e7-94d3-ae900a0e55db" containerName="nova-api-log" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.185089 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7225aed-20ef-47e7-94d3-ae900a0e55db" containerName="nova-api-api" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.185107 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7225aed-20ef-47e7-94d3-ae900a0e55db" containerName="nova-api-log" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.185121 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d317bcd-8815-4e1c-8821-630f492c0be6" containerName="nova-scheduler-scheduler" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.186110 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.188799 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.217166 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.220983 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x9gx\" (UniqueName: \"kubernetes.io/projected/3fd6c763-8b11-4c88-a9f5-b63830efe063-kube-api-access-4x9gx\") pod \"nova-api-0\" (UID: \"3fd6c763-8b11-4c88-a9f5-b63830efe063\") " pod="openstack/nova-api-0" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.221134 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3fd6c763-8b11-4c88-a9f5-b63830efe063-logs\") pod \"nova-api-0\" (UID: \"3fd6c763-8b11-4c88-a9f5-b63830efe063\") " pod="openstack/nova-api-0" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.221290 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fd6c763-8b11-4c88-a9f5-b63830efe063-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3fd6c763-8b11-4c88-a9f5-b63830efe063\") " pod="openstack/nova-api-0" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.221352 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fd6c763-8b11-4c88-a9f5-b63830efe063-config-data\") pod \"nova-api-0\" (UID: \"3fd6c763-8b11-4c88-a9f5-b63830efe063\") " pod="openstack/nova-api-0" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.221509 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d317bcd-8815-4e1c-8821-630f492c0be6-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.221542 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d317bcd-8815-4e1c-8821-630f492c0be6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.221559 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzjgk\" (UniqueName: \"kubernetes.io/projected/4d317bcd-8815-4e1c-8821-630f492c0be6-kube-api-access-vzjgk\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.322480 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x9gx\" (UniqueName: \"kubernetes.io/projected/3fd6c763-8b11-4c88-a9f5-b63830efe063-kube-api-access-4x9gx\") pod \"nova-api-0\" (UID: \"3fd6c763-8b11-4c88-a9f5-b63830efe063\") " pod="openstack/nova-api-0" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.322558 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3fd6c763-8b11-4c88-a9f5-b63830efe063-logs\") pod \"nova-api-0\" (UID: \"3fd6c763-8b11-4c88-a9f5-b63830efe063\") " pod="openstack/nova-api-0" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.322646 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fd6c763-8b11-4c88-a9f5-b63830efe063-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3fd6c763-8b11-4c88-a9f5-b63830efe063\") " pod="openstack/nova-api-0" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.322670 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fd6c763-8b11-4c88-a9f5-b63830efe063-config-data\") pod \"nova-api-0\" (UID: \"3fd6c763-8b11-4c88-a9f5-b63830efe063\") " pod="openstack/nova-api-0" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.323165 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3fd6c763-8b11-4c88-a9f5-b63830efe063-logs\") pod \"nova-api-0\" (UID: \"3fd6c763-8b11-4c88-a9f5-b63830efe063\") " pod="openstack/nova-api-0" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.326182 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fd6c763-8b11-4c88-a9f5-b63830efe063-config-data\") pod \"nova-api-0\" (UID: \"3fd6c763-8b11-4c88-a9f5-b63830efe063\") " pod="openstack/nova-api-0" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.337330 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fd6c763-8b11-4c88-a9f5-b63830efe063-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3fd6c763-8b11-4c88-a9f5-b63830efe063\") " pod="openstack/nova-api-0" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.340093 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x9gx\" (UniqueName: \"kubernetes.io/projected/3fd6c763-8b11-4c88-a9f5-b63830efe063-kube-api-access-4x9gx\") pod \"nova-api-0\" (UID: \"3fd6c763-8b11-4c88-a9f5-b63830efe063\") " pod="openstack/nova-api-0" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.523964 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.676958 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7225aed-20ef-47e7-94d3-ae900a0e55db" path="/var/lib/kubelet/pods/d7225aed-20ef-47e7-94d3-ae900a0e55db/volumes" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.866426 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"f182a7e0-ebd8-4258-9269-43a662e39af8","Type":"ContainerStarted","Data":"e5a40e8fd3cdfddb16f8d8578ce8ae7e8bb1b0df1b91c68efe717a27f4fc382c"} Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.867125 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.880586 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4d317bcd-8815-4e1c-8821-630f492c0be6","Type":"ContainerDied","Data":"fa1b3ef6775cf9907debba3e4222e31db218f181b044a1d5dd7e858afa87a615"} Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.880686 4776 scope.go:117] "RemoveContainer" containerID="9425209167e4619730c55377e996dde9de2b26c83eb3409f8adad3bcd2cd854a" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.880876 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.909916 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.909892492 podStartE2EDuration="2.909892492s" podCreationTimestamp="2025-11-25 09:46:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:46:29.903636575 +0000 UTC m=+1334.944696128" watchObservedRunningTime="2025-11-25 09:46:29.909892492 +0000 UTC m=+1334.950952045" Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.960700 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.971228 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.988619 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.996929 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 09:46:29 crc kubenswrapper[4776]: I1125 09:46:29.998425 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 09:46:30 crc kubenswrapper[4776]: I1125 09:46:30.011566 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 25 09:46:30 crc kubenswrapper[4776]: I1125 09:46:30.023773 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 09:46:30 crc kubenswrapper[4776]: I1125 09:46:30.039839 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ck9c2\" (UniqueName: \"kubernetes.io/projected/66011de8-4f1e-4c9d-9bfb-5d0fb37daa14-kube-api-access-ck9c2\") pod \"nova-scheduler-0\" (UID: \"66011de8-4f1e-4c9d-9bfb-5d0fb37daa14\") " pod="openstack/nova-scheduler-0" Nov 25 09:46:30 crc kubenswrapper[4776]: I1125 09:46:30.039941 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66011de8-4f1e-4c9d-9bfb-5d0fb37daa14-config-data\") pod \"nova-scheduler-0\" (UID: \"66011de8-4f1e-4c9d-9bfb-5d0fb37daa14\") " pod="openstack/nova-scheduler-0" Nov 25 09:46:30 crc kubenswrapper[4776]: I1125 09:46:30.040448 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66011de8-4f1e-4c9d-9bfb-5d0fb37daa14-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"66011de8-4f1e-4c9d-9bfb-5d0fb37daa14\") " pod="openstack/nova-scheduler-0" Nov 25 09:46:30 crc kubenswrapper[4776]: I1125 09:46:30.142032 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66011de8-4f1e-4c9d-9bfb-5d0fb37daa14-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"66011de8-4f1e-4c9d-9bfb-5d0fb37daa14\") " pod="openstack/nova-scheduler-0" Nov 25 09:46:30 crc kubenswrapper[4776]: I1125 09:46:30.142140 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ck9c2\" (UniqueName: \"kubernetes.io/projected/66011de8-4f1e-4c9d-9bfb-5d0fb37daa14-kube-api-access-ck9c2\") pod \"nova-scheduler-0\" (UID: \"66011de8-4f1e-4c9d-9bfb-5d0fb37daa14\") " pod="openstack/nova-scheduler-0" Nov 25 09:46:30 crc kubenswrapper[4776]: I1125 09:46:30.142171 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66011de8-4f1e-4c9d-9bfb-5d0fb37daa14-config-data\") pod \"nova-scheduler-0\" (UID: \"66011de8-4f1e-4c9d-9bfb-5d0fb37daa14\") " pod="openstack/nova-scheduler-0" Nov 25 09:46:30 crc kubenswrapper[4776]: I1125 09:46:30.152858 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66011de8-4f1e-4c9d-9bfb-5d0fb37daa14-config-data\") pod \"nova-scheduler-0\" (UID: \"66011de8-4f1e-4c9d-9bfb-5d0fb37daa14\") " pod="openstack/nova-scheduler-0" Nov 25 09:46:30 crc kubenswrapper[4776]: I1125 09:46:30.160754 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66011de8-4f1e-4c9d-9bfb-5d0fb37daa14-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"66011de8-4f1e-4c9d-9bfb-5d0fb37daa14\") " pod="openstack/nova-scheduler-0" Nov 25 09:46:30 crc kubenswrapper[4776]: I1125 09:46:30.163408 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ck9c2\" (UniqueName: \"kubernetes.io/projected/66011de8-4f1e-4c9d-9bfb-5d0fb37daa14-kube-api-access-ck9c2\") pod \"nova-scheduler-0\" (UID: \"66011de8-4f1e-4c9d-9bfb-5d0fb37daa14\") " pod="openstack/nova-scheduler-0" Nov 25 09:46:30 crc kubenswrapper[4776]: I1125 09:46:30.423669 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 09:46:30 crc kubenswrapper[4776]: I1125 09:46:30.896885 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3fd6c763-8b11-4c88-a9f5-b63830efe063","Type":"ContainerStarted","Data":"4105d1d897ad666a132559e4e85151fd91afa80102510d21f72cb2b3a19982c3"} Nov 25 09:46:30 crc kubenswrapper[4776]: I1125 09:46:30.897381 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3fd6c763-8b11-4c88-a9f5-b63830efe063","Type":"ContainerStarted","Data":"fa19ba9ebb2d7a99775499f75792a50cc747196875d92699aae892c04c1f55a8"} Nov 25 09:46:30 crc kubenswrapper[4776]: I1125 09:46:30.897412 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3fd6c763-8b11-4c88-a9f5-b63830efe063","Type":"ContainerStarted","Data":"76d1f24d20681720c94d57bd91f56f0cf1f52e52e4978c0a73d49076b00311aa"} Nov 25 09:46:30 crc kubenswrapper[4776]: I1125 09:46:30.928583 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 09:46:30 crc kubenswrapper[4776]: I1125 09:46:30.932965 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.932947772 podStartE2EDuration="1.932947772s" podCreationTimestamp="2025-11-25 09:46:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:46:30.924275435 +0000 UTC m=+1335.965334988" watchObservedRunningTime="2025-11-25 09:46:30.932947772 +0000 UTC m=+1335.974007325" Nov 25 09:46:30 crc kubenswrapper[4776]: W1125 09:46:30.938609 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66011de8_4f1e_4c9d_9bfb_5d0fb37daa14.slice/crio-59461b1548a1266871b57d75ed7ae85873ac8ace14a6e0c6084bfd58575a8c0f WatchSource:0}: Error finding container 59461b1548a1266871b57d75ed7ae85873ac8ace14a6e0c6084bfd58575a8c0f: Status 404 returned error can't find the container with id 59461b1548a1266871b57d75ed7ae85873ac8ace14a6e0c6084bfd58575a8c0f Nov 25 09:46:31 crc kubenswrapper[4776]: I1125 09:46:31.674741 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d317bcd-8815-4e1c-8821-630f492c0be6" path="/var/lib/kubelet/pods/4d317bcd-8815-4e1c-8821-630f492c0be6/volumes" Nov 25 09:46:31 crc kubenswrapper[4776]: I1125 09:46:31.907015 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"66011de8-4f1e-4c9d-9bfb-5d0fb37daa14","Type":"ContainerStarted","Data":"ab260ec7794ff3d66ff3c728ab9aab93670ce5d3ac37c667f843d07124606553"} Nov 25 09:46:31 crc kubenswrapper[4776]: I1125 09:46:31.907092 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"66011de8-4f1e-4c9d-9bfb-5d0fb37daa14","Type":"ContainerStarted","Data":"59461b1548a1266871b57d75ed7ae85873ac8ace14a6e0c6084bfd58575a8c0f"} Nov 25 09:46:31 crc kubenswrapper[4776]: I1125 09:46:31.935333 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.935307834 podStartE2EDuration="2.935307834s" podCreationTimestamp="2025-11-25 09:46:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:46:31.92716559 +0000 UTC m=+1336.968225153" watchObservedRunningTime="2025-11-25 09:46:31.935307834 +0000 UTC m=+1336.976367437" Nov 25 09:46:32 crc kubenswrapper[4776]: I1125 09:46:32.251782 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 09:46:32 crc kubenswrapper[4776]: I1125 09:46:32.253209 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 09:46:33 crc kubenswrapper[4776]: I1125 09:46:33.206897 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 25 09:46:35 crc kubenswrapper[4776]: I1125 09:46:35.424353 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 25 09:46:37 crc kubenswrapper[4776]: I1125 09:46:37.251815 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 09:46:37 crc kubenswrapper[4776]: I1125 09:46:37.252453 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 09:46:38 crc kubenswrapper[4776]: I1125 09:46:38.261570 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3a35a929-dab0-4c6c-ac0f-2199cb71f4d0" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 09:46:38 crc kubenswrapper[4776]: I1125 09:46:38.271360 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3a35a929-dab0-4c6c-ac0f-2199cb71f4d0" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 09:46:39 crc kubenswrapper[4776]: I1125 09:46:39.525304 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 09:46:39 crc kubenswrapper[4776]: I1125 09:46:39.525682 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 09:46:40 crc kubenswrapper[4776]: I1125 09:46:40.424147 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 25 09:46:40 crc kubenswrapper[4776]: I1125 09:46:40.454429 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 25 09:46:40 crc kubenswrapper[4776]: I1125 09:46:40.614102 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3fd6c763-8b11-4c88-a9f5-b63830efe063" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.195:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 09:46:40 crc kubenswrapper[4776]: I1125 09:46:40.614687 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3fd6c763-8b11-4c88-a9f5-b63830efe063" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.195:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 09:46:41 crc kubenswrapper[4776]: I1125 09:46:41.025030 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 25 09:46:47 crc kubenswrapper[4776]: I1125 09:46:47.259809 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 25 09:46:47 crc kubenswrapper[4776]: I1125 09:46:47.260500 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 25 09:46:47 crc kubenswrapper[4776]: I1125 09:46:47.266514 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 25 09:46:47 crc kubenswrapper[4776]: I1125 09:46:47.271556 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 25 09:46:47 crc kubenswrapper[4776]: I1125 09:46:47.998679 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.047218 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/603ba880-f1f5-4893-9e81-bd2031b42192-combined-ca-bundle\") pod \"603ba880-f1f5-4893-9e81-bd2031b42192\" (UID: \"603ba880-f1f5-4893-9e81-bd2031b42192\") " Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.047260 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnxhv\" (UniqueName: \"kubernetes.io/projected/603ba880-f1f5-4893-9e81-bd2031b42192-kube-api-access-mnxhv\") pod \"603ba880-f1f5-4893-9e81-bd2031b42192\" (UID: \"603ba880-f1f5-4893-9e81-bd2031b42192\") " Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.047298 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/603ba880-f1f5-4893-9e81-bd2031b42192-config-data\") pod \"603ba880-f1f5-4893-9e81-bd2031b42192\" (UID: \"603ba880-f1f5-4893-9e81-bd2031b42192\") " Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.047459 4776 generic.go:334] "Generic (PLEG): container finished" podID="603ba880-f1f5-4893-9e81-bd2031b42192" containerID="0ba1f6da1b94ee190b90d2a6e712f500ba7120cebfced993b8957549e5bc9481" exitCode=137 Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.047544 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.047598 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"603ba880-f1f5-4893-9e81-bd2031b42192","Type":"ContainerDied","Data":"0ba1f6da1b94ee190b90d2a6e712f500ba7120cebfced993b8957549e5bc9481"} Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.047650 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"603ba880-f1f5-4893-9e81-bd2031b42192","Type":"ContainerDied","Data":"077710e85b301d3997b57e5d2152740d26f14feda8819d0fece8eda2069c9d6d"} Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.047673 4776 scope.go:117] "RemoveContainer" containerID="0ba1f6da1b94ee190b90d2a6e712f500ba7120cebfced993b8957549e5bc9481" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.055316 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/603ba880-f1f5-4893-9e81-bd2031b42192-kube-api-access-mnxhv" (OuterVolumeSpecName: "kube-api-access-mnxhv") pod "603ba880-f1f5-4893-9e81-bd2031b42192" (UID: "603ba880-f1f5-4893-9e81-bd2031b42192"). InnerVolumeSpecName "kube-api-access-mnxhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.076597 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/603ba880-f1f5-4893-9e81-bd2031b42192-config-data" (OuterVolumeSpecName: "config-data") pod "603ba880-f1f5-4893-9e81-bd2031b42192" (UID: "603ba880-f1f5-4893-9e81-bd2031b42192"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.077416 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/603ba880-f1f5-4893-9e81-bd2031b42192-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "603ba880-f1f5-4893-9e81-bd2031b42192" (UID: "603ba880-f1f5-4893-9e81-bd2031b42192"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.149641 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/603ba880-f1f5-4893-9e81-bd2031b42192-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.149673 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnxhv\" (UniqueName: \"kubernetes.io/projected/603ba880-f1f5-4893-9e81-bd2031b42192-kube-api-access-mnxhv\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.149686 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/603ba880-f1f5-4893-9e81-bd2031b42192-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.151522 4776 scope.go:117] "RemoveContainer" containerID="0ba1f6da1b94ee190b90d2a6e712f500ba7120cebfced993b8957549e5bc9481" Nov 25 09:46:48 crc kubenswrapper[4776]: E1125 09:46:48.152277 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ba1f6da1b94ee190b90d2a6e712f500ba7120cebfced993b8957549e5bc9481\": container with ID starting with 0ba1f6da1b94ee190b90d2a6e712f500ba7120cebfced993b8957549e5bc9481 not found: ID does not exist" containerID="0ba1f6da1b94ee190b90d2a6e712f500ba7120cebfced993b8957549e5bc9481" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.152318 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ba1f6da1b94ee190b90d2a6e712f500ba7120cebfced993b8957549e5bc9481"} err="failed to get container status \"0ba1f6da1b94ee190b90d2a6e712f500ba7120cebfced993b8957549e5bc9481\": rpc error: code = NotFound desc = could not find container \"0ba1f6da1b94ee190b90d2a6e712f500ba7120cebfced993b8957549e5bc9481\": container with ID starting with 0ba1f6da1b94ee190b90d2a6e712f500ba7120cebfced993b8957549e5bc9481 not found: ID does not exist" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.386416 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.404042 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.428791 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 09:46:48 crc kubenswrapper[4776]: E1125 09:46:48.429867 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="603ba880-f1f5-4893-9e81-bd2031b42192" containerName="nova-cell1-novncproxy-novncproxy" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.429907 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="603ba880-f1f5-4893-9e81-bd2031b42192" containerName="nova-cell1-novncproxy-novncproxy" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.430540 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="603ba880-f1f5-4893-9e81-bd2031b42192" containerName="nova-cell1-novncproxy-novncproxy" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.431445 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.434643 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.434670 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.434831 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.442031 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.558231 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e09f9958-4faf-4a28-8214-c3ead146122c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e09f9958-4faf-4a28-8214-c3ead146122c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.558305 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e09f9958-4faf-4a28-8214-c3ead146122c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e09f9958-4faf-4a28-8214-c3ead146122c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.558366 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8b28\" (UniqueName: \"kubernetes.io/projected/e09f9958-4faf-4a28-8214-c3ead146122c-kube-api-access-s8b28\") pod \"nova-cell1-novncproxy-0\" (UID: \"e09f9958-4faf-4a28-8214-c3ead146122c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.558391 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e09f9958-4faf-4a28-8214-c3ead146122c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e09f9958-4faf-4a28-8214-c3ead146122c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.558445 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e09f9958-4faf-4a28-8214-c3ead146122c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e09f9958-4faf-4a28-8214-c3ead146122c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.660471 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e09f9958-4faf-4a28-8214-c3ead146122c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e09f9958-4faf-4a28-8214-c3ead146122c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.660542 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e09f9958-4faf-4a28-8214-c3ead146122c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e09f9958-4faf-4a28-8214-c3ead146122c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.660573 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8b28\" (UniqueName: \"kubernetes.io/projected/e09f9958-4faf-4a28-8214-c3ead146122c-kube-api-access-s8b28\") pod \"nova-cell1-novncproxy-0\" (UID: \"e09f9958-4faf-4a28-8214-c3ead146122c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.660603 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e09f9958-4faf-4a28-8214-c3ead146122c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e09f9958-4faf-4a28-8214-c3ead146122c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.660637 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e09f9958-4faf-4a28-8214-c3ead146122c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e09f9958-4faf-4a28-8214-c3ead146122c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.664701 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e09f9958-4faf-4a28-8214-c3ead146122c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e09f9958-4faf-4a28-8214-c3ead146122c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.665355 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e09f9958-4faf-4a28-8214-c3ead146122c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e09f9958-4faf-4a28-8214-c3ead146122c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.665695 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e09f9958-4faf-4a28-8214-c3ead146122c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e09f9958-4faf-4a28-8214-c3ead146122c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.666756 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e09f9958-4faf-4a28-8214-c3ead146122c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e09f9958-4faf-4a28-8214-c3ead146122c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.678720 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8b28\" (UniqueName: \"kubernetes.io/projected/e09f9958-4faf-4a28-8214-c3ead146122c-kube-api-access-s8b28\") pod \"nova-cell1-novncproxy-0\" (UID: \"e09f9958-4faf-4a28-8214-c3ead146122c\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:48 crc kubenswrapper[4776]: I1125 09:46:48.755537 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:49 crc kubenswrapper[4776]: I1125 09:46:49.200860 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 09:46:49 crc kubenswrapper[4776]: I1125 09:46:49.528544 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 09:46:49 crc kubenswrapper[4776]: I1125 09:46:49.529578 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 09:46:49 crc kubenswrapper[4776]: I1125 09:46:49.529656 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 09:46:49 crc kubenswrapper[4776]: I1125 09:46:49.539610 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 09:46:49 crc kubenswrapper[4776]: I1125 09:46:49.674336 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="603ba880-f1f5-4893-9e81-bd2031b42192" path="/var/lib/kubelet/pods/603ba880-f1f5-4893-9e81-bd2031b42192/volumes" Nov 25 09:46:50 crc kubenswrapper[4776]: I1125 09:46:50.070165 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e09f9958-4faf-4a28-8214-c3ead146122c","Type":"ContainerStarted","Data":"e056e61a1527b24d984169499b3cde97a02136661c62174d6fe28407dfd3287a"} Nov 25 09:46:50 crc kubenswrapper[4776]: I1125 09:46:50.070216 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e09f9958-4faf-4a28-8214-c3ead146122c","Type":"ContainerStarted","Data":"184f9329ca4bc16ca9830f25689b95cf6dba545b914093958315b55ece208bbd"} Nov 25 09:46:50 crc kubenswrapper[4776]: I1125 09:46:50.070679 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 09:46:50 crc kubenswrapper[4776]: I1125 09:46:50.080435 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 09:46:50 crc kubenswrapper[4776]: I1125 09:46:50.090853 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.090830292 podStartE2EDuration="2.090830292s" podCreationTimestamp="2025-11-25 09:46:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:46:50.090005441 +0000 UTC m=+1355.131064994" watchObservedRunningTime="2025-11-25 09:46:50.090830292 +0000 UTC m=+1355.131889845" Nov 25 09:46:50 crc kubenswrapper[4776]: I1125 09:46:50.284794 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d7f54fb65-td75r"] Nov 25 09:46:50 crc kubenswrapper[4776]: I1125 09:46:50.293650 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" Nov 25 09:46:50 crc kubenswrapper[4776]: I1125 09:46:50.302116 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d7f54fb65-td75r"] Nov 25 09:46:50 crc kubenswrapper[4776]: I1125 09:46:50.400246 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-dns-swift-storage-0\") pod \"dnsmasq-dns-5d7f54fb65-td75r\" (UID: \"9ab346c9-f70f-4663-9b53-67b4d66b3112\") " pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" Nov 25 09:46:50 crc kubenswrapper[4776]: I1125 09:46:50.400562 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-ovsdbserver-nb\") pod \"dnsmasq-dns-5d7f54fb65-td75r\" (UID: \"9ab346c9-f70f-4663-9b53-67b4d66b3112\") " pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" Nov 25 09:46:50 crc kubenswrapper[4776]: I1125 09:46:50.400590 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-dns-svc\") pod \"dnsmasq-dns-5d7f54fb65-td75r\" (UID: \"9ab346c9-f70f-4663-9b53-67b4d66b3112\") " pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" Nov 25 09:46:50 crc kubenswrapper[4776]: I1125 09:46:50.400651 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-config\") pod \"dnsmasq-dns-5d7f54fb65-td75r\" (UID: \"9ab346c9-f70f-4663-9b53-67b4d66b3112\") " pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" Nov 25 09:46:50 crc kubenswrapper[4776]: I1125 09:46:50.400677 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7g6c\" (UniqueName: \"kubernetes.io/projected/9ab346c9-f70f-4663-9b53-67b4d66b3112-kube-api-access-p7g6c\") pod \"dnsmasq-dns-5d7f54fb65-td75r\" (UID: \"9ab346c9-f70f-4663-9b53-67b4d66b3112\") " pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" Nov 25 09:46:50 crc kubenswrapper[4776]: I1125 09:46:50.401148 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-ovsdbserver-sb\") pod \"dnsmasq-dns-5d7f54fb65-td75r\" (UID: \"9ab346c9-f70f-4663-9b53-67b4d66b3112\") " pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" Nov 25 09:46:50 crc kubenswrapper[4776]: I1125 09:46:50.503530 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-dns-swift-storage-0\") pod \"dnsmasq-dns-5d7f54fb65-td75r\" (UID: \"9ab346c9-f70f-4663-9b53-67b4d66b3112\") " pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" Nov 25 09:46:50 crc kubenswrapper[4776]: I1125 09:46:50.503589 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-ovsdbserver-nb\") pod \"dnsmasq-dns-5d7f54fb65-td75r\" (UID: \"9ab346c9-f70f-4663-9b53-67b4d66b3112\") " pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" Nov 25 09:46:50 crc kubenswrapper[4776]: I1125 09:46:50.503616 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-dns-svc\") pod \"dnsmasq-dns-5d7f54fb65-td75r\" (UID: \"9ab346c9-f70f-4663-9b53-67b4d66b3112\") " pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" Nov 25 09:46:50 crc kubenswrapper[4776]: I1125 09:46:50.504737 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-ovsdbserver-nb\") pod \"dnsmasq-dns-5d7f54fb65-td75r\" (UID: \"9ab346c9-f70f-4663-9b53-67b4d66b3112\") " pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" Nov 25 09:46:50 crc kubenswrapper[4776]: I1125 09:46:50.504755 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-dns-svc\") pod \"dnsmasq-dns-5d7f54fb65-td75r\" (UID: \"9ab346c9-f70f-4663-9b53-67b4d66b3112\") " pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" Nov 25 09:46:50 crc kubenswrapper[4776]: I1125 09:46:50.504802 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-config\") pod \"dnsmasq-dns-5d7f54fb65-td75r\" (UID: \"9ab346c9-f70f-4663-9b53-67b4d66b3112\") " pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" Nov 25 09:46:50 crc kubenswrapper[4776]: I1125 09:46:50.504858 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7g6c\" (UniqueName: \"kubernetes.io/projected/9ab346c9-f70f-4663-9b53-67b4d66b3112-kube-api-access-p7g6c\") pod \"dnsmasq-dns-5d7f54fb65-td75r\" (UID: \"9ab346c9-f70f-4663-9b53-67b4d66b3112\") " pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" Nov 25 09:46:50 crc kubenswrapper[4776]: I1125 09:46:50.504922 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-dns-swift-storage-0\") pod \"dnsmasq-dns-5d7f54fb65-td75r\" (UID: \"9ab346c9-f70f-4663-9b53-67b4d66b3112\") " pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" Nov 25 09:46:50 crc kubenswrapper[4776]: I1125 09:46:50.505351 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-config\") pod \"dnsmasq-dns-5d7f54fb65-td75r\" (UID: \"9ab346c9-f70f-4663-9b53-67b4d66b3112\") " pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" Nov 25 09:46:50 crc kubenswrapper[4776]: I1125 09:46:50.505526 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-ovsdbserver-sb\") pod \"dnsmasq-dns-5d7f54fb65-td75r\" (UID: \"9ab346c9-f70f-4663-9b53-67b4d66b3112\") " pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" Nov 25 09:46:50 crc kubenswrapper[4776]: I1125 09:46:50.506235 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-ovsdbserver-sb\") pod \"dnsmasq-dns-5d7f54fb65-td75r\" (UID: \"9ab346c9-f70f-4663-9b53-67b4d66b3112\") " pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" Nov 25 09:46:50 crc kubenswrapper[4776]: I1125 09:46:50.532572 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7g6c\" (UniqueName: \"kubernetes.io/projected/9ab346c9-f70f-4663-9b53-67b4d66b3112-kube-api-access-p7g6c\") pod \"dnsmasq-dns-5d7f54fb65-td75r\" (UID: \"9ab346c9-f70f-4663-9b53-67b4d66b3112\") " pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" Nov 25 09:46:50 crc kubenswrapper[4776]: I1125 09:46:50.618680 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" Nov 25 09:46:51 crc kubenswrapper[4776]: I1125 09:46:51.176987 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d7f54fb65-td75r"] Nov 25 09:46:51 crc kubenswrapper[4776]: W1125 09:46:51.188702 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ab346c9_f70f_4663_9b53_67b4d66b3112.slice/crio-e2ea8ecd0422a4e5db4d8eaedcbaa39313dd266172bfa7165de56b2e091c7147 WatchSource:0}: Error finding container e2ea8ecd0422a4e5db4d8eaedcbaa39313dd266172bfa7165de56b2e091c7147: Status 404 returned error can't find the container with id e2ea8ecd0422a4e5db4d8eaedcbaa39313dd266172bfa7165de56b2e091c7147 Nov 25 09:46:52 crc kubenswrapper[4776]: I1125 09:46:52.094273 4776 generic.go:334] "Generic (PLEG): container finished" podID="9ab346c9-f70f-4663-9b53-67b4d66b3112" containerID="59870a401f6a788fbbb94bfbf471e78e459d6e2cbdf6f95acf0bf39015b4d0ff" exitCode=0 Nov 25 09:46:52 crc kubenswrapper[4776]: I1125 09:46:52.095328 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" event={"ID":"9ab346c9-f70f-4663-9b53-67b4d66b3112","Type":"ContainerDied","Data":"59870a401f6a788fbbb94bfbf471e78e459d6e2cbdf6f95acf0bf39015b4d0ff"} Nov 25 09:46:52 crc kubenswrapper[4776]: I1125 09:46:52.095357 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" event={"ID":"9ab346c9-f70f-4663-9b53-67b4d66b3112","Type":"ContainerStarted","Data":"e2ea8ecd0422a4e5db4d8eaedcbaa39313dd266172bfa7165de56b2e091c7147"} Nov 25 09:46:52 crc kubenswrapper[4776]: I1125 09:46:52.862070 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:46:52 crc kubenswrapper[4776]: I1125 09:46:52.862787 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f1e0bba8-06f9-481c-a780-bfc17ec5222c" containerName="ceilometer-central-agent" containerID="cri-o://a61faed17a75f0fadba92e5ee69f5a2f1d3c2259dffec60af1df198f74f2f188" gracePeriod=30 Nov 25 09:46:52 crc kubenswrapper[4776]: I1125 09:46:52.862888 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f1e0bba8-06f9-481c-a780-bfc17ec5222c" containerName="sg-core" containerID="cri-o://b6d6e1b8d62ded83fbd5439c8721571564758b292c4ad38aae991861620d9cfb" gracePeriod=30 Nov 25 09:46:52 crc kubenswrapper[4776]: I1125 09:46:52.862925 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f1e0bba8-06f9-481c-a780-bfc17ec5222c" containerName="proxy-httpd" containerID="cri-o://30c6ca260ebb54bae727c7b09725297576062adf42235ac89f7f0b9f8fe54531" gracePeriod=30 Nov 25 09:46:52 crc kubenswrapper[4776]: I1125 09:46:52.863051 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f1e0bba8-06f9-481c-a780-bfc17ec5222c" containerName="ceilometer-notification-agent" containerID="cri-o://bf1ec9e8754cbd3f26c19b9836d8b9f855007d555bfd31191664e08b436499e8" gracePeriod=30 Nov 25 09:46:52 crc kubenswrapper[4776]: I1125 09:46:52.972126 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 09:46:53 crc kubenswrapper[4776]: I1125 09:46:53.105299 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" event={"ID":"9ab346c9-f70f-4663-9b53-67b4d66b3112","Type":"ContainerStarted","Data":"2b9a66a95781ceb011f5c668fc0302d4aa355e258935b0bc8a74577b415062f6"} Nov 25 09:46:53 crc kubenswrapper[4776]: I1125 09:46:53.106525 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" Nov 25 09:46:53 crc kubenswrapper[4776]: I1125 09:46:53.107515 4776 generic.go:334] "Generic (PLEG): container finished" podID="f1e0bba8-06f9-481c-a780-bfc17ec5222c" containerID="30c6ca260ebb54bae727c7b09725297576062adf42235ac89f7f0b9f8fe54531" exitCode=0 Nov 25 09:46:53 crc kubenswrapper[4776]: I1125 09:46:53.107553 4776 generic.go:334] "Generic (PLEG): container finished" podID="f1e0bba8-06f9-481c-a780-bfc17ec5222c" containerID="b6d6e1b8d62ded83fbd5439c8721571564758b292c4ad38aae991861620d9cfb" exitCode=2 Nov 25 09:46:53 crc kubenswrapper[4776]: I1125 09:46:53.107580 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f1e0bba8-06f9-481c-a780-bfc17ec5222c","Type":"ContainerDied","Data":"30c6ca260ebb54bae727c7b09725297576062adf42235ac89f7f0b9f8fe54531"} Nov 25 09:46:53 crc kubenswrapper[4776]: I1125 09:46:53.107612 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f1e0bba8-06f9-481c-a780-bfc17ec5222c","Type":"ContainerDied","Data":"b6d6e1b8d62ded83fbd5439c8721571564758b292c4ad38aae991861620d9cfb"} Nov 25 09:46:53 crc kubenswrapper[4776]: I1125 09:46:53.107762 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3fd6c763-8b11-4c88-a9f5-b63830efe063" containerName="nova-api-log" containerID="cri-o://fa19ba9ebb2d7a99775499f75792a50cc747196875d92699aae892c04c1f55a8" gracePeriod=30 Nov 25 09:46:53 crc kubenswrapper[4776]: I1125 09:46:53.107858 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3fd6c763-8b11-4c88-a9f5-b63830efe063" containerName="nova-api-api" containerID="cri-o://4105d1d897ad666a132559e4e85151fd91afa80102510d21f72cb2b3a19982c3" gracePeriod=30 Nov 25 09:46:53 crc kubenswrapper[4776]: I1125 09:46:53.146146 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" podStartSLOduration=3.146124106 podStartE2EDuration="3.146124106s" podCreationTimestamp="2025-11-25 09:46:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:46:53.138249689 +0000 UTC m=+1358.179309302" watchObservedRunningTime="2025-11-25 09:46:53.146124106 +0000 UTC m=+1358.187183669" Nov 25 09:46:53 crc kubenswrapper[4776]: I1125 09:46:53.755763 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.122770 4776 generic.go:334] "Generic (PLEG): container finished" podID="f1e0bba8-06f9-481c-a780-bfc17ec5222c" containerID="bf1ec9e8754cbd3f26c19b9836d8b9f855007d555bfd31191664e08b436499e8" exitCode=0 Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.122810 4776 generic.go:334] "Generic (PLEG): container finished" podID="f1e0bba8-06f9-481c-a780-bfc17ec5222c" containerID="a61faed17a75f0fadba92e5ee69f5a2f1d3c2259dffec60af1df198f74f2f188" exitCode=0 Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.122856 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f1e0bba8-06f9-481c-a780-bfc17ec5222c","Type":"ContainerDied","Data":"bf1ec9e8754cbd3f26c19b9836d8b9f855007d555bfd31191664e08b436499e8"} Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.122911 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f1e0bba8-06f9-481c-a780-bfc17ec5222c","Type":"ContainerDied","Data":"a61faed17a75f0fadba92e5ee69f5a2f1d3c2259dffec60af1df198f74f2f188"} Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.124647 4776 generic.go:334] "Generic (PLEG): container finished" podID="3fd6c763-8b11-4c88-a9f5-b63830efe063" containerID="fa19ba9ebb2d7a99775499f75792a50cc747196875d92699aae892c04c1f55a8" exitCode=143 Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.125222 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3fd6c763-8b11-4c88-a9f5-b63830efe063","Type":"ContainerDied","Data":"fa19ba9ebb2d7a99775499f75792a50cc747196875d92699aae892c04c1f55a8"} Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.503247 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.593060 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-combined-ca-bundle\") pod \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.593179 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-sg-core-conf-yaml\") pod \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.593271 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-ceilometer-tls-certs\") pod \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.593342 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzwnc\" (UniqueName: \"kubernetes.io/projected/f1e0bba8-06f9-481c-a780-bfc17ec5222c-kube-api-access-pzwnc\") pod \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.593392 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f1e0bba8-06f9-481c-a780-bfc17ec5222c-log-httpd\") pod \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.593522 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f1e0bba8-06f9-481c-a780-bfc17ec5222c-run-httpd\") pod \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.593593 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-config-data\") pod \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.593697 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-scripts\") pod \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\" (UID: \"f1e0bba8-06f9-481c-a780-bfc17ec5222c\") " Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.597492 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1e0bba8-06f9-481c-a780-bfc17ec5222c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f1e0bba8-06f9-481c-a780-bfc17ec5222c" (UID: "f1e0bba8-06f9-481c-a780-bfc17ec5222c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.598797 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1e0bba8-06f9-481c-a780-bfc17ec5222c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f1e0bba8-06f9-481c-a780-bfc17ec5222c" (UID: "f1e0bba8-06f9-481c-a780-bfc17ec5222c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.613490 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1e0bba8-06f9-481c-a780-bfc17ec5222c-kube-api-access-pzwnc" (OuterVolumeSpecName: "kube-api-access-pzwnc") pod "f1e0bba8-06f9-481c-a780-bfc17ec5222c" (UID: "f1e0bba8-06f9-481c-a780-bfc17ec5222c"). InnerVolumeSpecName "kube-api-access-pzwnc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.656302 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-scripts" (OuterVolumeSpecName: "scripts") pod "f1e0bba8-06f9-481c-a780-bfc17ec5222c" (UID: "f1e0bba8-06f9-481c-a780-bfc17ec5222c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.704393 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzwnc\" (UniqueName: \"kubernetes.io/projected/f1e0bba8-06f9-481c-a780-bfc17ec5222c-kube-api-access-pzwnc\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.704429 4776 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f1e0bba8-06f9-481c-a780-bfc17ec5222c-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.704440 4776 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f1e0bba8-06f9-481c-a780-bfc17ec5222c-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.704451 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.721291 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f1e0bba8-06f9-481c-a780-bfc17ec5222c" (UID: "f1e0bba8-06f9-481c-a780-bfc17ec5222c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.803313 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "f1e0bba8-06f9-481c-a780-bfc17ec5222c" (UID: "f1e0bba8-06f9-481c-a780-bfc17ec5222c"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.809527 4776 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.809566 4776 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.818356 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f1e0bba8-06f9-481c-a780-bfc17ec5222c" (UID: "f1e0bba8-06f9-481c-a780-bfc17ec5222c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.847765 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-config-data" (OuterVolumeSpecName: "config-data") pod "f1e0bba8-06f9-481c-a780-bfc17ec5222c" (UID: "f1e0bba8-06f9-481c-a780-bfc17ec5222c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.911352 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:54 crc kubenswrapper[4776]: I1125 09:46:54.911854 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1e0bba8-06f9-481c-a780-bfc17ec5222c-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.135049 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f1e0bba8-06f9-481c-a780-bfc17ec5222c","Type":"ContainerDied","Data":"0c10415e998a38a6825c9564b5c9eb05f9577d6610db0f9e10b2b4a9fa196edd"} Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.135133 4776 scope.go:117] "RemoveContainer" containerID="30c6ca260ebb54bae727c7b09725297576062adf42235ac89f7f0b9f8fe54531" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.135083 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.158106 4776 scope.go:117] "RemoveContainer" containerID="b6d6e1b8d62ded83fbd5439c8721571564758b292c4ad38aae991861620d9cfb" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.173083 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.181063 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.183741 4776 scope.go:117] "RemoveContainer" containerID="bf1ec9e8754cbd3f26c19b9836d8b9f855007d555bfd31191664e08b436499e8" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.207935 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:46:55 crc kubenswrapper[4776]: E1125 09:46:55.208485 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1e0bba8-06f9-481c-a780-bfc17ec5222c" containerName="ceilometer-central-agent" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.208511 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1e0bba8-06f9-481c-a780-bfc17ec5222c" containerName="ceilometer-central-agent" Nov 25 09:46:55 crc kubenswrapper[4776]: E1125 09:46:55.208530 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1e0bba8-06f9-481c-a780-bfc17ec5222c" containerName="ceilometer-notification-agent" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.208539 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1e0bba8-06f9-481c-a780-bfc17ec5222c" containerName="ceilometer-notification-agent" Nov 25 09:46:55 crc kubenswrapper[4776]: E1125 09:46:55.208547 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1e0bba8-06f9-481c-a780-bfc17ec5222c" containerName="proxy-httpd" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.208555 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1e0bba8-06f9-481c-a780-bfc17ec5222c" containerName="proxy-httpd" Nov 25 09:46:55 crc kubenswrapper[4776]: E1125 09:46:55.208573 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1e0bba8-06f9-481c-a780-bfc17ec5222c" containerName="sg-core" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.208581 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1e0bba8-06f9-481c-a780-bfc17ec5222c" containerName="sg-core" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.208900 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1e0bba8-06f9-481c-a780-bfc17ec5222c" containerName="ceilometer-notification-agent" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.208924 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1e0bba8-06f9-481c-a780-bfc17ec5222c" containerName="proxy-httpd" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.208937 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1e0bba8-06f9-481c-a780-bfc17ec5222c" containerName="ceilometer-central-agent" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.208989 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1e0bba8-06f9-481c-a780-bfc17ec5222c" containerName="sg-core" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.211566 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.214399 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.217398 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.228312 4776 scope.go:117] "RemoveContainer" containerID="a61faed17a75f0fadba92e5ee69f5a2f1d3c2259dffec60af1df198f74f2f188" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.230931 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.233724 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.294626 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:46:55 crc kubenswrapper[4776]: E1125 09:46:55.295347 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceilometer-tls-certs combined-ca-bundle config-data kube-api-access-mttgr log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="55ccde5c-1bbe-4c35-b0a5-521dff19c0d8" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.325976 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-run-httpd\") pod \"ceilometer-0\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " pod="openstack/ceilometer-0" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.326022 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " pod="openstack/ceilometer-0" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.326041 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-config-data\") pod \"ceilometer-0\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " pod="openstack/ceilometer-0" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.326079 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " pod="openstack/ceilometer-0" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.326125 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " pod="openstack/ceilometer-0" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.326201 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mttgr\" (UniqueName: \"kubernetes.io/projected/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-kube-api-access-mttgr\") pod \"ceilometer-0\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " pod="openstack/ceilometer-0" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.326232 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-log-httpd\") pod \"ceilometer-0\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " pod="openstack/ceilometer-0" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.326249 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-scripts\") pod \"ceilometer-0\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " pod="openstack/ceilometer-0" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.428245 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-run-httpd\") pod \"ceilometer-0\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " pod="openstack/ceilometer-0" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.428307 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " pod="openstack/ceilometer-0" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.428333 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-config-data\") pod \"ceilometer-0\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " pod="openstack/ceilometer-0" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.428388 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " pod="openstack/ceilometer-0" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.428435 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " pod="openstack/ceilometer-0" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.428478 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mttgr\" (UniqueName: \"kubernetes.io/projected/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-kube-api-access-mttgr\") pod \"ceilometer-0\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " pod="openstack/ceilometer-0" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.428513 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-log-httpd\") pod \"ceilometer-0\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " pod="openstack/ceilometer-0" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.428531 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-scripts\") pod \"ceilometer-0\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " pod="openstack/ceilometer-0" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.428881 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-run-httpd\") pod \"ceilometer-0\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " pod="openstack/ceilometer-0" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.429561 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-log-httpd\") pod \"ceilometer-0\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " pod="openstack/ceilometer-0" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.433401 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " pod="openstack/ceilometer-0" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.433795 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-scripts\") pod \"ceilometer-0\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " pod="openstack/ceilometer-0" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.433904 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " pod="openstack/ceilometer-0" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.438962 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-config-data\") pod \"ceilometer-0\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " pod="openstack/ceilometer-0" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.440291 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " pod="openstack/ceilometer-0" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.447375 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mttgr\" (UniqueName: \"kubernetes.io/projected/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-kube-api-access-mttgr\") pod \"ceilometer-0\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " pod="openstack/ceilometer-0" Nov 25 09:46:55 crc kubenswrapper[4776]: I1125 09:46:55.674717 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1e0bba8-06f9-481c-a780-bfc17ec5222c" path="/var/lib/kubelet/pods/f1e0bba8-06f9-481c-a780-bfc17ec5222c/volumes" Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.144843 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.159807 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.242305 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-log-httpd\") pod \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.242481 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-sg-core-conf-yaml\") pod \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.242516 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-combined-ca-bundle\") pod \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.242537 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-ceilometer-tls-certs\") pod \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.242593 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-scripts\") pod \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.242625 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-run-httpd\") pod \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.243180 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "55ccde5c-1bbe-4c35-b0a5-521dff19c0d8" (UID: "55ccde5c-1bbe-4c35-b0a5-521dff19c0d8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.243243 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mttgr\" (UniqueName: \"kubernetes.io/projected/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-kube-api-access-mttgr\") pod \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.243324 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-config-data\") pod \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\" (UID: \"55ccde5c-1bbe-4c35-b0a5-521dff19c0d8\") " Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.243612 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "55ccde5c-1bbe-4c35-b0a5-521dff19c0d8" (UID: "55ccde5c-1bbe-4c35-b0a5-521dff19c0d8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.244174 4776 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.244197 4776 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.249366 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-kube-api-access-mttgr" (OuterVolumeSpecName: "kube-api-access-mttgr") pod "55ccde5c-1bbe-4c35-b0a5-521dff19c0d8" (UID: "55ccde5c-1bbe-4c35-b0a5-521dff19c0d8"). InnerVolumeSpecName "kube-api-access-mttgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.249475 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-scripts" (OuterVolumeSpecName: "scripts") pod "55ccde5c-1bbe-4c35-b0a5-521dff19c0d8" (UID: "55ccde5c-1bbe-4c35-b0a5-521dff19c0d8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.251130 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "55ccde5c-1bbe-4c35-b0a5-521dff19c0d8" (UID: "55ccde5c-1bbe-4c35-b0a5-521dff19c0d8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.251232 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "55ccde5c-1bbe-4c35-b0a5-521dff19c0d8" (UID: "55ccde5c-1bbe-4c35-b0a5-521dff19c0d8"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.253616 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "55ccde5c-1bbe-4c35-b0a5-521dff19c0d8" (UID: "55ccde5c-1bbe-4c35-b0a5-521dff19c0d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.254638 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-config-data" (OuterVolumeSpecName: "config-data") pod "55ccde5c-1bbe-4c35-b0a5-521dff19c0d8" (UID: "55ccde5c-1bbe-4c35-b0a5-521dff19c0d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.345501 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.345539 4776 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.345553 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.345566 4776 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.345578 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.345587 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mttgr\" (UniqueName: \"kubernetes.io/projected/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8-kube-api-access-mttgr\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.674879 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.754307 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4x9gx\" (UniqueName: \"kubernetes.io/projected/3fd6c763-8b11-4c88-a9f5-b63830efe063-kube-api-access-4x9gx\") pod \"3fd6c763-8b11-4c88-a9f5-b63830efe063\" (UID: \"3fd6c763-8b11-4c88-a9f5-b63830efe063\") " Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.754394 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fd6c763-8b11-4c88-a9f5-b63830efe063-combined-ca-bundle\") pod \"3fd6c763-8b11-4c88-a9f5-b63830efe063\" (UID: \"3fd6c763-8b11-4c88-a9f5-b63830efe063\") " Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.754499 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3fd6c763-8b11-4c88-a9f5-b63830efe063-logs\") pod \"3fd6c763-8b11-4c88-a9f5-b63830efe063\" (UID: \"3fd6c763-8b11-4c88-a9f5-b63830efe063\") " Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.754527 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fd6c763-8b11-4c88-a9f5-b63830efe063-config-data\") pod \"3fd6c763-8b11-4c88-a9f5-b63830efe063\" (UID: \"3fd6c763-8b11-4c88-a9f5-b63830efe063\") " Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.754963 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fd6c763-8b11-4c88-a9f5-b63830efe063-logs" (OuterVolumeSpecName: "logs") pod "3fd6c763-8b11-4c88-a9f5-b63830efe063" (UID: "3fd6c763-8b11-4c88-a9f5-b63830efe063"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.759264 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fd6c763-8b11-4c88-a9f5-b63830efe063-kube-api-access-4x9gx" (OuterVolumeSpecName: "kube-api-access-4x9gx") pod "3fd6c763-8b11-4c88-a9f5-b63830efe063" (UID: "3fd6c763-8b11-4c88-a9f5-b63830efe063"). InnerVolumeSpecName "kube-api-access-4x9gx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.793376 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fd6c763-8b11-4c88-a9f5-b63830efe063-config-data" (OuterVolumeSpecName: "config-data") pod "3fd6c763-8b11-4c88-a9f5-b63830efe063" (UID: "3fd6c763-8b11-4c88-a9f5-b63830efe063"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.821262 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fd6c763-8b11-4c88-a9f5-b63830efe063-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3fd6c763-8b11-4c88-a9f5-b63830efe063" (UID: "3fd6c763-8b11-4c88-a9f5-b63830efe063"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.858605 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4x9gx\" (UniqueName: \"kubernetes.io/projected/3fd6c763-8b11-4c88-a9f5-b63830efe063-kube-api-access-4x9gx\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.858681 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fd6c763-8b11-4c88-a9f5-b63830efe063-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.858697 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3fd6c763-8b11-4c88-a9f5-b63830efe063-logs\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:56 crc kubenswrapper[4776]: I1125 09:46:56.858709 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3fd6c763-8b11-4c88-a9f5-b63830efe063-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.160211 4776 generic.go:334] "Generic (PLEG): container finished" podID="3fd6c763-8b11-4c88-a9f5-b63830efe063" containerID="4105d1d897ad666a132559e4e85151fd91afa80102510d21f72cb2b3a19982c3" exitCode=0 Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.160531 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.161184 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.161829 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3fd6c763-8b11-4c88-a9f5-b63830efe063","Type":"ContainerDied","Data":"4105d1d897ad666a132559e4e85151fd91afa80102510d21f72cb2b3a19982c3"} Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.161884 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3fd6c763-8b11-4c88-a9f5-b63830efe063","Type":"ContainerDied","Data":"76d1f24d20681720c94d57bd91f56f0cf1f52e52e4978c0a73d49076b00311aa"} Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.161908 4776 scope.go:117] "RemoveContainer" containerID="4105d1d897ad666a132559e4e85151fd91afa80102510d21f72cb2b3a19982c3" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.229546 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.244465 4776 scope.go:117] "RemoveContainer" containerID="fa19ba9ebb2d7a99775499f75792a50cc747196875d92699aae892c04c1f55a8" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.244659 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.255944 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.267880 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.278151 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:46:57 crc kubenswrapper[4776]: E1125 09:46:57.278671 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fd6c763-8b11-4c88-a9f5-b63830efe063" containerName="nova-api-api" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.278692 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fd6c763-8b11-4c88-a9f5-b63830efe063" containerName="nova-api-api" Nov 25 09:46:57 crc kubenswrapper[4776]: E1125 09:46:57.278734 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fd6c763-8b11-4c88-a9f5-b63830efe063" containerName="nova-api-log" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.278743 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fd6c763-8b11-4c88-a9f5-b63830efe063" containerName="nova-api-log" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.278971 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fd6c763-8b11-4c88-a9f5-b63830efe063" containerName="nova-api-api" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.278992 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fd6c763-8b11-4c88-a9f5-b63830efe063" containerName="nova-api-log" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.281681 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.283230 4776 scope.go:117] "RemoveContainer" containerID="4105d1d897ad666a132559e4e85151fd91afa80102510d21f72cb2b3a19982c3" Nov 25 09:46:57 crc kubenswrapper[4776]: E1125 09:46:57.286470 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4105d1d897ad666a132559e4e85151fd91afa80102510d21f72cb2b3a19982c3\": container with ID starting with 4105d1d897ad666a132559e4e85151fd91afa80102510d21f72cb2b3a19982c3 not found: ID does not exist" containerID="4105d1d897ad666a132559e4e85151fd91afa80102510d21f72cb2b3a19982c3" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.286935 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4105d1d897ad666a132559e4e85151fd91afa80102510d21f72cb2b3a19982c3"} err="failed to get container status \"4105d1d897ad666a132559e4e85151fd91afa80102510d21f72cb2b3a19982c3\": rpc error: code = NotFound desc = could not find container \"4105d1d897ad666a132559e4e85151fd91afa80102510d21f72cb2b3a19982c3\": container with ID starting with 4105d1d897ad666a132559e4e85151fd91afa80102510d21f72cb2b3a19982c3 not found: ID does not exist" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.286978 4776 scope.go:117] "RemoveContainer" containerID="fa19ba9ebb2d7a99775499f75792a50cc747196875d92699aae892c04c1f55a8" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.287117 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.287783 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.288018 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.288173 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:46:57 crc kubenswrapper[4776]: E1125 09:46:57.291621 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa19ba9ebb2d7a99775499f75792a50cc747196875d92699aae892c04c1f55a8\": container with ID starting with fa19ba9ebb2d7a99775499f75792a50cc747196875d92699aae892c04c1f55a8 not found: ID does not exist" containerID="fa19ba9ebb2d7a99775499f75792a50cc747196875d92699aae892c04c1f55a8" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.291671 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa19ba9ebb2d7a99775499f75792a50cc747196875d92699aae892c04c1f55a8"} err="failed to get container status \"fa19ba9ebb2d7a99775499f75792a50cc747196875d92699aae892c04c1f55a8\": rpc error: code = NotFound desc = could not find container \"fa19ba9ebb2d7a99775499f75792a50cc747196875d92699aae892c04c1f55a8\": container with ID starting with fa19ba9ebb2d7a99775499f75792a50cc747196875d92699aae892c04c1f55a8 not found: ID does not exist" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.305048 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.313432 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.315690 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.317133 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.317952 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.331146 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.367821 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " pod="openstack/ceilometer-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.367883 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " pod="openstack/ceilometer-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.367918 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b90b0bf-a4b6-4336-9e49-33792ead4f07-public-tls-certs\") pod \"nova-api-0\" (UID: \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\") " pod="openstack/nova-api-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.367964 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n56sr\" (UniqueName: \"kubernetes.io/projected/2b90b0bf-a4b6-4336-9e49-33792ead4f07-kube-api-access-n56sr\") pod \"nova-api-0\" (UID: \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\") " pod="openstack/nova-api-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.367991 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aab7d356-ec67-4e5d-9c25-f26e39786f79-log-httpd\") pod \"ceilometer-0\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " pod="openstack/ceilometer-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.368021 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b90b0bf-a4b6-4336-9e49-33792ead4f07-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\") " pod="openstack/nova-api-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.368046 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aab7d356-ec67-4e5d-9c25-f26e39786f79-run-httpd\") pod \"ceilometer-0\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " pod="openstack/ceilometer-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.368179 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxgd8\" (UniqueName: \"kubernetes.io/projected/aab7d356-ec67-4e5d-9c25-f26e39786f79-kube-api-access-dxgd8\") pod \"ceilometer-0\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " pod="openstack/ceilometer-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.368280 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-config-data\") pod \"ceilometer-0\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " pod="openstack/ceilometer-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.368316 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b90b0bf-a4b6-4336-9e49-33792ead4f07-config-data\") pod \"nova-api-0\" (UID: \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\") " pod="openstack/nova-api-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.368338 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b90b0bf-a4b6-4336-9e49-33792ead4f07-logs\") pod \"nova-api-0\" (UID: \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\") " pod="openstack/nova-api-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.368361 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-scripts\") pod \"ceilometer-0\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " pod="openstack/ceilometer-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.368567 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b90b0bf-a4b6-4336-9e49-33792ead4f07-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\") " pod="openstack/nova-api-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.368617 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " pod="openstack/ceilometer-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.470876 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b90b0bf-a4b6-4336-9e49-33792ead4f07-logs\") pod \"nova-api-0\" (UID: \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\") " pod="openstack/nova-api-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.471468 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-scripts\") pod \"ceilometer-0\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " pod="openstack/ceilometer-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.472364 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b90b0bf-a4b6-4336-9e49-33792ead4f07-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\") " pod="openstack/nova-api-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.472406 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " pod="openstack/ceilometer-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.472507 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " pod="openstack/ceilometer-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.471385 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b90b0bf-a4b6-4336-9e49-33792ead4f07-logs\") pod \"nova-api-0\" (UID: \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\") " pod="openstack/nova-api-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.472542 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " pod="openstack/ceilometer-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.472834 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b90b0bf-a4b6-4336-9e49-33792ead4f07-public-tls-certs\") pod \"nova-api-0\" (UID: \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\") " pod="openstack/nova-api-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.472931 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n56sr\" (UniqueName: \"kubernetes.io/projected/2b90b0bf-a4b6-4336-9e49-33792ead4f07-kube-api-access-n56sr\") pod \"nova-api-0\" (UID: \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\") " pod="openstack/nova-api-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.472966 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aab7d356-ec67-4e5d-9c25-f26e39786f79-log-httpd\") pod \"ceilometer-0\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " pod="openstack/ceilometer-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.472998 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b90b0bf-a4b6-4336-9e49-33792ead4f07-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\") " pod="openstack/nova-api-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.473024 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aab7d356-ec67-4e5d-9c25-f26e39786f79-run-httpd\") pod \"ceilometer-0\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " pod="openstack/ceilometer-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.473108 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxgd8\" (UniqueName: \"kubernetes.io/projected/aab7d356-ec67-4e5d-9c25-f26e39786f79-kube-api-access-dxgd8\") pod \"ceilometer-0\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " pod="openstack/ceilometer-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.473183 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-config-data\") pod \"ceilometer-0\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " pod="openstack/ceilometer-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.473224 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b90b0bf-a4b6-4336-9e49-33792ead4f07-config-data\") pod \"nova-api-0\" (UID: \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\") " pod="openstack/nova-api-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.473537 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aab7d356-ec67-4e5d-9c25-f26e39786f79-log-httpd\") pod \"ceilometer-0\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " pod="openstack/ceilometer-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.473760 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aab7d356-ec67-4e5d-9c25-f26e39786f79-run-httpd\") pod \"ceilometer-0\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " pod="openstack/ceilometer-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.477238 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " pod="openstack/ceilometer-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.478436 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b90b0bf-a4b6-4336-9e49-33792ead4f07-public-tls-certs\") pod \"nova-api-0\" (UID: \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\") " pod="openstack/nova-api-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.478615 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b90b0bf-a4b6-4336-9e49-33792ead4f07-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\") " pod="openstack/nova-api-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.478995 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-scripts\") pod \"ceilometer-0\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " pod="openstack/ceilometer-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.479345 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b90b0bf-a4b6-4336-9e49-33792ead4f07-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\") " pod="openstack/nova-api-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.480617 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " pod="openstack/ceilometer-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.480707 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b90b0bf-a4b6-4336-9e49-33792ead4f07-config-data\") pod \"nova-api-0\" (UID: \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\") " pod="openstack/nova-api-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.481424 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " pod="openstack/ceilometer-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.484895 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-config-data\") pod \"ceilometer-0\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " pod="openstack/ceilometer-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.489599 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n56sr\" (UniqueName: \"kubernetes.io/projected/2b90b0bf-a4b6-4336-9e49-33792ead4f07-kube-api-access-n56sr\") pod \"nova-api-0\" (UID: \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\") " pod="openstack/nova-api-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.494016 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxgd8\" (UniqueName: \"kubernetes.io/projected/aab7d356-ec67-4e5d-9c25-f26e39786f79-kube-api-access-dxgd8\") pod \"ceilometer-0\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " pod="openstack/ceilometer-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.617576 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.639433 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.675840 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fd6c763-8b11-4c88-a9f5-b63830efe063" path="/var/lib/kubelet/pods/3fd6c763-8b11-4c88-a9f5-b63830efe063/volumes" Nov 25 09:46:57 crc kubenswrapper[4776]: I1125 09:46:57.676501 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55ccde5c-1bbe-4c35-b0a5-521dff19c0d8" path="/var/lib/kubelet/pods/55ccde5c-1bbe-4c35-b0a5-521dff19c0d8/volumes" Nov 25 09:46:58 crc kubenswrapper[4776]: I1125 09:46:58.085634 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:46:58 crc kubenswrapper[4776]: I1125 09:46:58.174037 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aab7d356-ec67-4e5d-9c25-f26e39786f79","Type":"ContainerStarted","Data":"c5e4a4b031984a16cc9d6b7895ceb3f8ccac6ee47ecc06363cdffe20b5013c94"} Nov 25 09:46:58 crc kubenswrapper[4776]: I1125 09:46:58.191108 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 09:46:58 crc kubenswrapper[4776]: W1125 09:46:58.204911 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b90b0bf_a4b6_4336_9e49_33792ead4f07.slice/crio-f8d31df8b7a3a4d8d3f2899d59443769fe63e3a693b75de4adf3f81dbbfdc256 WatchSource:0}: Error finding container f8d31df8b7a3a4d8d3f2899d59443769fe63e3a693b75de4adf3f81dbbfdc256: Status 404 returned error can't find the container with id f8d31df8b7a3a4d8d3f2899d59443769fe63e3a693b75de4adf3f81dbbfdc256 Nov 25 09:46:58 crc kubenswrapper[4776]: I1125 09:46:58.757883 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:58 crc kubenswrapper[4776]: I1125 09:46:58.785976 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:59 crc kubenswrapper[4776]: I1125 09:46:59.186130 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2b90b0bf-a4b6-4336-9e49-33792ead4f07","Type":"ContainerStarted","Data":"79cbebf3e9c8cb7e6ea34200c99cf5ad70ddabde13fac5d58d25f07b09c5e055"} Nov 25 09:46:59 crc kubenswrapper[4776]: I1125 09:46:59.187186 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2b90b0bf-a4b6-4336-9e49-33792ead4f07","Type":"ContainerStarted","Data":"5a25e29978f0284d518a88b255bee88fa125d11b08d1eaeb950230e38ea74769"} Nov 25 09:46:59 crc kubenswrapper[4776]: I1125 09:46:59.187305 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2b90b0bf-a4b6-4336-9e49-33792ead4f07","Type":"ContainerStarted","Data":"f8d31df8b7a3a4d8d3f2899d59443769fe63e3a693b75de4adf3f81dbbfdc256"} Nov 25 09:46:59 crc kubenswrapper[4776]: I1125 09:46:59.190325 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aab7d356-ec67-4e5d-9c25-f26e39786f79","Type":"ContainerStarted","Data":"a4df2757871f92f27d85077692ce959d3142d7496b28c90b9062a0ed0d5d450d"} Nov 25 09:46:59 crc kubenswrapper[4776]: I1125 09:46:59.211135 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:46:59 crc kubenswrapper[4776]: I1125 09:46:59.216344 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.216326891 podStartE2EDuration="2.216326891s" podCreationTimestamp="2025-11-25 09:46:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:46:59.20675708 +0000 UTC m=+1364.247816633" watchObservedRunningTime="2025-11-25 09:46:59.216326891 +0000 UTC m=+1364.257386434" Nov 25 09:46:59 crc kubenswrapper[4776]: I1125 09:46:59.366225 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-lpc2f"] Nov 25 09:46:59 crc kubenswrapper[4776]: I1125 09:46:59.367884 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-lpc2f" Nov 25 09:46:59 crc kubenswrapper[4776]: I1125 09:46:59.371905 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 25 09:46:59 crc kubenswrapper[4776]: I1125 09:46:59.372178 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 25 09:46:59 crc kubenswrapper[4776]: I1125 09:46:59.386193 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-lpc2f"] Nov 25 09:46:59 crc kubenswrapper[4776]: I1125 09:46:59.411767 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc2a0874-1575-4605-afb9-01a79c958ba5-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-lpc2f\" (UID: \"cc2a0874-1575-4605-afb9-01a79c958ba5\") " pod="openstack/nova-cell1-cell-mapping-lpc2f" Nov 25 09:46:59 crc kubenswrapper[4776]: I1125 09:46:59.411834 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc2a0874-1575-4605-afb9-01a79c958ba5-config-data\") pod \"nova-cell1-cell-mapping-lpc2f\" (UID: \"cc2a0874-1575-4605-afb9-01a79c958ba5\") " pod="openstack/nova-cell1-cell-mapping-lpc2f" Nov 25 09:46:59 crc kubenswrapper[4776]: I1125 09:46:59.411973 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc2a0874-1575-4605-afb9-01a79c958ba5-scripts\") pod \"nova-cell1-cell-mapping-lpc2f\" (UID: \"cc2a0874-1575-4605-afb9-01a79c958ba5\") " pod="openstack/nova-cell1-cell-mapping-lpc2f" Nov 25 09:46:59 crc kubenswrapper[4776]: I1125 09:46:59.412037 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-242lk\" (UniqueName: \"kubernetes.io/projected/cc2a0874-1575-4605-afb9-01a79c958ba5-kube-api-access-242lk\") pod \"nova-cell1-cell-mapping-lpc2f\" (UID: \"cc2a0874-1575-4605-afb9-01a79c958ba5\") " pod="openstack/nova-cell1-cell-mapping-lpc2f" Nov 25 09:46:59 crc kubenswrapper[4776]: I1125 09:46:59.514588 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc2a0874-1575-4605-afb9-01a79c958ba5-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-lpc2f\" (UID: \"cc2a0874-1575-4605-afb9-01a79c958ba5\") " pod="openstack/nova-cell1-cell-mapping-lpc2f" Nov 25 09:46:59 crc kubenswrapper[4776]: I1125 09:46:59.514662 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc2a0874-1575-4605-afb9-01a79c958ba5-config-data\") pod \"nova-cell1-cell-mapping-lpc2f\" (UID: \"cc2a0874-1575-4605-afb9-01a79c958ba5\") " pod="openstack/nova-cell1-cell-mapping-lpc2f" Nov 25 09:46:59 crc kubenswrapper[4776]: I1125 09:46:59.514801 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc2a0874-1575-4605-afb9-01a79c958ba5-scripts\") pod \"nova-cell1-cell-mapping-lpc2f\" (UID: \"cc2a0874-1575-4605-afb9-01a79c958ba5\") " pod="openstack/nova-cell1-cell-mapping-lpc2f" Nov 25 09:46:59 crc kubenswrapper[4776]: I1125 09:46:59.514858 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-242lk\" (UniqueName: \"kubernetes.io/projected/cc2a0874-1575-4605-afb9-01a79c958ba5-kube-api-access-242lk\") pod \"nova-cell1-cell-mapping-lpc2f\" (UID: \"cc2a0874-1575-4605-afb9-01a79c958ba5\") " pod="openstack/nova-cell1-cell-mapping-lpc2f" Nov 25 09:46:59 crc kubenswrapper[4776]: I1125 09:46:59.522816 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc2a0874-1575-4605-afb9-01a79c958ba5-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-lpc2f\" (UID: \"cc2a0874-1575-4605-afb9-01a79c958ba5\") " pod="openstack/nova-cell1-cell-mapping-lpc2f" Nov 25 09:46:59 crc kubenswrapper[4776]: I1125 09:46:59.523232 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc2a0874-1575-4605-afb9-01a79c958ba5-config-data\") pod \"nova-cell1-cell-mapping-lpc2f\" (UID: \"cc2a0874-1575-4605-afb9-01a79c958ba5\") " pod="openstack/nova-cell1-cell-mapping-lpc2f" Nov 25 09:46:59 crc kubenswrapper[4776]: I1125 09:46:59.527141 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc2a0874-1575-4605-afb9-01a79c958ba5-scripts\") pod \"nova-cell1-cell-mapping-lpc2f\" (UID: \"cc2a0874-1575-4605-afb9-01a79c958ba5\") " pod="openstack/nova-cell1-cell-mapping-lpc2f" Nov 25 09:46:59 crc kubenswrapper[4776]: I1125 09:46:59.530681 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-242lk\" (UniqueName: \"kubernetes.io/projected/cc2a0874-1575-4605-afb9-01a79c958ba5-kube-api-access-242lk\") pod \"nova-cell1-cell-mapping-lpc2f\" (UID: \"cc2a0874-1575-4605-afb9-01a79c958ba5\") " pod="openstack/nova-cell1-cell-mapping-lpc2f" Nov 25 09:46:59 crc kubenswrapper[4776]: I1125 09:46:59.764167 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-lpc2f" Nov 25 09:47:00 crc kubenswrapper[4776]: I1125 09:47:00.201012 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aab7d356-ec67-4e5d-9c25-f26e39786f79","Type":"ContainerStarted","Data":"c99e1a9b60f21bc937fd065f74ab82371110943731ae37f120fc60b0c71336da"} Nov 25 09:47:00 crc kubenswrapper[4776]: I1125 09:47:00.201119 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aab7d356-ec67-4e5d-9c25-f26e39786f79","Type":"ContainerStarted","Data":"a2a4c7e738a8387346bc8cb061fcaffab37035f8b7356aba874420f641555783"} Nov 25 09:47:00 crc kubenswrapper[4776]: I1125 09:47:00.237506 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-lpc2f"] Nov 25 09:47:00 crc kubenswrapper[4776]: I1125 09:47:00.620817 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" Nov 25 09:47:00 crc kubenswrapper[4776]: I1125 09:47:00.682904 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dd7c4987f-4kphp"] Nov 25 09:47:00 crc kubenswrapper[4776]: I1125 09:47:00.683170 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" podUID="d7d021ba-cc92-4ef8-a471-26aa495ea6ab" containerName="dnsmasq-dns" containerID="cri-o://f7079fb542df2e880a7d48696479a3e07c7298972160f08d1c04ea9542b544ab" gracePeriod=10 Nov 25 09:47:01 crc kubenswrapper[4776]: I1125 09:47:01.216927 4776 generic.go:334] "Generic (PLEG): container finished" podID="d7d021ba-cc92-4ef8-a471-26aa495ea6ab" containerID="f7079fb542df2e880a7d48696479a3e07c7298972160f08d1c04ea9542b544ab" exitCode=0 Nov 25 09:47:01 crc kubenswrapper[4776]: I1125 09:47:01.217105 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" event={"ID":"d7d021ba-cc92-4ef8-a471-26aa495ea6ab","Type":"ContainerDied","Data":"f7079fb542df2e880a7d48696479a3e07c7298972160f08d1c04ea9542b544ab"} Nov 25 09:47:01 crc kubenswrapper[4776]: I1125 09:47:01.217394 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" event={"ID":"d7d021ba-cc92-4ef8-a471-26aa495ea6ab","Type":"ContainerDied","Data":"2c91e3bfc1cfec95c2ce46dfe176d7106a508078b2cf1affec2fe790903f8b1a"} Nov 25 09:47:01 crc kubenswrapper[4776]: I1125 09:47:01.217412 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c91e3bfc1cfec95c2ce46dfe176d7106a508078b2cf1affec2fe790903f8b1a" Nov 25 09:47:01 crc kubenswrapper[4776]: I1125 09:47:01.225736 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-lpc2f" event={"ID":"cc2a0874-1575-4605-afb9-01a79c958ba5","Type":"ContainerStarted","Data":"716b38eadb9be8e7ee2b70b2dd192c0196f35d29a4682d29db4b4d8f9899ba52"} Nov 25 09:47:01 crc kubenswrapper[4776]: I1125 09:47:01.225780 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-lpc2f" event={"ID":"cc2a0874-1575-4605-afb9-01a79c958ba5","Type":"ContainerStarted","Data":"6e48671ac5eb3a1cb85a2d6140a32179cd8af5e9bd47e934e03e9e044c3c8b98"} Nov 25 09:47:01 crc kubenswrapper[4776]: I1125 09:47:01.239805 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-lpc2f" podStartSLOduration=2.239784108 podStartE2EDuration="2.239784108s" podCreationTimestamp="2025-11-25 09:46:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:47:01.237396988 +0000 UTC m=+1366.278456541" watchObservedRunningTime="2025-11-25 09:47:01.239784108 +0000 UTC m=+1366.280843671" Nov 25 09:47:01 crc kubenswrapper[4776]: I1125 09:47:01.290851 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" Nov 25 09:47:01 crc kubenswrapper[4776]: I1125 09:47:01.359187 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-ovsdbserver-sb\") pod \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\" (UID: \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\") " Nov 25 09:47:01 crc kubenswrapper[4776]: I1125 09:47:01.359233 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-dns-swift-storage-0\") pod \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\" (UID: \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\") " Nov 25 09:47:01 crc kubenswrapper[4776]: I1125 09:47:01.359380 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-ovsdbserver-nb\") pod \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\" (UID: \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\") " Nov 25 09:47:01 crc kubenswrapper[4776]: I1125 09:47:01.359399 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fqn5\" (UniqueName: \"kubernetes.io/projected/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-kube-api-access-6fqn5\") pod \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\" (UID: \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\") " Nov 25 09:47:01 crc kubenswrapper[4776]: I1125 09:47:01.359426 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-dns-svc\") pod \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\" (UID: \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\") " Nov 25 09:47:01 crc kubenswrapper[4776]: I1125 09:47:01.359459 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-config\") pod \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\" (UID: \"d7d021ba-cc92-4ef8-a471-26aa495ea6ab\") " Nov 25 09:47:01 crc kubenswrapper[4776]: I1125 09:47:01.369999 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-kube-api-access-6fqn5" (OuterVolumeSpecName: "kube-api-access-6fqn5") pod "d7d021ba-cc92-4ef8-a471-26aa495ea6ab" (UID: "d7d021ba-cc92-4ef8-a471-26aa495ea6ab"). InnerVolumeSpecName "kube-api-access-6fqn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:47:01 crc kubenswrapper[4776]: I1125 09:47:01.446928 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d7d021ba-cc92-4ef8-a471-26aa495ea6ab" (UID: "d7d021ba-cc92-4ef8-a471-26aa495ea6ab"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:47:01 crc kubenswrapper[4776]: I1125 09:47:01.460524 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d7d021ba-cc92-4ef8-a471-26aa495ea6ab" (UID: "d7d021ba-cc92-4ef8-a471-26aa495ea6ab"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:47:01 crc kubenswrapper[4776]: I1125 09:47:01.461752 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:01 crc kubenswrapper[4776]: I1125 09:47:01.461771 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fqn5\" (UniqueName: \"kubernetes.io/projected/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-kube-api-access-6fqn5\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:01 crc kubenswrapper[4776]: I1125 09:47:01.461781 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:01 crc kubenswrapper[4776]: I1125 09:47:01.467564 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-config" (OuterVolumeSpecName: "config") pod "d7d021ba-cc92-4ef8-a471-26aa495ea6ab" (UID: "d7d021ba-cc92-4ef8-a471-26aa495ea6ab"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:47:01 crc kubenswrapper[4776]: I1125 09:47:01.471467 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d7d021ba-cc92-4ef8-a471-26aa495ea6ab" (UID: "d7d021ba-cc92-4ef8-a471-26aa495ea6ab"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:47:01 crc kubenswrapper[4776]: I1125 09:47:01.478074 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d7d021ba-cc92-4ef8-a471-26aa495ea6ab" (UID: "d7d021ba-cc92-4ef8-a471-26aa495ea6ab"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:47:01 crc kubenswrapper[4776]: I1125 09:47:01.563310 4776 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:01 crc kubenswrapper[4776]: I1125 09:47:01.563350 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:01 crc kubenswrapper[4776]: I1125 09:47:01.563363 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7d021ba-cc92-4ef8-a471-26aa495ea6ab-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:02 crc kubenswrapper[4776]: I1125 09:47:02.235893 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dd7c4987f-4kphp" Nov 25 09:47:02 crc kubenswrapper[4776]: I1125 09:47:02.236263 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aab7d356-ec67-4e5d-9c25-f26e39786f79","Type":"ContainerStarted","Data":"ec693b92c2b81399b4a2d57c1980de7b988f0cc59e2cef0ec21c34b5dff21bd2"} Nov 25 09:47:02 crc kubenswrapper[4776]: I1125 09:47:02.236711 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 09:47:02 crc kubenswrapper[4776]: I1125 09:47:02.269622 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.187769414 podStartE2EDuration="5.269601041s" podCreationTimestamp="2025-11-25 09:46:57 +0000 UTC" firstStartedPulling="2025-11-25 09:46:58.086173596 +0000 UTC m=+1363.127233149" lastFinishedPulling="2025-11-25 09:47:01.168005213 +0000 UTC m=+1366.209064776" observedRunningTime="2025-11-25 09:47:02.263247592 +0000 UTC m=+1367.304307155" watchObservedRunningTime="2025-11-25 09:47:02.269601041 +0000 UTC m=+1367.310660594" Nov 25 09:47:02 crc kubenswrapper[4776]: I1125 09:47:02.288981 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dd7c4987f-4kphp"] Nov 25 09:47:02 crc kubenswrapper[4776]: I1125 09:47:02.297610 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5dd7c4987f-4kphp"] Nov 25 09:47:03 crc kubenswrapper[4776]: I1125 09:47:03.692031 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7d021ba-cc92-4ef8-a471-26aa495ea6ab" path="/var/lib/kubelet/pods/d7d021ba-cc92-4ef8-a471-26aa495ea6ab/volumes" Nov 25 09:47:06 crc kubenswrapper[4776]: I1125 09:47:06.268613 4776 generic.go:334] "Generic (PLEG): container finished" podID="cc2a0874-1575-4605-afb9-01a79c958ba5" containerID="716b38eadb9be8e7ee2b70b2dd192c0196f35d29a4682d29db4b4d8f9899ba52" exitCode=0 Nov 25 09:47:06 crc kubenswrapper[4776]: I1125 09:47:06.268726 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-lpc2f" event={"ID":"cc2a0874-1575-4605-afb9-01a79c958ba5","Type":"ContainerDied","Data":"716b38eadb9be8e7ee2b70b2dd192c0196f35d29a4682d29db4b4d8f9899ba52"} Nov 25 09:47:07 crc kubenswrapper[4776]: I1125 09:47:07.633448 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-lpc2f" Nov 25 09:47:07 crc kubenswrapper[4776]: I1125 09:47:07.640093 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 09:47:07 crc kubenswrapper[4776]: I1125 09:47:07.640143 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 09:47:07 crc kubenswrapper[4776]: I1125 09:47:07.810793 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc2a0874-1575-4605-afb9-01a79c958ba5-scripts\") pod \"cc2a0874-1575-4605-afb9-01a79c958ba5\" (UID: \"cc2a0874-1575-4605-afb9-01a79c958ba5\") " Nov 25 09:47:07 crc kubenswrapper[4776]: I1125 09:47:07.810863 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc2a0874-1575-4605-afb9-01a79c958ba5-combined-ca-bundle\") pod \"cc2a0874-1575-4605-afb9-01a79c958ba5\" (UID: \"cc2a0874-1575-4605-afb9-01a79c958ba5\") " Nov 25 09:47:07 crc kubenswrapper[4776]: I1125 09:47:07.810961 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc2a0874-1575-4605-afb9-01a79c958ba5-config-data\") pod \"cc2a0874-1575-4605-afb9-01a79c958ba5\" (UID: \"cc2a0874-1575-4605-afb9-01a79c958ba5\") " Nov 25 09:47:07 crc kubenswrapper[4776]: I1125 09:47:07.811090 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-242lk\" (UniqueName: \"kubernetes.io/projected/cc2a0874-1575-4605-afb9-01a79c958ba5-kube-api-access-242lk\") pod \"cc2a0874-1575-4605-afb9-01a79c958ba5\" (UID: \"cc2a0874-1575-4605-afb9-01a79c958ba5\") " Nov 25 09:47:07 crc kubenswrapper[4776]: I1125 09:47:07.816810 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc2a0874-1575-4605-afb9-01a79c958ba5-scripts" (OuterVolumeSpecName: "scripts") pod "cc2a0874-1575-4605-afb9-01a79c958ba5" (UID: "cc2a0874-1575-4605-afb9-01a79c958ba5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:47:07 crc kubenswrapper[4776]: I1125 09:47:07.828743 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc2a0874-1575-4605-afb9-01a79c958ba5-kube-api-access-242lk" (OuterVolumeSpecName: "kube-api-access-242lk") pod "cc2a0874-1575-4605-afb9-01a79c958ba5" (UID: "cc2a0874-1575-4605-afb9-01a79c958ba5"). InnerVolumeSpecName "kube-api-access-242lk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:47:07 crc kubenswrapper[4776]: I1125 09:47:07.839766 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc2a0874-1575-4605-afb9-01a79c958ba5-config-data" (OuterVolumeSpecName: "config-data") pod "cc2a0874-1575-4605-afb9-01a79c958ba5" (UID: "cc2a0874-1575-4605-afb9-01a79c958ba5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:47:07 crc kubenswrapper[4776]: I1125 09:47:07.846935 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc2a0874-1575-4605-afb9-01a79c958ba5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cc2a0874-1575-4605-afb9-01a79c958ba5" (UID: "cc2a0874-1575-4605-afb9-01a79c958ba5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:47:07 crc kubenswrapper[4776]: I1125 09:47:07.914038 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-242lk\" (UniqueName: \"kubernetes.io/projected/cc2a0874-1575-4605-afb9-01a79c958ba5-kube-api-access-242lk\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:07 crc kubenswrapper[4776]: I1125 09:47:07.914080 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc2a0874-1575-4605-afb9-01a79c958ba5-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:07 crc kubenswrapper[4776]: I1125 09:47:07.914092 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc2a0874-1575-4605-afb9-01a79c958ba5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:07 crc kubenswrapper[4776]: I1125 09:47:07.914101 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc2a0874-1575-4605-afb9-01a79c958ba5-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:08 crc kubenswrapper[4776]: I1125 09:47:08.289224 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-lpc2f" event={"ID":"cc2a0874-1575-4605-afb9-01a79c958ba5","Type":"ContainerDied","Data":"6e48671ac5eb3a1cb85a2d6140a32179cd8af5e9bd47e934e03e9e044c3c8b98"} Nov 25 09:47:08 crc kubenswrapper[4776]: I1125 09:47:08.289276 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e48671ac5eb3a1cb85a2d6140a32179cd8af5e9bd47e934e03e9e044c3c8b98" Nov 25 09:47:08 crc kubenswrapper[4776]: I1125 09:47:08.289364 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-lpc2f" Nov 25 09:47:08 crc kubenswrapper[4776]: I1125 09:47:08.482981 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 09:47:08 crc kubenswrapper[4776]: I1125 09:47:08.483290 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="66011de8-4f1e-4c9d-9bfb-5d0fb37daa14" containerName="nova-scheduler-scheduler" containerID="cri-o://ab260ec7794ff3d66ff3c728ab9aab93670ce5d3ac37c667f843d07124606553" gracePeriod=30 Nov 25 09:47:08 crc kubenswrapper[4776]: I1125 09:47:08.500552 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 09:47:08 crc kubenswrapper[4776]: I1125 09:47:08.500828 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2b90b0bf-a4b6-4336-9e49-33792ead4f07" containerName="nova-api-log" containerID="cri-o://5a25e29978f0284d518a88b255bee88fa125d11b08d1eaeb950230e38ea74769" gracePeriod=30 Nov 25 09:47:08 crc kubenswrapper[4776]: I1125 09:47:08.500964 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2b90b0bf-a4b6-4336-9e49-33792ead4f07" containerName="nova-api-api" containerID="cri-o://79cbebf3e9c8cb7e6ea34200c99cf5ad70ddabde13fac5d58d25f07b09c5e055" gracePeriod=30 Nov 25 09:47:08 crc kubenswrapper[4776]: I1125 09:47:08.506599 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2b90b0bf-a4b6-4336-9e49-33792ead4f07" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": EOF" Nov 25 09:47:08 crc kubenswrapper[4776]: I1125 09:47:08.506599 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2b90b0bf-a4b6-4336-9e49-33792ead4f07" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.201:8774/\": EOF" Nov 25 09:47:08 crc kubenswrapper[4776]: I1125 09:47:08.519254 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 09:47:08 crc kubenswrapper[4776]: I1125 09:47:08.519532 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3a35a929-dab0-4c6c-ac0f-2199cb71f4d0" containerName="nova-metadata-log" containerID="cri-o://0b13b46852bb397571c1533f5e6b6a73573a684637b6d0c2af1f64451e8b0640" gracePeriod=30 Nov 25 09:47:08 crc kubenswrapper[4776]: I1125 09:47:08.519930 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3a35a929-dab0-4c6c-ac0f-2199cb71f4d0" containerName="nova-metadata-metadata" containerID="cri-o://8746f8f433ab92b6be995d01ccc400704b41df632f434effb9df0790d7335ffb" gracePeriod=30 Nov 25 09:47:09 crc kubenswrapper[4776]: I1125 09:47:09.299314 4776 generic.go:334] "Generic (PLEG): container finished" podID="2b90b0bf-a4b6-4336-9e49-33792ead4f07" containerID="5a25e29978f0284d518a88b255bee88fa125d11b08d1eaeb950230e38ea74769" exitCode=143 Nov 25 09:47:09 crc kubenswrapper[4776]: I1125 09:47:09.299528 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2b90b0bf-a4b6-4336-9e49-33792ead4f07","Type":"ContainerDied","Data":"5a25e29978f0284d518a88b255bee88fa125d11b08d1eaeb950230e38ea74769"} Nov 25 09:47:09 crc kubenswrapper[4776]: I1125 09:47:09.301097 4776 generic.go:334] "Generic (PLEG): container finished" podID="66011de8-4f1e-4c9d-9bfb-5d0fb37daa14" containerID="ab260ec7794ff3d66ff3c728ab9aab93670ce5d3ac37c667f843d07124606553" exitCode=0 Nov 25 09:47:09 crc kubenswrapper[4776]: I1125 09:47:09.301143 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"66011de8-4f1e-4c9d-9bfb-5d0fb37daa14","Type":"ContainerDied","Data":"ab260ec7794ff3d66ff3c728ab9aab93670ce5d3ac37c667f843d07124606553"} Nov 25 09:47:09 crc kubenswrapper[4776]: I1125 09:47:09.303531 4776 generic.go:334] "Generic (PLEG): container finished" podID="3a35a929-dab0-4c6c-ac0f-2199cb71f4d0" containerID="0b13b46852bb397571c1533f5e6b6a73573a684637b6d0c2af1f64451e8b0640" exitCode=143 Nov 25 09:47:09 crc kubenswrapper[4776]: I1125 09:47:09.303560 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0","Type":"ContainerDied","Data":"0b13b46852bb397571c1533f5e6b6a73573a684637b6d0c2af1f64451e8b0640"} Nov 25 09:47:09 crc kubenswrapper[4776]: I1125 09:47:09.695214 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 09:47:09 crc kubenswrapper[4776]: I1125 09:47:09.764298 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ck9c2\" (UniqueName: \"kubernetes.io/projected/66011de8-4f1e-4c9d-9bfb-5d0fb37daa14-kube-api-access-ck9c2\") pod \"66011de8-4f1e-4c9d-9bfb-5d0fb37daa14\" (UID: \"66011de8-4f1e-4c9d-9bfb-5d0fb37daa14\") " Nov 25 09:47:09 crc kubenswrapper[4776]: I1125 09:47:09.764427 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66011de8-4f1e-4c9d-9bfb-5d0fb37daa14-combined-ca-bundle\") pod \"66011de8-4f1e-4c9d-9bfb-5d0fb37daa14\" (UID: \"66011de8-4f1e-4c9d-9bfb-5d0fb37daa14\") " Nov 25 09:47:09 crc kubenswrapper[4776]: I1125 09:47:09.764478 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66011de8-4f1e-4c9d-9bfb-5d0fb37daa14-config-data\") pod \"66011de8-4f1e-4c9d-9bfb-5d0fb37daa14\" (UID: \"66011de8-4f1e-4c9d-9bfb-5d0fb37daa14\") " Nov 25 09:47:09 crc kubenswrapper[4776]: I1125 09:47:09.772189 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66011de8-4f1e-4c9d-9bfb-5d0fb37daa14-kube-api-access-ck9c2" (OuterVolumeSpecName: "kube-api-access-ck9c2") pod "66011de8-4f1e-4c9d-9bfb-5d0fb37daa14" (UID: "66011de8-4f1e-4c9d-9bfb-5d0fb37daa14"). InnerVolumeSpecName "kube-api-access-ck9c2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:47:09 crc kubenswrapper[4776]: I1125 09:47:09.797149 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66011de8-4f1e-4c9d-9bfb-5d0fb37daa14-config-data" (OuterVolumeSpecName: "config-data") pod "66011de8-4f1e-4c9d-9bfb-5d0fb37daa14" (UID: "66011de8-4f1e-4c9d-9bfb-5d0fb37daa14"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:47:09 crc kubenswrapper[4776]: I1125 09:47:09.815320 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66011de8-4f1e-4c9d-9bfb-5d0fb37daa14-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "66011de8-4f1e-4c9d-9bfb-5d0fb37daa14" (UID: "66011de8-4f1e-4c9d-9bfb-5d0fb37daa14"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:47:09 crc kubenswrapper[4776]: I1125 09:47:09.866722 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ck9c2\" (UniqueName: \"kubernetes.io/projected/66011de8-4f1e-4c9d-9bfb-5d0fb37daa14-kube-api-access-ck9c2\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:09 crc kubenswrapper[4776]: I1125 09:47:09.866764 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66011de8-4f1e-4c9d-9bfb-5d0fb37daa14-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:09 crc kubenswrapper[4776]: I1125 09:47:09.866773 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66011de8-4f1e-4c9d-9bfb-5d0fb37daa14-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:10 crc kubenswrapper[4776]: I1125 09:47:10.312434 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"66011de8-4f1e-4c9d-9bfb-5d0fb37daa14","Type":"ContainerDied","Data":"59461b1548a1266871b57d75ed7ae85873ac8ace14a6e0c6084bfd58575a8c0f"} Nov 25 09:47:10 crc kubenswrapper[4776]: I1125 09:47:10.312488 4776 scope.go:117] "RemoveContainer" containerID="ab260ec7794ff3d66ff3c728ab9aab93670ce5d3ac37c667f843d07124606553" Nov 25 09:47:10 crc kubenswrapper[4776]: I1125 09:47:10.312508 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 09:47:10 crc kubenswrapper[4776]: I1125 09:47:10.348153 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 09:47:10 crc kubenswrapper[4776]: I1125 09:47:10.361428 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 09:47:10 crc kubenswrapper[4776]: I1125 09:47:10.371494 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 09:47:10 crc kubenswrapper[4776]: E1125 09:47:10.371986 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66011de8-4f1e-4c9d-9bfb-5d0fb37daa14" containerName="nova-scheduler-scheduler" Nov 25 09:47:10 crc kubenswrapper[4776]: I1125 09:47:10.372006 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="66011de8-4f1e-4c9d-9bfb-5d0fb37daa14" containerName="nova-scheduler-scheduler" Nov 25 09:47:10 crc kubenswrapper[4776]: E1125 09:47:10.372042 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc2a0874-1575-4605-afb9-01a79c958ba5" containerName="nova-manage" Nov 25 09:47:10 crc kubenswrapper[4776]: I1125 09:47:10.372050 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc2a0874-1575-4605-afb9-01a79c958ba5" containerName="nova-manage" Nov 25 09:47:10 crc kubenswrapper[4776]: E1125 09:47:10.372098 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7d021ba-cc92-4ef8-a471-26aa495ea6ab" containerName="dnsmasq-dns" Nov 25 09:47:10 crc kubenswrapper[4776]: I1125 09:47:10.372107 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7d021ba-cc92-4ef8-a471-26aa495ea6ab" containerName="dnsmasq-dns" Nov 25 09:47:10 crc kubenswrapper[4776]: E1125 09:47:10.372143 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7d021ba-cc92-4ef8-a471-26aa495ea6ab" containerName="init" Nov 25 09:47:10 crc kubenswrapper[4776]: I1125 09:47:10.372153 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7d021ba-cc92-4ef8-a471-26aa495ea6ab" containerName="init" Nov 25 09:47:10 crc kubenswrapper[4776]: I1125 09:47:10.372383 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7d021ba-cc92-4ef8-a471-26aa495ea6ab" containerName="dnsmasq-dns" Nov 25 09:47:10 crc kubenswrapper[4776]: I1125 09:47:10.372404 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc2a0874-1575-4605-afb9-01a79c958ba5" containerName="nova-manage" Nov 25 09:47:10 crc kubenswrapper[4776]: I1125 09:47:10.372421 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="66011de8-4f1e-4c9d-9bfb-5d0fb37daa14" containerName="nova-scheduler-scheduler" Nov 25 09:47:10 crc kubenswrapper[4776]: I1125 09:47:10.373215 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 09:47:10 crc kubenswrapper[4776]: I1125 09:47:10.375396 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 25 09:47:10 crc kubenswrapper[4776]: I1125 09:47:10.394728 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 09:47:10 crc kubenswrapper[4776]: I1125 09:47:10.477283 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bae8251-3d5d-490b-9116-73769dc18101-config-data\") pod \"nova-scheduler-0\" (UID: \"8bae8251-3d5d-490b-9116-73769dc18101\") " pod="openstack/nova-scheduler-0" Nov 25 09:47:10 crc kubenswrapper[4776]: I1125 09:47:10.477495 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vr9x\" (UniqueName: \"kubernetes.io/projected/8bae8251-3d5d-490b-9116-73769dc18101-kube-api-access-6vr9x\") pod \"nova-scheduler-0\" (UID: \"8bae8251-3d5d-490b-9116-73769dc18101\") " pod="openstack/nova-scheduler-0" Nov 25 09:47:10 crc kubenswrapper[4776]: I1125 09:47:10.477837 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bae8251-3d5d-490b-9116-73769dc18101-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8bae8251-3d5d-490b-9116-73769dc18101\") " pod="openstack/nova-scheduler-0" Nov 25 09:47:10 crc kubenswrapper[4776]: I1125 09:47:10.579906 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vr9x\" (UniqueName: \"kubernetes.io/projected/8bae8251-3d5d-490b-9116-73769dc18101-kube-api-access-6vr9x\") pod \"nova-scheduler-0\" (UID: \"8bae8251-3d5d-490b-9116-73769dc18101\") " pod="openstack/nova-scheduler-0" Nov 25 09:47:10 crc kubenswrapper[4776]: I1125 09:47:10.580028 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bae8251-3d5d-490b-9116-73769dc18101-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8bae8251-3d5d-490b-9116-73769dc18101\") " pod="openstack/nova-scheduler-0" Nov 25 09:47:10 crc kubenswrapper[4776]: I1125 09:47:10.580108 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bae8251-3d5d-490b-9116-73769dc18101-config-data\") pod \"nova-scheduler-0\" (UID: \"8bae8251-3d5d-490b-9116-73769dc18101\") " pod="openstack/nova-scheduler-0" Nov 25 09:47:10 crc kubenswrapper[4776]: I1125 09:47:10.589149 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bae8251-3d5d-490b-9116-73769dc18101-config-data\") pod \"nova-scheduler-0\" (UID: \"8bae8251-3d5d-490b-9116-73769dc18101\") " pod="openstack/nova-scheduler-0" Nov 25 09:47:10 crc kubenswrapper[4776]: I1125 09:47:10.603691 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bae8251-3d5d-490b-9116-73769dc18101-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8bae8251-3d5d-490b-9116-73769dc18101\") " pod="openstack/nova-scheduler-0" Nov 25 09:47:10 crc kubenswrapper[4776]: I1125 09:47:10.622689 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vr9x\" (UniqueName: \"kubernetes.io/projected/8bae8251-3d5d-490b-9116-73769dc18101-kube-api-access-6vr9x\") pod \"nova-scheduler-0\" (UID: \"8bae8251-3d5d-490b-9116-73769dc18101\") " pod="openstack/nova-scheduler-0" Nov 25 09:47:10 crc kubenswrapper[4776]: I1125 09:47:10.692569 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 09:47:11 crc kubenswrapper[4776]: I1125 09:47:11.135159 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 09:47:11 crc kubenswrapper[4776]: I1125 09:47:11.325135 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8bae8251-3d5d-490b-9116-73769dc18101","Type":"ContainerStarted","Data":"257f4e7f047aaf2041def20b6caaee331d83d82f30d019096ad21bfd241e42a8"} Nov 25 09:47:11 crc kubenswrapper[4776]: I1125 09:47:11.325451 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8bae8251-3d5d-490b-9116-73769dc18101","Type":"ContainerStarted","Data":"8ab35a66b3792c7af7863e852b1699430b7994109c81a79b1dd2c13a6886c7cb"} Nov 25 09:47:11 crc kubenswrapper[4776]: I1125 09:47:11.346922 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.3469053 podStartE2EDuration="1.3469053s" podCreationTimestamp="2025-11-25 09:47:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:47:11.342678293 +0000 UTC m=+1376.383737846" watchObservedRunningTime="2025-11-25 09:47:11.3469053 +0000 UTC m=+1376.387964843" Nov 25 09:47:11 crc kubenswrapper[4776]: I1125 09:47:11.675419 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66011de8-4f1e-4c9d-9bfb-5d0fb37daa14" path="/var/lib/kubelet/pods/66011de8-4f1e-4c9d-9bfb-5d0fb37daa14/volumes" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.110396 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.207115 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-logs\") pod \"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0\" (UID: \"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0\") " Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.207218 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-combined-ca-bundle\") pod \"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0\" (UID: \"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0\") " Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.207368 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rb2tv\" (UniqueName: \"kubernetes.io/projected/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-kube-api-access-rb2tv\") pod \"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0\" (UID: \"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0\") " Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.207404 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-nova-metadata-tls-certs\") pod \"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0\" (UID: \"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0\") " Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.207473 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-config-data\") pod \"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0\" (UID: \"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0\") " Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.207745 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-logs" (OuterVolumeSpecName: "logs") pod "3a35a929-dab0-4c6c-ac0f-2199cb71f4d0" (UID: "3a35a929-dab0-4c6c-ac0f-2199cb71f4d0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.208025 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-logs\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.225449 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-kube-api-access-rb2tv" (OuterVolumeSpecName: "kube-api-access-rb2tv") pod "3a35a929-dab0-4c6c-ac0f-2199cb71f4d0" (UID: "3a35a929-dab0-4c6c-ac0f-2199cb71f4d0"). InnerVolumeSpecName "kube-api-access-rb2tv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.241026 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-config-data" (OuterVolumeSpecName: "config-data") pod "3a35a929-dab0-4c6c-ac0f-2199cb71f4d0" (UID: "3a35a929-dab0-4c6c-ac0f-2199cb71f4d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.249261 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a35a929-dab0-4c6c-ac0f-2199cb71f4d0" (UID: "3a35a929-dab0-4c6c-ac0f-2199cb71f4d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.270965 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "3a35a929-dab0-4c6c-ac0f-2199cb71f4d0" (UID: "3a35a929-dab0-4c6c-ac0f-2199cb71f4d0"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.309474 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rb2tv\" (UniqueName: \"kubernetes.io/projected/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-kube-api-access-rb2tv\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.309511 4776 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.309522 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.309532 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.336516 4776 generic.go:334] "Generic (PLEG): container finished" podID="3a35a929-dab0-4c6c-ac0f-2199cb71f4d0" containerID="8746f8f433ab92b6be995d01ccc400704b41df632f434effb9df0790d7335ffb" exitCode=0 Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.336591 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.336639 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0","Type":"ContainerDied","Data":"8746f8f433ab92b6be995d01ccc400704b41df632f434effb9df0790d7335ffb"} Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.336678 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3a35a929-dab0-4c6c-ac0f-2199cb71f4d0","Type":"ContainerDied","Data":"3d239108af315c66806c7a34b54f33898ff8d6979b8f77e45e3c4a31afd64786"} Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.336700 4776 scope.go:117] "RemoveContainer" containerID="8746f8f433ab92b6be995d01ccc400704b41df632f434effb9df0790d7335ffb" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.372732 4776 scope.go:117] "RemoveContainer" containerID="0b13b46852bb397571c1533f5e6b6a73573a684637b6d0c2af1f64451e8b0640" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.373570 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.382218 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.396269 4776 scope.go:117] "RemoveContainer" containerID="8746f8f433ab92b6be995d01ccc400704b41df632f434effb9df0790d7335ffb" Nov 25 09:47:12 crc kubenswrapper[4776]: E1125 09:47:12.397359 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8746f8f433ab92b6be995d01ccc400704b41df632f434effb9df0790d7335ffb\": container with ID starting with 8746f8f433ab92b6be995d01ccc400704b41df632f434effb9df0790d7335ffb not found: ID does not exist" containerID="8746f8f433ab92b6be995d01ccc400704b41df632f434effb9df0790d7335ffb" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.397395 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8746f8f433ab92b6be995d01ccc400704b41df632f434effb9df0790d7335ffb"} err="failed to get container status \"8746f8f433ab92b6be995d01ccc400704b41df632f434effb9df0790d7335ffb\": rpc error: code = NotFound desc = could not find container \"8746f8f433ab92b6be995d01ccc400704b41df632f434effb9df0790d7335ffb\": container with ID starting with 8746f8f433ab92b6be995d01ccc400704b41df632f434effb9df0790d7335ffb not found: ID does not exist" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.397419 4776 scope.go:117] "RemoveContainer" containerID="0b13b46852bb397571c1533f5e6b6a73573a684637b6d0c2af1f64451e8b0640" Nov 25 09:47:12 crc kubenswrapper[4776]: E1125 09:47:12.397859 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b13b46852bb397571c1533f5e6b6a73573a684637b6d0c2af1f64451e8b0640\": container with ID starting with 0b13b46852bb397571c1533f5e6b6a73573a684637b6d0c2af1f64451e8b0640 not found: ID does not exist" containerID="0b13b46852bb397571c1533f5e6b6a73573a684637b6d0c2af1f64451e8b0640" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.397912 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b13b46852bb397571c1533f5e6b6a73573a684637b6d0c2af1f64451e8b0640"} err="failed to get container status \"0b13b46852bb397571c1533f5e6b6a73573a684637b6d0c2af1f64451e8b0640\": rpc error: code = NotFound desc = could not find container \"0b13b46852bb397571c1533f5e6b6a73573a684637b6d0c2af1f64451e8b0640\": container with ID starting with 0b13b46852bb397571c1533f5e6b6a73573a684637b6d0c2af1f64451e8b0640 not found: ID does not exist" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.409751 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 25 09:47:12 crc kubenswrapper[4776]: E1125 09:47:12.410144 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a35a929-dab0-4c6c-ac0f-2199cb71f4d0" containerName="nova-metadata-log" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.410159 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a35a929-dab0-4c6c-ac0f-2199cb71f4d0" containerName="nova-metadata-log" Nov 25 09:47:12 crc kubenswrapper[4776]: E1125 09:47:12.410185 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a35a929-dab0-4c6c-ac0f-2199cb71f4d0" containerName="nova-metadata-metadata" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.410191 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a35a929-dab0-4c6c-ac0f-2199cb71f4d0" containerName="nova-metadata-metadata" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.410364 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a35a929-dab0-4c6c-ac0f-2199cb71f4d0" containerName="nova-metadata-log" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.410385 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a35a929-dab0-4c6c-ac0f-2199cb71f4d0" containerName="nova-metadata-metadata" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.411345 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.413773 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.415982 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.417845 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.512029 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-config-data\") pod \"nova-metadata-0\" (UID: \"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91\") " pod="openstack/nova-metadata-0" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.512095 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ftrk\" (UniqueName: \"kubernetes.io/projected/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-kube-api-access-9ftrk\") pod \"nova-metadata-0\" (UID: \"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91\") " pod="openstack/nova-metadata-0" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.512134 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-logs\") pod \"nova-metadata-0\" (UID: \"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91\") " pod="openstack/nova-metadata-0" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.512195 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91\") " pod="openstack/nova-metadata-0" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.512240 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91\") " pod="openstack/nova-metadata-0" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.613486 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-logs\") pod \"nova-metadata-0\" (UID: \"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91\") " pod="openstack/nova-metadata-0" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.613590 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91\") " pod="openstack/nova-metadata-0" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.613669 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91\") " pod="openstack/nova-metadata-0" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.613847 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-config-data\") pod \"nova-metadata-0\" (UID: \"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91\") " pod="openstack/nova-metadata-0" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.613892 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ftrk\" (UniqueName: \"kubernetes.io/projected/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-kube-api-access-9ftrk\") pod \"nova-metadata-0\" (UID: \"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91\") " pod="openstack/nova-metadata-0" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.613904 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-logs\") pod \"nova-metadata-0\" (UID: \"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91\") " pod="openstack/nova-metadata-0" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.617133 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91\") " pod="openstack/nova-metadata-0" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.617264 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91\") " pod="openstack/nova-metadata-0" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.617557 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-config-data\") pod \"nova-metadata-0\" (UID: \"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91\") " pod="openstack/nova-metadata-0" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.630841 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ftrk\" (UniqueName: \"kubernetes.io/projected/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-kube-api-access-9ftrk\") pod \"nova-metadata-0\" (UID: \"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91\") " pod="openstack/nova-metadata-0" Nov 25 09:47:12 crc kubenswrapper[4776]: I1125 09:47:12.731379 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 09:47:13 crc kubenswrapper[4776]: I1125 09:47:13.167588 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 09:47:13 crc kubenswrapper[4776]: W1125 09:47:13.179636 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a0ab5a5_2d62_4457_be9d_fe8242ab7c91.slice/crio-fe6b9828c965a764d125db4efe76edee643df6bf78aabbb24e0fbbd1bfeabd1f WatchSource:0}: Error finding container fe6b9828c965a764d125db4efe76edee643df6bf78aabbb24e0fbbd1bfeabd1f: Status 404 returned error can't find the container with id fe6b9828c965a764d125db4efe76edee643df6bf78aabbb24e0fbbd1bfeabd1f Nov 25 09:47:13 crc kubenswrapper[4776]: I1125 09:47:13.346693 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91","Type":"ContainerStarted","Data":"fe6b9828c965a764d125db4efe76edee643df6bf78aabbb24e0fbbd1bfeabd1f"} Nov 25 09:47:13 crc kubenswrapper[4776]: I1125 09:47:13.673769 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a35a929-dab0-4c6c-ac0f-2199cb71f4d0" path="/var/lib/kubelet/pods/3a35a929-dab0-4c6c-ac0f-2199cb71f4d0/volumes" Nov 25 09:47:14 crc kubenswrapper[4776]: I1125 09:47:14.357435 4776 generic.go:334] "Generic (PLEG): container finished" podID="2b90b0bf-a4b6-4336-9e49-33792ead4f07" containerID="79cbebf3e9c8cb7e6ea34200c99cf5ad70ddabde13fac5d58d25f07b09c5e055" exitCode=0 Nov 25 09:47:14 crc kubenswrapper[4776]: I1125 09:47:14.357506 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2b90b0bf-a4b6-4336-9e49-33792ead4f07","Type":"ContainerDied","Data":"79cbebf3e9c8cb7e6ea34200c99cf5ad70ddabde13fac5d58d25f07b09c5e055"} Nov 25 09:47:14 crc kubenswrapper[4776]: I1125 09:47:14.358809 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2b90b0bf-a4b6-4336-9e49-33792ead4f07","Type":"ContainerDied","Data":"f8d31df8b7a3a4d8d3f2899d59443769fe63e3a693b75de4adf3f81dbbfdc256"} Nov 25 09:47:14 crc kubenswrapper[4776]: I1125 09:47:14.358875 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8d31df8b7a3a4d8d3f2899d59443769fe63e3a693b75de4adf3f81dbbfdc256" Nov 25 09:47:14 crc kubenswrapper[4776]: I1125 09:47:14.361459 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91","Type":"ContainerStarted","Data":"7927ef3d5f7eb88faf6e7c414da603268c27589afcdfe2331cf08fabea8c8c3c"} Nov 25 09:47:14 crc kubenswrapper[4776]: I1125 09:47:14.361501 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91","Type":"ContainerStarted","Data":"6388f255053948e8f439f0e43bfe9480c34d62fbb564522cf02851e2a2aef43a"} Nov 25 09:47:14 crc kubenswrapper[4776]: I1125 09:47:14.383428 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.383404439 podStartE2EDuration="2.383404439s" podCreationTimestamp="2025-11-25 09:47:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:47:14.378195998 +0000 UTC m=+1379.419255551" watchObservedRunningTime="2025-11-25 09:47:14.383404439 +0000 UTC m=+1379.424463982" Nov 25 09:47:14 crc kubenswrapper[4776]: I1125 09:47:14.393676 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 09:47:14 crc kubenswrapper[4776]: I1125 09:47:14.552493 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b90b0bf-a4b6-4336-9e49-33792ead4f07-public-tls-certs\") pod \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\" (UID: \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\") " Nov 25 09:47:14 crc kubenswrapper[4776]: I1125 09:47:14.552737 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b90b0bf-a4b6-4336-9e49-33792ead4f07-logs\") pod \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\" (UID: \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\") " Nov 25 09:47:14 crc kubenswrapper[4776]: I1125 09:47:14.552800 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b90b0bf-a4b6-4336-9e49-33792ead4f07-config-data\") pod \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\" (UID: \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\") " Nov 25 09:47:14 crc kubenswrapper[4776]: I1125 09:47:14.552851 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n56sr\" (UniqueName: \"kubernetes.io/projected/2b90b0bf-a4b6-4336-9e49-33792ead4f07-kube-api-access-n56sr\") pod \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\" (UID: \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\") " Nov 25 09:47:14 crc kubenswrapper[4776]: I1125 09:47:14.552885 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b90b0bf-a4b6-4336-9e49-33792ead4f07-internal-tls-certs\") pod \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\" (UID: \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\") " Nov 25 09:47:14 crc kubenswrapper[4776]: I1125 09:47:14.552953 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b90b0bf-a4b6-4336-9e49-33792ead4f07-combined-ca-bundle\") pod \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\" (UID: \"2b90b0bf-a4b6-4336-9e49-33792ead4f07\") " Nov 25 09:47:14 crc kubenswrapper[4776]: I1125 09:47:14.553736 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b90b0bf-a4b6-4336-9e49-33792ead4f07-logs" (OuterVolumeSpecName: "logs") pod "2b90b0bf-a4b6-4336-9e49-33792ead4f07" (UID: "2b90b0bf-a4b6-4336-9e49-33792ead4f07"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:47:14 crc kubenswrapper[4776]: I1125 09:47:14.555726 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b90b0bf-a4b6-4336-9e49-33792ead4f07-logs\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:14 crc kubenswrapper[4776]: I1125 09:47:14.576264 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b90b0bf-a4b6-4336-9e49-33792ead4f07-kube-api-access-n56sr" (OuterVolumeSpecName: "kube-api-access-n56sr") pod "2b90b0bf-a4b6-4336-9e49-33792ead4f07" (UID: "2b90b0bf-a4b6-4336-9e49-33792ead4f07"). InnerVolumeSpecName "kube-api-access-n56sr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:47:14 crc kubenswrapper[4776]: I1125 09:47:14.584849 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b90b0bf-a4b6-4336-9e49-33792ead4f07-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2b90b0bf-a4b6-4336-9e49-33792ead4f07" (UID: "2b90b0bf-a4b6-4336-9e49-33792ead4f07"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:47:14 crc kubenswrapper[4776]: I1125 09:47:14.596757 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b90b0bf-a4b6-4336-9e49-33792ead4f07-config-data" (OuterVolumeSpecName: "config-data") pod "2b90b0bf-a4b6-4336-9e49-33792ead4f07" (UID: "2b90b0bf-a4b6-4336-9e49-33792ead4f07"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:47:14 crc kubenswrapper[4776]: I1125 09:47:14.612612 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b90b0bf-a4b6-4336-9e49-33792ead4f07-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2b90b0bf-a4b6-4336-9e49-33792ead4f07" (UID: "2b90b0bf-a4b6-4336-9e49-33792ead4f07"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:47:14 crc kubenswrapper[4776]: I1125 09:47:14.631577 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b90b0bf-a4b6-4336-9e49-33792ead4f07-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2b90b0bf-a4b6-4336-9e49-33792ead4f07" (UID: "2b90b0bf-a4b6-4336-9e49-33792ead4f07"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:47:14 crc kubenswrapper[4776]: I1125 09:47:14.656973 4776 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b90b0bf-a4b6-4336-9e49-33792ead4f07-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:14 crc kubenswrapper[4776]: I1125 09:47:14.657000 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b90b0bf-a4b6-4336-9e49-33792ead4f07-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:14 crc kubenswrapper[4776]: I1125 09:47:14.657010 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n56sr\" (UniqueName: \"kubernetes.io/projected/2b90b0bf-a4b6-4336-9e49-33792ead4f07-kube-api-access-n56sr\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:14 crc kubenswrapper[4776]: I1125 09:47:14.657021 4776 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b90b0bf-a4b6-4336-9e49-33792ead4f07-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:14 crc kubenswrapper[4776]: I1125 09:47:14.657029 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b90b0bf-a4b6-4336-9e49-33792ead4f07-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.371484 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.410442 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.416225 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.451539 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 09:47:15 crc kubenswrapper[4776]: E1125 09:47:15.452271 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b90b0bf-a4b6-4336-9e49-33792ead4f07" containerName="nova-api-api" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.452300 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b90b0bf-a4b6-4336-9e49-33792ead4f07" containerName="nova-api-api" Nov 25 09:47:15 crc kubenswrapper[4776]: E1125 09:47:15.452318 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b90b0bf-a4b6-4336-9e49-33792ead4f07" containerName="nova-api-log" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.452328 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b90b0bf-a4b6-4336-9e49-33792ead4f07" containerName="nova-api-log" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.452670 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b90b0bf-a4b6-4336-9e49-33792ead4f07" containerName="nova-api-log" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.452700 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b90b0bf-a4b6-4336-9e49-33792ead4f07" containerName="nova-api-api" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.454298 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.457608 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.458143 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.458377 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.476722 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffdd6\" (UniqueName: \"kubernetes.io/projected/6bf3edad-487a-4c68-9dbf-9789a94f8fba-kube-api-access-ffdd6\") pod \"nova-api-0\" (UID: \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\") " pod="openstack/nova-api-0" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.476764 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bf3edad-487a-4c68-9dbf-9789a94f8fba-public-tls-certs\") pod \"nova-api-0\" (UID: \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\") " pod="openstack/nova-api-0" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.476788 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bf3edad-487a-4c68-9dbf-9789a94f8fba-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\") " pod="openstack/nova-api-0" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.476838 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bf3edad-487a-4c68-9dbf-9789a94f8fba-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\") " pod="openstack/nova-api-0" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.476867 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6bf3edad-487a-4c68-9dbf-9789a94f8fba-logs\") pod \"nova-api-0\" (UID: \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\") " pod="openstack/nova-api-0" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.476959 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bf3edad-487a-4c68-9dbf-9789a94f8fba-config-data\") pod \"nova-api-0\" (UID: \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\") " pod="openstack/nova-api-0" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.481107 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.578628 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bf3edad-487a-4c68-9dbf-9789a94f8fba-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\") " pod="openstack/nova-api-0" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.578671 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6bf3edad-487a-4c68-9dbf-9789a94f8fba-logs\") pod \"nova-api-0\" (UID: \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\") " pod="openstack/nova-api-0" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.578772 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bf3edad-487a-4c68-9dbf-9789a94f8fba-config-data\") pod \"nova-api-0\" (UID: \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\") " pod="openstack/nova-api-0" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.578838 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffdd6\" (UniqueName: \"kubernetes.io/projected/6bf3edad-487a-4c68-9dbf-9789a94f8fba-kube-api-access-ffdd6\") pod \"nova-api-0\" (UID: \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\") " pod="openstack/nova-api-0" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.578861 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bf3edad-487a-4c68-9dbf-9789a94f8fba-public-tls-certs\") pod \"nova-api-0\" (UID: \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\") " pod="openstack/nova-api-0" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.578890 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bf3edad-487a-4c68-9dbf-9789a94f8fba-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\") " pod="openstack/nova-api-0" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.579627 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6bf3edad-487a-4c68-9dbf-9789a94f8fba-logs\") pod \"nova-api-0\" (UID: \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\") " pod="openstack/nova-api-0" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.584472 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bf3edad-487a-4c68-9dbf-9789a94f8fba-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\") " pod="openstack/nova-api-0" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.584908 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bf3edad-487a-4c68-9dbf-9789a94f8fba-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\") " pod="openstack/nova-api-0" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.586016 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bf3edad-487a-4c68-9dbf-9789a94f8fba-config-data\") pod \"nova-api-0\" (UID: \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\") " pod="openstack/nova-api-0" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.588852 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bf3edad-487a-4c68-9dbf-9789a94f8fba-public-tls-certs\") pod \"nova-api-0\" (UID: \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\") " pod="openstack/nova-api-0" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.596544 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffdd6\" (UniqueName: \"kubernetes.io/projected/6bf3edad-487a-4c68-9dbf-9789a94f8fba-kube-api-access-ffdd6\") pod \"nova-api-0\" (UID: \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\") " pod="openstack/nova-api-0" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.684038 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b90b0bf-a4b6-4336-9e49-33792ead4f07" path="/var/lib/kubelet/pods/2b90b0bf-a4b6-4336-9e49-33792ead4f07/volumes" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.693470 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 25 09:47:15 crc kubenswrapper[4776]: I1125 09:47:15.787112 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 09:47:16 crc kubenswrapper[4776]: I1125 09:47:16.291222 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 09:47:16 crc kubenswrapper[4776]: I1125 09:47:16.381672 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6bf3edad-487a-4c68-9dbf-9789a94f8fba","Type":"ContainerStarted","Data":"36817d6beb7f295e36cecc52ef61815f589eda60f500320fe51a837b338cf6db"} Nov 25 09:47:17 crc kubenswrapper[4776]: I1125 09:47:17.391246 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6bf3edad-487a-4c68-9dbf-9789a94f8fba","Type":"ContainerStarted","Data":"24a3083e873c7a16456c526fdfe2bc77dcf34c30f066c1f1af77953f2e1ecddd"} Nov 25 09:47:17 crc kubenswrapper[4776]: I1125 09:47:17.391570 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6bf3edad-487a-4c68-9dbf-9789a94f8fba","Type":"ContainerStarted","Data":"07109bfcbb050b4c8db424a410626fa31b2236c8d915084ff3b34d220f1c1cf0"} Nov 25 09:47:17 crc kubenswrapper[4776]: I1125 09:47:17.412872 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.412853441 podStartE2EDuration="2.412853441s" podCreationTimestamp="2025-11-25 09:47:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:47:17.409715042 +0000 UTC m=+1382.450774595" watchObservedRunningTime="2025-11-25 09:47:17.412853441 +0000 UTC m=+1382.453912994" Nov 25 09:47:17 crc kubenswrapper[4776]: I1125 09:47:17.733092 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 09:47:17 crc kubenswrapper[4776]: I1125 09:47:17.733135 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 09:47:20 crc kubenswrapper[4776]: I1125 09:47:20.693126 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 25 09:47:20 crc kubenswrapper[4776]: I1125 09:47:20.725437 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 25 09:47:21 crc kubenswrapper[4776]: I1125 09:47:21.453684 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 25 09:47:22 crc kubenswrapper[4776]: I1125 09:47:22.732773 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 09:47:22 crc kubenswrapper[4776]: I1125 09:47:22.733163 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 09:47:23 crc kubenswrapper[4776]: I1125 09:47:23.748383 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6a0ab5a5-2d62-4457-be9d-fe8242ab7c91" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 09:47:23 crc kubenswrapper[4776]: I1125 09:47:23.748392 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6a0ab5a5-2d62-4457-be9d-fe8242ab7c91" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 09:47:25 crc kubenswrapper[4776]: I1125 09:47:25.787663 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 09:47:25 crc kubenswrapper[4776]: I1125 09:47:25.787984 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 09:47:26 crc kubenswrapper[4776]: I1125 09:47:26.801688 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6bf3edad-487a-4c68-9dbf-9789a94f8fba" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.205:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 09:47:26 crc kubenswrapper[4776]: I1125 09:47:26.801823 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6bf3edad-487a-4c68-9dbf-9789a94f8fba" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.205:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 09:47:27 crc kubenswrapper[4776]: I1125 09:47:27.626113 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 25 09:47:32 crc kubenswrapper[4776]: I1125 09:47:32.736772 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 25 09:47:32 crc kubenswrapper[4776]: I1125 09:47:32.742776 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 25 09:47:32 crc kubenswrapper[4776]: I1125 09:47:32.744030 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 25 09:47:33 crc kubenswrapper[4776]: I1125 09:47:33.544349 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 25 09:47:35 crc kubenswrapper[4776]: I1125 09:47:35.795449 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 09:47:35 crc kubenswrapper[4776]: I1125 09:47:35.795760 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 09:47:35 crc kubenswrapper[4776]: I1125 09:47:35.796301 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 09:47:35 crc kubenswrapper[4776]: I1125 09:47:35.796357 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 09:47:35 crc kubenswrapper[4776]: I1125 09:47:35.801914 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 09:47:35 crc kubenswrapper[4776]: I1125 09:47:35.806992 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 09:47:41 crc kubenswrapper[4776]: I1125 09:47:41.382514 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5wmvt"] Nov 25 09:47:41 crc kubenswrapper[4776]: I1125 09:47:41.386028 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5wmvt" Nov 25 09:47:41 crc kubenswrapper[4776]: I1125 09:47:41.398988 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5wmvt"] Nov 25 09:47:41 crc kubenswrapper[4776]: I1125 09:47:41.503910 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9266e4bf-3e2e-4ed5-b799-02e56af27091-utilities\") pod \"redhat-operators-5wmvt\" (UID: \"9266e4bf-3e2e-4ed5-b799-02e56af27091\") " pod="openshift-marketplace/redhat-operators-5wmvt" Nov 25 09:47:41 crc kubenswrapper[4776]: I1125 09:47:41.504005 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hnxs\" (UniqueName: \"kubernetes.io/projected/9266e4bf-3e2e-4ed5-b799-02e56af27091-kube-api-access-7hnxs\") pod \"redhat-operators-5wmvt\" (UID: \"9266e4bf-3e2e-4ed5-b799-02e56af27091\") " pod="openshift-marketplace/redhat-operators-5wmvt" Nov 25 09:47:41 crc kubenswrapper[4776]: I1125 09:47:41.504040 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9266e4bf-3e2e-4ed5-b799-02e56af27091-catalog-content\") pod \"redhat-operators-5wmvt\" (UID: \"9266e4bf-3e2e-4ed5-b799-02e56af27091\") " pod="openshift-marketplace/redhat-operators-5wmvt" Nov 25 09:47:41 crc kubenswrapper[4776]: I1125 09:47:41.606038 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9266e4bf-3e2e-4ed5-b799-02e56af27091-utilities\") pod \"redhat-operators-5wmvt\" (UID: \"9266e4bf-3e2e-4ed5-b799-02e56af27091\") " pod="openshift-marketplace/redhat-operators-5wmvt" Nov 25 09:47:41 crc kubenswrapper[4776]: I1125 09:47:41.606128 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hnxs\" (UniqueName: \"kubernetes.io/projected/9266e4bf-3e2e-4ed5-b799-02e56af27091-kube-api-access-7hnxs\") pod \"redhat-operators-5wmvt\" (UID: \"9266e4bf-3e2e-4ed5-b799-02e56af27091\") " pod="openshift-marketplace/redhat-operators-5wmvt" Nov 25 09:47:41 crc kubenswrapper[4776]: I1125 09:47:41.606165 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9266e4bf-3e2e-4ed5-b799-02e56af27091-catalog-content\") pod \"redhat-operators-5wmvt\" (UID: \"9266e4bf-3e2e-4ed5-b799-02e56af27091\") " pod="openshift-marketplace/redhat-operators-5wmvt" Nov 25 09:47:41 crc kubenswrapper[4776]: I1125 09:47:41.606646 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9266e4bf-3e2e-4ed5-b799-02e56af27091-utilities\") pod \"redhat-operators-5wmvt\" (UID: \"9266e4bf-3e2e-4ed5-b799-02e56af27091\") " pod="openshift-marketplace/redhat-operators-5wmvt" Nov 25 09:47:41 crc kubenswrapper[4776]: I1125 09:47:41.607553 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9266e4bf-3e2e-4ed5-b799-02e56af27091-catalog-content\") pod \"redhat-operators-5wmvt\" (UID: \"9266e4bf-3e2e-4ed5-b799-02e56af27091\") " pod="openshift-marketplace/redhat-operators-5wmvt" Nov 25 09:47:41 crc kubenswrapper[4776]: I1125 09:47:41.630948 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hnxs\" (UniqueName: \"kubernetes.io/projected/9266e4bf-3e2e-4ed5-b799-02e56af27091-kube-api-access-7hnxs\") pod \"redhat-operators-5wmvt\" (UID: \"9266e4bf-3e2e-4ed5-b799-02e56af27091\") " pod="openshift-marketplace/redhat-operators-5wmvt" Nov 25 09:47:41 crc kubenswrapper[4776]: I1125 09:47:41.710381 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5wmvt" Nov 25 09:47:42 crc kubenswrapper[4776]: I1125 09:47:42.021440 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5wmvt"] Nov 25 09:47:42 crc kubenswrapper[4776]: I1125 09:47:42.626160 4776 generic.go:334] "Generic (PLEG): container finished" podID="9266e4bf-3e2e-4ed5-b799-02e56af27091" containerID="77f3f2bd4f6583ba506502ce3622fd8e1b8dd903ba70835674336a5c58d9241b" exitCode=0 Nov 25 09:47:42 crc kubenswrapper[4776]: I1125 09:47:42.626230 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5wmvt" event={"ID":"9266e4bf-3e2e-4ed5-b799-02e56af27091","Type":"ContainerDied","Data":"77f3f2bd4f6583ba506502ce3622fd8e1b8dd903ba70835674336a5c58d9241b"} Nov 25 09:47:42 crc kubenswrapper[4776]: I1125 09:47:42.626507 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5wmvt" event={"ID":"9266e4bf-3e2e-4ed5-b799-02e56af27091","Type":"ContainerStarted","Data":"ddff0921234b55aa271943735142330c5e32a5ff802fc3d2504f4b81f0a926c4"} Nov 25 09:47:42 crc kubenswrapper[4776]: I1125 09:47:42.629731 4776 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 09:47:45 crc kubenswrapper[4776]: I1125 09:47:45.658160 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5wmvt" event={"ID":"9266e4bf-3e2e-4ed5-b799-02e56af27091","Type":"ContainerStarted","Data":"5006a19b68d700e697cd2f465b73452fcf31deed84559cb5659fa9a712fb7a12"} Nov 25 09:47:47 crc kubenswrapper[4776]: I1125 09:47:47.684892 4776 generic.go:334] "Generic (PLEG): container finished" podID="9266e4bf-3e2e-4ed5-b799-02e56af27091" containerID="5006a19b68d700e697cd2f465b73452fcf31deed84559cb5659fa9a712fb7a12" exitCode=0 Nov 25 09:47:47 crc kubenswrapper[4776]: I1125 09:47:47.684963 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5wmvt" event={"ID":"9266e4bf-3e2e-4ed5-b799-02e56af27091","Type":"ContainerDied","Data":"5006a19b68d700e697cd2f465b73452fcf31deed84559cb5659fa9a712fb7a12"} Nov 25 09:47:47 crc kubenswrapper[4776]: I1125 09:47:47.817979 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:47:47 crc kubenswrapper[4776]: I1125 09:47:47.818051 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:47:48 crc kubenswrapper[4776]: I1125 09:47:48.697305 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5wmvt" event={"ID":"9266e4bf-3e2e-4ed5-b799-02e56af27091","Type":"ContainerStarted","Data":"2a2229b7cd7d8596847c255b5a632079e86a284eb29c3cd724e9e64edc2fcf8b"} Nov 25 09:47:48 crc kubenswrapper[4776]: I1125 09:47:48.721928 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5wmvt" podStartSLOduration=1.9495234 podStartE2EDuration="7.721907291s" podCreationTimestamp="2025-11-25 09:47:41 +0000 UTC" firstStartedPulling="2025-11-25 09:47:42.629491783 +0000 UTC m=+1407.670551326" lastFinishedPulling="2025-11-25 09:47:48.401875654 +0000 UTC m=+1413.442935217" observedRunningTime="2025-11-25 09:47:48.721454009 +0000 UTC m=+1413.762513582" watchObservedRunningTime="2025-11-25 09:47:48.721907291 +0000 UTC m=+1413.762966844" Nov 25 09:47:51 crc kubenswrapper[4776]: I1125 09:47:51.710731 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5wmvt" Nov 25 09:47:51 crc kubenswrapper[4776]: I1125 09:47:51.711375 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5wmvt" Nov 25 09:47:52 crc kubenswrapper[4776]: I1125 09:47:52.760578 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5wmvt" podUID="9266e4bf-3e2e-4ed5-b799-02e56af27091" containerName="registry-server" probeResult="failure" output=< Nov 25 09:47:52 crc kubenswrapper[4776]: timeout: failed to connect service ":50051" within 1s Nov 25 09:47:52 crc kubenswrapper[4776]: > Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.341916 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.342393 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="3fc11cf2-c6c7-4250-b7f4-9c68d78152eb" containerName="openstackclient" containerID="cri-o://5b70abe095c3919e69bde8512f0bf19cca9ec9011ee75d534001611515145f6c" gracePeriod=2 Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.360534 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.406945 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.407333 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="6e7d9d60-9538-4870-8f62-434fd9f1ab0d" containerName="openstack-network-exporter" containerID="cri-o://17cb0f155d2722e145dfbd65de7d057bce9f1ad8f11fcd4b4347b19f7b5ce218" gracePeriod=300 Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.550889 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.567262 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="6e7d9d60-9538-4870-8f62-434fd9f1ab0d" containerName="ovsdbserver-sb" containerID="cri-o://ce501a13d7bab63cc6c0c2608dcb59e6fdc92a7de550426b2ee3b8ad29264e30" gracePeriod=300 Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.634337 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.635219 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df" containerName="openstack-network-exporter" containerID="cri-o://ee2853983a149e403a868c5541f34466a714bc2a4620f7f4820109d6bf8f4f6c" gracePeriod=300 Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.650446 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placemente176-account-delete-wbgz5"] Nov 25 09:47:54 crc kubenswrapper[4776]: E1125 09:47:54.650952 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fc11cf2-c6c7-4250-b7f4-9c68d78152eb" containerName="openstackclient" Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.651515 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fc11cf2-c6c7-4250-b7f4-9c68d78152eb" containerName="openstackclient" Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.652454 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fc11cf2-c6c7-4250-b7f4-9c68d78152eb" containerName="openstackclient" Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.658295 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placemente176-account-delete-wbgz5" Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.699926 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placemente176-account-delete-wbgz5"] Nov 25 09:47:54 crc kubenswrapper[4776]: E1125 09:47:54.720263 4776 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 25 09:47:54 crc kubenswrapper[4776]: E1125 09:47:54.720346 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f32041a0-a8fc-48a5-afab-476baff89e8c-config-data podName:f32041a0-a8fc-48a5-afab-476baff89e8c nodeName:}" failed. No retries permitted until 2025-11-25 09:47:55.220322714 +0000 UTC m=+1420.261382267 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/f32041a0-a8fc-48a5-afab-476baff89e8c-config-data") pod "rabbitmq-cell1-server-0" (UID: "f32041a0-a8fc-48a5-afab-476baff89e8c") : configmap "rabbitmq-cell1-config-data" not found Nov 25 09:47:54 crc kubenswrapper[4776]: E1125 09:47:54.738561 4776 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.193:37512->38.102.83.193:36915: write tcp 38.102.83.193:37512->38.102.83.193:36915: write: broken pipe Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.778513 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df" containerName="ovsdbserver-nb" containerID="cri-o://215609fe2a7a4ad5088201866359a4ff41f487eaa766e53084fae555576c9ef0" gracePeriod=300 Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.795799 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glanceb9b5-account-delete-tjrnr"] Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.797243 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glanceb9b5-account-delete-tjrnr" Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.814183 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glanceb9b5-account-delete-tjrnr"] Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.821905 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6be099e6-89d6-45a0-9d0f-4dca4d91701a-operator-scripts\") pod \"placemente176-account-delete-wbgz5\" (UID: \"6be099e6-89d6-45a0-9d0f-4dca4d91701a\") " pod="openstack/placemente176-account-delete-wbgz5" Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.822027 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb29h\" (UniqueName: \"kubernetes.io/projected/6be099e6-89d6-45a0-9d0f-4dca4d91701a-kube-api-access-wb29h\") pod \"placemente176-account-delete-wbgz5\" (UID: \"6be099e6-89d6-45a0-9d0f-4dca4d91701a\") " pod="openstack/placemente176-account-delete-wbgz5" Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.859468 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_6e7d9d60-9538-4870-8f62-434fd9f1ab0d/ovsdbserver-sb/0.log" Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.859518 4776 generic.go:334] "Generic (PLEG): container finished" podID="6e7d9d60-9538-4870-8f62-434fd9f1ab0d" containerID="17cb0f155d2722e145dfbd65de7d057bce9f1ad8f11fcd4b4347b19f7b5ce218" exitCode=2 Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.859536 4776 generic.go:334] "Generic (PLEG): container finished" podID="6e7d9d60-9538-4870-8f62-434fd9f1ab0d" containerID="ce501a13d7bab63cc6c0c2608dcb59e6fdc92a7de550426b2ee3b8ad29264e30" exitCode=143 Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.859559 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6e7d9d60-9538-4870-8f62-434fd9f1ab0d","Type":"ContainerDied","Data":"17cb0f155d2722e145dfbd65de7d057bce9f1ad8f11fcd4b4347b19f7b5ce218"} Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.859589 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6e7d9d60-9538-4870-8f62-434fd9f1ab0d","Type":"ContainerDied","Data":"ce501a13d7bab63cc6c0c2608dcb59e6fdc92a7de550426b2ee3b8ad29264e30"} Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.872312 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbicandb0b-account-delete-xtvnn"] Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.876336 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbicandb0b-account-delete-xtvnn" Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.923548 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbicandb0b-account-delete-xtvnn"] Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.924578 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfmhc\" (UniqueName: \"kubernetes.io/projected/bcaec0f3-13af-4f73-8e7f-7e575b8e2008-kube-api-access-vfmhc\") pod \"glanceb9b5-account-delete-tjrnr\" (UID: \"bcaec0f3-13af-4f73-8e7f-7e575b8e2008\") " pod="openstack/glanceb9b5-account-delete-tjrnr" Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.924658 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6be099e6-89d6-45a0-9d0f-4dca4d91701a-operator-scripts\") pod \"placemente176-account-delete-wbgz5\" (UID: \"6be099e6-89d6-45a0-9d0f-4dca4d91701a\") " pod="openstack/placemente176-account-delete-wbgz5" Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.924698 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcaec0f3-13af-4f73-8e7f-7e575b8e2008-operator-scripts\") pod \"glanceb9b5-account-delete-tjrnr\" (UID: \"bcaec0f3-13af-4f73-8e7f-7e575b8e2008\") " pod="openstack/glanceb9b5-account-delete-tjrnr" Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.924764 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wb29h\" (UniqueName: \"kubernetes.io/projected/6be099e6-89d6-45a0-9d0f-4dca4d91701a-kube-api-access-wb29h\") pod \"placemente176-account-delete-wbgz5\" (UID: \"6be099e6-89d6-45a0-9d0f-4dca4d91701a\") " pod="openstack/placemente176-account-delete-wbgz5" Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.925697 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6be099e6-89d6-45a0-9d0f-4dca4d91701a-operator-scripts\") pod \"placemente176-account-delete-wbgz5\" (UID: \"6be099e6-89d6-45a0-9d0f-4dca4d91701a\") " pod="openstack/placemente176-account-delete-wbgz5" Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.960087 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder774d-account-delete-b29n5"] Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.960746 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wb29h\" (UniqueName: \"kubernetes.io/projected/6be099e6-89d6-45a0-9d0f-4dca4d91701a-kube-api-access-wb29h\") pod \"placemente176-account-delete-wbgz5\" (UID: \"6be099e6-89d6-45a0-9d0f-4dca4d91701a\") " pod="openstack/placemente176-account-delete-wbgz5" Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.961728 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder774d-account-delete-b29n5" Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.986576 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.986810 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="7dadc68f-0b89-41ba-84f8-e9bcde16ca64" containerName="ovn-northd" containerID="cri-o://1ed6008959fa3368ff11289c84925032c1050a0bd2f104218d9763b1be776a55" gracePeriod=30 Nov 25 09:47:54 crc kubenswrapper[4776]: I1125 09:47:54.986998 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="7dadc68f-0b89-41ba-84f8-e9bcde16ca64" containerName="openstack-network-exporter" containerID="cri-o://048ec6577ec600c7421727ef4cb3662a66e77cfa268385a725211a51ef3249cd" gracePeriod=30 Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.007665 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder774d-account-delete-b29n5"] Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.029481 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqcbj\" (UniqueName: \"kubernetes.io/projected/ef85f90f-c377-4273-9fff-451420d6008a-kube-api-access-gqcbj\") pod \"cinder774d-account-delete-b29n5\" (UID: \"ef85f90f-c377-4273-9fff-451420d6008a\") " pod="openstack/cinder774d-account-delete-b29n5" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.029535 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sd9gw\" (UniqueName: \"kubernetes.io/projected/6d40516c-5fcc-4b24-918f-95bf79cb94b4-kube-api-access-sd9gw\") pod \"barbicandb0b-account-delete-xtvnn\" (UID: \"6d40516c-5fcc-4b24-918f-95bf79cb94b4\") " pod="openstack/barbicandb0b-account-delete-xtvnn" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.029583 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcaec0f3-13af-4f73-8e7f-7e575b8e2008-operator-scripts\") pod \"glanceb9b5-account-delete-tjrnr\" (UID: \"bcaec0f3-13af-4f73-8e7f-7e575b8e2008\") " pod="openstack/glanceb9b5-account-delete-tjrnr" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.029644 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d40516c-5fcc-4b24-918f-95bf79cb94b4-operator-scripts\") pod \"barbicandb0b-account-delete-xtvnn\" (UID: \"6d40516c-5fcc-4b24-918f-95bf79cb94b4\") " pod="openstack/barbicandb0b-account-delete-xtvnn" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.031657 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfmhc\" (UniqueName: \"kubernetes.io/projected/bcaec0f3-13af-4f73-8e7f-7e575b8e2008-kube-api-access-vfmhc\") pod \"glanceb9b5-account-delete-tjrnr\" (UID: \"bcaec0f3-13af-4f73-8e7f-7e575b8e2008\") " pod="openstack/glanceb9b5-account-delete-tjrnr" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.031716 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef85f90f-c377-4273-9fff-451420d6008a-operator-scripts\") pod \"cinder774d-account-delete-b29n5\" (UID: \"ef85f90f-c377-4273-9fff-451420d6008a\") " pod="openstack/cinder774d-account-delete-b29n5" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.032798 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcaec0f3-13af-4f73-8e7f-7e575b8e2008-operator-scripts\") pod \"glanceb9b5-account-delete-tjrnr\" (UID: \"bcaec0f3-13af-4f73-8e7f-7e575b8e2008\") " pod="openstack/glanceb9b5-account-delete-tjrnr" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.068910 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfmhc\" (UniqueName: \"kubernetes.io/projected/bcaec0f3-13af-4f73-8e7f-7e575b8e2008-kube-api-access-vfmhc\") pod \"glanceb9b5-account-delete-tjrnr\" (UID: \"bcaec0f3-13af-4f73-8e7f-7e575b8e2008\") " pod="openstack/glanceb9b5-account-delete-tjrnr" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.136148 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef85f90f-c377-4273-9fff-451420d6008a-operator-scripts\") pod \"cinder774d-account-delete-b29n5\" (UID: \"ef85f90f-c377-4273-9fff-451420d6008a\") " pod="openstack/cinder774d-account-delete-b29n5" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.136237 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqcbj\" (UniqueName: \"kubernetes.io/projected/ef85f90f-c377-4273-9fff-451420d6008a-kube-api-access-gqcbj\") pod \"cinder774d-account-delete-b29n5\" (UID: \"ef85f90f-c377-4273-9fff-451420d6008a\") " pod="openstack/cinder774d-account-delete-b29n5" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.136258 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sd9gw\" (UniqueName: \"kubernetes.io/projected/6d40516c-5fcc-4b24-918f-95bf79cb94b4-kube-api-access-sd9gw\") pod \"barbicandb0b-account-delete-xtvnn\" (UID: \"6d40516c-5fcc-4b24-918f-95bf79cb94b4\") " pod="openstack/barbicandb0b-account-delete-xtvnn" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.136312 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d40516c-5fcc-4b24-918f-95bf79cb94b4-operator-scripts\") pod \"barbicandb0b-account-delete-xtvnn\" (UID: \"6d40516c-5fcc-4b24-918f-95bf79cb94b4\") " pod="openstack/barbicandb0b-account-delete-xtvnn" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.137045 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef85f90f-c377-4273-9fff-451420d6008a-operator-scripts\") pod \"cinder774d-account-delete-b29n5\" (UID: \"ef85f90f-c377-4273-9fff-451420d6008a\") " pod="openstack/cinder774d-account-delete-b29n5" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.137136 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d40516c-5fcc-4b24-918f-95bf79cb94b4-operator-scripts\") pod \"barbicandb0b-account-delete-xtvnn\" (UID: \"6d40516c-5fcc-4b24-918f-95bf79cb94b4\") " pod="openstack/barbicandb0b-account-delete-xtvnn" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.156602 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placemente176-account-delete-wbgz5" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.183345 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqcbj\" (UniqueName: \"kubernetes.io/projected/ef85f90f-c377-4273-9fff-451420d6008a-kube-api-access-gqcbj\") pod \"cinder774d-account-delete-b29n5\" (UID: \"ef85f90f-c377-4273-9fff-451420d6008a\") " pod="openstack/cinder774d-account-delete-b29n5" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.184289 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glanceb9b5-account-delete-tjrnr" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.190410 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sd9gw\" (UniqueName: \"kubernetes.io/projected/6d40516c-5fcc-4b24-918f-95bf79cb94b4-kube-api-access-sd9gw\") pod \"barbicandb0b-account-delete-xtvnn\" (UID: \"6d40516c-5fcc-4b24-918f-95bf79cb94b4\") " pod="openstack/barbicandb0b-account-delete-xtvnn" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.205513 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbicandb0b-account-delete-xtvnn" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.225067 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell09dc4-account-delete-w5xcj"] Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.244768 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell09dc4-account-delete-w5xcj"] Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.274695 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.245880 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell09dc4-account-delete-w5xcj" Nov 25 09:47:55 crc kubenswrapper[4776]: E1125 09:47:55.250262 4776 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 25 09:47:55 crc kubenswrapper[4776]: E1125 09:47:55.275779 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f32041a0-a8fc-48a5-afab-476baff89e8c-config-data podName:f32041a0-a8fc-48a5-afab-476baff89e8c nodeName:}" failed. No retries permitted until 2025-11-25 09:47:56.27575286 +0000 UTC m=+1421.316812413 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/f32041a0-a8fc-48a5-afab-476baff89e8c-config-data") pod "rabbitmq-cell1-server-0" (UID: "f32041a0-a8fc-48a5-afab-476baff89e8c") : configmap "rabbitmq-cell1-config-data" not found Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.312851 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapi3b46-account-delete-9w5r9"] Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.316155 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi3b46-account-delete-9w5r9" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.318755 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapi3b46-account-delete-9w5r9"] Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.361229 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/946ea713-bfba-449f-82ab-28e915469938-operator-scripts\") pod \"novacell09dc4-account-delete-w5xcj\" (UID: \"946ea713-bfba-449f-82ab-28e915469938\") " pod="openstack/novacell09dc4-account-delete-w5xcj" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.361279 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59dkv\" (UniqueName: \"kubernetes.io/projected/946ea713-bfba-449f-82ab-28e915469938-kube-api-access-59dkv\") pod \"novacell09dc4-account-delete-w5xcj\" (UID: \"946ea713-bfba-449f-82ab-28e915469938\") " pod="openstack/novacell09dc4-account-delete-w5xcj" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.366846 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-ttw4k"] Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.383436 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-ttw4k"] Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.416198 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder774d-account-delete-b29n5" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.423110 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-2bs2f"] Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.439381 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-7fgjf"] Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.456623 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron5151-account-delete-8vbjg"] Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.463557 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/14b51bba-0b15-469b-bdea-90fae8d9b976-operator-scripts\") pod \"novaapi3b46-account-delete-9w5r9\" (UID: \"14b51bba-0b15-469b-bdea-90fae8d9b976\") " pod="openstack/novaapi3b46-account-delete-9w5r9" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.463632 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/946ea713-bfba-449f-82ab-28e915469938-operator-scripts\") pod \"novacell09dc4-account-delete-w5xcj\" (UID: \"946ea713-bfba-449f-82ab-28e915469938\") " pod="openstack/novacell09dc4-account-delete-w5xcj" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.463655 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59dkv\" (UniqueName: \"kubernetes.io/projected/946ea713-bfba-449f-82ab-28e915469938-kube-api-access-59dkv\") pod \"novacell09dc4-account-delete-w5xcj\" (UID: \"946ea713-bfba-449f-82ab-28e915469938\") " pod="openstack/novacell09dc4-account-delete-w5xcj" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.463679 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcs9z\" (UniqueName: \"kubernetes.io/projected/14b51bba-0b15-469b-bdea-90fae8d9b976-kube-api-access-gcs9z\") pod \"novaapi3b46-account-delete-9w5r9\" (UID: \"14b51bba-0b15-469b-bdea-90fae8d9b976\") " pod="openstack/novaapi3b46-account-delete-9w5r9" Nov 25 09:47:55 crc kubenswrapper[4776]: E1125 09:47:55.464389 4776 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 25 09:47:55 crc kubenswrapper[4776]: E1125 09:47:55.464450 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4512cf4e-6c75-493e-a80a-4b0accd750a8-config-data podName:4512cf4e-6c75-493e-a80a-4b0accd750a8 nodeName:}" failed. No retries permitted until 2025-11-25 09:47:55.964415674 +0000 UTC m=+1421.005475227 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/4512cf4e-6c75-493e-a80a-4b0accd750a8-config-data") pod "rabbitmq-server-0" (UID: "4512cf4e-6c75-493e-a80a-4b0accd750a8") : configmap "rabbitmq-config-data" not found Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.464567 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron5151-account-delete-8vbjg" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.474493 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/946ea713-bfba-449f-82ab-28e915469938-operator-scripts\") pod \"novacell09dc4-account-delete-w5xcj\" (UID: \"946ea713-bfba-449f-82ab-28e915469938\") " pod="openstack/novacell09dc4-account-delete-w5xcj" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.500740 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59dkv\" (UniqueName: \"kubernetes.io/projected/946ea713-bfba-449f-82ab-28e915469938-kube-api-access-59dkv\") pod \"novacell09dc4-account-delete-w5xcj\" (UID: \"946ea713-bfba-449f-82ab-28e915469938\") " pod="openstack/novacell09dc4-account-delete-w5xcj" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.554303 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-2mts9"] Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.554533 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-2mts9" podUID="cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd" containerName="openstack-network-exporter" containerID="cri-o://29696e80dd0a6c71a81e7417db086e3acded64a814a8e862ba2fc208da55d802" gracePeriod=30 Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.566531 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/14b51bba-0b15-469b-bdea-90fae8d9b976-operator-scripts\") pod \"novaapi3b46-account-delete-9w5r9\" (UID: \"14b51bba-0b15-469b-bdea-90fae8d9b976\") " pod="openstack/novaapi3b46-account-delete-9w5r9" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.566598 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73f2c2ab-6cb6-4544-a3f7-20722e5b6400-operator-scripts\") pod \"neutron5151-account-delete-8vbjg\" (UID: \"73f2c2ab-6cb6-4544-a3f7-20722e5b6400\") " pod="openstack/neutron5151-account-delete-8vbjg" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.566673 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcs9z\" (UniqueName: \"kubernetes.io/projected/14b51bba-0b15-469b-bdea-90fae8d9b976-kube-api-access-gcs9z\") pod \"novaapi3b46-account-delete-9w5r9\" (UID: \"14b51bba-0b15-469b-bdea-90fae8d9b976\") " pod="openstack/novaapi3b46-account-delete-9w5r9" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.566708 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92frh\" (UniqueName: \"kubernetes.io/projected/73f2c2ab-6cb6-4544-a3f7-20722e5b6400-kube-api-access-92frh\") pod \"neutron5151-account-delete-8vbjg\" (UID: \"73f2c2ab-6cb6-4544-a3f7-20722e5b6400\") " pod="openstack/neutron5151-account-delete-8vbjg" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.567536 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/14b51bba-0b15-469b-bdea-90fae8d9b976-operator-scripts\") pod \"novaapi3b46-account-delete-9w5r9\" (UID: \"14b51bba-0b15-469b-bdea-90fae8d9b976\") " pod="openstack/novaapi3b46-account-delete-9w5r9" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.613849 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell09dc4-account-delete-w5xcj" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.637897 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron5151-account-delete-8vbjg"] Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.642470 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcs9z\" (UniqueName: \"kubernetes.io/projected/14b51bba-0b15-469b-bdea-90fae8d9b976-kube-api-access-gcs9z\") pod \"novaapi3b46-account-delete-9w5r9\" (UID: \"14b51bba-0b15-469b-bdea-90fae8d9b976\") " pod="openstack/novaapi3b46-account-delete-9w5r9" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.670588 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73f2c2ab-6cb6-4544-a3f7-20722e5b6400-operator-scripts\") pod \"neutron5151-account-delete-8vbjg\" (UID: \"73f2c2ab-6cb6-4544-a3f7-20722e5b6400\") " pod="openstack/neutron5151-account-delete-8vbjg" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.678901 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92frh\" (UniqueName: \"kubernetes.io/projected/73f2c2ab-6cb6-4544-a3f7-20722e5b6400-kube-api-access-92frh\") pod \"neutron5151-account-delete-8vbjg\" (UID: \"73f2c2ab-6cb6-4544-a3f7-20722e5b6400\") " pod="openstack/neutron5151-account-delete-8vbjg" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.676031 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73f2c2ab-6cb6-4544-a3f7-20722e5b6400-operator-scripts\") pod \"neutron5151-account-delete-8vbjg\" (UID: \"73f2c2ab-6cb6-4544-a3f7-20722e5b6400\") " pod="openstack/neutron5151-account-delete-8vbjg" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.671010 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi3b46-account-delete-9w5r9" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.791593 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92frh\" (UniqueName: \"kubernetes.io/projected/73f2c2ab-6cb6-4544-a3f7-20722e5b6400-kube-api-access-92frh\") pod \"neutron5151-account-delete-8vbjg\" (UID: \"73f2c2ab-6cb6-4544-a3f7-20722e5b6400\") " pod="openstack/neutron5151-account-delete-8vbjg" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.863125 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron5151-account-delete-8vbjg" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.910609 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c18590c-cf10-4c05-8a97-b9f1d091868a" path="/var/lib/kubelet/pods/7c18590c-cf10-4c05-8a97-b9f1d091868a/volumes" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.911937 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-4b82s"] Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.926222 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-4b82s"] Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.926271 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-4htz7"] Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.926297 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-4htz7"] Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.926311 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d7f54fb65-td75r"] Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.926660 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" podUID="9ab346c9-f70f-4663-9b53-67b4d66b3112" containerName="dnsmasq-dns" containerID="cri-o://2b9a66a95781ceb011f5c668fc0302d4aa355e258935b0bc8a74577b415062f6" gracePeriod=10 Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.933653 4776 generic.go:334] "Generic (PLEG): container finished" podID="7dadc68f-0b89-41ba-84f8-e9bcde16ca64" containerID="048ec6577ec600c7421727ef4cb3662a66e77cfa268385a725211a51ef3249cd" exitCode=2 Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.933773 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7dadc68f-0b89-41ba-84f8-e9bcde16ca64","Type":"ContainerDied","Data":"048ec6577ec600c7421727ef4cb3662a66e77cfa268385a725211a51ef3249cd"} Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.943923 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df/ovsdbserver-nb/0.log" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.943974 4776 generic.go:334] "Generic (PLEG): container finished" podID="7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df" containerID="ee2853983a149e403a868c5541f34466a714bc2a4620f7f4820109d6bf8f4f6c" exitCode=2 Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.944009 4776 generic.go:334] "Generic (PLEG): container finished" podID="7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df" containerID="215609fe2a7a4ad5088201866359a4ff41f487eaa766e53084fae555576c9ef0" exitCode=143 Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.952811 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df","Type":"ContainerDied","Data":"ee2853983a149e403a868c5541f34466a714bc2a4620f7f4820109d6bf8f4f6c"} Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.952877 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df","Type":"ContainerDied","Data":"215609fe2a7a4ad5088201866359a4ff41f487eaa766e53084fae555576c9ef0"} Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.962412 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-bdhw6"] Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.983059 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-2mts9_cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd/openstack-network-exporter/0.log" Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.983131 4776 generic.go:334] "Generic (PLEG): container finished" podID="cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd" containerID="29696e80dd0a6c71a81e7417db086e3acded64a814a8e862ba2fc208da55d802" exitCode=2 Nov 25 09:47:55 crc kubenswrapper[4776]: I1125 09:47:55.983169 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-2mts9" event={"ID":"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd","Type":"ContainerDied","Data":"29696e80dd0a6c71a81e7417db086e3acded64a814a8e862ba2fc208da55d802"} Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.014994 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-bdhw6"] Nov 25 09:47:56 crc kubenswrapper[4776]: E1125 09:47:56.017506 4776 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 25 09:47:56 crc kubenswrapper[4776]: E1125 09:47:56.017572 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4512cf4e-6c75-493e-a80a-4b0accd750a8-config-data podName:4512cf4e-6c75-493e-a80a-4b0accd750a8 nodeName:}" failed. No retries permitted until 2025-11-25 09:47:57.017556582 +0000 UTC m=+1422.058616125 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/4512cf4e-6c75-493e-a80a-4b0accd750a8-config-data") pod "rabbitmq-server-0" (UID: "4512cf4e-6c75-493e-a80a-4b0accd750a8") : configmap "rabbitmq-config-data" not found Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.071910 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-djwk2"] Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.088274 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-djwk2"] Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.111039 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.111723 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="account-server" containerID="cri-o://9e7cf1b3a65acbcaaedff95080720317c88059b1b6b759f4434c871973a12bc3" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.111763 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="object-auditor" containerID="cri-o://670e0e074022b9b3b944ddaeb78954a059e4c4b8e2f7af4e2f91000daddc1263" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.111793 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="container-replicator" containerID="cri-o://a4e815cd4ca77d4505a242b624a83fcda104ea6e85674eb4c15ddc064562ffd6" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.111856 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="object-replicator" containerID="cri-o://a5e5c6ad114abc906cc7868aa09dc3d1a6a83ffb494d14de45deba5107b01b02" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.111876 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="account-reaper" containerID="cri-o://cd073bb8b2d7b566c99b6e569d6dbfa6b16b08d5db2e84e47a8d2d85eb563f1b" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.111919 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="swift-recon-cron" containerID="cri-o://faa52779c1f67996100c644db4f91149b383bf2ef821c114ac1e92b3b697987f" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.111935 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="account-replicator" containerID="cri-o://05b65f4bed8c7914cf657df0f540c0d19593a2c84c93c9ae6c4f106530470ec4" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.111859 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="container-server" containerID="cri-o://e2072d3b3f2b295702755794df1a3f5a2a130db157825a15a28d4ac1c3a1c674" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.111922 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="account-auditor" containerID="cri-o://6ccd151880ed401c0372e024b6b953cdc0eba4d9d9e8cdd47435d059001cbc30" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.111962 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="container-updater" containerID="cri-o://f0cdb71ed47fdaaa7eabd7f81c93a293616c6c022e1c4d8cef28b88de0dfb8d5" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.111809 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="object-server" containerID="cri-o://2c713a54dd52d79687c4437832031c9afbb02bd373076897a51654ae360ff22a" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.112033 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="object-expirer" containerID="cri-o://8e84b91c7719175c8908ad8fa6987a175f46744663fe91972c8bbb602ae871d8" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.112053 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="object-updater" containerID="cri-o://49bb08566118dc9e19698c8156598960f4778c7e09f832e3eeeaa9b6d4fc9a62" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.112049 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="container-auditor" containerID="cri-o://434eabc5a25e34093e38beb759273c9fac1b8758c3a03273e4ddb5f564fcdc08" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.111988 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="rsync" containerID="cri-o://152026f7bf460a5258df8a4110443a6e694c32ee892ab2dcb9469e4391545baa" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.131424 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-lpc2f"] Nov 25 09:47:56 crc kubenswrapper[4776]: E1125 09:47:56.136702 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 215609fe2a7a4ad5088201866359a4ff41f487eaa766e53084fae555576c9ef0 is running failed: container process not found" containerID="215609fe2a7a4ad5088201866359a4ff41f487eaa766e53084fae555576c9ef0" cmd=["/usr/bin/pidof","ovsdb-server"] Nov 25 09:47:56 crc kubenswrapper[4776]: E1125 09:47:56.159826 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 215609fe2a7a4ad5088201866359a4ff41f487eaa766e53084fae555576c9ef0 is running failed: container process not found" containerID="215609fe2a7a4ad5088201866359a4ff41f487eaa766e53084fae555576c9ef0" cmd=["/usr/bin/pidof","ovsdb-server"] Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.160009 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-lpc2f"] Nov 25 09:47:56 crc kubenswrapper[4776]: E1125 09:47:56.163084 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 215609fe2a7a4ad5088201866359a4ff41f487eaa766e53084fae555576c9ef0 is running failed: container process not found" containerID="215609fe2a7a4ad5088201866359a4ff41f487eaa766e53084fae555576c9ef0" cmd=["/usr/bin/pidof","ovsdb-server"] Nov 25 09:47:56 crc kubenswrapper[4776]: E1125 09:47:56.163137 4776 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 215609fe2a7a4ad5088201866359a4ff41f487eaa766e53084fae555576c9ef0 is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-nb-0" podUID="7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df" containerName="ovsdbserver-nb" Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.196323 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-69dpf"] Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.233205 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-69dpf"] Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.254010 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-lgdgb"] Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.281273 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-lgdgb"] Nov 25 09:47:56 crc kubenswrapper[4776]: E1125 09:47:56.332991 4776 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 25 09:47:56 crc kubenswrapper[4776]: E1125 09:47:56.333069 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f32041a0-a8fc-48a5-afab-476baff89e8c-config-data podName:f32041a0-a8fc-48a5-afab-476baff89e8c nodeName:}" failed. No retries permitted until 2025-11-25 09:47:58.333048345 +0000 UTC m=+1423.374107898 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/f32041a0-a8fc-48a5-afab-476baff89e8c-config-data") pod "rabbitmq-cell1-server-0" (UID: "f32041a0-a8fc-48a5-afab-476baff89e8c") : configmap "rabbitmq-cell1-config-data" not found Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.347191 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_6e7d9d60-9538-4870-8f62-434fd9f1ab0d/ovsdbserver-sb/0.log" Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.347538 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.352893 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-667769d556-lqtlp"] Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.353201 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-667769d556-lqtlp" podUID="a0b97845-4578-4801-896a-281fdafdb351" containerName="placement-log" containerID="cri-o://5fac5a59b29d2ae80cf45865e0db91319f65183c12c9b736093c4452a98a4355" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.353392 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-667769d556-lqtlp" podUID="a0b97845-4578-4801-896a-281fdafdb351" containerName="placement-api" containerID="cri-o://856d610eaa67921f1c8d09ca91236aafc5070c66f02e6f2c2f6ee1982f26930e" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.392059 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.392365 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="b708b100-166c-49b1-a47b-76d47ba2c6e5" containerName="glance-log" containerID="cri-o://7f99367ae1299a4af7ca7bdb87418a9b017add504f1ee9377bbcccc241edd6fb" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.392537 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="b708b100-166c-49b1-a47b-76d47ba2c6e5" containerName="glance-httpd" containerID="cri-o://5919b100098159f61f76d181a1ae66bfafcb2916a56d7fe4fe1cee93ad3c32d0" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.434605 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxscj\" (UniqueName: \"kubernetes.io/projected/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-kube-api-access-pxscj\") pod \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.434647 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-metrics-certs-tls-certs\") pod \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.434695 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-ovsdbserver-sb-tls-certs\") pod \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.434739 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-scripts\") pod \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.434831 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-combined-ca-bundle\") pod \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.434897 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-config\") pod \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.434927 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.435006 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-ovsdb-rundir\") pod \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\" (UID: \"6e7d9d60-9538-4870-8f62-434fd9f1ab0d\") " Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.436892 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-scripts" (OuterVolumeSpecName: "scripts") pod "6e7d9d60-9538-4870-8f62-434fd9f1ab0d" (UID: "6e7d9d60-9538-4870-8f62-434fd9f1ab0d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.437752 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-config" (OuterVolumeSpecName: "config") pod "6e7d9d60-9538-4870-8f62-434fd9f1ab0d" (UID: "6e7d9d60-9538-4870-8f62-434fd9f1ab0d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.449392 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "6e7d9d60-9538-4870-8f62-434fd9f1ab0d" (UID: "6e7d9d60-9538-4870-8f62-434fd9f1ab0d"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.501106 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.501399 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d2cda6de-ef12-42f4-a6e5-824c2db3cd01" containerName="glance-log" containerID="cri-o://0bd7925c5eadaefd62b355bbd7aebe47ce493b779918483df9c72dc9a141944b" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.501843 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d2cda6de-ef12-42f4-a6e5-824c2db3cd01" containerName="glance-httpd" containerID="cri-o://9fddb69738d8e8fd199d1f7a677b07c2d8ceff1639dbe9784b2c6fd247da4ae1" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.506678 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-kube-api-access-pxscj" (OuterVolumeSpecName: "kube-api-access-pxscj") pod "6e7d9d60-9538-4870-8f62-434fd9f1ab0d" (UID: "6e7d9d60-9538-4870-8f62-434fd9f1ab0d"). InnerVolumeSpecName "kube-api-access-pxscj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.549330 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.553380 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="bc920d71-4308-4d90-a891-8b760651ccc9" containerName="cinder-scheduler" containerID="cri-o://6bffff98e55f5e2fe0110f4a9d127c1dff4b4c552606101a667c0fb53d5e9b16" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.553902 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="bc920d71-4308-4d90-a891-8b760651ccc9" containerName="probe" containerID="cri-o://c9f1b69c74818d42cf402965ae75f33a682f61bb14e9874a7f7904a0e8298e5e" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.571198 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.572184 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.572196 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.572207 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxscj\" (UniqueName: \"kubernetes.io/projected/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-kube-api-access-pxscj\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.571319 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.572508 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="86e320e0-80ee-40df-b3a1-a48cb810a435" containerName="cinder-api-log" containerID="cri-o://643bd96dcd87a61d61b437895cd783519ab3ff279196abf17121bbe3e4d485dd" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.573002 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="86e320e0-80ee-40df-b3a1-a48cb810a435" containerName="cinder-api" containerID="cri-o://968312305c48cd2a75a04c3e2a43d37d9ed640ee14d080081ad79c323f2cf0f5" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.582459 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "6e7d9d60-9538-4870-8f62-434fd9f1ab0d" (UID: "6e7d9d60-9538-4870-8f62-434fd9f1ab0d"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.611276 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6e7d9d60-9538-4870-8f62-434fd9f1ab0d" (UID: "6e7d9d60-9538-4870-8f62-434fd9f1ab0d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.620165 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.620455 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6bf3edad-487a-4c68-9dbf-9789a94f8fba" containerName="nova-api-log" containerID="cri-o://07109bfcbb050b4c8db424a410626fa31b2236c8d915084ff3b34d220f1c1cf0" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.620969 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6bf3edad-487a-4c68-9dbf-9789a94f8fba" containerName="nova-api-api" containerID="cri-o://24a3083e873c7a16456c526fdfe2bc77dcf34c30f066c1f1af77953f2e1ecddd" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.646684 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.662553 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.662858 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6a0ab5a5-2d62-4457-be9d-fe8242ab7c91" containerName="nova-metadata-log" containerID="cri-o://6388f255053948e8f439f0e43bfe9480c34d62fbb564522cf02851e2a2aef43a" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.663629 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6a0ab5a5-2d62-4457-be9d-fe8242ab7c91" containerName="nova-metadata-metadata" containerID="cri-o://7927ef3d5f7eb88faf6e7c414da603268c27589afcdfe2331cf08fabea8c8c3c" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.674165 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.674201 4776 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.687573 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.728396 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-a02b-account-create-dpj2t"] Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.743476 4776 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.763984 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-hc2rc"] Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.769477 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "6e7d9d60-9538-4870-8f62-434fd9f1ab0d" (UID: "6e7d9d60-9538-4870-8f62-434fd9f1ab0d"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.775951 4776 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.775978 4776 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.778724 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="f32041a0-a8fc-48a5-afab-476baff89e8c" containerName="rabbitmq" containerID="cri-o://7b42621ff074baca099513ddb94523b955d55200c16323a7bb88a078926fca25" gracePeriod=604800 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.781727 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-a02b-account-create-dpj2t"] Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.794730 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-hc2rc"] Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.818939 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "6e7d9d60-9538-4870-8f62-434fd9f1ab0d" (UID: "6e7d9d60-9538-4870-8f62-434fd9f1ab0d"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.833747 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-54dbffbb57-kmqfb"] Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.834115 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-54dbffbb57-kmqfb" podUID="fe3ff67b-01ff-480f-be91-7c8235593c97" containerName="neutron-api" containerID="cri-o://94e2c5021a1689b990cd8fd474841522443a046e3012b95cb3f5f1f9b90efdd1" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.835494 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-54dbffbb57-kmqfb" podUID="fe3ff67b-01ff-480f-be91-7c8235593c97" containerName="neutron-httpd" containerID="cri-o://5c1129db350942a00317a1332a9a8db16c63fbd966ba3d63890450c144964ddb" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.884570 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-2mts9_cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd/openstack-network-exporter/0.log" Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.884868 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-2mts9" Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.888453 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e7d9d60-9538-4870-8f62-434fd9f1ab0d-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:56 crc kubenswrapper[4776]: E1125 09:47:56.912965 4776 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Nov 25 09:47:56 crc kubenswrapper[4776]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Nov 25 09:47:56 crc kubenswrapper[4776]: + source /usr/local/bin/container-scripts/functions Nov 25 09:47:56 crc kubenswrapper[4776]: ++ OVNBridge=br-int Nov 25 09:47:56 crc kubenswrapper[4776]: ++ OVNRemote=tcp:localhost:6642 Nov 25 09:47:56 crc kubenswrapper[4776]: ++ OVNEncapType=geneve Nov 25 09:47:56 crc kubenswrapper[4776]: ++ OVNAvailabilityZones= Nov 25 09:47:56 crc kubenswrapper[4776]: ++ EnableChassisAsGateway=true Nov 25 09:47:56 crc kubenswrapper[4776]: ++ PhysicalNetworks= Nov 25 09:47:56 crc kubenswrapper[4776]: ++ OVNHostName= Nov 25 09:47:56 crc kubenswrapper[4776]: ++ DB_FILE=/etc/openvswitch/conf.db Nov 25 09:47:56 crc kubenswrapper[4776]: ++ ovs_dir=/var/lib/openvswitch Nov 25 09:47:56 crc kubenswrapper[4776]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Nov 25 09:47:56 crc kubenswrapper[4776]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Nov 25 09:47:56 crc kubenswrapper[4776]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 25 09:47:56 crc kubenswrapper[4776]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 25 09:47:56 crc kubenswrapper[4776]: + sleep 0.5 Nov 25 09:47:56 crc kubenswrapper[4776]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 25 09:47:56 crc kubenswrapper[4776]: + sleep 0.5 Nov 25 09:47:56 crc kubenswrapper[4776]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 25 09:47:56 crc kubenswrapper[4776]: + cleanup_ovsdb_server_semaphore Nov 25 09:47:56 crc kubenswrapper[4776]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 25 09:47:56 crc kubenswrapper[4776]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Nov 25 09:47:56 crc kubenswrapper[4776]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-7fgjf" message=< Nov 25 09:47:56 crc kubenswrapper[4776]: Exiting ovsdb-server (5) [ OK ] Nov 25 09:47:56 crc kubenswrapper[4776]: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Nov 25 09:47:56 crc kubenswrapper[4776]: + source /usr/local/bin/container-scripts/functions Nov 25 09:47:56 crc kubenswrapper[4776]: ++ OVNBridge=br-int Nov 25 09:47:56 crc kubenswrapper[4776]: ++ OVNRemote=tcp:localhost:6642 Nov 25 09:47:56 crc kubenswrapper[4776]: ++ OVNEncapType=geneve Nov 25 09:47:56 crc kubenswrapper[4776]: ++ OVNAvailabilityZones= Nov 25 09:47:56 crc kubenswrapper[4776]: ++ EnableChassisAsGateway=true Nov 25 09:47:56 crc kubenswrapper[4776]: ++ PhysicalNetworks= Nov 25 09:47:56 crc kubenswrapper[4776]: ++ OVNHostName= Nov 25 09:47:56 crc kubenswrapper[4776]: ++ DB_FILE=/etc/openvswitch/conf.db Nov 25 09:47:56 crc kubenswrapper[4776]: ++ ovs_dir=/var/lib/openvswitch Nov 25 09:47:56 crc kubenswrapper[4776]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Nov 25 09:47:56 crc kubenswrapper[4776]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Nov 25 09:47:56 crc kubenswrapper[4776]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 25 09:47:56 crc kubenswrapper[4776]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 25 09:47:56 crc kubenswrapper[4776]: + sleep 0.5 Nov 25 09:47:56 crc kubenswrapper[4776]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 25 09:47:56 crc kubenswrapper[4776]: + sleep 0.5 Nov 25 09:47:56 crc kubenswrapper[4776]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 25 09:47:56 crc kubenswrapper[4776]: + cleanup_ovsdb_server_semaphore Nov 25 09:47:56 crc kubenswrapper[4776]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 25 09:47:56 crc kubenswrapper[4776]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Nov 25 09:47:56 crc kubenswrapper[4776]: > Nov 25 09:47:56 crc kubenswrapper[4776]: E1125 09:47:56.913005 4776 kuberuntime_container.go:691] "PreStop hook failed" err=< Nov 25 09:47:56 crc kubenswrapper[4776]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Nov 25 09:47:56 crc kubenswrapper[4776]: + source /usr/local/bin/container-scripts/functions Nov 25 09:47:56 crc kubenswrapper[4776]: ++ OVNBridge=br-int Nov 25 09:47:56 crc kubenswrapper[4776]: ++ OVNRemote=tcp:localhost:6642 Nov 25 09:47:56 crc kubenswrapper[4776]: ++ OVNEncapType=geneve Nov 25 09:47:56 crc kubenswrapper[4776]: ++ OVNAvailabilityZones= Nov 25 09:47:56 crc kubenswrapper[4776]: ++ EnableChassisAsGateway=true Nov 25 09:47:56 crc kubenswrapper[4776]: ++ PhysicalNetworks= Nov 25 09:47:56 crc kubenswrapper[4776]: ++ OVNHostName= Nov 25 09:47:56 crc kubenswrapper[4776]: ++ DB_FILE=/etc/openvswitch/conf.db Nov 25 09:47:56 crc kubenswrapper[4776]: ++ ovs_dir=/var/lib/openvswitch Nov 25 09:47:56 crc kubenswrapper[4776]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Nov 25 09:47:56 crc kubenswrapper[4776]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Nov 25 09:47:56 crc kubenswrapper[4776]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 25 09:47:56 crc kubenswrapper[4776]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 25 09:47:56 crc kubenswrapper[4776]: + sleep 0.5 Nov 25 09:47:56 crc kubenswrapper[4776]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 25 09:47:56 crc kubenswrapper[4776]: + sleep 0.5 Nov 25 09:47:56 crc kubenswrapper[4776]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Nov 25 09:47:56 crc kubenswrapper[4776]: + cleanup_ovsdb_server_semaphore Nov 25 09:47:56 crc kubenswrapper[4776]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Nov 25 09:47:56 crc kubenswrapper[4776]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Nov 25 09:47:56 crc kubenswrapper[4776]: > pod="openstack/ovn-controller-ovs-7fgjf" podUID="10fe3dc2-2bde-4380-8ad3-d71832e31095" containerName="ovsdb-server" containerID="cri-o://eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06" Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.913043 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-7fgjf" podUID="10fe3dc2-2bde-4380-8ad3-d71832e31095" containerName="ovsdb-server" containerID="cri-o://eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06" gracePeriod=29 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.913939 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-66cdf565f-6lzqw"] Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.914252 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-66cdf565f-6lzqw" podUID="1d1e9a91-e344-442f-8bbc-90b1f98e2e1f" containerName="barbican-worker-log" containerID="cri-o://87f8cbf578187a46d6e2d2c1082d1ed105b039c2788db71db6340e4c1b8457d4" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.914467 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-66cdf565f-6lzqw" podUID="1d1e9a91-e344-442f-8bbc-90b1f98e2e1f" containerName="barbican-worker" containerID="cri-o://9779df52088b267add1650137fc3fc3f92ad6d3ea2e75a34659668d2ef3813bb" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.919778 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-56988fbb4-nlxs9"] Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.920062 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-56988fbb4-nlxs9" podUID="e51292cd-00f5-4e0a-ba2b-6319bd8a5e45" containerName="barbican-keystone-listener-log" containerID="cri-o://a1a93104c41ae0baf6a1781b8d29302ccef9fef9da152e6274b2e92e08df9a8c" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.920163 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-56988fbb4-nlxs9" podUID="e51292cd-00f5-4e0a-ba2b-6319bd8a5e45" containerName="barbican-keystone-listener" containerID="cri-o://63cb3f8363fb0824de5241eef2c1f027163930b39c0a67a62ab2eadf418c85d7" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.926349 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.926617 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="e09f9958-4faf-4a28-8214-c3ead146122c" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://e056e61a1527b24d984169499b3cde97a02136661c62174d6fe28407dfd3287a" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.935937 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-769ddf488d-46ssj"] Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.937279 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-769ddf488d-46ssj" podUID="4702d881-d701-41a4-b36f-2d063a8ae246" containerName="barbican-api-log" containerID="cri-o://d45871f326a3dbb0f683b3d995f92f15682fd356f7dd9ec10b42a715df612c77" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.937890 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-769ddf488d-46ssj" podUID="4702d881-d701-41a4-b36f-2d063a8ae246" containerName="barbican-api" containerID="cri-o://4e9d85a2660614a7e89218283e8f24a0ddfcb9cad63c01ab8119453aceaad099" gracePeriod=30 Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.946315 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.959846 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 09:47:56 crc kubenswrapper[4776]: I1125 09:47:56.960172 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="8bae8251-3d5d-490b-9116-73769dc18101" containerName="nova-scheduler-scheduler" containerID="cri-o://257f4e7f047aaf2041def20b6caaee331d83d82f30d019096ad21bfd241e42a8" gracePeriod=30 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:56.992934 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-config\") pod \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\" (UID: \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\") " Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:56.993089 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-ovs-rundir\") pod \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\" (UID: \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\") " Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:56.993247 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xbmh\" (UniqueName: \"kubernetes.io/projected/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-kube-api-access-8xbmh\") pod \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\" (UID: \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\") " Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:56.993283 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-combined-ca-bundle\") pod \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\" (UID: \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\") " Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:56.993309 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-ovn-rundir\") pod \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\" (UID: \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\") " Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:56.993345 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-metrics-certs-tls-certs\") pod \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\" (UID: \"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd\") " Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.001149 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd" (UID: "cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.001211 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-config" (OuterVolumeSpecName: "config") pod "cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd" (UID: "cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.001244 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd" (UID: "cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.010304 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-7fgjf" podUID="10fe3dc2-2bde-4380-8ad3-d71832e31095" containerName="ovs-vswitchd" containerID="cri-o://5002006f906f0b17afbf25b23648688a842975cfd3b385d57e3191040607fb57" gracePeriod=29 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.015568 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-kube-api-access-8xbmh" (OuterVolumeSpecName: "kube-api-access-8xbmh") pod "cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd" (UID: "cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd"). InnerVolumeSpecName "kube-api-access-8xbmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.017693 4776 generic.go:334] "Generic (PLEG): container finished" podID="6a0ab5a5-2d62-4457-be9d-fe8242ab7c91" containerID="6388f255053948e8f439f0e43bfe9480c34d62fbb564522cf02851e2a2aef43a" exitCode=143 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.017758 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91","Type":"ContainerDied","Data":"6388f255053948e8f439f0e43bfe9480c34d62fbb564522cf02851e2a2aef43a"} Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.021553 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.027593 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df/ovsdbserver-nb/0.log" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.027666 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.041535 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="4512cf4e-6c75-493e-a80a-4b0accd750a8" containerName="rabbitmq" containerID="cri-o://a5fdd3d6b6a35387501cbc30834b6e26eb181b138a16ef46ae65f71952655401" gracePeriod=604800 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.093317 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd" (UID: "cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.098125 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-config\") pod \"9ab346c9-f70f-4663-9b53-67b4d66b3112\" (UID: \"9ab346c9-f70f-4663-9b53-67b4d66b3112\") " Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.098249 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-dns-svc\") pod \"9ab346c9-f70f-4663-9b53-67b4d66b3112\" (UID: \"9ab346c9-f70f-4663-9b53-67b4d66b3112\") " Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.098316 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-scripts\") pod \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.098365 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-ovsdbserver-sb\") pod \"9ab346c9-f70f-4663-9b53-67b4d66b3112\" (UID: \"9ab346c9-f70f-4663-9b53-67b4d66b3112\") " Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.098389 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-config\") pod \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.098424 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.098447 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-combined-ca-bundle\") pod \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.098471 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-ovsdb-rundir\") pod \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.098506 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-ovsdbserver-nb\") pod \"9ab346c9-f70f-4663-9b53-67b4d66b3112\" (UID: \"9ab346c9-f70f-4663-9b53-67b4d66b3112\") " Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.098538 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-ovsdbserver-nb-tls-certs\") pod \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.098623 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-metrics-certs-tls-certs\") pod \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.098650 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7g6c\" (UniqueName: \"kubernetes.io/projected/9ab346c9-f70f-4663-9b53-67b4d66b3112-kube-api-access-p7g6c\") pod \"9ab346c9-f70f-4663-9b53-67b4d66b3112\" (UID: \"9ab346c9-f70f-4663-9b53-67b4d66b3112\") " Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.098687 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvrvj\" (UniqueName: \"kubernetes.io/projected/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-kube-api-access-wvrvj\") pod \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\" (UID: \"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df\") " Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.098706 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-dns-swift-storage-0\") pod \"9ab346c9-f70f-4663-9b53-67b4d66b3112\" (UID: \"9ab346c9-f70f-4663-9b53-67b4d66b3112\") " Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.099264 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xbmh\" (UniqueName: \"kubernetes.io/projected/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-kube-api-access-8xbmh\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.099285 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.099297 4776 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-ovn-rundir\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.099308 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.099319 4776 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-ovs-rundir\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:57 crc kubenswrapper[4776]: E1125 09:47:57.099404 4776 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 25 09:47:57 crc kubenswrapper[4776]: E1125 09:47:57.099458 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4512cf4e-6c75-493e-a80a-4b0accd750a8-config-data podName:4512cf4e-6c75-493e-a80a-4b0accd750a8 nodeName:}" failed. No retries permitted until 2025-11-25 09:47:59.099439194 +0000 UTC m=+1424.140498747 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/4512cf4e-6c75-493e-a80a-4b0accd750a8-config-data") pod "rabbitmq-server-0" (UID: "4512cf4e-6c75-493e-a80a-4b0accd750a8") : configmap "rabbitmq-config-data" not found Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.100265 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-scripts" (OuterVolumeSpecName: "scripts") pod "7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df" (UID: "7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.100307 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df" (UID: "7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.103934 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-config" (OuterVolumeSpecName: "config") pod "7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df" (UID: "7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.117617 4776 generic.go:334] "Generic (PLEG): container finished" podID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerID="152026f7bf460a5258df8a4110443a6e694c32ee892ab2dcb9469e4391545baa" exitCode=0 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.118206 4776 generic.go:334] "Generic (PLEG): container finished" podID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerID="8e84b91c7719175c8908ad8fa6987a175f46744663fe91972c8bbb602ae871d8" exitCode=0 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.118221 4776 generic.go:334] "Generic (PLEG): container finished" podID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerID="49bb08566118dc9e19698c8156598960f4778c7e09f832e3eeeaa9b6d4fc9a62" exitCode=0 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.118227 4776 generic.go:334] "Generic (PLEG): container finished" podID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerID="670e0e074022b9b3b944ddaeb78954a059e4c4b8e2f7af4e2f91000daddc1263" exitCode=0 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.118236 4776 generic.go:334] "Generic (PLEG): container finished" podID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerID="a5e5c6ad114abc906cc7868aa09dc3d1a6a83ffb494d14de45deba5107b01b02" exitCode=0 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.118263 4776 generic.go:334] "Generic (PLEG): container finished" podID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerID="2c713a54dd52d79687c4437832031c9afbb02bd373076897a51654ae360ff22a" exitCode=0 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.118289 4776 generic.go:334] "Generic (PLEG): container finished" podID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerID="f0cdb71ed47fdaaa7eabd7f81c93a293616c6c022e1c4d8cef28b88de0dfb8d5" exitCode=0 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.118295 4776 generic.go:334] "Generic (PLEG): container finished" podID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerID="434eabc5a25e34093e38beb759273c9fac1b8758c3a03273e4ddb5f564fcdc08" exitCode=0 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.118301 4776 generic.go:334] "Generic (PLEG): container finished" podID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerID="a4e815cd4ca77d4505a242b624a83fcda104ea6e85674eb4c15ddc064562ffd6" exitCode=0 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.118308 4776 generic.go:334] "Generic (PLEG): container finished" podID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerID="e2072d3b3f2b295702755794df1a3f5a2a130db157825a15a28d4ac1c3a1c674" exitCode=0 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.118313 4776 generic.go:334] "Generic (PLEG): container finished" podID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerID="cd073bb8b2d7b566c99b6e569d6dbfa6b16b08d5db2e84e47a8d2d85eb563f1b" exitCode=0 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.118320 4776 generic.go:334] "Generic (PLEG): container finished" podID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerID="6ccd151880ed401c0372e024b6b953cdc0eba4d9d9e8cdd47435d059001cbc30" exitCode=0 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.118343 4776 generic.go:334] "Generic (PLEG): container finished" podID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerID="05b65f4bed8c7914cf657df0f540c0d19593a2c84c93c9ae6c4f106530470ec4" exitCode=0 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.118351 4776 generic.go:334] "Generic (PLEG): container finished" podID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerID="9e7cf1b3a65acbcaaedff95080720317c88059b1b6b759f4434c871973a12bc3" exitCode=0 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.117903 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerDied","Data":"152026f7bf460a5258df8a4110443a6e694c32ee892ab2dcb9469e4391545baa"} Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.118454 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerDied","Data":"8e84b91c7719175c8908ad8fa6987a175f46744663fe91972c8bbb602ae871d8"} Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.118467 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerDied","Data":"49bb08566118dc9e19698c8156598960f4778c7e09f832e3eeeaa9b6d4fc9a62"} Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.118477 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerDied","Data":"670e0e074022b9b3b944ddaeb78954a059e4c4b8e2f7af4e2f91000daddc1263"} Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.118485 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerDied","Data":"a5e5c6ad114abc906cc7868aa09dc3d1a6a83ffb494d14de45deba5107b01b02"} Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.118516 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerDied","Data":"2c713a54dd52d79687c4437832031c9afbb02bd373076897a51654ae360ff22a"} Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.118527 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerDied","Data":"f0cdb71ed47fdaaa7eabd7f81c93a293616c6c022e1c4d8cef28b88de0dfb8d5"} Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.118534 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerDied","Data":"434eabc5a25e34093e38beb759273c9fac1b8758c3a03273e4ddb5f564fcdc08"} Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.118544 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerDied","Data":"a4e815cd4ca77d4505a242b624a83fcda104ea6e85674eb4c15ddc064562ffd6"} Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.118553 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerDied","Data":"e2072d3b3f2b295702755794df1a3f5a2a130db157825a15a28d4ac1c3a1c674"} Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.118562 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerDied","Data":"cd073bb8b2d7b566c99b6e569d6dbfa6b16b08d5db2e84e47a8d2d85eb563f1b"} Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.118571 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerDied","Data":"6ccd151880ed401c0372e024b6b953cdc0eba4d9d9e8cdd47435d059001cbc30"} Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.118602 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerDied","Data":"05b65f4bed8c7914cf657df0f540c0d19593a2c84c93c9ae6c4f106530470ec4"} Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.118611 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerDied","Data":"9e7cf1b3a65acbcaaedff95080720317c88059b1b6b759f4434c871973a12bc3"} Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.131201 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.136904 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df" (UID: "7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.140590 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-kube-api-access-wvrvj" (OuterVolumeSpecName: "kube-api-access-wvrvj") pod "7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df" (UID: "7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df"). InnerVolumeSpecName "kube-api-access-wvrvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.140944 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ab346c9-f70f-4663-9b53-67b4d66b3112-kube-api-access-p7g6c" (OuterVolumeSpecName: "kube-api-access-p7g6c") pod "9ab346c9-f70f-4663-9b53-67b4d66b3112" (UID: "9ab346c9-f70f-4663-9b53-67b4d66b3112"). InnerVolumeSpecName "kube-api-access-p7g6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.145045 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df/ovsdbserver-nb/0.log" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.145132 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df","Type":"ContainerDied","Data":"82692ea71c99491423198e874e369d17ef5e820a3bebef5b7b1bc955ce2bf61e"} Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.145167 4776 scope.go:117] "RemoveContainer" containerID="ee2853983a149e403a868c5541f34466a714bc2a4620f7f4820109d6bf8f4f6c" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.145285 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.156167 4776 generic.go:334] "Generic (PLEG): container finished" podID="a0b97845-4578-4801-896a-281fdafdb351" containerID="5fac5a59b29d2ae80cf45865e0db91319f65183c12c9b736093c4452a98a4355" exitCode=143 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.156239 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-667769d556-lqtlp" event={"ID":"a0b97845-4578-4801-896a-281fdafdb351","Type":"ContainerDied","Data":"5fac5a59b29d2ae80cf45865e0db91319f65183c12c9b736093c4452a98a4355"} Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.171929 4776 generic.go:334] "Generic (PLEG): container finished" podID="86e320e0-80ee-40df-b3a1-a48cb810a435" containerID="643bd96dcd87a61d61b437895cd783519ab3ff279196abf17121bbe3e4d485dd" exitCode=143 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.172009 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"86e320e0-80ee-40df-b3a1-a48cb810a435","Type":"ContainerDied","Data":"643bd96dcd87a61d61b437895cd783519ab3ff279196abf17121bbe3e4d485dd"} Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.186470 4776 scope.go:117] "RemoveContainer" containerID="215609fe2a7a4ad5088201866359a4ff41f487eaa766e53084fae555576c9ef0" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.193365 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df" (UID: "7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.198882 4776 generic.go:334] "Generic (PLEG): container finished" podID="d2cda6de-ef12-42f4-a6e5-824c2db3cd01" containerID="0bd7925c5eadaefd62b355bbd7aebe47ce493b779918483df9c72dc9a141944b" exitCode=143 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.199287 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d2cda6de-ef12-42f4-a6e5-824c2db3cd01","Type":"ContainerDied","Data":"0bd7925c5eadaefd62b355bbd7aebe47ce493b779918483df9c72dc9a141944b"} Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.201261 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3fc11cf2-c6c7-4250-b7f4-9c68d78152eb-openstack-config-secret\") pod \"3fc11cf2-c6c7-4250-b7f4-9c68d78152eb\" (UID: \"3fc11cf2-c6c7-4250-b7f4-9c68d78152eb\") " Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.202312 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3fc11cf2-c6c7-4250-b7f4-9c68d78152eb-openstack-config\") pod \"3fc11cf2-c6c7-4250-b7f4-9c68d78152eb\" (UID: \"3fc11cf2-c6c7-4250-b7f4-9c68d78152eb\") " Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.202885 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xk879\" (UniqueName: \"kubernetes.io/projected/3fc11cf2-c6c7-4250-b7f4-9c68d78152eb-kube-api-access-xk879\") pod \"3fc11cf2-c6c7-4250-b7f4-9c68d78152eb\" (UID: \"3fc11cf2-c6c7-4250-b7f4-9c68d78152eb\") " Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.203007 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fc11cf2-c6c7-4250-b7f4-9c68d78152eb-combined-ca-bundle\") pod \"3fc11cf2-c6c7-4250-b7f4-9c68d78152eb\" (UID: \"3fc11cf2-c6c7-4250-b7f4-9c68d78152eb\") " Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.204930 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.205006 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.205038 4776 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.208073 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.208116 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.208128 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7g6c\" (UniqueName: \"kubernetes.io/projected/9ab346c9-f70f-4663-9b53-67b4d66b3112-kube-api-access-p7g6c\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.208144 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvrvj\" (UniqueName: \"kubernetes.io/projected/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-kube-api-access-wvrvj\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.211773 4776 generic.go:334] "Generic (PLEG): container finished" podID="3fc11cf2-c6c7-4250-b7f4-9c68d78152eb" containerID="5b70abe095c3919e69bde8512f0bf19cca9ec9011ee75d534001611515145f6c" exitCode=137 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.212018 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.233756 4776 scope.go:117] "RemoveContainer" containerID="5b70abe095c3919e69bde8512f0bf19cca9ec9011ee75d534001611515145f6c" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.234663 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fc11cf2-c6c7-4250-b7f4-9c68d78152eb-kube-api-access-xk879" (OuterVolumeSpecName: "kube-api-access-xk879") pod "3fc11cf2-c6c7-4250-b7f4-9c68d78152eb" (UID: "3fc11cf2-c6c7-4250-b7f4-9c68d78152eb"). InnerVolumeSpecName "kube-api-access-xk879". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.245896 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7nsh6"] Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.251178 4776 generic.go:334] "Generic (PLEG): container finished" podID="b708b100-166c-49b1-a47b-76d47ba2c6e5" containerID="7f99367ae1299a4af7ca7bdb87418a9b017add504f1ee9377bbcccc241edd6fb" exitCode=143 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.251256 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b708b100-166c-49b1-a47b-76d47ba2c6e5","Type":"ContainerDied","Data":"7f99367ae1299a4af7ca7bdb87418a9b017add504f1ee9377bbcccc241edd6fb"} Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.258532 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.258764 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="f182a7e0-ebd8-4258-9269-43a662e39af8" containerName="nova-cell1-conductor-conductor" containerID="cri-o://e5a40e8fd3cdfddb16f8d8578ce8ae7e8bb1b0df1b91c68efe717a27f4fc382c" gracePeriod=30 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.269480 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_6e7d9d60-9538-4870-8f62-434fd9f1ab0d/ovsdbserver-sb/0.log" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.276279 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd" (UID: "cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.276459 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6e7d9d60-9538-4870-8f62-434fd9f1ab0d","Type":"ContainerDied","Data":"c5c65c194be6d770911d8e35680f49902817102fc8b9409f79be2693b39814bd"} Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.276544 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.281899 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9ab346c9-f70f-4663-9b53-67b4d66b3112" (UID: "9ab346c9-f70f-4663-9b53-67b4d66b3112"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.286441 4776 scope.go:117] "RemoveContainer" containerID="17cb0f155d2722e145dfbd65de7d057bce9f1ad8f11fcd4b4347b19f7b5ce218" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.286461 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7nsh6"] Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.298549 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-2mts9_cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd/openstack-network-exporter/0.log" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.298673 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-2mts9" event={"ID":"cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd","Type":"ContainerDied","Data":"01da0db65f363a5a31d740c6ee7385aca9b9ea4986f9368086fdf9521ab0b01a"} Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.298827 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-2mts9" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.319664 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.319712 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-config" (OuterVolumeSpecName: "config") pod "9ab346c9-f70f-4663-9b53-67b4d66b3112" (UID: "9ab346c9-f70f-4663-9b53-67b4d66b3112"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.320198 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4" containerName="nova-cell0-conductor-conductor" containerID="cri-o://51468cb0acece8ce1a386bdf49793e6322b08cddac5cf23fc46abef1f5330e22" gracePeriod=30 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.320897 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.320921 4776 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.320931 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.320945 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xk879\" (UniqueName: \"kubernetes.io/projected/3fc11cf2-c6c7-4250-b7f4-9c68d78152eb-kube-api-access-xk879\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.326393 4776 scope.go:117] "RemoveContainer" containerID="ce501a13d7bab63cc6c0c2608dcb59e6fdc92a7de550426b2ee3b8ad29264e30" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.327694 4776 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.327765 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jfmbd"] Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.333387 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9ab346c9-f70f-4663-9b53-67b4d66b3112" (UID: "9ab346c9-f70f-4663-9b53-67b4d66b3112"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.342010 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jfmbd"] Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.347962 4776 generic.go:334] "Generic (PLEG): container finished" podID="6bf3edad-487a-4c68-9dbf-9789a94f8fba" containerID="07109bfcbb050b4c8db424a410626fa31b2236c8d915084ff3b34d220f1c1cf0" exitCode=143 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.348119 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6bf3edad-487a-4c68-9dbf-9789a94f8fba","Type":"ContainerDied","Data":"07109bfcbb050b4c8db424a410626fa31b2236c8d915084ff3b34d220f1c1cf0"} Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.352655 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="5920f152-04cd-4366-b479-665a1ce8937c" containerName="galera" containerID="cri-o://fb166b129931ff672d247f77f4c3b881346f25b0410d683eb6bb28f72d3af0a4" gracePeriod=30 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.355631 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.361013 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9ab346c9-f70f-4663-9b53-67b4d66b3112" (UID: "9ab346c9-f70f-4663-9b53-67b4d66b3112"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.362323 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.362633 4776 generic.go:334] "Generic (PLEG): container finished" podID="9ab346c9-f70f-4663-9b53-67b4d66b3112" containerID="2b9a66a95781ceb011f5c668fc0302d4aa355e258935b0bc8a74577b415062f6" exitCode=0 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.362682 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" event={"ID":"9ab346c9-f70f-4663-9b53-67b4d66b3112","Type":"ContainerDied","Data":"2b9a66a95781ceb011f5c668fc0302d4aa355e258935b0bc8a74577b415062f6"} Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.362706 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" event={"ID":"9ab346c9-f70f-4663-9b53-67b4d66b3112","Type":"ContainerDied","Data":"e2ea8ecd0422a4e5db4d8eaedcbaa39313dd266172bfa7165de56b2e091c7147"} Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.362800 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7f54fb65-td75r" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.363510 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fc11cf2-c6c7-4250-b7f4-9c68d78152eb-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "3fc11cf2-c6c7-4250-b7f4-9c68d78152eb" (UID: "3fc11cf2-c6c7-4250-b7f4-9c68d78152eb"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.367253 4776 scope.go:117] "RemoveContainer" containerID="29696e80dd0a6c71a81e7417db086e3acded64a814a8e862ba2fc208da55d802" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.392945 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fc11cf2-c6c7-4250-b7f4-9c68d78152eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3fc11cf2-c6c7-4250-b7f4-9c68d78152eb" (UID: "3fc11cf2-c6c7-4250-b7f4-9c68d78152eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.423350 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.423402 4776 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3fc11cf2-c6c7-4250-b7f4-9c68d78152eb-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.423411 4776 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.423419 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fc11cf2-c6c7-4250-b7f4-9c68d78152eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.423427 4776 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.424237 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3fc11cf2-c6c7-4250-b7f4-9c68d78152eb-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "3fc11cf2-c6c7-4250-b7f4-9c68d78152eb" (UID: "3fc11cf2-c6c7-4250-b7f4-9c68d78152eb"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.426254 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-2mts9"] Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.440823 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-2mts9"] Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.442199 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9ab346c9-f70f-4663-9b53-67b4d66b3112" (UID: "9ab346c9-f70f-4663-9b53-67b4d66b3112"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.447226 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df" (UID: "7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.648222 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-5864b7d7fc-g9z2r"] Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.648497 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-5864b7d7fc-g9z2r" podUID="f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9" containerName="proxy-httpd" containerID="cri-o://8c4015a72b16a36f9de326c38c3b6c1efa071b368ac1bacb90a14ea00d371022" gracePeriod=30 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.648937 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-5864b7d7fc-g9z2r" podUID="f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9" containerName="proxy-server" containerID="cri-o://94b41979aa9c10d3bd3737ef33206925359824fb82b712cb4748abcf10f1e055" gracePeriod=30 Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.661617 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ab346c9-f70f-4663-9b53-67b4d66b3112-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.661657 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.661674 4776 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3fc11cf2-c6c7-4250-b7f4-9c68d78152eb-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.680831 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19f20b56-a02b-4831-932c-0a58f91544da" path="/var/lib/kubelet/pods/19f20b56-a02b-4831-932c-0a58f91544da/volumes" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.681568 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22bb0550-ee44-46e6-b45e-1158524bfa3e" path="/var/lib/kubelet/pods/22bb0550-ee44-46e6-b45e-1158524bfa3e/volumes" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.682177 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fc11cf2-c6c7-4250-b7f4-9c68d78152eb" path="/var/lib/kubelet/pods/3fc11cf2-c6c7-4250-b7f4-9c68d78152eb/volumes" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.683650 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ff2a086-8bce-46e2-b294-3b3f31e39b4e" path="/var/lib/kubelet/pods/3ff2a086-8bce-46e2-b294-3b3f31e39b4e/volumes" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.684400 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="610a6a76-1c03-49a0-850f-ee5131308f7b" path="/var/lib/kubelet/pods/610a6a76-1c03-49a0-850f-ee5131308f7b/volumes" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.685300 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e7d9d60-9538-4870-8f62-434fd9f1ab0d" path="/var/lib/kubelet/pods/6e7d9d60-9538-4870-8f62-434fd9f1ab0d/volumes" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.686918 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="737e5e32-1fa7-4140-bfad-1bc3e10ee390" path="/var/lib/kubelet/pods/737e5e32-1fa7-4140-bfad-1bc3e10ee390/volumes" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.687503 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="758472ef-bebc-469a-87a0-e7f1a83ae279" path="/var/lib/kubelet/pods/758472ef-bebc-469a-87a0-e7f1a83ae279/volumes" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.688166 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cae7ff0c-7548-455d-85c8-3e68554658ce" path="/var/lib/kubelet/pods/cae7ff0c-7548-455d-85c8-3e68554658ce/volumes" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.688717 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515" path="/var/lib/kubelet/pods/cb65c0ff-3657-4a53-ac0b-ca0b5a9fb515/volumes" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.689752 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc2a0874-1575-4605-afb9-01a79c958ba5" path="/var/lib/kubelet/pods/cc2a0874-1575-4605-afb9-01a79c958ba5/volumes" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.690445 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd" path="/var/lib/kubelet/pods/cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd/volumes" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.690999 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d32cd325-5a9d-4a3b-873f-fccab96da880" path="/var/lib/kubelet/pods/d32cd325-5a9d-4a3b-873f-fccab96da880/volumes" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.692176 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e522431f-0956-49f7-9c46-b0202824bd56" path="/var/lib/kubelet/pods/e522431f-0956-49f7-9c46-b0202824bd56/volumes" Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.712616 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d7f54fb65-td75r"] Nov 25 09:47:57 crc kubenswrapper[4776]: I1125 09:47:57.727814 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d7f54fb65-td75r"] Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.254891 4776 scope.go:117] "RemoveContainer" containerID="2b9a66a95781ceb011f5c668fc0302d4aa355e258935b0bc8a74577b415062f6" Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.262736 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df" (UID: "7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:47:58 crc kubenswrapper[4776]: E1125 09:47:58.263353 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e5a40e8fd3cdfddb16f8d8578ce8ae7e8bb1b0df1b91c68efe717a27f4fc382c" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 25 09:47:58 crc kubenswrapper[4776]: E1125 09:47:58.267292 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e5a40e8fd3cdfddb16f8d8578ce8ae7e8bb1b0df1b91c68efe717a27f4fc382c" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.292188 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbicandb0b-account-delete-xtvnn"] Nov 25 09:47:58 crc kubenswrapper[4776]: E1125 09:47:58.304022 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e5a40e8fd3cdfddb16f8d8578ce8ae7e8bb1b0df1b91c68efe717a27f4fc382c" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 25 09:47:58 crc kubenswrapper[4776]: E1125 09:47:58.304114 4776 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="f182a7e0-ebd8-4258-9269-43a662e39af8" containerName="nova-cell1-conductor-conductor" Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.306510 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placemente176-account-delete-wbgz5"] Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.345174 4776 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:58 crc kubenswrapper[4776]: E1125 09:47:58.353735 4776 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 25 09:47:58 crc kubenswrapper[4776]: E1125 09:47:58.353821 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f32041a0-a8fc-48a5-afab-476baff89e8c-config-data podName:f32041a0-a8fc-48a5-afab-476baff89e8c nodeName:}" failed. No retries permitted until 2025-11-25 09:48:02.353799014 +0000 UTC m=+1427.394858567 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/f32041a0-a8fc-48a5-afab-476baff89e8c-config-data") pod "rabbitmq-cell1-server-0" (UID: "f32041a0-a8fc-48a5-afab-476baff89e8c") : configmap "rabbitmq-cell1-config-data" not found Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.380812 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell09dc4-account-delete-w5xcj"] Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.418956 4776 generic.go:334] "Generic (PLEG): container finished" podID="bc920d71-4308-4d90-a891-8b760651ccc9" containerID="c9f1b69c74818d42cf402965ae75f33a682f61bb14e9874a7f7904a0e8298e5e" exitCode=0 Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.419061 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bc920d71-4308-4d90-a891-8b760651ccc9","Type":"ContainerDied","Data":"c9f1b69c74818d42cf402965ae75f33a682f61bb14e9874a7f7904a0e8298e5e"} Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.428566 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder774d-account-delete-b29n5"] Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.429939 4776 generic.go:334] "Generic (PLEG): container finished" podID="e51292cd-00f5-4e0a-ba2b-6319bd8a5e45" containerID="a1a93104c41ae0baf6a1781b8d29302ccef9fef9da152e6274b2e92e08df9a8c" exitCode=143 Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.430006 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-56988fbb4-nlxs9" event={"ID":"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45","Type":"ContainerDied","Data":"a1a93104c41ae0baf6a1781b8d29302ccef9fef9da152e6274b2e92e08df9a8c"} Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.432305 4776 generic.go:334] "Generic (PLEG): container finished" podID="4702d881-d701-41a4-b36f-2d063a8ae246" containerID="d45871f326a3dbb0f683b3d995f92f15682fd356f7dd9ec10b42a715df612c77" exitCode=143 Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.432351 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-769ddf488d-46ssj" event={"ID":"4702d881-d701-41a4-b36f-2d063a8ae246","Type":"ContainerDied","Data":"d45871f326a3dbb0f683b3d995f92f15682fd356f7dd9ec10b42a715df612c77"} Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.443450 4776 generic.go:334] "Generic (PLEG): container finished" podID="1d1e9a91-e344-442f-8bbc-90b1f98e2e1f" containerID="87f8cbf578187a46d6e2d2c1082d1ed105b039c2788db71db6340e4c1b8457d4" exitCode=143 Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.443657 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-66cdf565f-6lzqw" event={"ID":"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f","Type":"ContainerDied","Data":"87f8cbf578187a46d6e2d2c1082d1ed105b039c2788db71db6340e4c1b8457d4"} Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.463858 4776 generic.go:334] "Generic (PLEG): container finished" podID="10fe3dc2-2bde-4380-8ad3-d71832e31095" containerID="eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06" exitCode=0 Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.463897 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7fgjf" event={"ID":"10fe3dc2-2bde-4380-8ad3-d71832e31095","Type":"ContainerDied","Data":"eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06"} Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.471896 4776 generic.go:334] "Generic (PLEG): container finished" podID="fe3ff67b-01ff-480f-be91-7c8235593c97" containerID="5c1129db350942a00317a1332a9a8db16c63fbd966ba3d63890450c144964ddb" exitCode=0 Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.471930 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54dbffbb57-kmqfb" event={"ID":"fe3ff67b-01ff-480f-be91-7c8235593c97","Type":"ContainerDied","Data":"5c1129db350942a00317a1332a9a8db16c63fbd966ba3d63890450c144964ddb"} Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.476650 4776 generic.go:334] "Generic (PLEG): container finished" podID="f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9" containerID="94b41979aa9c10d3bd3737ef33206925359824fb82b712cb4748abcf10f1e055" exitCode=0 Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.476704 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5864b7d7fc-g9z2r" event={"ID":"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9","Type":"ContainerDied","Data":"94b41979aa9c10d3bd3737ef33206925359824fb82b712cb4748abcf10f1e055"} Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.480172 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glanceb9b5-account-delete-tjrnr"] Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.497333 4776 generic.go:334] "Generic (PLEG): container finished" podID="e09f9958-4faf-4a28-8214-c3ead146122c" containerID="e056e61a1527b24d984169499b3cde97a02136661c62174d6fe28407dfd3287a" exitCode=0 Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.497389 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e09f9958-4faf-4a28-8214-c3ead146122c","Type":"ContainerDied","Data":"e056e61a1527b24d984169499b3cde97a02136661c62174d6fe28407dfd3287a"} Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.518137 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron5151-account-delete-8vbjg"] Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.530049 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapi3b46-account-delete-9w5r9"] Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.561121 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.579002 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 25 09:47:58 crc kubenswrapper[4776]: E1125 09:47:58.625776 4776 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf89d44cf_1ca0_4581_aa8c_e86e0f3c61d9.slice/crio-94b41979aa9c10d3bd3737ef33206925359824fb82b712cb4748abcf10f1e055.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a8cfa42_c498_4a9c_bb2b_cd7939c7a8df.slice/crio-82692ea71c99491423198e874e369d17ef5e820a3bebef5b7b1bc955ce2bf61e\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode09f9958_4faf_4a28_8214_c3ead146122c.slice/crio-conmon-e056e61a1527b24d984169499b3cde97a02136661c62174d6fe28407dfd3287a.scope\": RecentStats: unable to find data in memory cache]" Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.671922 4776 scope.go:117] "RemoveContainer" containerID="59870a401f6a788fbbb94bfbf471e78e459d6e2cbdf6f95acf0bf39015b4d0ff" Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.900816 4776 scope.go:117] "RemoveContainer" containerID="2b9a66a95781ceb011f5c668fc0302d4aa355e258935b0bc8a74577b415062f6" Nov 25 09:47:58 crc kubenswrapper[4776]: E1125 09:47:58.901612 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b9a66a95781ceb011f5c668fc0302d4aa355e258935b0bc8a74577b415062f6\": container with ID starting with 2b9a66a95781ceb011f5c668fc0302d4aa355e258935b0bc8a74577b415062f6 not found: ID does not exist" containerID="2b9a66a95781ceb011f5c668fc0302d4aa355e258935b0bc8a74577b415062f6" Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.901659 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b9a66a95781ceb011f5c668fc0302d4aa355e258935b0bc8a74577b415062f6"} err="failed to get container status \"2b9a66a95781ceb011f5c668fc0302d4aa355e258935b0bc8a74577b415062f6\": rpc error: code = NotFound desc = could not find container \"2b9a66a95781ceb011f5c668fc0302d4aa355e258935b0bc8a74577b415062f6\": container with ID starting with 2b9a66a95781ceb011f5c668fc0302d4aa355e258935b0bc8a74577b415062f6 not found: ID does not exist" Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.901686 4776 scope.go:117] "RemoveContainer" containerID="59870a401f6a788fbbb94bfbf471e78e459d6e2cbdf6f95acf0bf39015b4d0ff" Nov 25 09:47:58 crc kubenswrapper[4776]: E1125 09:47:58.902357 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59870a401f6a788fbbb94bfbf471e78e459d6e2cbdf6f95acf0bf39015b4d0ff\": container with ID starting with 59870a401f6a788fbbb94bfbf471e78e459d6e2cbdf6f95acf0bf39015b4d0ff not found: ID does not exist" containerID="59870a401f6a788fbbb94bfbf471e78e459d6e2cbdf6f95acf0bf39015b4d0ff" Nov 25 09:47:58 crc kubenswrapper[4776]: I1125 09:47:58.902463 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59870a401f6a788fbbb94bfbf471e78e459d6e2cbdf6f95acf0bf39015b4d0ff"} err="failed to get container status \"59870a401f6a788fbbb94bfbf471e78e459d6e2cbdf6f95acf0bf39015b4d0ff\": rpc error: code = NotFound desc = could not find container \"59870a401f6a788fbbb94bfbf471e78e459d6e2cbdf6f95acf0bf39015b4d0ff\": container with ID starting with 59870a401f6a788fbbb94bfbf471e78e459d6e2cbdf6f95acf0bf39015b4d0ff not found: ID does not exist" Nov 25 09:47:58 crc kubenswrapper[4776]: E1125 09:47:58.910359 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1ed6008959fa3368ff11289c84925032c1050a0bd2f104218d9763b1be776a55" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 25 09:47:58 crc kubenswrapper[4776]: E1125 09:47:58.914854 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1ed6008959fa3368ff11289c84925032c1050a0bd2f104218d9763b1be776a55" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 25 09:47:58 crc kubenswrapper[4776]: E1125 09:47:58.916667 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1ed6008959fa3368ff11289c84925032c1050a0bd2f104218d9763b1be776a55" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Nov 25 09:47:58 crc kubenswrapper[4776]: E1125 09:47:58.916705 4776 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="7dadc68f-0b89-41ba-84f8-e9bcde16ca64" containerName="ovn-northd" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.036950 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-5864b7d7fc-g9z2r" podUID="f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9" containerName="proxy-server" probeResult="failure" output="Get \"https://10.217.0.168:8080/healthcheck\": read tcp 10.217.0.2:53932->10.217.0.168:8080: read: connection reset by peer" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.037002 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-5864b7d7fc-g9z2r" podUID="f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.168:8080/healthcheck\": read tcp 10.217.0.2:53948->10.217.0.168:8080: read: connection reset by peer" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.043426 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.062566 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e09f9958-4faf-4a28-8214-c3ead146122c-vencrypt-tls-certs\") pod \"e09f9958-4faf-4a28-8214-c3ead146122c\" (UID: \"e09f9958-4faf-4a28-8214-c3ead146122c\") " Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.062779 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8b28\" (UniqueName: \"kubernetes.io/projected/e09f9958-4faf-4a28-8214-c3ead146122c-kube-api-access-s8b28\") pod \"e09f9958-4faf-4a28-8214-c3ead146122c\" (UID: \"e09f9958-4faf-4a28-8214-c3ead146122c\") " Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.062960 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e09f9958-4faf-4a28-8214-c3ead146122c-config-data\") pod \"e09f9958-4faf-4a28-8214-c3ead146122c\" (UID: \"e09f9958-4faf-4a28-8214-c3ead146122c\") " Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.063035 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e09f9958-4faf-4a28-8214-c3ead146122c-combined-ca-bundle\") pod \"e09f9958-4faf-4a28-8214-c3ead146122c\" (UID: \"e09f9958-4faf-4a28-8214-c3ead146122c\") " Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.063149 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e09f9958-4faf-4a28-8214-c3ead146122c-nova-novncproxy-tls-certs\") pod \"e09f9958-4faf-4a28-8214-c3ead146122c\" (UID: \"e09f9958-4faf-4a28-8214-c3ead146122c\") " Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.068745 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e09f9958-4faf-4a28-8214-c3ead146122c-kube-api-access-s8b28" (OuterVolumeSpecName: "kube-api-access-s8b28") pod "e09f9958-4faf-4a28-8214-c3ead146122c" (UID: "e09f9958-4faf-4a28-8214-c3ead146122c"). InnerVolumeSpecName "kube-api-access-s8b28". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.100965 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e09f9958-4faf-4a28-8214-c3ead146122c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e09f9958-4faf-4a28-8214-c3ead146122c" (UID: "e09f9958-4faf-4a28-8214-c3ead146122c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.109558 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e09f9958-4faf-4a28-8214-c3ead146122c-config-data" (OuterVolumeSpecName: "config-data") pod "e09f9958-4faf-4a28-8214-c3ead146122c" (UID: "e09f9958-4faf-4a28-8214-c3ead146122c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.130783 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e09f9958-4faf-4a28-8214-c3ead146122c-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "e09f9958-4faf-4a28-8214-c3ead146122c" (UID: "e09f9958-4faf-4a28-8214-c3ead146122c"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.160383 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e09f9958-4faf-4a28-8214-c3ead146122c-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "e09f9958-4faf-4a28-8214-c3ead146122c" (UID: "e09f9958-4faf-4a28-8214-c3ead146122c"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.194745 4776 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e09f9958-4faf-4a28-8214-c3ead146122c-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.194784 4776 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e09f9958-4faf-4a28-8214-c3ead146122c-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.194794 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8b28\" (UniqueName: \"kubernetes.io/projected/e09f9958-4faf-4a28-8214-c3ead146122c-kube-api-access-s8b28\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.194807 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e09f9958-4faf-4a28-8214-c3ead146122c-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.194818 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e09f9958-4faf-4a28-8214-c3ead146122c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:59 crc kubenswrapper[4776]: E1125 09:47:59.194897 4776 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 25 09:47:59 crc kubenswrapper[4776]: E1125 09:47:59.194948 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4512cf4e-6c75-493e-a80a-4b0accd750a8-config-data podName:4512cf4e-6c75-493e-a80a-4b0accd750a8 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:03.194934113 +0000 UTC m=+1428.235993666 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/4512cf4e-6c75-493e-a80a-4b0accd750a8-config-data") pod "rabbitmq-server-0" (UID: "4512cf4e-6c75-493e-a80a-4b0accd750a8") : configmap "rabbitmq-config-data" not found Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.215741 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.295673 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bc920d71-4308-4d90-a891-8b760651ccc9-etc-machine-id\") pod \"bc920d71-4308-4d90-a891-8b760651ccc9\" (UID: \"bc920d71-4308-4d90-a891-8b760651ccc9\") " Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.295737 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc920d71-4308-4d90-a891-8b760651ccc9-combined-ca-bundle\") pod \"bc920d71-4308-4d90-a891-8b760651ccc9\" (UID: \"bc920d71-4308-4d90-a891-8b760651ccc9\") " Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.295758 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc920d71-4308-4d90-a891-8b760651ccc9-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "bc920d71-4308-4d90-a891-8b760651ccc9" (UID: "bc920d71-4308-4d90-a891-8b760651ccc9"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.295838 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc920d71-4308-4d90-a891-8b760651ccc9-scripts\") pod \"bc920d71-4308-4d90-a891-8b760651ccc9\" (UID: \"bc920d71-4308-4d90-a891-8b760651ccc9\") " Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.295882 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc920d71-4308-4d90-a891-8b760651ccc9-config-data\") pod \"bc920d71-4308-4d90-a891-8b760651ccc9\" (UID: \"bc920d71-4308-4d90-a891-8b760651ccc9\") " Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.295974 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9686x\" (UniqueName: \"kubernetes.io/projected/bc920d71-4308-4d90-a891-8b760651ccc9-kube-api-access-9686x\") pod \"bc920d71-4308-4d90-a891-8b760651ccc9\" (UID: \"bc920d71-4308-4d90-a891-8b760651ccc9\") " Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.296003 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bc920d71-4308-4d90-a891-8b760651ccc9-config-data-custom\") pod \"bc920d71-4308-4d90-a891-8b760651ccc9\" (UID: \"bc920d71-4308-4d90-a891-8b760651ccc9\") " Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.296540 4776 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bc920d71-4308-4d90-a891-8b760651ccc9-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.302572 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc920d71-4308-4d90-a891-8b760651ccc9-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "bc920d71-4308-4d90-a891-8b760651ccc9" (UID: "bc920d71-4308-4d90-a891-8b760651ccc9"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.304270 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc920d71-4308-4d90-a891-8b760651ccc9-scripts" (OuterVolumeSpecName: "scripts") pod "bc920d71-4308-4d90-a891-8b760651ccc9" (UID: "bc920d71-4308-4d90-a891-8b760651ccc9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.306183 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc920d71-4308-4d90-a891-8b760651ccc9-kube-api-access-9686x" (OuterVolumeSpecName: "kube-api-access-9686x") pod "bc920d71-4308-4d90-a891-8b760651ccc9" (UID: "bc920d71-4308-4d90-a891-8b760651ccc9"). InnerVolumeSpecName "kube-api-access-9686x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.398010 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc920d71-4308-4d90-a891-8b760651ccc9-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.398052 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9686x\" (UniqueName: \"kubernetes.io/projected/bc920d71-4308-4d90-a891-8b760651ccc9-kube-api-access-9686x\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.398077 4776 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bc920d71-4308-4d90-a891-8b760651ccc9-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.571999 4776 generic.go:334] "Generic (PLEG): container finished" podID="bc920d71-4308-4d90-a891-8b760651ccc9" containerID="6bffff98e55f5e2fe0110f4a9d127c1dff4b4c552606101a667c0fb53d5e9b16" exitCode=0 Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.572591 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.573879 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bc920d71-4308-4d90-a891-8b760651ccc9","Type":"ContainerDied","Data":"6bffff98e55f5e2fe0110f4a9d127c1dff4b4c552606101a667c0fb53d5e9b16"} Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.573919 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bc920d71-4308-4d90-a891-8b760651ccc9","Type":"ContainerDied","Data":"a26a73284ec7ca88f3d4569feec4777b80aa3ca2f76bd5f8814d4c904b763b0d"} Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.573943 4776 scope.go:117] "RemoveContainer" containerID="c9f1b69c74818d42cf402965ae75f33a682f61bb14e9874a7f7904a0e8298e5e" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.584259 4776 generic.go:334] "Generic (PLEG): container finished" podID="f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9" containerID="8c4015a72b16a36f9de326c38c3b6c1efa071b368ac1bacb90a14ea00d371022" exitCode=0 Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.584322 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5864b7d7fc-g9z2r" event={"ID":"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9","Type":"ContainerDied","Data":"8c4015a72b16a36f9de326c38c3b6c1efa071b368ac1bacb90a14ea00d371022"} Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.590647 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glanceb9b5-account-delete-tjrnr" event={"ID":"bcaec0f3-13af-4f73-8e7f-7e575b8e2008","Type":"ContainerStarted","Data":"4a42c4c2fcc9fc1a6adb53fc048a0c22c503282432d6521db1b327054c0ea919"} Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.590683 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glanceb9b5-account-delete-tjrnr" event={"ID":"bcaec0f3-13af-4f73-8e7f-7e575b8e2008","Type":"ContainerStarted","Data":"abe37879060260786a380d4efa643725c294e581dc2808c91e62da1b2dd37969"} Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.594431 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi3b46-account-delete-9w5r9" event={"ID":"14b51bba-0b15-469b-bdea-90fae8d9b976","Type":"ContainerStarted","Data":"5d66e9e059cc4ecbbbdd54624c88073c9a95bd315ce5454ae88b21b09a0588b0"} Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.598221 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e09f9958-4faf-4a28-8214-c3ead146122c","Type":"ContainerDied","Data":"184f9329ca4bc16ca9830f25689b95cf6dba545b914093958315b55ece208bbd"} Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.598343 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.600741 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron5151-account-delete-8vbjg" event={"ID":"73f2c2ab-6cb6-4544-a3f7-20722e5b6400","Type":"ContainerStarted","Data":"b773b52fed0b22bef9171161cc75302f0462620c1fb47cb91658f91785551dc1"} Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.616644 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder774d-account-delete-b29n5" event={"ID":"ef85f90f-c377-4273-9fff-451420d6008a","Type":"ContainerStarted","Data":"5e4df52a7d69593f81f35d33034b747cbf372fe6bdfff6a79400c8d265865396"} Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.624412 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.624718 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aab7d356-ec67-4e5d-9c25-f26e39786f79" containerName="ceilometer-central-agent" containerID="cri-o://a4df2757871f92f27d85077692ce959d3142d7496b28c90b9062a0ed0d5d450d" gracePeriod=30 Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.624732 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aab7d356-ec67-4e5d-9c25-f26e39786f79" containerName="sg-core" containerID="cri-o://c99e1a9b60f21bc937fd065f74ab82371110943731ae37f120fc60b0c71336da" gracePeriod=30 Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.624767 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aab7d356-ec67-4e5d-9c25-f26e39786f79" containerName="proxy-httpd" containerID="cri-o://ec693b92c2b81399b4a2d57c1980de7b988f0cc59e2cef0ec21c34b5dff21bd2" gracePeriod=30 Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.624814 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aab7d356-ec67-4e5d-9c25-f26e39786f79" containerName="ceilometer-notification-agent" containerID="cri-o://a2a4c7e738a8387346bc8cb061fcaffab37035f8b7356aba874420f641555783" gracePeriod=30 Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.634210 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glanceb9b5-account-delete-tjrnr" podStartSLOduration=5.634186578 podStartE2EDuration="5.634186578s" podCreationTimestamp="2025-11-25 09:47:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:47:59.616558275 +0000 UTC m=+1424.657617838" watchObservedRunningTime="2025-11-25 09:47:59.634186578 +0000 UTC m=+1424.675246131" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.643585 4776 generic.go:334] "Generic (PLEG): container finished" podID="6d40516c-5fcc-4b24-918f-95bf79cb94b4" containerID="4a672773256a5735e511e0d75fb11f455144c24c1ad5bbd0829cbe34523b873d" exitCode=0 Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.643929 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicandb0b-account-delete-xtvnn" event={"ID":"6d40516c-5fcc-4b24-918f-95bf79cb94b4","Type":"ContainerDied","Data":"4a672773256a5735e511e0d75fb11f455144c24c1ad5bbd0829cbe34523b873d"} Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.643983 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicandb0b-account-delete-xtvnn" event={"ID":"6d40516c-5fcc-4b24-918f-95bf79cb94b4","Type":"ContainerStarted","Data":"512b6cdcd78de19c4a57934ea8b56bff013e1d10be30032755b9e655ec66cdd1"} Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.649598 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placemente176-account-delete-wbgz5" event={"ID":"6be099e6-89d6-45a0-9d0f-4dca4d91701a","Type":"ContainerStarted","Data":"d7ddbba5ee21f93a892892dc7ab2ec9efca17632ed5b4c898a4b773f4b5b6b2c"} Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.676465 4776 generic.go:334] "Generic (PLEG): container finished" podID="5920f152-04cd-4366-b479-665a1ce8937c" containerID="fb166b129931ff672d247f77f4c3b881346f25b0410d683eb6bb28f72d3af0a4" exitCode=0 Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.698708 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron5151-account-delete-8vbjg" podStartSLOduration=4.69869143 podStartE2EDuration="4.69869143s" podCreationTimestamp="2025-11-25 09:47:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:47:59.658057688 +0000 UTC m=+1424.699117241" watchObservedRunningTime="2025-11-25 09:47:59.69869143 +0000 UTC m=+1424.739750983" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.728211 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df" path="/var/lib/kubelet/pods/7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df/volumes" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.737708 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placemente176-account-delete-wbgz5" podStartSLOduration=5.73768247 podStartE2EDuration="5.73768247s" podCreationTimestamp="2025-11-25 09:47:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:47:59.671163368 +0000 UTC m=+1424.712222921" watchObservedRunningTime="2025-11-25 09:47:59.73768247 +0000 UTC m=+1424.778742023" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.739113 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ab346c9-f70f-4663-9b53-67b4d66b3112" path="/var/lib/kubelet/pods/9ab346c9-f70f-4663-9b53-67b4d66b3112/volumes" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.910639 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-54dbffbb57-kmqfb" podUID="fe3ff67b-01ff-480f-be91-7c8235593c97" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.156:9696/\": dial tcp 10.217.0.156:9696: connect: connection refused" Nov 25 09:47:59 crc kubenswrapper[4776]: I1125 09:47:59.977678 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc920d71-4308-4d90-a891-8b760651ccc9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bc920d71-4308-4d90-a891-8b760651ccc9" (UID: "bc920d71-4308-4d90-a891-8b760651ccc9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:00 crc kubenswrapper[4776]: I1125 09:48:00.012033 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc920d71-4308-4d90-a891-8b760651ccc9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:00 crc kubenswrapper[4776]: I1125 09:48:00.189138 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="6a0ab5a5-2d62-4457-be9d-fe8242ab7c91" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": read tcp 10.217.0.2:46756->10.217.0.204:8775: read: connection reset by peer" Nov 25 09:48:00 crc kubenswrapper[4776]: I1125 09:48:00.191267 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="6a0ab5a5-2d62-4457-be9d-fe8242ab7c91" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": read tcp 10.217.0.2:46766->10.217.0.204:8775: read: connection reset by peer" Nov 25 09:48:00 crc kubenswrapper[4776]: I1125 09:48:00.291977 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc920d71-4308-4d90-a891-8b760651ccc9-config-data" (OuterVolumeSpecName: "config-data") pod "bc920d71-4308-4d90-a891-8b760651ccc9" (UID: "bc920d71-4308-4d90-a891-8b760651ccc9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:00 crc kubenswrapper[4776]: I1125 09:48:00.333413 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc920d71-4308-4d90-a891-8b760651ccc9-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:00 crc kubenswrapper[4776]: I1125 09:48:00.652239 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-769ddf488d-46ssj" podUID="4702d881-d701-41a4-b36f-2d063a8ae246" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.162:9311/healthcheck\": read tcp 10.217.0.2:47786->10.217.0.162:9311: read: connection reset by peer" Nov 25 09:48:00 crc kubenswrapper[4776]: I1125 09:48:00.652506 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-769ddf488d-46ssj" podUID="4702d881-d701-41a4-b36f-2d063a8ae246" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.162:9311/healthcheck\": read tcp 10.217.0.2:47792->10.217.0.162:9311: read: connection reset by peer" Nov 25 09:48:00 crc kubenswrapper[4776]: E1125 09:48:00.669618 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06 is running failed: container process not found" containerID="eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 25 09:48:00 crc kubenswrapper[4776]: E1125 09:48:00.670946 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06 is running failed: container process not found" containerID="eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 25 09:48:00 crc kubenswrapper[4776]: E1125 09:48:00.674249 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06 is running failed: container process not found" containerID="eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 25 09:48:00 crc kubenswrapper[4776]: E1125 09:48:00.674321 4776 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-7fgjf" podUID="10fe3dc2-2bde-4380-8ad3-d71832e31095" containerName="ovsdb-server" Nov 25 09:48:00 crc kubenswrapper[4776]: I1125 09:48:00.674845 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="4512cf4e-6c75-493e-a80a-4b0accd750a8" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.103:5671: connect: connection refused" Nov 25 09:48:00 crc kubenswrapper[4776]: I1125 09:48:00.682378 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-2bs2f" podUID="c59c3d1e-c5c3-4036-ae28-436585e303ab" containerName="ovn-controller" probeResult="failure" output=< Nov 25 09:48:00 crc kubenswrapper[4776]: ERROR - Failed to get connection status from ovn-controller, ovn-appctl exit status: 0 Nov 25 09:48:00 crc kubenswrapper[4776]: > Nov 25 09:48:00 crc kubenswrapper[4776]: E1125 09:48:00.720601 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="257f4e7f047aaf2041def20b6caaee331d83d82f30d019096ad21bfd241e42a8" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 25 09:48:00 crc kubenswrapper[4776]: E1125 09:48:00.755233 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="257f4e7f047aaf2041def20b6caaee331d83d82f30d019096ad21bfd241e42a8" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 25 09:48:00 crc kubenswrapper[4776]: E1125 09:48:00.755463 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5002006f906f0b17afbf25b23648688a842975cfd3b385d57e3191040607fb57" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 25 09:48:00 crc kubenswrapper[4776]: E1125 09:48:00.759214 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5002006f906f0b17afbf25b23648688a842975cfd3b385d57e3191040607fb57" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 25 09:48:00 crc kubenswrapper[4776]: I1125 09:48:00.759413 4776 generic.go:334] "Generic (PLEG): container finished" podID="86e320e0-80ee-40df-b3a1-a48cb810a435" containerID="968312305c48cd2a75a04c3e2a43d37d9ed640ee14d080081ad79c323f2cf0f5" exitCode=0 Nov 25 09:48:00 crc kubenswrapper[4776]: E1125 09:48:00.759461 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="257f4e7f047aaf2041def20b6caaee331d83d82f30d019096ad21bfd241e42a8" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 25 09:48:00 crc kubenswrapper[4776]: E1125 09:48:00.759483 4776 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="8bae8251-3d5d-490b-9116-73769dc18101" containerName="nova-scheduler-scheduler" Nov 25 09:48:00 crc kubenswrapper[4776]: E1125 09:48:00.765999 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5002006f906f0b17afbf25b23648688a842975cfd3b385d57e3191040607fb57" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 25 09:48:00 crc kubenswrapper[4776]: E1125 09:48:00.766046 4776 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-7fgjf" podUID="10fe3dc2-2bde-4380-8ad3-d71832e31095" containerName="ovs-vswitchd" Nov 25 09:48:00 crc kubenswrapper[4776]: E1125 09:48:00.790405 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="51468cb0acece8ce1a386bdf49793e6322b08cddac5cf23fc46abef1f5330e22" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 25 09:48:00 crc kubenswrapper[4776]: E1125 09:48:00.794137 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="51468cb0acece8ce1a386bdf49793e6322b08cddac5cf23fc46abef1f5330e22" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 25 09:48:00 crc kubenswrapper[4776]: E1125 09:48:00.795501 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="51468cb0acece8ce1a386bdf49793e6322b08cddac5cf23fc46abef1f5330e22" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 25 09:48:00 crc kubenswrapper[4776]: E1125 09:48:00.795543 4776 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4" containerName="nova-cell0-conductor-conductor" Nov 25 09:48:00 crc kubenswrapper[4776]: I1125 09:48:00.795770 4776 generic.go:334] "Generic (PLEG): container finished" podID="4702d881-d701-41a4-b36f-2d063a8ae246" containerID="4e9d85a2660614a7e89218283e8f24a0ddfcb9cad63c01ab8119453aceaad099" exitCode=0 Nov 25 09:48:00 crc kubenswrapper[4776]: I1125 09:48:00.800931 4776 generic.go:334] "Generic (PLEG): container finished" podID="6a0ab5a5-2d62-4457-be9d-fe8242ab7c91" containerID="7927ef3d5f7eb88faf6e7c414da603268c27589afcdfe2331cf08fabea8c8c3c" exitCode=0 Nov 25 09:48:00 crc kubenswrapper[4776]: I1125 09:48:00.804051 4776 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/neutron5151-account-delete-8vbjg" secret="" err="secret \"galera-openstack-dockercfg-2t654\" not found" Nov 25 09:48:00 crc kubenswrapper[4776]: I1125 09:48:00.813873 4776 generic.go:334] "Generic (PLEG): container finished" podID="b708b100-166c-49b1-a47b-76d47ba2c6e5" containerID="5919b100098159f61f76d181a1ae66bfafcb2916a56d7fe4fe1cee93ad3c32d0" exitCode=0 Nov 25 09:48:00 crc kubenswrapper[4776]: I1125 09:48:00.839547 4776 generic.go:334] "Generic (PLEG): container finished" podID="a0b97845-4578-4801-896a-281fdafdb351" containerID="856d610eaa67921f1c8d09ca91236aafc5070c66f02e6f2c2f6ee1982f26930e" exitCode=0 Nov 25 09:48:00 crc kubenswrapper[4776]: I1125 09:48:00.856305 4776 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novaapi3b46-account-delete-9w5r9" secret="" err="secret \"galera-openstack-dockercfg-2t654\" not found" Nov 25 09:48:00 crc kubenswrapper[4776]: I1125 09:48:00.861945 4776 generic.go:334] "Generic (PLEG): container finished" podID="aab7d356-ec67-4e5d-9c25-f26e39786f79" containerID="ec693b92c2b81399b4a2d57c1980de7b988f0cc59e2cef0ec21c34b5dff21bd2" exitCode=0 Nov 25 09:48:00 crc kubenswrapper[4776]: I1125 09:48:00.861970 4776 generic.go:334] "Generic (PLEG): container finished" podID="aab7d356-ec67-4e5d-9c25-f26e39786f79" containerID="c99e1a9b60f21bc937fd065f74ab82371110943731ae37f120fc60b0c71336da" exitCode=2 Nov 25 09:48:00 crc kubenswrapper[4776]: I1125 09:48:00.863598 4776 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novacell09dc4-account-delete-w5xcj" secret="" err="secret \"galera-openstack-dockercfg-2t654\" not found" Nov 25 09:48:00 crc kubenswrapper[4776]: I1125 09:48:00.873648 4776 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/placemente176-account-delete-wbgz5" secret="" err="secret \"galera-openstack-dockercfg-2t654\" not found" Nov 25 09:48:00 crc kubenswrapper[4776]: I1125 09:48:00.878772 4776 generic.go:334] "Generic (PLEG): container finished" podID="d2cda6de-ef12-42f4-a6e5-824c2db3cd01" containerID="9fddb69738d8e8fd199d1f7a677b07c2d8ceff1639dbe9784b2c6fd247da4ae1" exitCode=0 Nov 25 09:48:00 crc kubenswrapper[4776]: I1125 09:48:00.880660 4776 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/cinder774d-account-delete-b29n5" secret="" err="secret \"galera-openstack-dockercfg-2t654\" not found" Nov 25 09:48:00 crc kubenswrapper[4776]: I1125 09:48:00.887685 4776 generic.go:334] "Generic (PLEG): container finished" podID="6bf3edad-487a-4c68-9dbf-9789a94f8fba" containerID="24a3083e873c7a16456c526fdfe2bc77dcf34c30f066c1f1af77953f2e1ecddd" exitCode=0 Nov 25 09:48:00 crc kubenswrapper[4776]: I1125 09:48:00.888391 4776 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/glanceb9b5-account-delete-tjrnr" secret="" err="secret \"galera-openstack-dockercfg-2t654\" not found" Nov 25 09:48:00 crc kubenswrapper[4776]: I1125 09:48:00.893125 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novaapi3b46-account-delete-9w5r9" podStartSLOduration=5.893101521 podStartE2EDuration="5.893101521s" podCreationTimestamp="2025-11-25 09:47:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:48:00.871097838 +0000 UTC m=+1425.912157391" watchObservedRunningTime="2025-11-25 09:48:00.893101521 +0000 UTC m=+1425.934161074" Nov 25 09:48:00 crc kubenswrapper[4776]: I1125 09:48:00.919209 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novacell09dc4-account-delete-w5xcj" podStartSLOduration=5.9191862870000005 podStartE2EDuration="5.919186287s" podCreationTimestamp="2025-11-25 09:47:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:48:00.887847269 +0000 UTC m=+1425.928906822" watchObservedRunningTime="2025-11-25 09:48:00.919186287 +0000 UTC m=+1425.960245850" Nov 25 09:48:00 crc kubenswrapper[4776]: I1125 09:48:00.937155 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder774d-account-delete-b29n5" podStartSLOduration=6.937128728 podStartE2EDuration="6.937128728s" podCreationTimestamp="2025-11-25 09:47:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:48:00.903028451 +0000 UTC m=+1425.944088004" watchObservedRunningTime="2025-11-25 09:48:00.937128728 +0000 UTC m=+1425.978188281" Nov 25 09:48:00 crc kubenswrapper[4776]: E1125 09:48:00.980373 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:00 crc kubenswrapper[4776]: E1125 09:48:00.980455 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/14b51bba-0b15-469b-bdea-90fae8d9b976-operator-scripts podName:14b51bba-0b15-469b-bdea-90fae8d9b976 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:01.480433487 +0000 UTC m=+1426.521493040 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/14b51bba-0b15-469b-bdea-90fae8d9b976-operator-scripts") pod "novaapi3b46-account-delete-9w5r9" (UID: "14b51bba-0b15-469b-bdea-90fae8d9b976") : configmap "openstack-scripts" not found Nov 25 09:48:00 crc kubenswrapper[4776]: E1125 09:48:00.980712 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:00 crc kubenswrapper[4776]: E1125 09:48:00.980744 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ef85f90f-c377-4273-9fff-451420d6008a-operator-scripts podName:ef85f90f-c377-4273-9fff-451420d6008a nodeName:}" failed. No retries permitted until 2025-11-25 09:48:01.480734885 +0000 UTC m=+1426.521794438 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/ef85f90f-c377-4273-9fff-451420d6008a-operator-scripts") pod "cinder774d-account-delete-b29n5" (UID: "ef85f90f-c377-4273-9fff-451420d6008a") : configmap "openstack-scripts" not found Nov 25 09:48:00 crc kubenswrapper[4776]: E1125 09:48:00.980786 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:00 crc kubenswrapper[4776]: E1125 09:48:00.980815 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/946ea713-bfba-449f-82ab-28e915469938-operator-scripts podName:946ea713-bfba-449f-82ab-28e915469938 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:01.480806046 +0000 UTC m=+1426.521865599 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/946ea713-bfba-449f-82ab-28e915469938-operator-scripts") pod "novacell09dc4-account-delete-w5xcj" (UID: "946ea713-bfba-449f-82ab-28e915469938") : configmap "openstack-scripts" not found Nov 25 09:48:00 crc kubenswrapper[4776]: E1125 09:48:00.980846 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:00 crc kubenswrapper[4776]: E1125 09:48:00.980868 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6be099e6-89d6-45a0-9d0f-4dca4d91701a-operator-scripts podName:6be099e6-89d6-45a0-9d0f-4dca4d91701a nodeName:}" failed. No retries permitted until 2025-11-25 09:48:01.480860708 +0000 UTC m=+1426.521920261 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/6be099e6-89d6-45a0-9d0f-4dca4d91701a-operator-scripts") pod "placemente176-account-delete-wbgz5" (UID: "6be099e6-89d6-45a0-9d0f-4dca4d91701a") : configmap "openstack-scripts" not found Nov 25 09:48:00 crc kubenswrapper[4776]: E1125 09:48:00.981053 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:00 crc kubenswrapper[4776]: E1125 09:48:00.981172 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/73f2c2ab-6cb6-4544-a3f7-20722e5b6400-operator-scripts podName:73f2c2ab-6cb6-4544-a3f7-20722e5b6400 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:01.481153025 +0000 UTC m=+1426.522212578 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/73f2c2ab-6cb6-4544-a3f7-20722e5b6400-operator-scripts") pod "neutron5151-account-delete-8vbjg" (UID: "73f2c2ab-6cb6-4544-a3f7-20722e5b6400") : configmap "openstack-scripts" not found Nov 25 09:48:01 crc kubenswrapper[4776]: E1125 09:48:01.034772 4776 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.373s" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.034882 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"5920f152-04cd-4366-b479-665a1ce8937c","Type":"ContainerDied","Data":"fb166b129931ff672d247f77f4c3b881346f25b0410d683eb6bb28f72d3af0a4"} Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.035051 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.035430 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="09f0463c-f091-495e-9773-c07fe902c200" containerName="kube-state-metrics" containerID="cri-o://94684aa85a999df01587ff8336fcf45d1847b413f000fe304eae44dda233eae6" gracePeriod=30 Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.035138 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell09dc4-account-delete-w5xcj" event={"ID":"946ea713-bfba-449f-82ab-28e915469938","Type":"ContainerStarted","Data":"3426e4cb115f29a90f341756fdee32ff5ebae07b8b55d943e28a698f73f1eb3d"} Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037321 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"86e320e0-80ee-40df-b3a1-a48cb810a435","Type":"ContainerDied","Data":"968312305c48cd2a75a04c3e2a43d37d9ed640ee14d080081ad79c323f2cf0f5"} Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037343 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-769ddf488d-46ssj" event={"ID":"4702d881-d701-41a4-b36f-2d063a8ae246","Type":"ContainerDied","Data":"4e9d85a2660614a7e89218283e8f24a0ddfcb9cad63c01ab8119453aceaad099"} Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037402 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037423 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-dvd6n"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037434 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-jswxm"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037450 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-dvd6n"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037466 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-jswxm"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037553 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91","Type":"ContainerDied","Data":"7927ef3d5f7eb88faf6e7c414da603268c27589afcdfe2331cf08fabea8c8c3c"} Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037568 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron5151-account-delete-8vbjg" event={"ID":"73f2c2ab-6cb6-4544-a3f7-20722e5b6400","Type":"ContainerStarted","Data":"f44277236f6a634e8f611e825fd05aabebb686cac5979f168a7f50210ff4ef17"} Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037624 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b708b100-166c-49b1-a47b-76d47ba2c6e5","Type":"ContainerDied","Data":"5919b100098159f61f76d181a1ae66bfafcb2916a56d7fe4fe1cee93ad3c32d0"} Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037640 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-667769d556-lqtlp" event={"ID":"a0b97845-4578-4801-896a-281fdafdb351","Type":"ContainerDied","Data":"856d610eaa67921f1c8d09ca91236aafc5070c66f02e6f2c2f6ee1982f26930e"} Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037654 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"5920f152-04cd-4366-b479-665a1ce8937c","Type":"ContainerDied","Data":"e7a84137f7ca4e701b70532e3ecb35e1f3005afd8011d822ceb5e2e2fba6c8cb"} Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037777 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7a84137f7ca4e701b70532e3ecb35e1f3005afd8011d822ceb5e2e2fba6c8cb" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037791 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5864b7d7fc-g9z2r" event={"ID":"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9","Type":"ContainerDied","Data":"54bc9c507715dcff53a21f00cb2fd667946240a7b4842b710dfb49b81a04e354"} Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037803 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54bc9c507715dcff53a21f00cb2fd667946240a7b4842b710dfb49b81a04e354" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037817 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-6b75d6ddb7-rmtms"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037833 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037847 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-j6tqf"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037861 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-j6tqf"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037876 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi3b46-account-delete-9w5r9" event={"ID":"14b51bba-0b15-469b-bdea-90fae8d9b976","Type":"ContainerStarted","Data":"d3de2925044530f7fd248214ce715db43c6f7a123e71e4f5fd11e8af0b5cda0e"} Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037890 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aab7d356-ec67-4e5d-9c25-f26e39786f79","Type":"ContainerDied","Data":"ec693b92c2b81399b4a2d57c1980de7b988f0cc59e2cef0ec21c34b5dff21bd2"} Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037908 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder774d-account-delete-b29n5"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037923 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aab7d356-ec67-4e5d-9c25-f26e39786f79","Type":"ContainerDied","Data":"c99e1a9b60f21bc937fd065f74ab82371110943731ae37f120fc60b0c71336da"} Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037934 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell09dc4-account-delete-w5xcj" event={"ID":"946ea713-bfba-449f-82ab-28e915469938","Type":"ContainerStarted","Data":"ac0deb2776fd44dcbddf8b4d4bae71247d519e589f78423415504e15ea5a4a0d"} Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037946 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-774d-account-create-vbblg"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037957 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placemente176-account-delete-wbgz5" event={"ID":"6be099e6-89d6-45a0-9d0f-4dca4d91701a","Type":"ContainerStarted","Data":"52647c9b509f4a2e0420df93981105d258326ab2d8003e1be47968ae44e82e11"} Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037971 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-774d-account-create-vbblg"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037985 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-tzmv5"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.037997 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d2cda6de-ef12-42f4-a6e5-824c2db3cd01","Type":"ContainerDied","Data":"9fddb69738d8e8fd199d1f7a677b07c2d8ceff1639dbe9784b2c6fd247da4ae1"} Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.038012 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder774d-account-delete-b29n5" event={"ID":"ef85f90f-c377-4273-9fff-451420d6008a","Type":"ContainerStarted","Data":"02f077bb80ad2e06d150c6b44d4bfbe70a57d6d488294c69cf9fd29d74fce9a8"} Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.038023 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6bf3edad-487a-4c68-9dbf-9789a94f8fba","Type":"ContainerDied","Data":"24a3083e873c7a16456c526fdfe2bc77dcf34c30f066c1f1af77953f2e1ecddd"} Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.038039 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-tzmv5"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.038052 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-b60a-account-create-nf6k7"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.038091 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-b60a-account-create-nf6k7"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.038107 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-96l8q"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.038120 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-96l8q"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.038132 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell09dc4-account-delete-w5xcj"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.038146 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-9dc4-account-create-cj5jv"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.038159 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-9dc4-account-create-cj5jv"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.038172 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-t5rsg"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.038182 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-t5rsg"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.038194 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi3b46-account-delete-9w5r9"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.038207 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-3b46-account-create-kg8tk"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.038218 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-3b46-account-create-kg8tk"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.038230 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-mfnw2"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.038242 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-mfnw2"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.038254 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5151-account-create-866rm"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.038310 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron5151-account-delete-8vbjg"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.038329 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5151-account-create-866rm"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.038517 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="cd5ce048-5a09-45a4-9a7d-66712b326ab7" containerName="memcached" containerID="cri-o://6b83a886269bd5ccf82ce338326e0a5b6749a7805141d811c4f9ff42aa7148ca" gracePeriod=30 Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.038740 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-6b75d6ddb7-rmtms" podUID="83c080b8-d976-4e81-b103-2442bb2eafe1" containerName="keystone-api" containerID="cri-o://d35530a8d2f45153bec6fcc975ede0bc99a55fb6d60c6edac19a05f0939799d7" gracePeriod=30 Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.067313 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.073179 4776 scope.go:117] "RemoveContainer" containerID="6bffff98e55f5e2fe0110f4a9d127c1dff4b4c552606101a667c0fb53d5e9b16" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.073662 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 25 09:48:01 crc kubenswrapper[4776]: E1125 09:48:01.083330 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:01 crc kubenswrapper[4776]: E1125 09:48:01.083479 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/bcaec0f3-13af-4f73-8e7f-7e575b8e2008-operator-scripts podName:bcaec0f3-13af-4f73-8e7f-7e575b8e2008 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:01.583458758 +0000 UTC m=+1426.624518311 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/bcaec0f3-13af-4f73-8e7f-7e575b8e2008-operator-scripts") pod "glanceb9b5-account-delete-tjrnr" (UID: "bcaec0f3-13af-4f73-8e7f-7e575b8e2008") : configmap "openstack-scripts" not found Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.084294 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.085832 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-667769d556-lqtlp" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.095208 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.104909 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.123397 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.185338 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-run-httpd\") pod \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.185436 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zj7k\" (UniqueName: \"kubernetes.io/projected/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-kube-api-access-7zj7k\") pod \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.185481 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-log-httpd\") pod \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.185571 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5920f152-04cd-4366-b479-665a1ce8937c-operator-scripts\") pod \"5920f152-04cd-4366-b479-665a1ce8937c\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.185641 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-config-data\") pod \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.185674 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-config-data\") pod \"a0b97845-4578-4801-896a-281fdafdb351\" (UID: \"a0b97845-4578-4801-896a-281fdafdb351\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.185728 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5920f152-04cd-4366-b479-665a1ce8937c-combined-ca-bundle\") pod \"5920f152-04cd-4366-b479-665a1ce8937c\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.185754 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"5920f152-04cd-4366-b479-665a1ce8937c\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.185767 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9" (UID: "f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.185804 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-etc-swift\") pod \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.185894 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-combined-ca-bundle\") pod \"a0b97845-4578-4801-896a-281fdafdb351\" (UID: \"a0b97845-4578-4801-896a-281fdafdb351\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.185960 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5920f152-04cd-4366-b479-665a1ce8937c-config-data-generated\") pod \"5920f152-04cd-4366-b479-665a1ce8937c\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.185987 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-264xx\" (UniqueName: \"kubernetes.io/projected/a0b97845-4578-4801-896a-281fdafdb351-kube-api-access-264xx\") pod \"a0b97845-4578-4801-896a-281fdafdb351\" (UID: \"a0b97845-4578-4801-896a-281fdafdb351\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.186115 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-combined-ca-bundle\") pod \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.186189 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-scripts\") pod \"a0b97845-4578-4801-896a-281fdafdb351\" (UID: \"a0b97845-4578-4801-896a-281fdafdb351\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.186226 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5920f152-04cd-4366-b479-665a1ce8937c-galera-tls-certs\") pod \"5920f152-04cd-4366-b479-665a1ce8937c\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.186289 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5920f152-04cd-4366-b479-665a1ce8937c-config-data-default\") pod \"5920f152-04cd-4366-b479-665a1ce8937c\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.186349 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-internal-tls-certs\") pod \"a0b97845-4578-4801-896a-281fdafdb351\" (UID: \"a0b97845-4578-4801-896a-281fdafdb351\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.186378 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5920f152-04cd-4366-b479-665a1ce8937c-kolla-config\") pod \"5920f152-04cd-4366-b479-665a1ce8937c\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.186401 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0b97845-4578-4801-896a-281fdafdb351-logs\") pod \"a0b97845-4578-4801-896a-281fdafdb351\" (UID: \"a0b97845-4578-4801-896a-281fdafdb351\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.186469 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-internal-tls-certs\") pod \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.186518 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5b4m\" (UniqueName: \"kubernetes.io/projected/5920f152-04cd-4366-b479-665a1ce8937c-kube-api-access-p5b4m\") pod \"5920f152-04cd-4366-b479-665a1ce8937c\" (UID: \"5920f152-04cd-4366-b479-665a1ce8937c\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.186562 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-public-tls-certs\") pod \"a0b97845-4578-4801-896a-281fdafdb351\" (UID: \"a0b97845-4578-4801-896a-281fdafdb351\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.186615 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-public-tls-certs\") pod \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\" (UID: \"f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.187106 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9" (UID: "f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.187329 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0b97845-4578-4801-896a-281fdafdb351-logs" (OuterVolumeSpecName: "logs") pod "a0b97845-4578-4801-896a-281fdafdb351" (UID: "a0b97845-4578-4801-896a-281fdafdb351"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.187659 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5920f152-04cd-4366-b479-665a1ce8937c-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "5920f152-04cd-4366-b479-665a1ce8937c" (UID: "5920f152-04cd-4366-b479-665a1ce8937c"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.188125 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5920f152-04cd-4366-b479-665a1ce8937c-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "5920f152-04cd-4366-b479-665a1ce8937c" (UID: "5920f152-04cd-4366-b479-665a1ce8937c"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.188216 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5920f152-04cd-4366-b479-665a1ce8937c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5920f152-04cd-4366-b479-665a1ce8937c" (UID: "5920f152-04cd-4366-b479-665a1ce8937c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.192390 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5920f152-04cd-4366-b479-665a1ce8937c-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "5920f152-04cd-4366-b479-665a1ce8937c" (UID: "5920f152-04cd-4366-b479-665a1ce8937c"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.192899 4776 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5920f152-04cd-4366-b479-665a1ce8937c-config-data-default\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.193018 4776 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5920f152-04cd-4366-b479-665a1ce8937c-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.193127 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0b97845-4578-4801-896a-281fdafdb351-logs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.193208 4776 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.193280 4776 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.193352 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5920f152-04cd-4366-b479-665a1ce8937c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.193424 4776 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5920f152-04cd-4366-b479-665a1ce8937c-config-data-generated\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.206277 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9" (UID: "f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.206530 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-scripts" (OuterVolumeSpecName: "scripts") pod "a0b97845-4578-4801-896a-281fdafdb351" (UID: "a0b97845-4578-4801-896a-281fdafdb351"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.207228 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-kube-api-access-7zj7k" (OuterVolumeSpecName: "kube-api-access-7zj7k") pod "f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9" (UID: "f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9"). InnerVolumeSpecName "kube-api-access-7zj7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.217019 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5920f152-04cd-4366-b479-665a1ce8937c-kube-api-access-p5b4m" (OuterVolumeSpecName: "kube-api-access-p5b4m") pod "5920f152-04cd-4366-b479-665a1ce8937c" (UID: "5920f152-04cd-4366-b479-665a1ce8937c"). InnerVolumeSpecName "kube-api-access-p5b4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.227535 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.244455 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "mysql-db") pod "5920f152-04cd-4366-b479-665a1ce8937c" (UID: "5920f152-04cd-4366-b479-665a1ce8937c"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.246396 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0b97845-4578-4801-896a-281fdafdb351-kube-api-access-264xx" (OuterVolumeSpecName: "kube-api-access-264xx") pod "a0b97845-4578-4801-896a-281fdafdb351" (UID: "a0b97845-4578-4801-896a-281fdafdb351"). InnerVolumeSpecName "kube-api-access-264xx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.256241 4776 scope.go:117] "RemoveContainer" containerID="c9f1b69c74818d42cf402965ae75f33a682f61bb14e9874a7f7904a0e8298e5e" Nov 25 09:48:01 crc kubenswrapper[4776]: E1125 09:48:01.256737 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9f1b69c74818d42cf402965ae75f33a682f61bb14e9874a7f7904a0e8298e5e\": container with ID starting with c9f1b69c74818d42cf402965ae75f33a682f61bb14e9874a7f7904a0e8298e5e not found: ID does not exist" containerID="c9f1b69c74818d42cf402965ae75f33a682f61bb14e9874a7f7904a0e8298e5e" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.256826 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9f1b69c74818d42cf402965ae75f33a682f61bb14e9874a7f7904a0e8298e5e"} err="failed to get container status \"c9f1b69c74818d42cf402965ae75f33a682f61bb14e9874a7f7904a0e8298e5e\": rpc error: code = NotFound desc = could not find container \"c9f1b69c74818d42cf402965ae75f33a682f61bb14e9874a7f7904a0e8298e5e\": container with ID starting with c9f1b69c74818d42cf402965ae75f33a682f61bb14e9874a7f7904a0e8298e5e not found: ID does not exist" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.256920 4776 scope.go:117] "RemoveContainer" containerID="6bffff98e55f5e2fe0110f4a9d127c1dff4b4c552606101a667c0fb53d5e9b16" Nov 25 09:48:01 crc kubenswrapper[4776]: E1125 09:48:01.258637 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bffff98e55f5e2fe0110f4a9d127c1dff4b4c552606101a667c0fb53d5e9b16\": container with ID starting with 6bffff98e55f5e2fe0110f4a9d127c1dff4b4c552606101a667c0fb53d5e9b16 not found: ID does not exist" containerID="6bffff98e55f5e2fe0110f4a9d127c1dff4b4c552606101a667c0fb53d5e9b16" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.258812 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bffff98e55f5e2fe0110f4a9d127c1dff4b4c552606101a667c0fb53d5e9b16"} err="failed to get container status \"6bffff98e55f5e2fe0110f4a9d127c1dff4b4c552606101a667c0fb53d5e9b16\": rpc error: code = NotFound desc = could not find container \"6bffff98e55f5e2fe0110f4a9d127c1dff4b4c552606101a667c0fb53d5e9b16\": container with ID starting with 6bffff98e55f5e2fe0110f4a9d127c1dff4b4c552606101a667c0fb53d5e9b16 not found: ID does not exist" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.258902 4776 scope.go:117] "RemoveContainer" containerID="e056e61a1527b24d984169499b3cde97a02136661c62174d6fe28407dfd3287a" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.272918 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5920f152-04cd-4366-b479-665a1ce8937c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5920f152-04cd-4366-b479-665a1ce8937c" (UID: "5920f152-04cd-4366-b479-665a1ce8937c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.298298 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5920f152-04cd-4366-b479-665a1ce8937c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.298543 4776 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.298617 4776 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.298701 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-264xx\" (UniqueName: \"kubernetes.io/projected/a0b97845-4578-4801-896a-281fdafdb351-kube-api-access-264xx\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.298770 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.298856 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5b4m\" (UniqueName: \"kubernetes.io/projected/5920f152-04cd-4366-b479-665a1ce8937c-kube-api-access-p5b4m\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.298922 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zj7k\" (UniqueName: \"kubernetes.io/projected/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-kube-api-access-7zj7k\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.306770 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-config-data" (OuterVolumeSpecName: "config-data") pod "a0b97845-4578-4801-896a-281fdafdb351" (UID: "a0b97845-4578-4801-896a-281fdafdb351"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.315227 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="167c5bc5-b21e-4c51-8765-1dcbf290294b" containerName="galera" containerID="cri-o://2f383d0debcf43b01564a0db854854b92a294770e346344bcd8dce403c1e0251" gracePeriod=30 Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.319902 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9" (UID: "f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.325292 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-config-data" (OuterVolumeSpecName: "config-data") pod "f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9" (UID: "f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.349964 4776 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.369276 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9" (UID: "f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.369415 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5920f152-04cd-4366-b479-665a1ce8937c-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "5920f152-04cd-4366-b479-665a1ce8937c" (UID: "5920f152-04cd-4366-b479-665a1ce8937c"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.384255 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9" (UID: "f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.401707 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-scripts\") pod \"86e320e0-80ee-40df-b3a1-a48cb810a435\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.401771 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-config-data-custom\") pod \"86e320e0-80ee-40df-b3a1-a48cb810a435\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.401793 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-combined-ca-bundle\") pod \"86e320e0-80ee-40df-b3a1-a48cb810a435\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.401824 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-public-tls-certs\") pod \"86e320e0-80ee-40df-b3a1-a48cb810a435\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.401899 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86e320e0-80ee-40df-b3a1-a48cb810a435-logs\") pod \"86e320e0-80ee-40df-b3a1-a48cb810a435\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.401923 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-internal-tls-certs\") pod \"86e320e0-80ee-40df-b3a1-a48cb810a435\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.401991 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-config-data\") pod \"86e320e0-80ee-40df-b3a1-a48cb810a435\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.402039 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/86e320e0-80ee-40df-b3a1-a48cb810a435-etc-machine-id\") pod \"86e320e0-80ee-40df-b3a1-a48cb810a435\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.402236 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k277j\" (UniqueName: \"kubernetes.io/projected/86e320e0-80ee-40df-b3a1-a48cb810a435-kube-api-access-k277j\") pod \"86e320e0-80ee-40df-b3a1-a48cb810a435\" (UID: \"86e320e0-80ee-40df-b3a1-a48cb810a435\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.403666 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.403689 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.403701 4776 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.403710 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.403720 4776 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5920f152-04cd-4366-b479-665a1ce8937c-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.403728 4776 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.403736 4776 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.404911 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86e320e0-80ee-40df-b3a1-a48cb810a435-logs" (OuterVolumeSpecName: "logs") pod "86e320e0-80ee-40df-b3a1-a48cb810a435" (UID: "86e320e0-80ee-40df-b3a1-a48cb810a435"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.406961 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/86e320e0-80ee-40df-b3a1-a48cb810a435-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "86e320e0-80ee-40df-b3a1-a48cb810a435" (UID: "86e320e0-80ee-40df-b3a1-a48cb810a435"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.407121 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86e320e0-80ee-40df-b3a1-a48cb810a435-kube-api-access-k277j" (OuterVolumeSpecName: "kube-api-access-k277j") pod "86e320e0-80ee-40df-b3a1-a48cb810a435" (UID: "86e320e0-80ee-40df-b3a1-a48cb810a435"). InnerVolumeSpecName "kube-api-access-k277j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.418546 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-scripts" (OuterVolumeSpecName: "scripts") pod "86e320e0-80ee-40df-b3a1-a48cb810a435" (UID: "86e320e0-80ee-40df-b3a1-a48cb810a435"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.418586 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "86e320e0-80ee-40df-b3a1-a48cb810a435" (UID: "86e320e0-80ee-40df-b3a1-a48cb810a435"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.435883 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a0b97845-4578-4801-896a-281fdafdb351" (UID: "a0b97845-4578-4801-896a-281fdafdb351"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.504945 4776 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/86e320e0-80ee-40df-b3a1-a48cb810a435-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.505305 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k277j\" (UniqueName: \"kubernetes.io/projected/86e320e0-80ee-40df-b3a1-a48cb810a435-kube-api-access-k277j\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.505325 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.505336 4776 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.505348 4776 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.505357 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86e320e0-80ee-40df-b3a1-a48cb810a435-logs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: E1125 09:48:01.505233 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:01 crc kubenswrapper[4776]: E1125 09:48:01.505412 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/14b51bba-0b15-469b-bdea-90fae8d9b976-operator-scripts podName:14b51bba-0b15-469b-bdea-90fae8d9b976 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:02.505395517 +0000 UTC m=+1427.546455070 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/14b51bba-0b15-469b-bdea-90fae8d9b976-operator-scripts") pod "novaapi3b46-account-delete-9w5r9" (UID: "14b51bba-0b15-469b-bdea-90fae8d9b976") : configmap "openstack-scripts" not found Nov 25 09:48:01 crc kubenswrapper[4776]: E1125 09:48:01.505267 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:01 crc kubenswrapper[4776]: E1125 09:48:01.505451 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ef85f90f-c377-4273-9fff-451420d6008a-operator-scripts podName:ef85f90f-c377-4273-9fff-451420d6008a nodeName:}" failed. No retries permitted until 2025-11-25 09:48:02.505439558 +0000 UTC m=+1427.546499111 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/ef85f90f-c377-4273-9fff-451420d6008a-operator-scripts") pod "cinder774d-account-delete-b29n5" (UID: "ef85f90f-c377-4273-9fff-451420d6008a") : configmap "openstack-scripts" not found Nov 25 09:48:01 crc kubenswrapper[4776]: E1125 09:48:01.505321 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:01 crc kubenswrapper[4776]: E1125 09:48:01.505477 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6be099e6-89d6-45a0-9d0f-4dca4d91701a-operator-scripts podName:6be099e6-89d6-45a0-9d0f-4dca4d91701a nodeName:}" failed. No retries permitted until 2025-11-25 09:48:02.505469909 +0000 UTC m=+1427.546529462 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/6be099e6-89d6-45a0-9d0f-4dca4d91701a-operator-scripts") pod "placemente176-account-delete-wbgz5" (UID: "6be099e6-89d6-45a0-9d0f-4dca4d91701a") : configmap "openstack-scripts" not found Nov 25 09:48:01 crc kubenswrapper[4776]: E1125 09:48:01.505357 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:01 crc kubenswrapper[4776]: E1125 09:48:01.505505 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/73f2c2ab-6cb6-4544-a3f7-20722e5b6400-operator-scripts podName:73f2c2ab-6cb6-4544-a3f7-20722e5b6400 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:02.505496109 +0000 UTC m=+1427.546555662 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/73f2c2ab-6cb6-4544-a3f7-20722e5b6400-operator-scripts") pod "neutron5151-account-delete-8vbjg" (UID: "73f2c2ab-6cb6-4544-a3f7-20722e5b6400") : configmap "openstack-scripts" not found Nov 25 09:48:01 crc kubenswrapper[4776]: E1125 09:48:01.505561 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:01 crc kubenswrapper[4776]: E1125 09:48:01.505585 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/946ea713-bfba-449f-82ab-28e915469938-operator-scripts podName:946ea713-bfba-449f-82ab-28e915469938 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:02.505577832 +0000 UTC m=+1427.546637385 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/946ea713-bfba-449f-82ab-28e915469938-operator-scripts") pod "novacell09dc4-account-delete-w5xcj" (UID: "946ea713-bfba-449f-82ab-28e915469938") : configmap "openstack-scripts" not found Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.506122 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.511112 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.519368 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a0b97845-4578-4801-896a-281fdafdb351" (UID: "a0b97845-4578-4801-896a-281fdafdb351"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.567620 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.574481 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.606766 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-combined-ca-bundle\") pod \"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91\" (UID: \"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.606804 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-config-data\") pod \"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91\" (UID: \"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.606822 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkxgs\" (UniqueName: \"kubernetes.io/projected/b708b100-166c-49b1-a47b-76d47ba2c6e5-kube-api-access-dkxgs\") pod \"b708b100-166c-49b1-a47b-76d47ba2c6e5\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.606840 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-config-data-custom\") pod \"4702d881-d701-41a4-b36f-2d063a8ae246\" (UID: \"4702d881-d701-41a4-b36f-2d063a8ae246\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.606863 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b708b100-166c-49b1-a47b-76d47ba2c6e5-config-data\") pod \"b708b100-166c-49b1-a47b-76d47ba2c6e5\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.606890 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-internal-tls-certs\") pod \"4702d881-d701-41a4-b36f-2d063a8ae246\" (UID: \"4702d881-d701-41a4-b36f-2d063a8ae246\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.607210 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-nova-metadata-tls-certs\") pod \"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91\" (UID: \"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.607251 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-combined-ca-bundle\") pod \"4702d881-d701-41a4-b36f-2d063a8ae246\" (UID: \"4702d881-d701-41a4-b36f-2d063a8ae246\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.607277 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b708b100-166c-49b1-a47b-76d47ba2c6e5-internal-tls-certs\") pod \"b708b100-166c-49b1-a47b-76d47ba2c6e5\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.607293 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-public-tls-certs\") pod \"4702d881-d701-41a4-b36f-2d063a8ae246\" (UID: \"4702d881-d701-41a4-b36f-2d063a8ae246\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.607311 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bf3edad-487a-4c68-9dbf-9789a94f8fba-public-tls-certs\") pod \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\" (UID: \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.607334 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"b708b100-166c-49b1-a47b-76d47ba2c6e5\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.607351 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldgn5\" (UniqueName: \"kubernetes.io/projected/4702d881-d701-41a4-b36f-2d063a8ae246-kube-api-access-ldgn5\") pod \"4702d881-d701-41a4-b36f-2d063a8ae246\" (UID: \"4702d881-d701-41a4-b36f-2d063a8ae246\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.607371 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4702d881-d701-41a4-b36f-2d063a8ae246-logs\") pod \"4702d881-d701-41a4-b36f-2d063a8ae246\" (UID: \"4702d881-d701-41a4-b36f-2d063a8ae246\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.607390 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bf3edad-487a-4c68-9dbf-9789a94f8fba-config-data\") pod \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\" (UID: \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.607406 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b708b100-166c-49b1-a47b-76d47ba2c6e5-scripts\") pod \"b708b100-166c-49b1-a47b-76d47ba2c6e5\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.607421 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffdd6\" (UniqueName: \"kubernetes.io/projected/6bf3edad-487a-4c68-9dbf-9789a94f8fba-kube-api-access-ffdd6\") pod \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\" (UID: \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.607437 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6bf3edad-487a-4c68-9dbf-9789a94f8fba-logs\") pod \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\" (UID: \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.607454 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b708b100-166c-49b1-a47b-76d47ba2c6e5-combined-ca-bundle\") pod \"b708b100-166c-49b1-a47b-76d47ba2c6e5\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.607506 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-logs\") pod \"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91\" (UID: \"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.607526 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ftrk\" (UniqueName: \"kubernetes.io/projected/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-kube-api-access-9ftrk\") pod \"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91\" (UID: \"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.607546 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b708b100-166c-49b1-a47b-76d47ba2c6e5-httpd-run\") pod \"b708b100-166c-49b1-a47b-76d47ba2c6e5\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.607571 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bf3edad-487a-4c68-9dbf-9789a94f8fba-combined-ca-bundle\") pod \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\" (UID: \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.607593 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-config-data\") pod \"4702d881-d701-41a4-b36f-2d063a8ae246\" (UID: \"4702d881-d701-41a4-b36f-2d063a8ae246\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.607610 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b708b100-166c-49b1-a47b-76d47ba2c6e5-logs\") pod \"b708b100-166c-49b1-a47b-76d47ba2c6e5\" (UID: \"b708b100-166c-49b1-a47b-76d47ba2c6e5\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.607629 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bf3edad-487a-4c68-9dbf-9789a94f8fba-internal-tls-certs\") pod \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\" (UID: \"6bf3edad-487a-4c68-9dbf-9789a94f8fba\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.607936 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: E1125 09:48:01.607982 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:01 crc kubenswrapper[4776]: E1125 09:48:01.608084 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/bcaec0f3-13af-4f73-8e7f-7e575b8e2008-operator-scripts podName:bcaec0f3-13af-4f73-8e7f-7e575b8e2008 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:02.608040628 +0000 UTC m=+1427.649100171 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/bcaec0f3-13af-4f73-8e7f-7e575b8e2008-operator-scripts") pod "glanceb9b5-account-delete-tjrnr" (UID: "bcaec0f3-13af-4f73-8e7f-7e575b8e2008") : configmap "openstack-scripts" not found Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.608240 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.617932 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbicandb0b-account-delete-xtvnn" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.624577 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4702d881-d701-41a4-b36f-2d063a8ae246-logs" (OuterVolumeSpecName: "logs") pod "4702d881-d701-41a4-b36f-2d063a8ae246" (UID: "4702d881-d701-41a4-b36f-2d063a8ae246"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.626957 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-logs" (OuterVolumeSpecName: "logs") pod "6a0ab5a5-2d62-4457-be9d-fe8242ab7c91" (UID: "6a0ab5a5-2d62-4457-be9d-fe8242ab7c91"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.642991 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b708b100-166c-49b1-a47b-76d47ba2c6e5-logs" (OuterVolumeSpecName: "logs") pod "b708b100-166c-49b1-a47b-76d47ba2c6e5" (UID: "b708b100-166c-49b1-a47b-76d47ba2c6e5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.650012 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6bf3edad-487a-4c68-9dbf-9789a94f8fba-logs" (OuterVolumeSpecName: "logs") pod "6bf3edad-487a-4c68-9dbf-9789a94f8fba" (UID: "6bf3edad-487a-4c68-9dbf-9789a94f8fba"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.652979 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b708b100-166c-49b1-a47b-76d47ba2c6e5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b708b100-166c-49b1-a47b-76d47ba2c6e5" (UID: "b708b100-166c-49b1-a47b-76d47ba2c6e5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.655838 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b708b100-166c-49b1-a47b-76d47ba2c6e5-kube-api-access-dkxgs" (OuterVolumeSpecName: "kube-api-access-dkxgs") pod "b708b100-166c-49b1-a47b-76d47ba2c6e5" (UID: "b708b100-166c-49b1-a47b-76d47ba2c6e5"). InnerVolumeSpecName "kube-api-access-dkxgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.655997 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-kube-api-access-9ftrk" (OuterVolumeSpecName: "kube-api-access-9ftrk") pod "6a0ab5a5-2d62-4457-be9d-fe8242ab7c91" (UID: "6a0ab5a5-2d62-4457-be9d-fe8242ab7c91"). InnerVolumeSpecName "kube-api-access-9ftrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.657995 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4702d881-d701-41a4-b36f-2d063a8ae246-kube-api-access-ldgn5" (OuterVolumeSpecName: "kube-api-access-ldgn5") pod "4702d881-d701-41a4-b36f-2d063a8ae246" (UID: "4702d881-d701-41a4-b36f-2d063a8ae246"). InnerVolumeSpecName "kube-api-access-ldgn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.662698 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4702d881-d701-41a4-b36f-2d063a8ae246" (UID: "4702d881-d701-41a4-b36f-2d063a8ae246"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.680214 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "86e320e0-80ee-40df-b3a1-a48cb810a435" (UID: "86e320e0-80ee-40df-b3a1-a48cb810a435"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.706102 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15bbd95b-2071-4988-83c7-8411c8a974fb" path="/var/lib/kubelet/pods/15bbd95b-2071-4988-83c7-8411c8a974fb/volumes" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.706900 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1de7f06d-24e7-4114-966b-d62df3366006" path="/var/lib/kubelet/pods/1de7f06d-24e7-4114-966b-d62df3366006/volumes" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.710004 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ftrk\" (UniqueName: \"kubernetes.io/projected/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-kube-api-access-9ftrk\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.710032 4776 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b708b100-166c-49b1-a47b-76d47ba2c6e5-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.710042 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.710053 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b708b100-166c-49b1-a47b-76d47ba2c6e5-logs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.710063 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkxgs\" (UniqueName: \"kubernetes.io/projected/b708b100-166c-49b1-a47b-76d47ba2c6e5-kube-api-access-dkxgs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.710087 4776 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.710095 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldgn5\" (UniqueName: \"kubernetes.io/projected/4702d881-d701-41a4-b36f-2d063a8ae246-kube-api-access-ldgn5\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.710103 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4702d881-d701-41a4-b36f-2d063a8ae246-logs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.710112 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6bf3edad-487a-4c68-9dbf-9789a94f8fba-logs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.710119 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-logs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.712871 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bf3edad-487a-4c68-9dbf-9789a94f8fba-kube-api-access-ffdd6" (OuterVolumeSpecName: "kube-api-access-ffdd6") pod "6bf3edad-487a-4c68-9dbf-9789a94f8fba" (UID: "6bf3edad-487a-4c68-9dbf-9789a94f8fba"). InnerVolumeSpecName "kube-api-access-ffdd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.712998 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b708b100-166c-49b1-a47b-76d47ba2c6e5-scripts" (OuterVolumeSpecName: "scripts") pod "b708b100-166c-49b1-a47b-76d47ba2c6e5" (UID: "b708b100-166c-49b1-a47b-76d47ba2c6e5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.713502 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="f32041a0-a8fc-48a5-afab-476baff89e8c" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.104:5671: connect: connection refused" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.714782 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b686729-83aa-48c6-9cce-cd28b26dd4b4" path="/var/lib/kubelet/pods/2b686729-83aa-48c6-9cce-cd28b26dd4b4/volumes" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.715390 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41e5f9c4-a3fc-4d92-bfee-09ee2fade2be" path="/var/lib/kubelet/pods/41e5f9c4-a3fc-4d92-bfee-09ee2fade2be/volumes" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.715981 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5705e80c-320b-49da-8c10-743627c8dbdb" path="/var/lib/kubelet/pods/5705e80c-320b-49da-8c10-743627c8dbdb/volumes" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.717351 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="789daf57-0bbd-4e06-84d4-37e9ca59e390" path="/var/lib/kubelet/pods/789daf57-0bbd-4e06-84d4-37e9ca59e390/volumes" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.717961 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83e76214-d644-4e9b-ab3a-2cbcae40cb23" path="/var/lib/kubelet/pods/83e76214-d644-4e9b-ab3a-2cbcae40cb23/volumes" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.718585 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e7d1dce-289c-409f-a68e-b214f8f4cc14" path="/var/lib/kubelet/pods/9e7d1dce-289c-409f-a68e-b214f8f4cc14/volumes" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.719640 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc920d71-4308-4d90-a891-8b760651ccc9" path="/var/lib/kubelet/pods/bc920d71-4308-4d90-a891-8b760651ccc9/volumes" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.729592 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "b708b100-166c-49b1-a47b-76d47ba2c6e5" (UID: "b708b100-166c-49b1-a47b-76d47ba2c6e5"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.735179 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca5a2dc7-b729-4038-9028-f08d37442b18" path="/var/lib/kubelet/pods/ca5a2dc7-b729-4038-9028-f08d37442b18/volumes" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.736252 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e09f9958-4faf-4a28-8214-c3ead146122c" path="/var/lib/kubelet/pods/e09f9958-4faf-4a28-8214-c3ead146122c/volumes" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.739842 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e107d317-acb4-46a2-8f19-836572e8904e" path="/var/lib/kubelet/pods/e107d317-acb4-46a2-8f19-836572e8904e/volumes" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.744455 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eadc8c7e-32f4-4bbf-8ac3-51ebbe471718" path="/var/lib/kubelet/pods/eadc8c7e-32f4-4bbf-8ac3-51ebbe471718/volumes" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.745970 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e" path="/var/lib/kubelet/pods/ed9fc3ce-a167-4ddf-b5b6-a4eb5c49638e/volumes" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.813637 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.813745 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-scripts\") pod \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.813840 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-public-tls-certs\") pod \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.813922 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-logs\") pod \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.814038 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-combined-ca-bundle\") pod \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.814113 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jznnd\" (UniqueName: \"kubernetes.io/projected/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-kube-api-access-jznnd\") pod \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.814177 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-config-data\") pod \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.814267 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d40516c-5fcc-4b24-918f-95bf79cb94b4-operator-scripts\") pod \"6d40516c-5fcc-4b24-918f-95bf79cb94b4\" (UID: \"6d40516c-5fcc-4b24-918f-95bf79cb94b4\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.814333 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-httpd-run\") pod \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\" (UID: \"d2cda6de-ef12-42f4-a6e5-824c2db3cd01\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.814379 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sd9gw\" (UniqueName: \"kubernetes.io/projected/6d40516c-5fcc-4b24-918f-95bf79cb94b4-kube-api-access-sd9gw\") pod \"6d40516c-5fcc-4b24-918f-95bf79cb94b4\" (UID: \"6d40516c-5fcc-4b24-918f-95bf79cb94b4\") " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.817297 4776 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.817362 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b708b100-166c-49b1-a47b-76d47ba2c6e5-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.817384 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffdd6\" (UniqueName: \"kubernetes.io/projected/6bf3edad-487a-4c68-9dbf-9789a94f8fba-kube-api-access-ffdd6\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.820329 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-config-data" (OuterVolumeSpecName: "config-data") pod "6a0ab5a5-2d62-4457-be9d-fe8242ab7c91" (UID: "6a0ab5a5-2d62-4457-be9d-fe8242ab7c91"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.820775 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d40516c-5fcc-4b24-918f-95bf79cb94b4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6d40516c-5fcc-4b24-918f-95bf79cb94b4" (UID: "6d40516c-5fcc-4b24-918f-95bf79cb94b4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.827707 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d2cda6de-ef12-42f4-a6e5-824c2db3cd01" (UID: "d2cda6de-ef12-42f4-a6e5-824c2db3cd01"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.833689 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-logs" (OuterVolumeSpecName: "logs") pod "d2cda6de-ef12-42f4-a6e5-824c2db3cd01" (UID: "d2cda6de-ef12-42f4-a6e5-824c2db3cd01"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.836858 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-scripts" (OuterVolumeSpecName: "scripts") pod "d2cda6de-ef12-42f4-a6e5-824c2db3cd01" (UID: "d2cda6de-ef12-42f4-a6e5-824c2db3cd01"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.839446 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-kube-api-access-jznnd" (OuterVolumeSpecName: "kube-api-access-jznnd") pod "d2cda6de-ef12-42f4-a6e5-824c2db3cd01" (UID: "d2cda6de-ef12-42f4-a6e5-824c2db3cd01"). InnerVolumeSpecName "kube-api-access-jznnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.906926 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "d2cda6de-ef12-42f4-a6e5-824c2db3cd01" (UID: "d2cda6de-ef12-42f4-a6e5-824c2db3cd01"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.907042 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d40516c-5fcc-4b24-918f-95bf79cb94b4-kube-api-access-sd9gw" (OuterVolumeSpecName: "kube-api-access-sd9gw") pod "6d40516c-5fcc-4b24-918f-95bf79cb94b4" (UID: "6d40516c-5fcc-4b24-918f-95bf79cb94b4"). InnerVolumeSpecName "kube-api-access-sd9gw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.908406 4776 generic.go:334] "Generic (PLEG): container finished" podID="09f0463c-f091-495e-9773-c07fe902c200" containerID="94684aa85a999df01587ff8336fcf45d1847b413f000fe304eae44dda233eae6" exitCode=2 Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.917425 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-769ddf488d-46ssj" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.918743 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.918767 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jznnd\" (UniqueName: \"kubernetes.io/projected/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-kube-api-access-jznnd\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.918778 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d40516c-5fcc-4b24-918f-95bf79cb94b4-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.918787 4776 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.918795 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sd9gw\" (UniqueName: \"kubernetes.io/projected/6d40516c-5fcc-4b24-918f-95bf79cb94b4-kube-api-access-sd9gw\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.918815 4776 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.918824 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.918832 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-logs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.922551 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.923838 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "86e320e0-80ee-40df-b3a1-a48cb810a435" (UID: "86e320e0-80ee-40df-b3a1-a48cb810a435"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.926807 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.928027 4776 generic.go:334] "Generic (PLEG): container finished" podID="1d1e9a91-e344-442f-8bbc-90b1f98e2e1f" containerID="9779df52088b267add1650137fc3fc3f92ad6d3ea2e75a34659668d2ef3813bb" exitCode=0 Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.929130 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a0ab5a5-2d62-4457-be9d-fe8242ab7c91" (UID: "6a0ab5a5-2d62-4457-be9d-fe8242ab7c91"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.937906 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-667769d556-lqtlp" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.946889 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.977287 4776 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.989668 4776 generic.go:334] "Generic (PLEG): container finished" podID="aab7d356-ec67-4e5d-9c25-f26e39786f79" containerID="a4df2757871f92f27d85077692ce959d3142d7496b28c90b9062a0ed0d5d450d" exitCode=0 Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.992012 4776 generic.go:334] "Generic (PLEG): container finished" podID="8bae8251-3d5d-490b-9116-73769dc18101" containerID="257f4e7f047aaf2041def20b6caaee331d83d82f30d019096ad21bfd241e42a8" exitCode=0 Nov 25 09:48:01 crc kubenswrapper[4776]: I1125 09:48:01.993769 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.004265 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b708b100-166c-49b1-a47b-76d47ba2c6e5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b708b100-166c-49b1-a47b-76d47ba2c6e5" (UID: "b708b100-166c-49b1-a47b-76d47ba2c6e5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.006368 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.010826 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bf3edad-487a-4c68-9dbf-9789a94f8fba-config-data" (OuterVolumeSpecName: "config-data") pod "6bf3edad-487a-4c68-9dbf-9789a94f8fba" (UID: "6bf3edad-487a-4c68-9dbf-9789a94f8fba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.017956 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbicandb0b-account-delete-xtvnn" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.019727 4776 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.019746 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.019758 4776 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.019768 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bf3edad-487a-4c68-9dbf-9789a94f8fba-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.019776 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b708b100-166c-49b1-a47b-76d47ba2c6e5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.020771 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "86e320e0-80ee-40df-b3a1-a48cb810a435" (UID: "86e320e0-80ee-40df-b3a1-a48cb810a435"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.021199 4776 generic.go:334] "Generic (PLEG): container finished" podID="f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4" containerID="51468cb0acece8ce1a386bdf49793e6322b08cddac5cf23fc46abef1f5330e22" exitCode=0 Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.021294 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5864b7d7fc-g9z2r" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.021370 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.028628 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder774d-account-delete-b29n5" podUID="ef85f90f-c377-4273-9fff-451420d6008a" containerName="mariadb-account-delete" containerID="cri-o://02f077bb80ad2e06d150c6b44d4bfbe70a57d6d488294c69cf9fd29d74fce9a8" gracePeriod=30 Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.032467 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron5151-account-delete-8vbjg" podUID="73f2c2ab-6cb6-4544-a3f7-20722e5b6400" containerName="mariadb-account-delete" containerID="cri-o://f44277236f6a634e8f611e825fd05aabebb686cac5979f168a7f50210ff4ef17" gracePeriod=30 Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.032641 4776 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/placemente176-account-delete-wbgz5" secret="" err="secret \"galera-openstack-dockercfg-2t654\" not found" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.032778 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/novaapi3b46-account-delete-9w5r9" podUID="14b51bba-0b15-469b-bdea-90fae8d9b976" containerName="mariadb-account-delete" containerID="cri-o://d3de2925044530f7fd248214ce715db43c6f7a123e71e4f5fd11e8af0b5cda0e" gracePeriod=30 Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.039367 4776 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novacell09dc4-account-delete-w5xcj" secret="" err="secret \"galera-openstack-dockercfg-2t654\" not found" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.058593 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bf3edad-487a-4c68-9dbf-9789a94f8fba-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6bf3edad-487a-4c68-9dbf-9789a94f8fba" (UID: "6bf3edad-487a-4c68-9dbf-9789a94f8fba"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.062954 4776 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.065962 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-config-data" (OuterVolumeSpecName: "config-data") pod "86e320e0-80ee-40df-b3a1-a48cb810a435" (UID: "86e320e0-80ee-40df-b3a1-a48cb810a435"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.077724 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d2cda6de-ef12-42f4-a6e5-824c2db3cd01" (UID: "d2cda6de-ef12-42f4-a6e5-824c2db3cd01"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.093919 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a0b97845-4578-4801-896a-281fdafdb351" (UID: "a0b97845-4578-4801-896a-281fdafdb351"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.118950 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4702d881-d701-41a4-b36f-2d063a8ae246" (UID: "4702d881-d701-41a4-b36f-2d063a8ae246"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.123559 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.123581 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.123591 4776 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.123600 4776 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/86e320e0-80ee-40df-b3a1-a48cb810a435-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.123609 4776 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bf3edad-487a-4c68-9dbf-9789a94f8fba-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.123617 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.123625 4776 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0b97845-4578-4801-896a-281fdafdb351-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.153596 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b708b100-166c-49b1-a47b-76d47ba2c6e5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "b708b100-166c-49b1-a47b-76d47ba2c6e5" (UID: "b708b100-166c-49b1-a47b-76d47ba2c6e5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.156902 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d2cda6de-ef12-42f4-a6e5-824c2db3cd01" (UID: "d2cda6de-ef12-42f4-a6e5-824c2db3cd01"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.158757 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "6a0ab5a5-2d62-4457-be9d-fe8242ab7c91" (UID: "6a0ab5a5-2d62-4457-be9d-fe8242ab7c91"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.159262 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b708b100-166c-49b1-a47b-76d47ba2c6e5-config-data" (OuterVolumeSpecName: "config-data") pod "b708b100-166c-49b1-a47b-76d47ba2c6e5" (UID: "b708b100-166c-49b1-a47b-76d47ba2c6e5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.162301 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-config-data" (OuterVolumeSpecName: "config-data") pod "4702d881-d701-41a4-b36f-2d063a8ae246" (UID: "4702d881-d701-41a4-b36f-2d063a8ae246"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.163077 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bf3edad-487a-4c68-9dbf-9789a94f8fba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6bf3edad-487a-4c68-9dbf-9789a94f8fba" (UID: "6bf3edad-487a-4c68-9dbf-9789a94f8fba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.172735 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4702d881-d701-41a4-b36f-2d063a8ae246" (UID: "4702d881-d701-41a4-b36f-2d063a8ae246"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.187289 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4702d881-d701-41a4-b36f-2d063a8ae246" (UID: "4702d881-d701-41a4-b36f-2d063a8ae246"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.195598 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bf3edad-487a-4c68-9dbf-9789a94f8fba-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6bf3edad-487a-4c68-9dbf-9789a94f8fba" (UID: "6bf3edad-487a-4c68-9dbf-9789a94f8fba"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.224661 4776 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.224687 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bf3edad-487a-4c68-9dbf-9789a94f8fba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.224699 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.224708 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b708b100-166c-49b1-a47b-76d47ba2c6e5-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.224716 4776 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.224724 4776 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.224732 4776 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b708b100-166c-49b1-a47b-76d47ba2c6e5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.224741 4776 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4702d881-d701-41a4-b36f-2d063a8ae246-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.224751 4776 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6bf3edad-487a-4c68-9dbf-9789a94f8fba-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.226834 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-config-data" (OuterVolumeSpecName: "config-data") pod "d2cda6de-ef12-42f4-a6e5-824c2db3cd01" (UID: "d2cda6de-ef12-42f4-a6e5-824c2db3cd01"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.236521 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"09f0463c-f091-495e-9773-c07fe902c200","Type":"ContainerDied","Data":"94684aa85a999df01587ff8336fcf45d1847b413f000fe304eae44dda233eae6"} Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.236650 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5wmvt" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.236665 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-769ddf488d-46ssj" event={"ID":"4702d881-d701-41a4-b36f-2d063a8ae246","Type":"ContainerDied","Data":"df59e2ce88d3eada192f146759b343022c729c3eedcbc0cb1bf5166b45a8f96f"} Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.236688 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a0ab5a5-2d62-4457-be9d-fe8242ab7c91","Type":"ContainerDied","Data":"fe6b9828c965a764d125db4efe76edee643df6bf78aabbb24e0fbbd1bfeabd1f"} Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.236703 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b708b100-166c-49b1-a47b-76d47ba2c6e5","Type":"ContainerDied","Data":"928800eb5246288399c4e346cf14e1b06defe30c42280c8cbd4ccdd157393af7"} Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.236726 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-66cdf565f-6lzqw" event={"ID":"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f","Type":"ContainerDied","Data":"9779df52088b267add1650137fc3fc3f92ad6d3ea2e75a34659668d2ef3813bb"} Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.236743 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-66cdf565f-6lzqw" event={"ID":"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f","Type":"ContainerDied","Data":"b690b96a14b9b4d3e66ad1db4ce9b8289a92134e04d4486aa5aa883c1148be9b"} Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.236755 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b690b96a14b9b4d3e66ad1db4ce9b8289a92134e04d4486aa5aa883c1148be9b" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.236765 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-667769d556-lqtlp" event={"ID":"a0b97845-4578-4801-896a-281fdafdb351","Type":"ContainerDied","Data":"422d5dbcd2c0da803fdd9482a7272e0f11941f78aaba66e0811b7ff8c628f5f6"} Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.236778 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"86e320e0-80ee-40df-b3a1-a48cb810a435","Type":"ContainerDied","Data":"e533e247825031f6a8cc6b6a24f05322c2eb862b637377dc7167d71281aadbac"} Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.236794 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aab7d356-ec67-4e5d-9c25-f26e39786f79","Type":"ContainerDied","Data":"a4df2757871f92f27d85077692ce959d3142d7496b28c90b9062a0ed0d5d450d"} Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.236809 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8bae8251-3d5d-490b-9116-73769dc18101","Type":"ContainerDied","Data":"257f4e7f047aaf2041def20b6caaee331d83d82f30d019096ad21bfd241e42a8"} Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.236822 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d2cda6de-ef12-42f4-a6e5-824c2db3cd01","Type":"ContainerDied","Data":"42441d71e06e7f040dbf095b2f3e3d0ef18caf0085fa5e84a1f401d5ef9342a1"} Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.236836 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6bf3edad-487a-4c68-9dbf-9789a94f8fba","Type":"ContainerDied","Data":"36817d6beb7f295e36cecc52ef61815f589eda60f500320fe51a837b338cf6db"} Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.236848 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicandb0b-account-delete-xtvnn" event={"ID":"6d40516c-5fcc-4b24-918f-95bf79cb94b4","Type":"ContainerDied","Data":"512b6cdcd78de19c4a57934ea8b56bff013e1d10be30032755b9e655ec66cdd1"} Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.236859 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="512b6cdcd78de19c4a57934ea8b56bff013e1d10be30032755b9e655ec66cdd1" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.236868 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4","Type":"ContainerDied","Data":"51468cb0acece8ce1a386bdf49793e6322b08cddac5cf23fc46abef1f5330e22"} Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.236880 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4","Type":"ContainerDied","Data":"a7cce054b7bef15b7a4a6a15d9d6f8782b9044f4308d061712276043ffdd055d"} Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.236889 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7cce054b7bef15b7a4a6a15d9d6f8782b9044f4308d061712276043ffdd055d" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.236907 4776 scope.go:117] "RemoveContainer" containerID="4e9d85a2660614a7e89218283e8f24a0ddfcb9cad63c01ab8119453aceaad099" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.286325 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-66cdf565f-6lzqw" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.327213 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2cda6de-ef12-42f4-a6e5-824c2db3cd01-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.363401 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5wmvt" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.382953 4776 scope.go:117] "RemoveContainer" containerID="d45871f326a3dbb0f683b3d995f92f15682fd356f7dd9ec10b42a715df612c77" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.427934 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-config-data\") pod \"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f\" (UID: \"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f\") " Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.428116 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-logs\") pod \"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f\" (UID: \"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f\") " Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.428164 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-config-data-custom\") pod \"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f\" (UID: \"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f\") " Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.428275 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6b2d\" (UniqueName: \"kubernetes.io/projected/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-kube-api-access-w6b2d\") pod \"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f\" (UID: \"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f\") " Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.428315 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-combined-ca-bundle\") pod \"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f\" (UID: \"1d1e9a91-e344-442f-8bbc-90b1f98e2e1f\") " Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.429041 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-logs" (OuterVolumeSpecName: "logs") pod "1d1e9a91-e344-442f-8bbc-90b1f98e2e1f" (UID: "1d1e9a91-e344-442f-8bbc-90b1f98e2e1f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.429281 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-logs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: E1125 09:48:02.429426 4776 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Nov 25 09:48:02 crc kubenswrapper[4776]: E1125 09:48:02.429543 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f32041a0-a8fc-48a5-afab-476baff89e8c-config-data podName:f32041a0-a8fc-48a5-afab-476baff89e8c nodeName:}" failed. No retries permitted until 2025-11-25 09:48:10.429523023 +0000 UTC m=+1435.470582646 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/f32041a0-a8fc-48a5-afab-476baff89e8c-config-data") pod "rabbitmq-cell1-server-0" (UID: "f32041a0-a8fc-48a5-afab-476baff89e8c") : configmap "rabbitmq-cell1-config-data" not found Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.432552 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1d1e9a91-e344-442f-8bbc-90b1f98e2e1f" (UID: "1d1e9a91-e344-442f-8bbc-90b1f98e2e1f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.436601 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-kube-api-access-w6b2d" (OuterVolumeSpecName: "kube-api-access-w6b2d") pod "1d1e9a91-e344-442f-8bbc-90b1f98e2e1f" (UID: "1d1e9a91-e344-442f-8bbc-90b1f98e2e1f"). InnerVolumeSpecName "kube-api-access-w6b2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.465509 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d1e9a91-e344-442f-8bbc-90b1f98e2e1f" (UID: "1d1e9a91-e344-442f-8bbc-90b1f98e2e1f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.476107 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5wmvt"] Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.497085 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.523192 4776 scope.go:117] "RemoveContainer" containerID="7927ef3d5f7eb88faf6e7c414da603268c27589afcdfe2331cf08fabea8c8c3c" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.523450 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.524499 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 25 09:48:02 crc kubenswrapper[4776]: E1125 09:48:02.533217 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:02 crc kubenswrapper[4776]: E1125 09:48:02.533307 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/14b51bba-0b15-469b-bdea-90fae8d9b976-operator-scripts podName:14b51bba-0b15-469b-bdea-90fae8d9b976 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:04.533283802 +0000 UTC m=+1429.574343425 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/14b51bba-0b15-469b-bdea-90fae8d9b976-operator-scripts") pod "novaapi3b46-account-delete-9w5r9" (UID: "14b51bba-0b15-469b-bdea-90fae8d9b976") : configmap "openstack-scripts" not found Nov 25 09:48:02 crc kubenswrapper[4776]: E1125 09:48:02.533355 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:02 crc kubenswrapper[4776]: E1125 09:48:02.533411 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ef85f90f-c377-4273-9fff-451420d6008a-operator-scripts podName:ef85f90f-c377-4273-9fff-451420d6008a nodeName:}" failed. No retries permitted until 2025-11-25 09:48:04.533397565 +0000 UTC m=+1429.574457118 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/ef85f90f-c377-4273-9fff-451420d6008a-operator-scripts") pod "cinder774d-account-delete-b29n5" (UID: "ef85f90f-c377-4273-9fff-451420d6008a") : configmap "openstack-scripts" not found Nov 25 09:48:02 crc kubenswrapper[4776]: E1125 09:48:02.533448 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:02 crc kubenswrapper[4776]: E1125 09:48:02.533469 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6be099e6-89d6-45a0-9d0f-4dca4d91701a-operator-scripts podName:6be099e6-89d6-45a0-9d0f-4dca4d91701a nodeName:}" failed. No retries permitted until 2025-11-25 09:48:04.533463537 +0000 UTC m=+1429.574523090 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/6be099e6-89d6-45a0-9d0f-4dca4d91701a-operator-scripts") pod "placemente176-account-delete-wbgz5" (UID: "6be099e6-89d6-45a0-9d0f-4dca4d91701a") : configmap "openstack-scripts" not found Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.533491 4776 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.533503 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6b2d\" (UniqueName: \"kubernetes.io/projected/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-kube-api-access-w6b2d\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.533513 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: E1125 09:48:02.533539 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:02 crc kubenswrapper[4776]: E1125 09:48:02.533559 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/73f2c2ab-6cb6-4544-a3f7-20722e5b6400-operator-scripts podName:73f2c2ab-6cb6-4544-a3f7-20722e5b6400 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:04.533553539 +0000 UTC m=+1429.574613092 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/73f2c2ab-6cb6-4544-a3f7-20722e5b6400-operator-scripts") pod "neutron5151-account-delete-8vbjg" (UID: "73f2c2ab-6cb6-4544-a3f7-20722e5b6400") : configmap "openstack-scripts" not found Nov 25 09:48:02 crc kubenswrapper[4776]: E1125 09:48:02.533619 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:02 crc kubenswrapper[4776]: E1125 09:48:02.533714 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/946ea713-bfba-449f-82ab-28e915469938-operator-scripts podName:946ea713-bfba-449f-82ab-28e915469938 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:04.533706383 +0000 UTC m=+1429.574765936 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/946ea713-bfba-449f-82ab-28e915469938-operator-scripts") pod "novacell09dc4-account-delete-w5xcj" (UID: "946ea713-bfba-449f-82ab-28e915469938") : configmap "openstack-scripts" not found Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.564587 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.584020 4776 scope.go:117] "RemoveContainer" containerID="6388f255053948e8f439f0e43bfe9480c34d62fbb564522cf02851e2a2aef43a" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.590207 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-config-data" (OuterVolumeSpecName: "config-data") pod "1d1e9a91-e344-442f-8bbc-90b1f98e2e1f" (UID: "1d1e9a91-e344-442f-8bbc-90b1f98e2e1f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.590369 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.611019 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.623893 4776 scope.go:117] "RemoveContainer" containerID="5919b100098159f61f76d181a1ae66bfafcb2916a56d7fe4fe1cee93ad3c32d0" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.635954 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-5864b7d7fc-g9z2r"] Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.636478 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/09f0463c-f091-495e-9773-c07fe902c200-kube-state-metrics-tls-config\") pod \"09f0463c-f091-495e-9773-c07fe902c200\" (UID: \"09f0463c-f091-495e-9773-c07fe902c200\") " Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.636576 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/09f0463c-f091-495e-9773-c07fe902c200-kube-state-metrics-tls-certs\") pod \"09f0463c-f091-495e-9773-c07fe902c200\" (UID: \"09f0463c-f091-495e-9773-c07fe902c200\") " Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.636622 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09f0463c-f091-495e-9773-c07fe902c200-combined-ca-bundle\") pod \"09f0463c-f091-495e-9773-c07fe902c200\" (UID: \"09f0463c-f091-495e-9773-c07fe902c200\") " Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.636681 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4-config-data\") pod \"f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4\" (UID: \"f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4\") " Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.636742 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vr9x\" (UniqueName: \"kubernetes.io/projected/8bae8251-3d5d-490b-9116-73769dc18101-kube-api-access-6vr9x\") pod \"8bae8251-3d5d-490b-9116-73769dc18101\" (UID: \"8bae8251-3d5d-490b-9116-73769dc18101\") " Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.636804 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4-combined-ca-bundle\") pod \"f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4\" (UID: \"f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4\") " Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.636834 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf7l8\" (UniqueName: \"kubernetes.io/projected/f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4-kube-api-access-bf7l8\") pod \"f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4\" (UID: \"f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4\") " Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.636870 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfg9g\" (UniqueName: \"kubernetes.io/projected/09f0463c-f091-495e-9773-c07fe902c200-kube-api-access-tfg9g\") pod \"09f0463c-f091-495e-9773-c07fe902c200\" (UID: \"09f0463c-f091-495e-9773-c07fe902c200\") " Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.636895 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bae8251-3d5d-490b-9116-73769dc18101-config-data\") pod \"8bae8251-3d5d-490b-9116-73769dc18101\" (UID: \"8bae8251-3d5d-490b-9116-73769dc18101\") " Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.636923 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bae8251-3d5d-490b-9116-73769dc18101-combined-ca-bundle\") pod \"8bae8251-3d5d-490b-9116-73769dc18101\" (UID: \"8bae8251-3d5d-490b-9116-73769dc18101\") " Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.637417 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: E1125 09:48:02.637485 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:02 crc kubenswrapper[4776]: E1125 09:48:02.637531 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/bcaec0f3-13af-4f73-8e7f-7e575b8e2008-operator-scripts podName:bcaec0f3-13af-4f73-8e7f-7e575b8e2008 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:04.637514653 +0000 UTC m=+1429.678574206 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/bcaec0f3-13af-4f73-8e7f-7e575b8e2008-operator-scripts") pod "glanceb9b5-account-delete-tjrnr" (UID: "bcaec0f3-13af-4f73-8e7f-7e575b8e2008") : configmap "openstack-scripts" not found Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.661413 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4-kube-api-access-bf7l8" (OuterVolumeSpecName: "kube-api-access-bf7l8") pod "f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4" (UID: "f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4"). InnerVolumeSpecName "kube-api-access-bf7l8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.661854 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bae8251-3d5d-490b-9116-73769dc18101-kube-api-access-6vr9x" (OuterVolumeSpecName: "kube-api-access-6vr9x") pod "8bae8251-3d5d-490b-9116-73769dc18101" (UID: "8bae8251-3d5d-490b-9116-73769dc18101"). InnerVolumeSpecName "kube-api-access-6vr9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.662815 4776 scope.go:117] "RemoveContainer" containerID="7f99367ae1299a4af7ca7bdb87418a9b017add504f1ee9377bbcccc241edd6fb" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.678753 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09f0463c-f091-495e-9773-c07fe902c200-kube-api-access-tfg9g" (OuterVolumeSpecName: "kube-api-access-tfg9g") pod "09f0463c-f091-495e-9773-c07fe902c200" (UID: "09f0463c-f091-495e-9773-c07fe902c200"). InnerVolumeSpecName "kube-api-access-tfg9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.685496 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-5864b7d7fc-g9z2r"] Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.723325 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09f0463c-f091-495e-9773-c07fe902c200-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "09f0463c-f091-495e-9773-c07fe902c200" (UID: "09f0463c-f091-495e-9773-c07fe902c200"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.735552 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4" (UID: "f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.741508 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd5ce048-5a09-45a4-9a7d-66712b326ab7-memcached-tls-certs\") pod \"cd5ce048-5a09-45a4-9a7d-66712b326ab7\" (UID: \"cd5ce048-5a09-45a4-9a7d-66712b326ab7\") " Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.741655 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cd5ce048-5a09-45a4-9a7d-66712b326ab7-kolla-config\") pod \"cd5ce048-5a09-45a4-9a7d-66712b326ab7\" (UID: \"cd5ce048-5a09-45a4-9a7d-66712b326ab7\") " Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.741746 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd5ce048-5a09-45a4-9a7d-66712b326ab7-combined-ca-bundle\") pod \"cd5ce048-5a09-45a4-9a7d-66712b326ab7\" (UID: \"cd5ce048-5a09-45a4-9a7d-66712b326ab7\") " Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.741766 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgvgj\" (UniqueName: \"kubernetes.io/projected/cd5ce048-5a09-45a4-9a7d-66712b326ab7-kube-api-access-fgvgj\") pod \"cd5ce048-5a09-45a4-9a7d-66712b326ab7\" (UID: \"cd5ce048-5a09-45a4-9a7d-66712b326ab7\") " Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.741792 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd5ce048-5a09-45a4-9a7d-66712b326ab7-config-data\") pod \"cd5ce048-5a09-45a4-9a7d-66712b326ab7\" (UID: \"cd5ce048-5a09-45a4-9a7d-66712b326ab7\") " Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.742171 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf7l8\" (UniqueName: \"kubernetes.io/projected/f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4-kube-api-access-bf7l8\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.742184 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfg9g\" (UniqueName: \"kubernetes.io/projected/09f0463c-f091-495e-9773-c07fe902c200-kube-api-access-tfg9g\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.742194 4776 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/09f0463c-f091-495e-9773-c07fe902c200-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.742203 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vr9x\" (UniqueName: \"kubernetes.io/projected/8bae8251-3d5d-490b-9116-73769dc18101-kube-api-access-6vr9x\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.742215 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.742837 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd5ce048-5a09-45a4-9a7d-66712b326ab7-config-data" (OuterVolumeSpecName: "config-data") pod "cd5ce048-5a09-45a4-9a7d-66712b326ab7" (UID: "cd5ce048-5a09-45a4-9a7d-66712b326ab7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.748773 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd5ce048-5a09-45a4-9a7d-66712b326ab7-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "cd5ce048-5a09-45a4-9a7d-66712b326ab7" (UID: "cd5ce048-5a09-45a4-9a7d-66712b326ab7"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.753446 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bae8251-3d5d-490b-9116-73769dc18101-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8bae8251-3d5d-490b-9116-73769dc18101" (UID: "8bae8251-3d5d-490b-9116-73769dc18101"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.760945 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-769ddf488d-46ssj"] Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.762825 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4-config-data" (OuterVolumeSpecName: "config-data") pod "f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4" (UID: "f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.775196 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd5ce048-5a09-45a4-9a7d-66712b326ab7-kube-api-access-fgvgj" (OuterVolumeSpecName: "kube-api-access-fgvgj") pod "cd5ce048-5a09-45a4-9a7d-66712b326ab7" (UID: "cd5ce048-5a09-45a4-9a7d-66712b326ab7"). InnerVolumeSpecName "kube-api-access-fgvgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.790159 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-769ddf488d-46ssj"] Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.801778 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09f0463c-f091-495e-9773-c07fe902c200-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "09f0463c-f091-495e-9773-c07fe902c200" (UID: "09f0463c-f091-495e-9773-c07fe902c200"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.802126 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd5ce048-5a09-45a4-9a7d-66712b326ab7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd5ce048-5a09-45a4-9a7d-66712b326ab7" (UID: "cd5ce048-5a09-45a4-9a7d-66712b326ab7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.807019 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bae8251-3d5d-490b-9116-73769dc18101-config-data" (OuterVolumeSpecName: "config-data") pod "8bae8251-3d5d-490b-9116-73769dc18101" (UID: "8bae8251-3d5d-490b-9116-73769dc18101"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.829822 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd5ce048-5a09-45a4-9a7d-66712b326ab7-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "cd5ce048-5a09-45a4-9a7d-66712b326ab7" (UID: "cd5ce048-5a09-45a4-9a7d-66712b326ab7"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.832175 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09f0463c-f091-495e-9773-c07fe902c200-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "09f0463c-f091-495e-9773-c07fe902c200" (UID: "09f0463c-f091-495e-9773-c07fe902c200"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.837441 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.874381 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bae8251-3d5d-490b-9116-73769dc18101-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.874611 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bae8251-3d5d-490b-9116-73769dc18101-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.875211 4776 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cd5ce048-5a09-45a4-9a7d-66712b326ab7-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.875305 4776 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/09f0463c-f091-495e-9773-c07fe902c200-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.875421 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09f0463c-f091-495e-9773-c07fe902c200-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.875501 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd5ce048-5a09-45a4-9a7d-66712b326ab7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.875627 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgvgj\" (UniqueName: \"kubernetes.io/projected/cd5ce048-5a09-45a4-9a7d-66712b326ab7-kube-api-access-fgvgj\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.864421 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.876237 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.876312 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd5ce048-5a09-45a4-9a7d-66712b326ab7-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.876440 4776 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd5ce048-5a09-45a4-9a7d-66712b326ab7-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.885816 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.895948 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.901798 4776 scope.go:117] "RemoveContainer" containerID="856d610eaa67921f1c8d09ca91236aafc5070c66f02e6f2c2f6ee1982f26930e" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.912477 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.923421 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.929098 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.929249 4776 scope.go:117] "RemoveContainer" containerID="5fac5a59b29d2ae80cf45865e0db91319f65183c12c9b736093c4452a98a4355" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.938152 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.943617 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-667769d556-lqtlp"] Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.952509 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-667769d556-lqtlp"] Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.955404 4776 scope.go:117] "RemoveContainer" containerID="968312305c48cd2a75a04c3e2a43d37d9ed640ee14d080081ad79c323f2cf0f5" Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.959205 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 25 09:48:02 crc kubenswrapper[4776]: I1125 09:48:02.963267 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.043982 4776 generic.go:334] "Generic (PLEG): container finished" podID="cd5ce048-5a09-45a4-9a7d-66712b326ab7" containerID="6b83a886269bd5ccf82ce338326e0a5b6749a7805141d811c4f9ff42aa7148ca" exitCode=0 Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.044100 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.044123 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"cd5ce048-5a09-45a4-9a7d-66712b326ab7","Type":"ContainerDied","Data":"6b83a886269bd5ccf82ce338326e0a5b6749a7805141d811c4f9ff42aa7148ca"} Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.045250 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"cd5ce048-5a09-45a4-9a7d-66712b326ab7","Type":"ContainerDied","Data":"aa48600c469e2c4f1b609fa041f9ecd490af4b97a40bc93c0a48c6d44ea6bc8e"} Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.050332 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8bae8251-3d5d-490b-9116-73769dc18101","Type":"ContainerDied","Data":"8ab35a66b3792c7af7863e852b1699430b7994109c81a79b1dd2c13a6886c7cb"} Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.050439 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.059659 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"09f0463c-f091-495e-9773-c07fe902c200","Type":"ContainerDied","Data":"d1ffc8968dfe50f1ebb42b0a99c53bc306bdc7bc1023dd5d4b8c82859dbf7e39"} Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.059743 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.069323 4776 generic.go:334] "Generic (PLEG): container finished" podID="f32041a0-a8fc-48a5-afab-476baff89e8c" containerID="7b42621ff074baca099513ddb94523b955d55200c16323a7bb88a078926fca25" exitCode=0 Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.069401 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f32041a0-a8fc-48a5-afab-476baff89e8c","Type":"ContainerDied","Data":"7b42621ff074baca099513ddb94523b955d55200c16323a7bb88a078926fca25"} Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.071084 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.071084 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-66cdf565f-6lzqw" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.071543 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/novacell09dc4-account-delete-w5xcj" podUID="946ea713-bfba-449f-82ab-28e915469938" containerName="mariadb-account-delete" containerID="cri-o://ac0deb2776fd44dcbddf8b4d4bae71247d519e589f78423415504e15ea5a4a0d" gracePeriod=30 Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.098259 4776 scope.go:117] "RemoveContainer" containerID="643bd96dcd87a61d61b437895cd783519ab3ff279196abf17121bbe3e4d485dd" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.101104 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.108922 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Nov 25 09:48:03 crc kubenswrapper[4776]: E1125 09:48:03.117039 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2f383d0debcf43b01564a0db854854b92a294770e346344bcd8dce403c1e0251" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Nov 25 09:48:03 crc kubenswrapper[4776]: E1125 09:48:03.118602 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2f383d0debcf43b01564a0db854854b92a294770e346344bcd8dce403c1e0251" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Nov 25 09:48:03 crc kubenswrapper[4776]: E1125 09:48:03.120121 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2f383d0debcf43b01564a0db854854b92a294770e346344bcd8dce403c1e0251" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Nov 25 09:48:03 crc kubenswrapper[4776]: E1125 09:48:03.120199 4776 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="167c5bc5-b21e-4c51-8765-1dcbf290294b" containerName="galera" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.168410 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.182212 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 09:48:03 crc kubenswrapper[4776]: E1125 09:48:03.191036 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e5a40e8fd3cdfddb16f8d8578ce8ae7e8bb1b0df1b91c68efe717a27f4fc382c" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.201125 4776 scope.go:117] "RemoveContainer" containerID="9fddb69738d8e8fd199d1f7a677b07c2d8ceff1639dbe9784b2c6fd247da4ae1" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.203912 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.215642 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 09:48:03 crc kubenswrapper[4776]: E1125 09:48:03.215940 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e5a40e8fd3cdfddb16f8d8578ce8ae7e8bb1b0df1b91c68efe717a27f4fc382c" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 25 09:48:03 crc kubenswrapper[4776]: E1125 09:48:03.217552 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e5a40e8fd3cdfddb16f8d8578ce8ae7e8bb1b0df1b91c68efe717a27f4fc382c" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 25 09:48:03 crc kubenswrapper[4776]: E1125 09:48:03.217616 4776 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="f182a7e0-ebd8-4258-9269-43a662e39af8" containerName="nova-cell1-conductor-conductor" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.224739 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.231909 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.237831 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-66cdf565f-6lzqw"] Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.242979 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-66cdf565f-6lzqw"] Nov 25 09:48:03 crc kubenswrapper[4776]: E1125 09:48:03.284025 4776 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Nov 25 09:48:03 crc kubenswrapper[4776]: E1125 09:48:03.284199 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4512cf4e-6c75-493e-a80a-4b0accd750a8-config-data podName:4512cf4e-6c75-493e-a80a-4b0accd750a8 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:11.284168033 +0000 UTC m=+1436.325227586 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/4512cf4e-6c75-493e-a80a-4b0accd750a8-config-data") pod "rabbitmq-server-0" (UID: "4512cf4e-6c75-493e-a80a-4b0accd750a8") : configmap "rabbitmq-config-data" not found Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.305876 4776 scope.go:117] "RemoveContainer" containerID="0bd7925c5eadaefd62b355bbd7aebe47ce493b779918483df9c72dc9a141944b" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.418323 4776 scope.go:117] "RemoveContainer" containerID="24a3083e873c7a16456c526fdfe2bc77dcf34c30f066c1f1af77953f2e1ecddd" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.455284 4776 scope.go:117] "RemoveContainer" containerID="07109bfcbb050b4c8db424a410626fa31b2236c8d915084ff3b34d220f1c1cf0" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.494808 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.501745 4776 scope.go:117] "RemoveContainer" containerID="6b83a886269bd5ccf82ce338326e0a5b6749a7805141d811c4f9ff42aa7148ca" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.567837 4776 scope.go:117] "RemoveContainer" containerID="6b83a886269bd5ccf82ce338326e0a5b6749a7805141d811c4f9ff42aa7148ca" Nov 25 09:48:03 crc kubenswrapper[4776]: E1125 09:48:03.568391 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b83a886269bd5ccf82ce338326e0a5b6749a7805141d811c4f9ff42aa7148ca\": container with ID starting with 6b83a886269bd5ccf82ce338326e0a5b6749a7805141d811c4f9ff42aa7148ca not found: ID does not exist" containerID="6b83a886269bd5ccf82ce338326e0a5b6749a7805141d811c4f9ff42aa7148ca" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.568433 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b83a886269bd5ccf82ce338326e0a5b6749a7805141d811c4f9ff42aa7148ca"} err="failed to get container status \"6b83a886269bd5ccf82ce338326e0a5b6749a7805141d811c4f9ff42aa7148ca\": rpc error: code = NotFound desc = could not find container \"6b83a886269bd5ccf82ce338326e0a5b6749a7805141d811c4f9ff42aa7148ca\": container with ID starting with 6b83a886269bd5ccf82ce338326e0a5b6749a7805141d811c4f9ff42aa7148ca not found: ID does not exist" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.568462 4776 scope.go:117] "RemoveContainer" containerID="257f4e7f047aaf2041def20b6caaee331d83d82f30d019096ad21bfd241e42a8" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.594167 4776 scope.go:117] "RemoveContainer" containerID="94684aa85a999df01587ff8336fcf45d1847b413f000fe304eae44dda233eae6" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.627172 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_7dadc68f-0b89-41ba-84f8-e9bcde16ca64/ovn-northd/0.log" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.627254 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.679056 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09f0463c-f091-495e-9773-c07fe902c200" path="/var/lib/kubelet/pods/09f0463c-f091-495e-9773-c07fe902c200/volumes" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.680637 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d1e9a91-e344-442f-8bbc-90b1f98e2e1f" path="/var/lib/kubelet/pods/1d1e9a91-e344-442f-8bbc-90b1f98e2e1f/volumes" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.681513 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4702d881-d701-41a4-b36f-2d063a8ae246" path="/var/lib/kubelet/pods/4702d881-d701-41a4-b36f-2d063a8ae246/volumes" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.682827 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5920f152-04cd-4366-b479-665a1ce8937c" path="/var/lib/kubelet/pods/5920f152-04cd-4366-b479-665a1ce8937c/volumes" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.683526 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a0ab5a5-2d62-4457-be9d-fe8242ab7c91" path="/var/lib/kubelet/pods/6a0ab5a5-2d62-4457-be9d-fe8242ab7c91/volumes" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.684614 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bf3edad-487a-4c68-9dbf-9789a94f8fba" path="/var/lib/kubelet/pods/6bf3edad-487a-4c68-9dbf-9789a94f8fba/volumes" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.685388 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86e320e0-80ee-40df-b3a1-a48cb810a435" path="/var/lib/kubelet/pods/86e320e0-80ee-40df-b3a1-a48cb810a435/volumes" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.686213 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bae8251-3d5d-490b-9116-73769dc18101" path="/var/lib/kubelet/pods/8bae8251-3d5d-490b-9116-73769dc18101/volumes" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.687213 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0b97845-4578-4801-896a-281fdafdb351" path="/var/lib/kubelet/pods/a0b97845-4578-4801-896a-281fdafdb351/volumes" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.688108 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b708b100-166c-49b1-a47b-76d47ba2c6e5" path="/var/lib/kubelet/pods/b708b100-166c-49b1-a47b-76d47ba2c6e5/volumes" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.688980 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd5ce048-5a09-45a4-9a7d-66712b326ab7" path="/var/lib/kubelet/pods/cd5ce048-5a09-45a4-9a7d-66712b326ab7/volumes" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.690291 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2cda6de-ef12-42f4-a6e5-824c2db3cd01" path="/var/lib/kubelet/pods/d2cda6de-ef12-42f4-a6e5-824c2db3cd01/volumes" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.691101 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9" path="/var/lib/kubelet/pods/f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9/volumes" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.691765 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4" path="/var/lib/kubelet/pods/f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4/volumes" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.693039 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f32041a0-a8fc-48a5-afab-476baff89e8c-rabbitmq-plugins\") pod \"f32041a0-a8fc-48a5-afab-476baff89e8c\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.693126 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"f32041a0-a8fc-48a5-afab-476baff89e8c\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.693177 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f32041a0-a8fc-48a5-afab-476baff89e8c-pod-info\") pod \"f32041a0-a8fc-48a5-afab-476baff89e8c\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.693199 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f32041a0-a8fc-48a5-afab-476baff89e8c-plugins-conf\") pod \"f32041a0-a8fc-48a5-afab-476baff89e8c\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.693264 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfwhg\" (UniqueName: \"kubernetes.io/projected/f32041a0-a8fc-48a5-afab-476baff89e8c-kube-api-access-qfwhg\") pod \"f32041a0-a8fc-48a5-afab-476baff89e8c\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.693313 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f32041a0-a8fc-48a5-afab-476baff89e8c-rabbitmq-erlang-cookie\") pod \"f32041a0-a8fc-48a5-afab-476baff89e8c\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.693333 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f32041a0-a8fc-48a5-afab-476baff89e8c-rabbitmq-confd\") pod \"f32041a0-a8fc-48a5-afab-476baff89e8c\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.693362 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f32041a0-a8fc-48a5-afab-476baff89e8c-config-data\") pod \"f32041a0-a8fc-48a5-afab-476baff89e8c\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.693414 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f32041a0-a8fc-48a5-afab-476baff89e8c-erlang-cookie-secret\") pod \"f32041a0-a8fc-48a5-afab-476baff89e8c\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.693444 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f32041a0-a8fc-48a5-afab-476baff89e8c-server-conf\") pod \"f32041a0-a8fc-48a5-afab-476baff89e8c\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.693491 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f32041a0-a8fc-48a5-afab-476baff89e8c-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "f32041a0-a8fc-48a5-afab-476baff89e8c" (UID: "f32041a0-a8fc-48a5-afab-476baff89e8c"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.693525 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f32041a0-a8fc-48a5-afab-476baff89e8c-rabbitmq-tls\") pod \"f32041a0-a8fc-48a5-afab-476baff89e8c\" (UID: \"f32041a0-a8fc-48a5-afab-476baff89e8c\") " Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.693950 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f32041a0-a8fc-48a5-afab-476baff89e8c-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "f32041a0-a8fc-48a5-afab-476baff89e8c" (UID: "f32041a0-a8fc-48a5-afab-476baff89e8c"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.694354 4776 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f32041a0-a8fc-48a5-afab-476baff89e8c-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.694375 4776 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f32041a0-a8fc-48a5-afab-476baff89e8c-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.694493 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f32041a0-a8fc-48a5-afab-476baff89e8c-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "f32041a0-a8fc-48a5-afab-476baff89e8c" (UID: "f32041a0-a8fc-48a5-afab-476baff89e8c"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.702381 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f32041a0-a8fc-48a5-afab-476baff89e8c-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "f32041a0-a8fc-48a5-afab-476baff89e8c" (UID: "f32041a0-a8fc-48a5-afab-476baff89e8c"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.704544 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f32041a0-a8fc-48a5-afab-476baff89e8c-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "f32041a0-a8fc-48a5-afab-476baff89e8c" (UID: "f32041a0-a8fc-48a5-afab-476baff89e8c"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.711130 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "persistence") pod "f32041a0-a8fc-48a5-afab-476baff89e8c" (UID: "f32041a0-a8fc-48a5-afab-476baff89e8c"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.712907 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/f32041a0-a8fc-48a5-afab-476baff89e8c-pod-info" (OuterVolumeSpecName: "pod-info") pod "f32041a0-a8fc-48a5-afab-476baff89e8c" (UID: "f32041a0-a8fc-48a5-afab-476baff89e8c"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.736455 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f32041a0-a8fc-48a5-afab-476baff89e8c-kube-api-access-qfwhg" (OuterVolumeSpecName: "kube-api-access-qfwhg") pod "f32041a0-a8fc-48a5-afab-476baff89e8c" (UID: "f32041a0-a8fc-48a5-afab-476baff89e8c"). InnerVolumeSpecName "kube-api-access-qfwhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.754542 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f32041a0-a8fc-48a5-afab-476baff89e8c-config-data" (OuterVolumeSpecName: "config-data") pod "f32041a0-a8fc-48a5-afab-476baff89e8c" (UID: "f32041a0-a8fc-48a5-afab-476baff89e8c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.756223 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="e09f9958-4faf-4a28-8214-c3ead146122c" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"https://10.217.0.197:6080/vnc_lite.html\": dial tcp 10.217.0.197:6080: i/o timeout" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.795244 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-metrics-certs-tls-certs\") pod \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\" (UID: \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\") " Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.795315 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-ovn-rundir\") pod \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\" (UID: \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\") " Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.795362 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-ovn-northd-tls-certs\") pod \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\" (UID: \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\") " Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.795401 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sn8sd\" (UniqueName: \"kubernetes.io/projected/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-kube-api-access-sn8sd\") pod \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\" (UID: \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\") " Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.795435 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-config\") pod \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\" (UID: \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\") " Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.796105 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-combined-ca-bundle\") pod \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\" (UID: \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\") " Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.796158 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-scripts\") pod \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\" (UID: \"7dadc68f-0b89-41ba-84f8-e9bcde16ca64\") " Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.796399 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "7dadc68f-0b89-41ba-84f8-e9bcde16ca64" (UID: "7dadc68f-0b89-41ba-84f8-e9bcde16ca64"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.796769 4776 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-ovn-rundir\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.797114 4776 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.797198 4776 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f32041a0-a8fc-48a5-afab-476baff89e8c-pod-info\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.797327 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfwhg\" (UniqueName: \"kubernetes.io/projected/f32041a0-a8fc-48a5-afab-476baff89e8c-kube-api-access-qfwhg\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.797423 4776 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f32041a0-a8fc-48a5-afab-476baff89e8c-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.797502 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f32041a0-a8fc-48a5-afab-476baff89e8c-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.797588 4776 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f32041a0-a8fc-48a5-afab-476baff89e8c-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.797735 4776 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f32041a0-a8fc-48a5-afab-476baff89e8c-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.797807 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-scripts" (OuterVolumeSpecName: "scripts") pod "7dadc68f-0b89-41ba-84f8-e9bcde16ca64" (UID: "7dadc68f-0b89-41ba-84f8-e9bcde16ca64"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.799302 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-config" (OuterVolumeSpecName: "config") pod "7dadc68f-0b89-41ba-84f8-e9bcde16ca64" (UID: "7dadc68f-0b89-41ba-84f8-e9bcde16ca64"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.804731 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-kube-api-access-sn8sd" (OuterVolumeSpecName: "kube-api-access-sn8sd") pod "7dadc68f-0b89-41ba-84f8-e9bcde16ca64" (UID: "7dadc68f-0b89-41ba-84f8-e9bcde16ca64"). InnerVolumeSpecName "kube-api-access-sn8sd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.811088 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f32041a0-a8fc-48a5-afab-476baff89e8c-server-conf" (OuterVolumeSpecName: "server-conf") pod "f32041a0-a8fc-48a5-afab-476baff89e8c" (UID: "f32041a0-a8fc-48a5-afab-476baff89e8c"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.819179 4776 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.827006 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7dadc68f-0b89-41ba-84f8-e9bcde16ca64" (UID: "7dadc68f-0b89-41ba-84f8-e9bcde16ca64"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.865996 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.866234 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.868355 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "7dadc68f-0b89-41ba-84f8-e9bcde16ca64" (UID: "7dadc68f-0b89-41ba-84f8-e9bcde16ca64"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.870048 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "7dadc68f-0b89-41ba-84f8-e9bcde16ca64" (UID: "7dadc68f-0b89-41ba-84f8-e9bcde16ca64"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.879189 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f32041a0-a8fc-48a5-afab-476baff89e8c-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "f32041a0-a8fc-48a5-afab-476baff89e8c" (UID: "f32041a0-a8fc-48a5-afab-476baff89e8c"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.899237 4776 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.899268 4776 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.899278 4776 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.899286 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sn8sd\" (UniqueName: \"kubernetes.io/projected/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-kube-api-access-sn8sd\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.899294 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.899302 4776 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f32041a0-a8fc-48a5-afab-476baff89e8c-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.899310 4776 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f32041a0-a8fc-48a5-afab-476baff89e8c-server-conf\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.899320 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:03 crc kubenswrapper[4776]: I1125 09:48:03.899327 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7dadc68f-0b89-41ba-84f8-e9bcde16ca64-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.000510 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4512cf4e-6c75-493e-a80a-4b0accd750a8-rabbitmq-plugins\") pod \"4512cf4e-6c75-493e-a80a-4b0accd750a8\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.000585 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4512cf4e-6c75-493e-a80a-4b0accd750a8-rabbitmq-erlang-cookie\") pod \"4512cf4e-6c75-493e-a80a-4b0accd750a8\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.000670 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/167c5bc5-b21e-4c51-8765-1dcbf290294b-kolla-config\") pod \"167c5bc5-b21e-4c51-8765-1dcbf290294b\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.000691 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/167c5bc5-b21e-4c51-8765-1dcbf290294b-config-data-generated\") pod \"167c5bc5-b21e-4c51-8765-1dcbf290294b\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.000711 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4512cf4e-6c75-493e-a80a-4b0accd750a8-rabbitmq-tls\") pod \"4512cf4e-6c75-493e-a80a-4b0accd750a8\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.000731 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4512cf4e-6c75-493e-a80a-4b0accd750a8-plugins-conf\") pod \"4512cf4e-6c75-493e-a80a-4b0accd750a8\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.000758 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"167c5bc5-b21e-4c51-8765-1dcbf290294b\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.000777 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"4512cf4e-6c75-493e-a80a-4b0accd750a8\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.000800 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4512cf4e-6c75-493e-a80a-4b0accd750a8-rabbitmq-confd\") pod \"4512cf4e-6c75-493e-a80a-4b0accd750a8\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.000821 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4512cf4e-6c75-493e-a80a-4b0accd750a8-config-data\") pod \"4512cf4e-6c75-493e-a80a-4b0accd750a8\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.000848 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/167c5bc5-b21e-4c51-8765-1dcbf290294b-galera-tls-certs\") pod \"167c5bc5-b21e-4c51-8765-1dcbf290294b\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.000871 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4512cf4e-6c75-493e-a80a-4b0accd750a8-erlang-cookie-secret\") pod \"4512cf4e-6c75-493e-a80a-4b0accd750a8\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.000898 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4512cf4e-6c75-493e-a80a-4b0accd750a8-pod-info\") pod \"4512cf4e-6c75-493e-a80a-4b0accd750a8\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.000936 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5g6h4\" (UniqueName: \"kubernetes.io/projected/4512cf4e-6c75-493e-a80a-4b0accd750a8-kube-api-access-5g6h4\") pod \"4512cf4e-6c75-493e-a80a-4b0accd750a8\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.000979 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/167c5bc5-b21e-4c51-8765-1dcbf290294b-combined-ca-bundle\") pod \"167c5bc5-b21e-4c51-8765-1dcbf290294b\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.001011 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/167c5bc5-b21e-4c51-8765-1dcbf290294b-operator-scripts\") pod \"167c5bc5-b21e-4c51-8765-1dcbf290294b\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.001032 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqc28\" (UniqueName: \"kubernetes.io/projected/167c5bc5-b21e-4c51-8765-1dcbf290294b-kube-api-access-zqc28\") pod \"167c5bc5-b21e-4c51-8765-1dcbf290294b\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.001061 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4512cf4e-6c75-493e-a80a-4b0accd750a8-server-conf\") pod \"4512cf4e-6c75-493e-a80a-4b0accd750a8\" (UID: \"4512cf4e-6c75-493e-a80a-4b0accd750a8\") " Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.001096 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/167c5bc5-b21e-4c51-8765-1dcbf290294b-config-data-default\") pod \"167c5bc5-b21e-4c51-8765-1dcbf290294b\" (UID: \"167c5bc5-b21e-4c51-8765-1dcbf290294b\") " Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.000934 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4512cf4e-6c75-493e-a80a-4b0accd750a8-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "4512cf4e-6c75-493e-a80a-4b0accd750a8" (UID: "4512cf4e-6c75-493e-a80a-4b0accd750a8"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.002025 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/167c5bc5-b21e-4c51-8765-1dcbf290294b-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "167c5bc5-b21e-4c51-8765-1dcbf290294b" (UID: "167c5bc5-b21e-4c51-8765-1dcbf290294b"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.002155 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4512cf4e-6c75-493e-a80a-4b0accd750a8-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "4512cf4e-6c75-493e-a80a-4b0accd750a8" (UID: "4512cf4e-6c75-493e-a80a-4b0accd750a8"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.002176 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/167c5bc5-b21e-4c51-8765-1dcbf290294b-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "167c5bc5-b21e-4c51-8765-1dcbf290294b" (UID: "167c5bc5-b21e-4c51-8765-1dcbf290294b"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.002469 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/167c5bc5-b21e-4c51-8765-1dcbf290294b-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "167c5bc5-b21e-4c51-8765-1dcbf290294b" (UID: "167c5bc5-b21e-4c51-8765-1dcbf290294b"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.002707 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/167c5bc5-b21e-4c51-8765-1dcbf290294b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "167c5bc5-b21e-4c51-8765-1dcbf290294b" (UID: "167c5bc5-b21e-4c51-8765-1dcbf290294b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.003951 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4512cf4e-6c75-493e-a80a-4b0accd750a8-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "4512cf4e-6c75-493e-a80a-4b0accd750a8" (UID: "4512cf4e-6c75-493e-a80a-4b0accd750a8"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.006011 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4512cf4e-6c75-493e-a80a-4b0accd750a8-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "4512cf4e-6c75-493e-a80a-4b0accd750a8" (UID: "4512cf4e-6c75-493e-a80a-4b0accd750a8"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.006119 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/167c5bc5-b21e-4c51-8765-1dcbf290294b-kube-api-access-zqc28" (OuterVolumeSpecName: "kube-api-access-zqc28") pod "167c5bc5-b21e-4c51-8765-1dcbf290294b" (UID: "167c5bc5-b21e-4c51-8765-1dcbf290294b"). InnerVolumeSpecName "kube-api-access-zqc28". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.006144 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "4512cf4e-6c75-493e-a80a-4b0accd750a8" (UID: "4512cf4e-6c75-493e-a80a-4b0accd750a8"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.007154 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4512cf4e-6c75-493e-a80a-4b0accd750a8-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "4512cf4e-6c75-493e-a80a-4b0accd750a8" (UID: "4512cf4e-6c75-493e-a80a-4b0accd750a8"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.007281 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/4512cf4e-6c75-493e-a80a-4b0accd750a8-pod-info" (OuterVolumeSpecName: "pod-info") pod "4512cf4e-6c75-493e-a80a-4b0accd750a8" (UID: "4512cf4e-6c75-493e-a80a-4b0accd750a8"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.015770 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4512cf4e-6c75-493e-a80a-4b0accd750a8-kube-api-access-5g6h4" (OuterVolumeSpecName: "kube-api-access-5g6h4") pod "4512cf4e-6c75-493e-a80a-4b0accd750a8" (UID: "4512cf4e-6c75-493e-a80a-4b0accd750a8"). InnerVolumeSpecName "kube-api-access-5g6h4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:04 crc kubenswrapper[4776]: E1125 09:48:04.015777 4776 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Nov 25 09:48:04 crc kubenswrapper[4776]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-11-25T09:47:56Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Nov 25 09:48:04 crc kubenswrapper[4776]: /etc/init.d/functions: line 589: 431 Alarm clock "$@" Nov 25 09:48:04 crc kubenswrapper[4776]: > execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-2bs2f" message=< Nov 25 09:48:04 crc kubenswrapper[4776]: Exiting ovn-controller (1) [FAILED] Nov 25 09:48:04 crc kubenswrapper[4776]: Killing ovn-controller (1) [ OK ] Nov 25 09:48:04 crc kubenswrapper[4776]: Killing ovn-controller (1) with SIGKILL [ OK ] Nov 25 09:48:04 crc kubenswrapper[4776]: 2025-11-25T09:47:56Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Nov 25 09:48:04 crc kubenswrapper[4776]: /etc/init.d/functions: line 589: 431 Alarm clock "$@" Nov 25 09:48:04 crc kubenswrapper[4776]: > Nov 25 09:48:04 crc kubenswrapper[4776]: E1125 09:48:04.016331 4776 kuberuntime_container.go:691] "PreStop hook failed" err=< Nov 25 09:48:04 crc kubenswrapper[4776]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-11-25T09:47:56Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Nov 25 09:48:04 crc kubenswrapper[4776]: /etc/init.d/functions: line 589: 431 Alarm clock "$@" Nov 25 09:48:04 crc kubenswrapper[4776]: > pod="openstack/ovn-controller-2bs2f" podUID="c59c3d1e-c5c3-4036-ae28-436585e303ab" containerName="ovn-controller" containerID="cri-o://c1a79c25e3fcf070aa4efe449ff4827dd403b86266f770d2f2b4441e527d7adf" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.016531 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-2bs2f" podUID="c59c3d1e-c5c3-4036-ae28-436585e303ab" containerName="ovn-controller" containerID="cri-o://c1a79c25e3fcf070aa4efe449ff4827dd403b86266f770d2f2b4441e527d7adf" gracePeriod=22 Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.026899 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "mysql-db") pod "167c5bc5-b21e-4c51-8765-1dcbf290294b" (UID: "167c5bc5-b21e-4c51-8765-1dcbf290294b"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.040226 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/167c5bc5-b21e-4c51-8765-1dcbf290294b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "167c5bc5-b21e-4c51-8765-1dcbf290294b" (UID: "167c5bc5-b21e-4c51-8765-1dcbf290294b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.051220 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4512cf4e-6c75-493e-a80a-4b0accd750a8-config-data" (OuterVolumeSpecName: "config-data") pod "4512cf4e-6c75-493e-a80a-4b0accd750a8" (UID: "4512cf4e-6c75-493e-a80a-4b0accd750a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.062490 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/167c5bc5-b21e-4c51-8765-1dcbf290294b-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "167c5bc5-b21e-4c51-8765-1dcbf290294b" (UID: "167c5bc5-b21e-4c51-8765-1dcbf290294b"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.069121 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4512cf4e-6c75-493e-a80a-4b0accd750a8-server-conf" (OuterVolumeSpecName: "server-conf") pod "4512cf4e-6c75-493e-a80a-4b0accd750a8" (UID: "4512cf4e-6c75-493e-a80a-4b0accd750a8"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.080048 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-2bs2f_c59c3d1e-c5c3-4036-ae28-436585e303ab/ovn-controller/0.log" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.080129 4776 generic.go:334] "Generic (PLEG): container finished" podID="c59c3d1e-c5c3-4036-ae28-436585e303ab" containerID="c1a79c25e3fcf070aa4efe449ff4827dd403b86266f770d2f2b4441e527d7adf" exitCode=137 Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.080180 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2bs2f" event={"ID":"c59c3d1e-c5c3-4036-ae28-436585e303ab","Type":"ContainerDied","Data":"c1a79c25e3fcf070aa4efe449ff4827dd403b86266f770d2f2b4441e527d7adf"} Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.084648 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_7dadc68f-0b89-41ba-84f8-e9bcde16ca64/ovn-northd/0.log" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.084704 4776 generic.go:334] "Generic (PLEG): container finished" podID="7dadc68f-0b89-41ba-84f8-e9bcde16ca64" containerID="1ed6008959fa3368ff11289c84925032c1050a0bd2f104218d9763b1be776a55" exitCode=139 Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.084763 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.084826 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7dadc68f-0b89-41ba-84f8-e9bcde16ca64","Type":"ContainerDied","Data":"1ed6008959fa3368ff11289c84925032c1050a0bd2f104218d9763b1be776a55"} Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.084888 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7dadc68f-0b89-41ba-84f8-e9bcde16ca64","Type":"ContainerDied","Data":"7dc4a9f6af83ac0f7f0b95466453bfe79cb1509b6af1217c43f42e812f15390f"} Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.084912 4776 scope.go:117] "RemoveContainer" containerID="048ec6577ec600c7421727ef4cb3662a66e77cfa268385a725211a51ef3249cd" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.088064 4776 generic.go:334] "Generic (PLEG): container finished" podID="167c5bc5-b21e-4c51-8765-1dcbf290294b" containerID="2f383d0debcf43b01564a0db854854b92a294770e346344bcd8dce403c1e0251" exitCode=0 Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.088130 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"167c5bc5-b21e-4c51-8765-1dcbf290294b","Type":"ContainerDied","Data":"2f383d0debcf43b01564a0db854854b92a294770e346344bcd8dce403c1e0251"} Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.088153 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"167c5bc5-b21e-4c51-8765-1dcbf290294b","Type":"ContainerDied","Data":"522e57d9f9f345bed75f21b21c222210c137832a7c267a45a4a71baba0244de8"} Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.088218 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.098741 4776 generic.go:334] "Generic (PLEG): container finished" podID="4512cf4e-6c75-493e-a80a-4b0accd750a8" containerID="a5fdd3d6b6a35387501cbc30834b6e26eb181b138a16ef46ae65f71952655401" exitCode=0 Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.098775 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.098798 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4512cf4e-6c75-493e-a80a-4b0accd750a8","Type":"ContainerDied","Data":"a5fdd3d6b6a35387501cbc30834b6e26eb181b138a16ef46ae65f71952655401"} Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.098890 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4512cf4e-6c75-493e-a80a-4b0accd750a8","Type":"ContainerDied","Data":"43d7aec6fc9f74385bc4e597e383ea2a628e7dda0fd162984cfba0351a6e82b6"} Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.102920 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4512cf4e-6c75-493e-a80a-4b0accd750a8-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.102945 4776 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/167c5bc5-b21e-4c51-8765-1dcbf290294b-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.102954 4776 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4512cf4e-6c75-493e-a80a-4b0accd750a8-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.102963 4776 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4512cf4e-6c75-493e-a80a-4b0accd750a8-pod-info\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.102972 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5g6h4\" (UniqueName: \"kubernetes.io/projected/4512cf4e-6c75-493e-a80a-4b0accd750a8-kube-api-access-5g6h4\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.102980 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/167c5bc5-b21e-4c51-8765-1dcbf290294b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.102989 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/167c5bc5-b21e-4c51-8765-1dcbf290294b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.102997 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqc28\" (UniqueName: \"kubernetes.io/projected/167c5bc5-b21e-4c51-8765-1dcbf290294b-kube-api-access-zqc28\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.103005 4776 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4512cf4e-6c75-493e-a80a-4b0accd750a8-server-conf\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.103039 4776 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/167c5bc5-b21e-4c51-8765-1dcbf290294b-config-data-default\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.103051 4776 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4512cf4e-6c75-493e-a80a-4b0accd750a8-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.103061 4776 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4512cf4e-6c75-493e-a80a-4b0accd750a8-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.103112 4776 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/167c5bc5-b21e-4c51-8765-1dcbf290294b-kolla-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.103120 4776 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/167c5bc5-b21e-4c51-8765-1dcbf290294b-config-data-generated\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.103127 4776 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4512cf4e-6c75-493e-a80a-4b0accd750a8-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.103137 4776 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4512cf4e-6c75-493e-a80a-4b0accd750a8-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.103156 4776 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.103170 4776 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.123145 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5wmvt" podUID="9266e4bf-3e2e-4ed5-b799-02e56af27091" containerName="registry-server" containerID="cri-o://2a2229b7cd7d8596847c255b5a632079e86a284eb29c3cd724e9e64edc2fcf8b" gracePeriod=2 Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.125560 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.125632 4776 scope.go:117] "RemoveContainer" containerID="1ed6008959fa3368ff11289c84925032c1050a0bd2f104218d9763b1be776a55" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.125651 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f32041a0-a8fc-48a5-afab-476baff89e8c","Type":"ContainerDied","Data":"bd22466d1d55a4b40c3761c1eed08d859f1f3d4ccfc546edb124b702e0b8ba48"} Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.129879 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4512cf4e-6c75-493e-a80a-4b0accd750a8-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "4512cf4e-6c75-493e-a80a-4b0accd750a8" (UID: "4512cf4e-6c75-493e-a80a-4b0accd750a8"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.136907 4776 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.149917 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.162257 4776 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.164465 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.167421 4776 scope.go:117] "RemoveContainer" containerID="048ec6577ec600c7421727ef4cb3662a66e77cfa268385a725211a51ef3249cd" Nov 25 09:48:04 crc kubenswrapper[4776]: E1125 09:48:04.167932 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"048ec6577ec600c7421727ef4cb3662a66e77cfa268385a725211a51ef3249cd\": container with ID starting with 048ec6577ec600c7421727ef4cb3662a66e77cfa268385a725211a51ef3249cd not found: ID does not exist" containerID="048ec6577ec600c7421727ef4cb3662a66e77cfa268385a725211a51ef3249cd" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.167983 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"048ec6577ec600c7421727ef4cb3662a66e77cfa268385a725211a51ef3249cd"} err="failed to get container status \"048ec6577ec600c7421727ef4cb3662a66e77cfa268385a725211a51ef3249cd\": rpc error: code = NotFound desc = could not find container \"048ec6577ec600c7421727ef4cb3662a66e77cfa268385a725211a51ef3249cd\": container with ID starting with 048ec6577ec600c7421727ef4cb3662a66e77cfa268385a725211a51ef3249cd not found: ID does not exist" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.168012 4776 scope.go:117] "RemoveContainer" containerID="1ed6008959fa3368ff11289c84925032c1050a0bd2f104218d9763b1be776a55" Nov 25 09:48:04 crc kubenswrapper[4776]: E1125 09:48:04.168330 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ed6008959fa3368ff11289c84925032c1050a0bd2f104218d9763b1be776a55\": container with ID starting with 1ed6008959fa3368ff11289c84925032c1050a0bd2f104218d9763b1be776a55 not found: ID does not exist" containerID="1ed6008959fa3368ff11289c84925032c1050a0bd2f104218d9763b1be776a55" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.168368 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ed6008959fa3368ff11289c84925032c1050a0bd2f104218d9763b1be776a55"} err="failed to get container status \"1ed6008959fa3368ff11289c84925032c1050a0bd2f104218d9763b1be776a55\": rpc error: code = NotFound desc = could not find container \"1ed6008959fa3368ff11289c84925032c1050a0bd2f104218d9763b1be776a55\": container with ID starting with 1ed6008959fa3368ff11289c84925032c1050a0bd2f104218d9763b1be776a55 not found: ID does not exist" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.168392 4776 scope.go:117] "RemoveContainer" containerID="2f383d0debcf43b01564a0db854854b92a294770e346344bcd8dce403c1e0251" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.201720 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.206724 4776 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.206761 4776 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.206775 4776 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4512cf4e-6c75-493e-a80a-4b0accd750a8-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.214759 4776 scope.go:117] "RemoveContainer" containerID="f655f33165dfd7482166f931fb775ba2b424e5b9c57973932d34b6255eda7162" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.220776 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.227738 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.234558 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.277048 4776 scope.go:117] "RemoveContainer" containerID="2f383d0debcf43b01564a0db854854b92a294770e346344bcd8dce403c1e0251" Nov 25 09:48:04 crc kubenswrapper[4776]: E1125 09:48:04.277937 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f383d0debcf43b01564a0db854854b92a294770e346344bcd8dce403c1e0251\": container with ID starting with 2f383d0debcf43b01564a0db854854b92a294770e346344bcd8dce403c1e0251 not found: ID does not exist" containerID="2f383d0debcf43b01564a0db854854b92a294770e346344bcd8dce403c1e0251" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.277986 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f383d0debcf43b01564a0db854854b92a294770e346344bcd8dce403c1e0251"} err="failed to get container status \"2f383d0debcf43b01564a0db854854b92a294770e346344bcd8dce403c1e0251\": rpc error: code = NotFound desc = could not find container \"2f383d0debcf43b01564a0db854854b92a294770e346344bcd8dce403c1e0251\": container with ID starting with 2f383d0debcf43b01564a0db854854b92a294770e346344bcd8dce403c1e0251 not found: ID does not exist" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.278025 4776 scope.go:117] "RemoveContainer" containerID="f655f33165dfd7482166f931fb775ba2b424e5b9c57973932d34b6255eda7162" Nov 25 09:48:04 crc kubenswrapper[4776]: E1125 09:48:04.278638 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f655f33165dfd7482166f931fb775ba2b424e5b9c57973932d34b6255eda7162\": container with ID starting with f655f33165dfd7482166f931fb775ba2b424e5b9c57973932d34b6255eda7162 not found: ID does not exist" containerID="f655f33165dfd7482166f931fb775ba2b424e5b9c57973932d34b6255eda7162" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.278695 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f655f33165dfd7482166f931fb775ba2b424e5b9c57973932d34b6255eda7162"} err="failed to get container status \"f655f33165dfd7482166f931fb775ba2b424e5b9c57973932d34b6255eda7162\": rpc error: code = NotFound desc = could not find container \"f655f33165dfd7482166f931fb775ba2b424e5b9c57973932d34b6255eda7162\": container with ID starting with f655f33165dfd7482166f931fb775ba2b424e5b9c57973932d34b6255eda7162 not found: ID does not exist" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.278731 4776 scope.go:117] "RemoveContainer" containerID="a5fdd3d6b6a35387501cbc30834b6e26eb181b138a16ef46ae65f71952655401" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.410300 4776 scope.go:117] "RemoveContainer" containerID="e5d2c432a51c339c288f48c9a58a0ccdecc76c714979d2e84604d3bc00357881" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.470486 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.471342 4776 scope.go:117] "RemoveContainer" containerID="a5fdd3d6b6a35387501cbc30834b6e26eb181b138a16ef46ae65f71952655401" Nov 25 09:48:04 crc kubenswrapper[4776]: E1125 09:48:04.471759 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5fdd3d6b6a35387501cbc30834b6e26eb181b138a16ef46ae65f71952655401\": container with ID starting with a5fdd3d6b6a35387501cbc30834b6e26eb181b138a16ef46ae65f71952655401 not found: ID does not exist" containerID="a5fdd3d6b6a35387501cbc30834b6e26eb181b138a16ef46ae65f71952655401" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.471789 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5fdd3d6b6a35387501cbc30834b6e26eb181b138a16ef46ae65f71952655401"} err="failed to get container status \"a5fdd3d6b6a35387501cbc30834b6e26eb181b138a16ef46ae65f71952655401\": rpc error: code = NotFound desc = could not find container \"a5fdd3d6b6a35387501cbc30834b6e26eb181b138a16ef46ae65f71952655401\": container with ID starting with a5fdd3d6b6a35387501cbc30834b6e26eb181b138a16ef46ae65f71952655401 not found: ID does not exist" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.471810 4776 scope.go:117] "RemoveContainer" containerID="e5d2c432a51c339c288f48c9a58a0ccdecc76c714979d2e84604d3bc00357881" Nov 25 09:48:04 crc kubenswrapper[4776]: E1125 09:48:04.472505 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5d2c432a51c339c288f48c9a58a0ccdecc76c714979d2e84604d3bc00357881\": container with ID starting with e5d2c432a51c339c288f48c9a58a0ccdecc76c714979d2e84604d3bc00357881 not found: ID does not exist" containerID="e5d2c432a51c339c288f48c9a58a0ccdecc76c714979d2e84604d3bc00357881" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.472558 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5d2c432a51c339c288f48c9a58a0ccdecc76c714979d2e84604d3bc00357881"} err="failed to get container status \"e5d2c432a51c339c288f48c9a58a0ccdecc76c714979d2e84604d3bc00357881\": rpc error: code = NotFound desc = could not find container \"e5d2c432a51c339c288f48c9a58a0ccdecc76c714979d2e84604d3bc00357881\": container with ID starting with e5d2c432a51c339c288f48c9a58a0ccdecc76c714979d2e84604d3bc00357881 not found: ID does not exist" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.472578 4776 scope.go:117] "RemoveContainer" containerID="7b42621ff074baca099513ddb94523b955d55200c16323a7bb88a078926fca25" Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.493807 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.528105 4776 scope.go:117] "RemoveContainer" containerID="b75205641b6269bc97dcaf11905b1173d0238b983968101c11decb00e38566b3" Nov 25 09:48:04 crc kubenswrapper[4776]: E1125 09:48:04.551839 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:04 crc kubenswrapper[4776]: E1125 09:48:04.551923 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/946ea713-bfba-449f-82ab-28e915469938-operator-scripts podName:946ea713-bfba-449f-82ab-28e915469938 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:08.551902708 +0000 UTC m=+1433.592962261 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/946ea713-bfba-449f-82ab-28e915469938-operator-scripts") pod "novacell09dc4-account-delete-w5xcj" (UID: "946ea713-bfba-449f-82ab-28e915469938") : configmap "openstack-scripts" not found Nov 25 09:48:04 crc kubenswrapper[4776]: E1125 09:48:04.552355 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:04 crc kubenswrapper[4776]: E1125 09:48:04.552399 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ef85f90f-c377-4273-9fff-451420d6008a-operator-scripts podName:ef85f90f-c377-4273-9fff-451420d6008a nodeName:}" failed. No retries permitted until 2025-11-25 09:48:08.55238916 +0000 UTC m=+1433.593448713 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/ef85f90f-c377-4273-9fff-451420d6008a-operator-scripts") pod "cinder774d-account-delete-b29n5" (UID: "ef85f90f-c377-4273-9fff-451420d6008a") : configmap "openstack-scripts" not found Nov 25 09:48:04 crc kubenswrapper[4776]: E1125 09:48:04.552434 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:04 crc kubenswrapper[4776]: E1125 09:48:04.552460 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/14b51bba-0b15-469b-bdea-90fae8d9b976-operator-scripts podName:14b51bba-0b15-469b-bdea-90fae8d9b976 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:08.552451342 +0000 UTC m=+1433.593510895 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/14b51bba-0b15-469b-bdea-90fae8d9b976-operator-scripts") pod "novaapi3b46-account-delete-9w5r9" (UID: "14b51bba-0b15-469b-bdea-90fae8d9b976") : configmap "openstack-scripts" not found Nov 25 09:48:04 crc kubenswrapper[4776]: E1125 09:48:04.552492 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:04 crc kubenswrapper[4776]: E1125 09:48:04.552517 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6be099e6-89d6-45a0-9d0f-4dca4d91701a-operator-scripts podName:6be099e6-89d6-45a0-9d0f-4dca4d91701a nodeName:}" failed. No retries permitted until 2025-11-25 09:48:08.552509113 +0000 UTC m=+1433.593568686 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/6be099e6-89d6-45a0-9d0f-4dca4d91701a-operator-scripts") pod "placemente176-account-delete-wbgz5" (UID: "6be099e6-89d6-45a0-9d0f-4dca4d91701a") : configmap "openstack-scripts" not found Nov 25 09:48:04 crc kubenswrapper[4776]: E1125 09:48:04.552547 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:04 crc kubenswrapper[4776]: E1125 09:48:04.552571 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/73f2c2ab-6cb6-4544-a3f7-20722e5b6400-operator-scripts podName:73f2c2ab-6cb6-4544-a3f7-20722e5b6400 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:08.552562964 +0000 UTC m=+1433.593622517 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/73f2c2ab-6cb6-4544-a3f7-20722e5b6400-operator-scripts") pod "neutron5151-account-delete-8vbjg" (UID: "73f2c2ab-6cb6-4544-a3f7-20722e5b6400") : configmap "openstack-scripts" not found Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.629865 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-e176-account-create-5g442"] Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.640483 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-e176-account-create-5g442"] Nov 25 09:48:04 crc kubenswrapper[4776]: E1125 09:48:04.680188 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:04 crc kubenswrapper[4776]: E1125 09:48:04.680254 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/bcaec0f3-13af-4f73-8e7f-7e575b8e2008-operator-scripts podName:bcaec0f3-13af-4f73-8e7f-7e575b8e2008 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:08.680241515 +0000 UTC m=+1433.721301058 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/bcaec0f3-13af-4f73-8e7f-7e575b8e2008-operator-scripts") pod "glanceb9b5-account-delete-tjrnr" (UID: "bcaec0f3-13af-4f73-8e7f-7e575b8e2008") : configmap "openstack-scripts" not found Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.681248 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placemente176-account-delete-wbgz5"] Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.681351 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placemente176-account-delete-wbgz5" podUID="6be099e6-89d6-45a0-9d0f-4dca4d91701a" containerName="mariadb-account-delete" containerID="cri-o://52647c9b509f4a2e0420df93981105d258326ab2d8003e1be47968ae44e82e11" gracePeriod=30 Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.697537 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-7ff97"] Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.715401 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-7ff97"] Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.779810 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-5g4ql"] Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.791088 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-5g4ql"] Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.814143 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-b9b5-account-create-ldfzn"] Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.823510 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glanceb9b5-account-delete-tjrnr"] Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.823747 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glanceb9b5-account-delete-tjrnr" podUID="bcaec0f3-13af-4f73-8e7f-7e575b8e2008" containerName="mariadb-account-delete" containerID="cri-o://4a42c4c2fcc9fc1a6adb53fc048a0c22c503282432d6521db1b327054c0ea919" gracePeriod=30 Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.842430 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-b9b5-account-create-ldfzn"] Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.924444 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-6l262"] Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.929604 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-6l262"] Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.972287 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbicandb0b-account-delete-xtvnn"] Nov 25 09:48:04 crc kubenswrapper[4776]: I1125 09:48:04.990506 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db0b-account-create-hssb4"] Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.000857 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db0b-account-create-hssb4"] Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.002315 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbicandb0b-account-delete-xtvnn"] Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.145726 4776 generic.go:334] "Generic (PLEG): container finished" podID="e51292cd-00f5-4e0a-ba2b-6319bd8a5e45" containerID="63cb3f8363fb0824de5241eef2c1f027163930b39c0a67a62ab2eadf418c85d7" exitCode=0 Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.145828 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-56988fbb4-nlxs9" event={"ID":"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45","Type":"ContainerDied","Data":"63cb3f8363fb0824de5241eef2c1f027163930b39c0a67a62ab2eadf418c85d7"} Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.154532 4776 generic.go:334] "Generic (PLEG): container finished" podID="aab7d356-ec67-4e5d-9c25-f26e39786f79" containerID="a2a4c7e738a8387346bc8cb061fcaffab37035f8b7356aba874420f641555783" exitCode=0 Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.154557 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aab7d356-ec67-4e5d-9c25-f26e39786f79","Type":"ContainerDied","Data":"a2a4c7e738a8387346bc8cb061fcaffab37035f8b7356aba874420f641555783"} Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.162635 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6b75d6ddb7-rmtms" event={"ID":"83c080b8-d976-4e81-b103-2442bb2eafe1","Type":"ContainerDied","Data":"d35530a8d2f45153bec6fcc975ede0bc99a55fb6d60c6edac19a05f0939799d7"} Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.165373 4776 generic.go:334] "Generic (PLEG): container finished" podID="83c080b8-d976-4e81-b103-2442bb2eafe1" containerID="d35530a8d2f45153bec6fcc975ede0bc99a55fb6d60c6edac19a05f0939799d7" exitCode=0 Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.165582 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6b75d6ddb7-rmtms" event={"ID":"83c080b8-d976-4e81-b103-2442bb2eafe1","Type":"ContainerDied","Data":"2894d4dea69bb1d9b29dcb1eef2460fb6473db26f24cc542599755b582fcb965"} Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.165614 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2894d4dea69bb1d9b29dcb1eef2460fb6473db26f24cc542599755b582fcb965" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.171187 4776 generic.go:334] "Generic (PLEG): container finished" podID="9266e4bf-3e2e-4ed5-b799-02e56af27091" containerID="2a2229b7cd7d8596847c255b5a632079e86a284eb29c3cd724e9e64edc2fcf8b" exitCode=0 Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.171316 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5wmvt" event={"ID":"9266e4bf-3e2e-4ed5-b799-02e56af27091","Type":"ContainerDied","Data":"2a2229b7cd7d8596847c255b5a632079e86a284eb29c3cd724e9e64edc2fcf8b"} Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.171349 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5wmvt" event={"ID":"9266e4bf-3e2e-4ed5-b799-02e56af27091","Type":"ContainerDied","Data":"ddff0921234b55aa271943735142330c5e32a5ff802fc3d2504f4b81f0a926c4"} Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.171363 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ddff0921234b55aa271943735142330c5e32a5ff802fc3d2504f4b81f0a926c4" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.177664 4776 generic.go:334] "Generic (PLEG): container finished" podID="f182a7e0-ebd8-4258-9269-43a662e39af8" containerID="e5a40e8fd3cdfddb16f8d8578ce8ae7e8bb1b0df1b91c68efe717a27f4fc382c" exitCode=0 Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.177758 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"f182a7e0-ebd8-4258-9269-43a662e39af8","Type":"ContainerDied","Data":"e5a40e8fd3cdfddb16f8d8578ce8ae7e8bb1b0df1b91c68efe717a27f4fc382c"} Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.180276 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-2bs2f_c59c3d1e-c5c3-4036-ae28-436585e303ab/ovn-controller/0.log" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.180351 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-2bs2f" event={"ID":"c59c3d1e-c5c3-4036-ae28-436585e303ab","Type":"ContainerDied","Data":"35bee38875dfc28653ef000ae6a9a992ad8fcb6fb75884a44dd02aeef180c4e2"} Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.180368 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35bee38875dfc28653ef000ae6a9a992ad8fcb6fb75884a44dd02aeef180c4e2" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.184966 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5wmvt" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.196844 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.221667 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-2bs2f_c59c3d1e-c5c3-4036-ae28-436585e303ab/ovn-controller/0.log" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.221832 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2bs2f" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.293852 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9266e4bf-3e2e-4ed5-b799-02e56af27091-utilities\") pod \"9266e4bf-3e2e-4ed5-b799-02e56af27091\" (UID: \"9266e4bf-3e2e-4ed5-b799-02e56af27091\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.293906 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9266e4bf-3e2e-4ed5-b799-02e56af27091-catalog-content\") pod \"9266e4bf-3e2e-4ed5-b799-02e56af27091\" (UID: \"9266e4bf-3e2e-4ed5-b799-02e56af27091\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.294027 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hnxs\" (UniqueName: \"kubernetes.io/projected/9266e4bf-3e2e-4ed5-b799-02e56af27091-kube-api-access-7hnxs\") pod \"9266e4bf-3e2e-4ed5-b799-02e56af27091\" (UID: \"9266e4bf-3e2e-4ed5-b799-02e56af27091\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.295890 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9266e4bf-3e2e-4ed5-b799-02e56af27091-utilities" (OuterVolumeSpecName: "utilities") pod "9266e4bf-3e2e-4ed5-b799-02e56af27091" (UID: "9266e4bf-3e2e-4ed5-b799-02e56af27091"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.302331 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9266e4bf-3e2e-4ed5-b799-02e56af27091-kube-api-access-7hnxs" (OuterVolumeSpecName: "kube-api-access-7hnxs") pod "9266e4bf-3e2e-4ed5-b799-02e56af27091" (UID: "9266e4bf-3e2e-4ed5-b799-02e56af27091"). InnerVolumeSpecName "kube-api-access-7hnxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.356652 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-56988fbb4-nlxs9" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.360951 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.396616 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4466v\" (UniqueName: \"kubernetes.io/projected/83c080b8-d976-4e81-b103-2442bb2eafe1-kube-api-access-4466v\") pod \"83c080b8-d976-4e81-b103-2442bb2eafe1\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.396671 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-credential-keys\") pod \"83c080b8-d976-4e81-b103-2442bb2eafe1\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.396707 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c59c3d1e-c5c3-4036-ae28-436585e303ab-scripts\") pod \"c59c3d1e-c5c3-4036-ae28-436585e303ab\" (UID: \"c59c3d1e-c5c3-4036-ae28-436585e303ab\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.396769 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-combined-ca-bundle\") pod \"83c080b8-d976-4e81-b103-2442bb2eafe1\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.396796 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-internal-tls-certs\") pod \"83c080b8-d976-4e81-b103-2442bb2eafe1\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.396825 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-fernet-keys\") pod \"83c080b8-d976-4e81-b103-2442bb2eafe1\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.396847 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-scripts\") pod \"83c080b8-d976-4e81-b103-2442bb2eafe1\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.396883 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-config-data\") pod \"83c080b8-d976-4e81-b103-2442bb2eafe1\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.396897 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c59c3d1e-c5c3-4036-ae28-436585e303ab-var-run-ovn\") pod \"c59c3d1e-c5c3-4036-ae28-436585e303ab\" (UID: \"c59c3d1e-c5c3-4036-ae28-436585e303ab\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.396928 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c59c3d1e-c5c3-4036-ae28-436585e303ab-var-run\") pod \"c59c3d1e-c5c3-4036-ae28-436585e303ab\" (UID: \"c59c3d1e-c5c3-4036-ae28-436585e303ab\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.396958 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/c59c3d1e-c5c3-4036-ae28-436585e303ab-ovn-controller-tls-certs\") pod \"c59c3d1e-c5c3-4036-ae28-436585e303ab\" (UID: \"c59c3d1e-c5c3-4036-ae28-436585e303ab\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.396984 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c59c3d1e-c5c3-4036-ae28-436585e303ab-combined-ca-bundle\") pod \"c59c3d1e-c5c3-4036-ae28-436585e303ab\" (UID: \"c59c3d1e-c5c3-4036-ae28-436585e303ab\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.397024 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-public-tls-certs\") pod \"83c080b8-d976-4e81-b103-2442bb2eafe1\" (UID: \"83c080b8-d976-4e81-b103-2442bb2eafe1\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.397037 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c59c3d1e-c5c3-4036-ae28-436585e303ab-var-log-ovn\") pod \"c59c3d1e-c5c3-4036-ae28-436585e303ab\" (UID: \"c59c3d1e-c5c3-4036-ae28-436585e303ab\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.397080 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjvld\" (UniqueName: \"kubernetes.io/projected/c59c3d1e-c5c3-4036-ae28-436585e303ab-kube-api-access-kjvld\") pod \"c59c3d1e-c5c3-4036-ae28-436585e303ab\" (UID: \"c59c3d1e-c5c3-4036-ae28-436585e303ab\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.397487 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9266e4bf-3e2e-4ed5-b799-02e56af27091-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.397513 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hnxs\" (UniqueName: \"kubernetes.io/projected/9266e4bf-3e2e-4ed5-b799-02e56af27091-kube-api-access-7hnxs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.401835 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-scripts" (OuterVolumeSpecName: "scripts") pod "83c080b8-d976-4e81-b103-2442bb2eafe1" (UID: "83c080b8-d976-4e81-b103-2442bb2eafe1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.402631 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c59c3d1e-c5c3-4036-ae28-436585e303ab-scripts" (OuterVolumeSpecName: "scripts") pod "c59c3d1e-c5c3-4036-ae28-436585e303ab" (UID: "c59c3d1e-c5c3-4036-ae28-436585e303ab"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.403613 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c59c3d1e-c5c3-4036-ae28-436585e303ab-var-run" (OuterVolumeSpecName: "var-run") pod "c59c3d1e-c5c3-4036-ae28-436585e303ab" (UID: "c59c3d1e-c5c3-4036-ae28-436585e303ab"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.403680 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c59c3d1e-c5c3-4036-ae28-436585e303ab-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "c59c3d1e-c5c3-4036-ae28-436585e303ab" (UID: "c59c3d1e-c5c3-4036-ae28-436585e303ab"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.405032 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c59c3d1e-c5c3-4036-ae28-436585e303ab-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "c59c3d1e-c5c3-4036-ae28-436585e303ab" (UID: "c59c3d1e-c5c3-4036-ae28-436585e303ab"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.409542 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83c080b8-d976-4e81-b103-2442bb2eafe1-kube-api-access-4466v" (OuterVolumeSpecName: "kube-api-access-4466v") pod "83c080b8-d976-4e81-b103-2442bb2eafe1" (UID: "83c080b8-d976-4e81-b103-2442bb2eafe1"). InnerVolumeSpecName "kube-api-access-4466v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.412201 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c59c3d1e-c5c3-4036-ae28-436585e303ab-kube-api-access-kjvld" (OuterVolumeSpecName: "kube-api-access-kjvld") pod "c59c3d1e-c5c3-4036-ae28-436585e303ab" (UID: "c59c3d1e-c5c3-4036-ae28-436585e303ab"). InnerVolumeSpecName "kube-api-access-kjvld". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.416153 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "83c080b8-d976-4e81-b103-2442bb2eafe1" (UID: "83c080b8-d976-4e81-b103-2442bb2eafe1"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.422493 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "83c080b8-d976-4e81-b103-2442bb2eafe1" (UID: "83c080b8-d976-4e81-b103-2442bb2eafe1"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.440706 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "83c080b8-d976-4e81-b103-2442bb2eafe1" (UID: "83c080b8-d976-4e81-b103-2442bb2eafe1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.441728 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9266e4bf-3e2e-4ed5-b799-02e56af27091-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9266e4bf-3e2e-4ed5-b799-02e56af27091" (UID: "9266e4bf-3e2e-4ed5-b799-02e56af27091"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.473617 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.481275 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-config-data" (OuterVolumeSpecName: "config-data") pod "83c080b8-d976-4e81-b103-2442bb2eafe1" (UID: "83c080b8-d976-4e81-b103-2442bb2eafe1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.483550 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "83c080b8-d976-4e81-b103-2442bb2eafe1" (UID: "83c080b8-d976-4e81-b103-2442bb2eafe1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.486990 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "83c080b8-d976-4e81-b103-2442bb2eafe1" (UID: "83c080b8-d976-4e81-b103-2442bb2eafe1"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.488637 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c59c3d1e-c5c3-4036-ae28-436585e303ab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c59c3d1e-c5c3-4036-ae28-436585e303ab" (UID: "c59c3d1e-c5c3-4036-ae28-436585e303ab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.490666 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c59c3d1e-c5c3-4036-ae28-436585e303ab-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "c59c3d1e-c5c3-4036-ae28-436585e303ab" (UID: "c59c3d1e-c5c3-4036-ae28-436585e303ab"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.498881 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-logs\") pod \"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45\" (UID: \"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.499684 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aab7d356-ec67-4e5d-9c25-f26e39786f79-log-httpd\") pod \"aab7d356-ec67-4e5d-9c25-f26e39786f79\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.499750 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-config-data\") pod \"aab7d356-ec67-4e5d-9c25-f26e39786f79\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.499793 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-scripts\") pod \"aab7d356-ec67-4e5d-9c25-f26e39786f79\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.500467 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zm22g\" (UniqueName: \"kubernetes.io/projected/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-kube-api-access-zm22g\") pod \"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45\" (UID: \"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.500495 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-sg-core-conf-yaml\") pod \"aab7d356-ec67-4e5d-9c25-f26e39786f79\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.500535 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-combined-ca-bundle\") pod \"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45\" (UID: \"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.500591 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-ceilometer-tls-certs\") pod \"aab7d356-ec67-4e5d-9c25-f26e39786f79\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.500624 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-config-data\") pod \"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45\" (UID: \"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.500255 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aab7d356-ec67-4e5d-9c25-f26e39786f79-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "aab7d356-ec67-4e5d-9c25-f26e39786f79" (UID: "aab7d356-ec67-4e5d-9c25-f26e39786f79"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.500398 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-logs" (OuterVolumeSpecName: "logs") pod "e51292cd-00f5-4e0a-ba2b-6319bd8a5e45" (UID: "e51292cd-00f5-4e0a-ba2b-6319bd8a5e45"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.501380 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-combined-ca-bundle\") pod \"aab7d356-ec67-4e5d-9c25-f26e39786f79\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.501422 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aab7d356-ec67-4e5d-9c25-f26e39786f79-run-httpd\") pod \"aab7d356-ec67-4e5d-9c25-f26e39786f79\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.501444 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxgd8\" (UniqueName: \"kubernetes.io/projected/aab7d356-ec67-4e5d-9c25-f26e39786f79-kube-api-access-dxgd8\") pod \"aab7d356-ec67-4e5d-9c25-f26e39786f79\" (UID: \"aab7d356-ec67-4e5d-9c25-f26e39786f79\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.501475 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-config-data-custom\") pod \"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45\" (UID: \"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.501496 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f182a7e0-ebd8-4258-9269-43a662e39af8-config-data\") pod \"f182a7e0-ebd8-4258-9269-43a662e39af8\" (UID: \"f182a7e0-ebd8-4258-9269-43a662e39af8\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.501810 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-logs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.501825 4776 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aab7d356-ec67-4e5d-9c25-f26e39786f79-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.501834 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4466v\" (UniqueName: \"kubernetes.io/projected/83c080b8-d976-4e81-b103-2442bb2eafe1-kube-api-access-4466v\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.501843 4776 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.501852 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c59c3d1e-c5c3-4036-ae28-436585e303ab-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.501860 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9266e4bf-3e2e-4ed5-b799-02e56af27091-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.501869 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.501877 4776 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.501886 4776 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.501894 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.501902 4776 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c59c3d1e-c5c3-4036-ae28-436585e303ab-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.501911 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.501920 4776 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c59c3d1e-c5c3-4036-ae28-436585e303ab-var-run\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.501928 4776 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/c59c3d1e-c5c3-4036-ae28-436585e303ab-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.501938 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c59c3d1e-c5c3-4036-ae28-436585e303ab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.501946 4776 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/83c080b8-d976-4e81-b103-2442bb2eafe1-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.501954 4776 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c59c3d1e-c5c3-4036-ae28-436585e303ab-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.501963 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjvld\" (UniqueName: \"kubernetes.io/projected/c59c3d1e-c5c3-4036-ae28-436585e303ab-kube-api-access-kjvld\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.502031 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aab7d356-ec67-4e5d-9c25-f26e39786f79-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "aab7d356-ec67-4e5d-9c25-f26e39786f79" (UID: "aab7d356-ec67-4e5d-9c25-f26e39786f79"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.505092 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aab7d356-ec67-4e5d-9c25-f26e39786f79-kube-api-access-dxgd8" (OuterVolumeSpecName: "kube-api-access-dxgd8") pod "aab7d356-ec67-4e5d-9c25-f26e39786f79" (UID: "aab7d356-ec67-4e5d-9c25-f26e39786f79"). InnerVolumeSpecName "kube-api-access-dxgd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.509158 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-scripts" (OuterVolumeSpecName: "scripts") pod "aab7d356-ec67-4e5d-9c25-f26e39786f79" (UID: "aab7d356-ec67-4e5d-9c25-f26e39786f79"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.512968 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e51292cd-00f5-4e0a-ba2b-6319bd8a5e45" (UID: "e51292cd-00f5-4e0a-ba2b-6319bd8a5e45"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.526354 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-kube-api-access-zm22g" (OuterVolumeSpecName: "kube-api-access-zm22g") pod "e51292cd-00f5-4e0a-ba2b-6319bd8a5e45" (UID: "e51292cd-00f5-4e0a-ba2b-6319bd8a5e45"). InnerVolumeSpecName "kube-api-access-zm22g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.527158 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "aab7d356-ec67-4e5d-9c25-f26e39786f79" (UID: "aab7d356-ec67-4e5d-9c25-f26e39786f79"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.544680 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f182a7e0-ebd8-4258-9269-43a662e39af8-config-data" (OuterVolumeSpecName: "config-data") pod "f182a7e0-ebd8-4258-9269-43a662e39af8" (UID: "f182a7e0-ebd8-4258-9269-43a662e39af8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.549264 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e51292cd-00f5-4e0a-ba2b-6319bd8a5e45" (UID: "e51292cd-00f5-4e0a-ba2b-6319bd8a5e45"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.551480 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "aab7d356-ec67-4e5d-9c25-f26e39786f79" (UID: "aab7d356-ec67-4e5d-9c25-f26e39786f79"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.556479 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-config-data" (OuterVolumeSpecName: "config-data") pod "e51292cd-00f5-4e0a-ba2b-6319bd8a5e45" (UID: "e51292cd-00f5-4e0a-ba2b-6319bd8a5e45"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.576985 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aab7d356-ec67-4e5d-9c25-f26e39786f79" (UID: "aab7d356-ec67-4e5d-9c25-f26e39786f79"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.585180 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-config-data" (OuterVolumeSpecName: "config-data") pod "aab7d356-ec67-4e5d-9c25-f26e39786f79" (UID: "aab7d356-ec67-4e5d-9c25-f26e39786f79"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.602857 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvghd\" (UniqueName: \"kubernetes.io/projected/f182a7e0-ebd8-4258-9269-43a662e39af8-kube-api-access-fvghd\") pod \"f182a7e0-ebd8-4258-9269-43a662e39af8\" (UID: \"f182a7e0-ebd8-4258-9269-43a662e39af8\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.602910 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f182a7e0-ebd8-4258-9269-43a662e39af8-combined-ca-bundle\") pod \"f182a7e0-ebd8-4258-9269-43a662e39af8\" (UID: \"f182a7e0-ebd8-4258-9269-43a662e39af8\") " Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.603325 4776 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.603360 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f182a7e0-ebd8-4258-9269-43a662e39af8-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.603374 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.603386 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.603398 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zm22g\" (UniqueName: \"kubernetes.io/projected/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-kube-api-access-zm22g\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.603412 4776 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.603425 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.603437 4776 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.603449 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.603460 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aab7d356-ec67-4e5d-9c25-f26e39786f79-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.603472 4776 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aab7d356-ec67-4e5d-9c25-f26e39786f79-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.603486 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxgd8\" (UniqueName: \"kubernetes.io/projected/aab7d356-ec67-4e5d-9c25-f26e39786f79-kube-api-access-dxgd8\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.606057 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f182a7e0-ebd8-4258-9269-43a662e39af8-kube-api-access-fvghd" (OuterVolumeSpecName: "kube-api-access-fvghd") pod "f182a7e0-ebd8-4258-9269-43a662e39af8" (UID: "f182a7e0-ebd8-4258-9269-43a662e39af8"). InnerVolumeSpecName "kube-api-access-fvghd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.626774 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f182a7e0-ebd8-4258-9269-43a662e39af8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f182a7e0-ebd8-4258-9269-43a662e39af8" (UID: "f182a7e0-ebd8-4258-9269-43a662e39af8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:05 crc kubenswrapper[4776]: E1125 09:48:05.627441 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06 is running failed: container process not found" containerID="eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 25 09:48:05 crc kubenswrapper[4776]: E1125 09:48:05.627875 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06 is running failed: container process not found" containerID="eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 25 09:48:05 crc kubenswrapper[4776]: E1125 09:48:05.628265 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06 is running failed: container process not found" containerID="eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 25 09:48:05 crc kubenswrapper[4776]: E1125 09:48:05.628313 4776 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-7fgjf" podUID="10fe3dc2-2bde-4380-8ad3-d71832e31095" containerName="ovsdb-server" Nov 25 09:48:05 crc kubenswrapper[4776]: E1125 09:48:05.628866 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5002006f906f0b17afbf25b23648688a842975cfd3b385d57e3191040607fb57" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 25 09:48:05 crc kubenswrapper[4776]: E1125 09:48:05.630348 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5002006f906f0b17afbf25b23648688a842975cfd3b385d57e3191040607fb57" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 25 09:48:05 crc kubenswrapper[4776]: E1125 09:48:05.632104 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5002006f906f0b17afbf25b23648688a842975cfd3b385d57e3191040607fb57" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 25 09:48:05 crc kubenswrapper[4776]: E1125 09:48:05.632181 4776 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-7fgjf" podUID="10fe3dc2-2bde-4380-8ad3-d71832e31095" containerName="ovs-vswitchd" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.687781 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3" path="/var/lib/kubelet/pods/00fc6870-0ad9-4bb0-bebf-36cb0dd8e5e3/volumes" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.688856 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="167c5bc5-b21e-4c51-8765-1dcbf290294b" path="/var/lib/kubelet/pods/167c5bc5-b21e-4c51-8765-1dcbf290294b/volumes" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.689793 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="411d6157-0520-436e-b350-3a573e7ad984" path="/var/lib/kubelet/pods/411d6157-0520-436e-b350-3a573e7ad984/volumes" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.691555 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="417adc1f-2f5c-4446-8cf3-994edf3b5e55" path="/var/lib/kubelet/pods/417adc1f-2f5c-4446-8cf3-994edf3b5e55/volumes" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.692447 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4512cf4e-6c75-493e-a80a-4b0accd750a8" path="/var/lib/kubelet/pods/4512cf4e-6c75-493e-a80a-4b0accd750a8/volumes" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.693240 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6466d754-e356-4950-9ce9-8f56e68d6b97" path="/var/lib/kubelet/pods/6466d754-e356-4950-9ce9-8f56e68d6b97/volumes" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.694415 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d40516c-5fcc-4b24-918f-95bf79cb94b4" path="/var/lib/kubelet/pods/6d40516c-5fcc-4b24-918f-95bf79cb94b4/volumes" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.695154 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dadc68f-0b89-41ba-84f8-e9bcde16ca64" path="/var/lib/kubelet/pods/7dadc68f-0b89-41ba-84f8-e9bcde16ca64/volumes" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.696044 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89" path="/var/lib/kubelet/pods/83061fcd-d3a1-4eb7-8aed-c0bd0be0cc89/volumes" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.697473 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f32041a0-a8fc-48a5-afab-476baff89e8c" path="/var/lib/kubelet/pods/f32041a0-a8fc-48a5-afab-476baff89e8c/volumes" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.701536 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f72b0b94-97f1-4bbc-9be4-c8965143cd26" path="/var/lib/kubelet/pods/f72b0b94-97f1-4bbc-9be4-c8965143cd26/volumes" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.715544 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvghd\" (UniqueName: \"kubernetes.io/projected/f182a7e0-ebd8-4258-9269-43a662e39af8-kube-api-access-fvghd\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:05 crc kubenswrapper[4776]: I1125 09:48:05.715574 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f182a7e0-ebd8-4258-9269-43a662e39af8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.153080 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wvl56"] Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.153624 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dadc68f-0b89-41ba-84f8-e9bcde16ca64" containerName="ovn-northd" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.153663 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dadc68f-0b89-41ba-84f8-e9bcde16ca64" containerName="ovn-northd" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.153682 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bf3edad-487a-4c68-9dbf-9789a94f8fba" containerName="nova-api-api" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.153690 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bf3edad-487a-4c68-9dbf-9789a94f8fba" containerName="nova-api-api" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.153708 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc920d71-4308-4d90-a891-8b760651ccc9" containerName="probe" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.153716 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc920d71-4308-4d90-a891-8b760651ccc9" containerName="probe" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.153725 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d1e9a91-e344-442f-8bbc-90b1f98e2e1f" containerName="barbican-worker-log" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.153735 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d1e9a91-e344-442f-8bbc-90b1f98e2e1f" containerName="barbican-worker-log" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.153751 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c59c3d1e-c5c3-4036-ae28-436585e303ab" containerName="ovn-controller" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.153758 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="c59c3d1e-c5c3-4036-ae28-436585e303ab" containerName="ovn-controller" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.153767 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83c080b8-d976-4e81-b103-2442bb2eafe1" containerName="keystone-api" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.153773 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="83c080b8-d976-4e81-b103-2442bb2eafe1" containerName="keystone-api" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.153788 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e51292cd-00f5-4e0a-ba2b-6319bd8a5e45" containerName="barbican-keystone-listener-log" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.153796 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="e51292cd-00f5-4e0a-ba2b-6319bd8a5e45" containerName="barbican-keystone-listener-log" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.153809 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ab346c9-f70f-4663-9b53-67b4d66b3112" containerName="init" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.153817 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ab346c9-f70f-4663-9b53-67b4d66b3112" containerName="init" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.153826 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9266e4bf-3e2e-4ed5-b799-02e56af27091" containerName="extract-content" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.153834 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="9266e4bf-3e2e-4ed5-b799-02e56af27091" containerName="extract-content" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.153845 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a0ab5a5-2d62-4457-be9d-fe8242ab7c91" containerName="nova-metadata-metadata" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.153852 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a0ab5a5-2d62-4457-be9d-fe8242ab7c91" containerName="nova-metadata-metadata" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.153867 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4512cf4e-6c75-493e-a80a-4b0accd750a8" containerName="setup-container" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.153875 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="4512cf4e-6c75-493e-a80a-4b0accd750a8" containerName="setup-container" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.153886 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86e320e0-80ee-40df-b3a1-a48cb810a435" containerName="cinder-api-log" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.153894 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="86e320e0-80ee-40df-b3a1-a48cb810a435" containerName="cinder-api-log" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.153912 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f182a7e0-ebd8-4258-9269-43a662e39af8" containerName="nova-cell1-conductor-conductor" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.153919 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="f182a7e0-ebd8-4258-9269-43a662e39af8" containerName="nova-cell1-conductor-conductor" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.153935 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4512cf4e-6c75-493e-a80a-4b0accd750a8" containerName="rabbitmq" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.153943 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="4512cf4e-6c75-493e-a80a-4b0accd750a8" containerName="rabbitmq" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.153956 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9266e4bf-3e2e-4ed5-b799-02e56af27091" containerName="extract-utilities" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.153964 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="9266e4bf-3e2e-4ed5-b799-02e56af27091" containerName="extract-utilities" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.153973 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df" containerName="ovsdbserver-nb" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.153981 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df" containerName="ovsdbserver-nb" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154018 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4702d881-d701-41a4-b36f-2d063a8ae246" containerName="barbican-api" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154026 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="4702d881-d701-41a4-b36f-2d063a8ae246" containerName="barbican-api" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154035 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4702d881-d701-41a4-b36f-2d063a8ae246" containerName="barbican-api-log" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154041 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="4702d881-d701-41a4-b36f-2d063a8ae246" containerName="barbican-api-log" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154056 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b708b100-166c-49b1-a47b-76d47ba2c6e5" containerName="glance-log" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154087 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="b708b100-166c-49b1-a47b-76d47ba2c6e5" containerName="glance-log" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154095 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09f0463c-f091-495e-9773-c07fe902c200" containerName="kube-state-metrics" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154103 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="09f0463c-f091-495e-9773-c07fe902c200" containerName="kube-state-metrics" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154114 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd" containerName="openstack-network-exporter" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154121 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd" containerName="openstack-network-exporter" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154133 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9" containerName="proxy-server" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154142 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9" containerName="proxy-server" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154155 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e7d9d60-9538-4870-8f62-434fd9f1ab0d" containerName="ovsdbserver-sb" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154162 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e7d9d60-9538-4870-8f62-434fd9f1ab0d" containerName="ovsdbserver-sb" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154175 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b97845-4578-4801-896a-281fdafdb351" containerName="placement-log" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154183 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b97845-4578-4801-896a-281fdafdb351" containerName="placement-log" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154192 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bae8251-3d5d-490b-9116-73769dc18101" containerName="nova-scheduler-scheduler" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154199 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bae8251-3d5d-490b-9116-73769dc18101" containerName="nova-scheduler-scheduler" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154208 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bf3edad-487a-4c68-9dbf-9789a94f8fba" containerName="nova-api-log" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154214 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bf3edad-487a-4c68-9dbf-9789a94f8fba" containerName="nova-api-log" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154227 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df" containerName="openstack-network-exporter" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154236 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df" containerName="openstack-network-exporter" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154248 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5920f152-04cd-4366-b479-665a1ce8937c" containerName="mysql-bootstrap" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154255 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="5920f152-04cd-4366-b479-665a1ce8937c" containerName="mysql-bootstrap" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154265 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e7d9d60-9538-4870-8f62-434fd9f1ab0d" containerName="openstack-network-exporter" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154272 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e7d9d60-9538-4870-8f62-434fd9f1ab0d" containerName="openstack-network-exporter" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154279 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f32041a0-a8fc-48a5-afab-476baff89e8c" containerName="setup-container" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154287 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="f32041a0-a8fc-48a5-afab-476baff89e8c" containerName="setup-container" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154296 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d1e9a91-e344-442f-8bbc-90b1f98e2e1f" containerName="barbican-worker" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154303 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d1e9a91-e344-442f-8bbc-90b1f98e2e1f" containerName="barbican-worker" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154315 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86e320e0-80ee-40df-b3a1-a48cb810a435" containerName="cinder-api" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154321 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="86e320e0-80ee-40df-b3a1-a48cb810a435" containerName="cinder-api" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154332 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dadc68f-0b89-41ba-84f8-e9bcde16ca64" containerName="openstack-network-exporter" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154339 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dadc68f-0b89-41ba-84f8-e9bcde16ca64" containerName="openstack-network-exporter" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154350 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5920f152-04cd-4366-b479-665a1ce8937c" containerName="galera" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154357 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="5920f152-04cd-4366-b479-665a1ce8937c" containerName="galera" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154366 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ab346c9-f70f-4663-9b53-67b4d66b3112" containerName="dnsmasq-dns" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154372 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ab346c9-f70f-4663-9b53-67b4d66b3112" containerName="dnsmasq-dns" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154380 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b708b100-166c-49b1-a47b-76d47ba2c6e5" containerName="glance-httpd" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154388 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="b708b100-166c-49b1-a47b-76d47ba2c6e5" containerName="glance-httpd" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154401 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d40516c-5fcc-4b24-918f-95bf79cb94b4" containerName="mariadb-account-delete" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154408 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d40516c-5fcc-4b24-918f-95bf79cb94b4" containerName="mariadb-account-delete" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154421 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2cda6de-ef12-42f4-a6e5-824c2db3cd01" containerName="glance-httpd" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154428 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2cda6de-ef12-42f4-a6e5-824c2db3cd01" containerName="glance-httpd" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154439 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd5ce048-5a09-45a4-9a7d-66712b326ab7" containerName="memcached" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154446 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd5ce048-5a09-45a4-9a7d-66712b326ab7" containerName="memcached" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154456 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f32041a0-a8fc-48a5-afab-476baff89e8c" containerName="rabbitmq" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154464 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="f32041a0-a8fc-48a5-afab-476baff89e8c" containerName="rabbitmq" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154482 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="167c5bc5-b21e-4c51-8765-1dcbf290294b" containerName="mysql-bootstrap" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154494 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="167c5bc5-b21e-4c51-8765-1dcbf290294b" containerName="mysql-bootstrap" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154509 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aab7d356-ec67-4e5d-9c25-f26e39786f79" containerName="proxy-httpd" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154516 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="aab7d356-ec67-4e5d-9c25-f26e39786f79" containerName="proxy-httpd" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154528 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aab7d356-ec67-4e5d-9c25-f26e39786f79" containerName="ceilometer-central-agent" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154535 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="aab7d356-ec67-4e5d-9c25-f26e39786f79" containerName="ceilometer-central-agent" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154544 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9" containerName="proxy-httpd" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154551 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9" containerName="proxy-httpd" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154564 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e09f9958-4faf-4a28-8214-c3ead146122c" containerName="nova-cell1-novncproxy-novncproxy" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154573 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="e09f9958-4faf-4a28-8214-c3ead146122c" containerName="nova-cell1-novncproxy-novncproxy" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154581 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="167c5bc5-b21e-4c51-8765-1dcbf290294b" containerName="galera" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154587 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="167c5bc5-b21e-4c51-8765-1dcbf290294b" containerName="galera" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154601 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a0ab5a5-2d62-4457-be9d-fe8242ab7c91" containerName="nova-metadata-log" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154607 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a0ab5a5-2d62-4457-be9d-fe8242ab7c91" containerName="nova-metadata-log" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154620 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0b97845-4578-4801-896a-281fdafdb351" containerName="placement-api" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154628 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0b97845-4578-4801-896a-281fdafdb351" containerName="placement-api" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154636 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc920d71-4308-4d90-a891-8b760651ccc9" containerName="cinder-scheduler" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154644 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc920d71-4308-4d90-a891-8b760651ccc9" containerName="cinder-scheduler" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154655 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4" containerName="nova-cell0-conductor-conductor" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154667 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4" containerName="nova-cell0-conductor-conductor" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154678 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9266e4bf-3e2e-4ed5-b799-02e56af27091" containerName="registry-server" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154685 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="9266e4bf-3e2e-4ed5-b799-02e56af27091" containerName="registry-server" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154695 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aab7d356-ec67-4e5d-9c25-f26e39786f79" containerName="sg-core" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154703 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="aab7d356-ec67-4e5d-9c25-f26e39786f79" containerName="sg-core" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154713 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aab7d356-ec67-4e5d-9c25-f26e39786f79" containerName="ceilometer-notification-agent" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154721 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="aab7d356-ec67-4e5d-9c25-f26e39786f79" containerName="ceilometer-notification-agent" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154729 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2cda6de-ef12-42f4-a6e5-824c2db3cd01" containerName="glance-log" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154736 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2cda6de-ef12-42f4-a6e5-824c2db3cd01" containerName="glance-log" Nov 25 09:48:06 crc kubenswrapper[4776]: E1125 09:48:06.154744 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e51292cd-00f5-4e0a-ba2b-6319bd8a5e45" containerName="barbican-keystone-listener" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154751 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="e51292cd-00f5-4e0a-ba2b-6319bd8a5e45" containerName="barbican-keystone-listener" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154934 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="aab7d356-ec67-4e5d-9c25-f26e39786f79" containerName="ceilometer-central-agent" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154950 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="f182a7e0-ebd8-4258-9269-43a662e39af8" containerName="nova-cell1-conductor-conductor" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154961 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="b708b100-166c-49b1-a47b-76d47ba2c6e5" containerName="glance-httpd" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154976 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="aab7d356-ec67-4e5d-9c25-f26e39786f79" containerName="sg-core" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.154990 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="e51292cd-00f5-4e0a-ba2b-6319bd8a5e45" containerName="barbican-keystone-listener" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155001 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc920d71-4308-4d90-a891-8b760651ccc9" containerName="cinder-scheduler" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155011 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ab346c9-f70f-4663-9b53-67b4d66b3112" containerName="dnsmasq-dns" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155021 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a0ab5a5-2d62-4457-be9d-fe8242ab7c91" containerName="nova-metadata-log" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155030 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0b97845-4578-4801-896a-281fdafdb351" containerName="placement-log" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155037 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2cda6de-ef12-42f4-a6e5-824c2db3cd01" containerName="glance-httpd" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155050 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="aab7d356-ec67-4e5d-9c25-f26e39786f79" containerName="proxy-httpd" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155057 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df" containerName="openstack-network-exporter" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155090 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="167c5bc5-b21e-4c51-8765-1dcbf290294b" containerName="galera" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155101 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="83c080b8-d976-4e81-b103-2442bb2eafe1" containerName="keystone-api" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155111 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdba3e6f-7d96-4e87-95dd-bf6a91e6a8cd" containerName="openstack-network-exporter" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155124 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="4702d881-d701-41a4-b36f-2d063a8ae246" containerName="barbican-api" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155134 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9cc962d-25ca-4dcb-9fb3-06eaa655c3b4" containerName="nova-cell0-conductor-conductor" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155144 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="9266e4bf-3e2e-4ed5-b799-02e56af27091" containerName="registry-server" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155152 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="86e320e0-80ee-40df-b3a1-a48cb810a435" containerName="cinder-api" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155163 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0b97845-4578-4801-896a-281fdafdb351" containerName="placement-api" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155170 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="aab7d356-ec67-4e5d-9c25-f26e39786f79" containerName="ceilometer-notification-agent" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155182 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bf3edad-487a-4c68-9dbf-9789a94f8fba" containerName="nova-api-api" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155191 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bf3edad-487a-4c68-9dbf-9789a94f8fba" containerName="nova-api-log" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155204 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dadc68f-0b89-41ba-84f8-e9bcde16ca64" containerName="ovn-northd" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155215 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9" containerName="proxy-server" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155222 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bae8251-3d5d-490b-9116-73769dc18101" containerName="nova-scheduler-scheduler" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155231 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="e51292cd-00f5-4e0a-ba2b-6319bd8a5e45" containerName="barbican-keystone-listener-log" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155243 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="c59c3d1e-c5c3-4036-ae28-436585e303ab" containerName="ovn-controller" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155254 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d1e9a91-e344-442f-8bbc-90b1f98e2e1f" containerName="barbican-worker-log" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155268 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a8cfa42-c498-4a9c-bb2b-cd7939c7a8df" containerName="ovsdbserver-nb" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155277 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dadc68f-0b89-41ba-84f8-e9bcde16ca64" containerName="openstack-network-exporter" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155287 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="09f0463c-f091-495e-9773-c07fe902c200" containerName="kube-state-metrics" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155297 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="f32041a0-a8fc-48a5-afab-476baff89e8c" containerName="rabbitmq" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155309 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e7d9d60-9538-4870-8f62-434fd9f1ab0d" containerName="ovsdbserver-sb" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155321 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="e09f9958-4faf-4a28-8214-c3ead146122c" containerName="nova-cell1-novncproxy-novncproxy" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155330 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="b708b100-166c-49b1-a47b-76d47ba2c6e5" containerName="glance-log" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155340 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d1e9a91-e344-442f-8bbc-90b1f98e2e1f" containerName="barbican-worker" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155349 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="4702d881-d701-41a4-b36f-2d063a8ae246" containerName="barbican-api-log" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155358 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="5920f152-04cd-4366-b479-665a1ce8937c" containerName="galera" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155370 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="f89d44cf-1ca0-4581-aa8c-e86e0f3c61d9" containerName="proxy-httpd" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155380 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d40516c-5fcc-4b24-918f-95bf79cb94b4" containerName="mariadb-account-delete" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155393 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="86e320e0-80ee-40df-b3a1-a48cb810a435" containerName="cinder-api-log" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155405 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2cda6de-ef12-42f4-a6e5-824c2db3cd01" containerName="glance-log" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155418 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc920d71-4308-4d90-a891-8b760651ccc9" containerName="probe" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155425 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="4512cf4e-6c75-493e-a80a-4b0accd750a8" containerName="rabbitmq" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155432 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e7d9d60-9538-4870-8f62-434fd9f1ab0d" containerName="openstack-network-exporter" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155445 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd5ce048-5a09-45a4-9a7d-66712b326ab7" containerName="memcached" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.155460 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a0ab5a5-2d62-4457-be9d-fe8242ab7c91" containerName="nova-metadata-metadata" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.156887 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wvl56" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.167996 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wvl56"] Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.197506 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.197524 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"f182a7e0-ebd8-4258-9269-43a662e39af8","Type":"ContainerDied","Data":"35af05bcb3324eb2b85951a073a1335b4087f998234cc11e4721f613083d3288"} Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.197589 4776 scope.go:117] "RemoveContainer" containerID="e5a40e8fd3cdfddb16f8d8578ce8ae7e8bb1b0df1b91c68efe717a27f4fc382c" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.201899 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aab7d356-ec67-4e5d-9c25-f26e39786f79","Type":"ContainerDied","Data":"c5e4a4b031984a16cc9d6b7895ceb3f8ccac6ee47ecc06363cdffe20b5013c94"} Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.202041 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.209402 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-56988fbb4-nlxs9" event={"ID":"e51292cd-00f5-4e0a-ba2b-6319bd8a5e45","Type":"ContainerDied","Data":"ee48efd287e21fe11c445356b7f2fed53754f03152b8e09e0cec9fc4b8e05b3c"} Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.209467 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6b75d6ddb7-rmtms" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.209536 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-56988fbb4-nlxs9" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.209937 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5wmvt" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.210057 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-2bs2f" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.236531 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.240136 4776 scope.go:117] "RemoveContainer" containerID="ec693b92c2b81399b4a2d57c1980de7b988f0cc59e2cef0ec21c34b5dff21bd2" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.241510 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.260440 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5wmvt"] Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.268816 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5wmvt"] Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.273649 4776 scope.go:117] "RemoveContainer" containerID="c99e1a9b60f21bc937fd065f74ab82371110943731ae37f120fc60b0c71336da" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.277647 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.285396 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.290565 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-56988fbb4-nlxs9"] Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.297278 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-56988fbb4-nlxs9"] Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.306802 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-6b75d6ddb7-rmtms"] Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.308364 4776 scope.go:117] "RemoveContainer" containerID="a2a4c7e738a8387346bc8cb061fcaffab37035f8b7356aba874420f641555783" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.311882 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-6b75d6ddb7-rmtms"] Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.317737 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-2bs2f"] Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.322513 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-2bs2f"] Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.325590 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41c7c636-6457-4a7d-b0da-30c590b643c7-catalog-content\") pod \"redhat-marketplace-wvl56\" (UID: \"41c7c636-6457-4a7d-b0da-30c590b643c7\") " pod="openshift-marketplace/redhat-marketplace-wvl56" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.325754 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkkz5\" (UniqueName: \"kubernetes.io/projected/41c7c636-6457-4a7d-b0da-30c590b643c7-kube-api-access-hkkz5\") pod \"redhat-marketplace-wvl56\" (UID: \"41c7c636-6457-4a7d-b0da-30c590b643c7\") " pod="openshift-marketplace/redhat-marketplace-wvl56" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.325789 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41c7c636-6457-4a7d-b0da-30c590b643c7-utilities\") pod \"redhat-marketplace-wvl56\" (UID: \"41c7c636-6457-4a7d-b0da-30c590b643c7\") " pod="openshift-marketplace/redhat-marketplace-wvl56" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.334198 4776 scope.go:117] "RemoveContainer" containerID="a4df2757871f92f27d85077692ce959d3142d7496b28c90b9062a0ed0d5d450d" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.356893 4776 scope.go:117] "RemoveContainer" containerID="63cb3f8363fb0824de5241eef2c1f027163930b39c0a67a62ab2eadf418c85d7" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.389231 4776 scope.go:117] "RemoveContainer" containerID="a1a93104c41ae0baf6a1781b8d29302ccef9fef9da152e6274b2e92e08df9a8c" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.427361 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41c7c636-6457-4a7d-b0da-30c590b643c7-utilities\") pod \"redhat-marketplace-wvl56\" (UID: \"41c7c636-6457-4a7d-b0da-30c590b643c7\") " pod="openshift-marketplace/redhat-marketplace-wvl56" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.427475 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41c7c636-6457-4a7d-b0da-30c590b643c7-catalog-content\") pod \"redhat-marketplace-wvl56\" (UID: \"41c7c636-6457-4a7d-b0da-30c590b643c7\") " pod="openshift-marketplace/redhat-marketplace-wvl56" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.427581 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkkz5\" (UniqueName: \"kubernetes.io/projected/41c7c636-6457-4a7d-b0da-30c590b643c7-kube-api-access-hkkz5\") pod \"redhat-marketplace-wvl56\" (UID: \"41c7c636-6457-4a7d-b0da-30c590b643c7\") " pod="openshift-marketplace/redhat-marketplace-wvl56" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.428080 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41c7c636-6457-4a7d-b0da-30c590b643c7-utilities\") pod \"redhat-marketplace-wvl56\" (UID: \"41c7c636-6457-4a7d-b0da-30c590b643c7\") " pod="openshift-marketplace/redhat-marketplace-wvl56" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.428284 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41c7c636-6457-4a7d-b0da-30c590b643c7-catalog-content\") pod \"redhat-marketplace-wvl56\" (UID: \"41c7c636-6457-4a7d-b0da-30c590b643c7\") " pod="openshift-marketplace/redhat-marketplace-wvl56" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.452956 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkkz5\" (UniqueName: \"kubernetes.io/projected/41c7c636-6457-4a7d-b0da-30c590b643c7-kube-api-access-hkkz5\") pod \"redhat-marketplace-wvl56\" (UID: \"41c7c636-6457-4a7d-b0da-30c590b643c7\") " pod="openshift-marketplace/redhat-marketplace-wvl56" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.494717 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wvl56" Nov 25 09:48:06 crc kubenswrapper[4776]: I1125 09:48:06.968709 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wvl56"] Nov 25 09:48:07 crc kubenswrapper[4776]: I1125 09:48:07.222150 4776 generic.go:334] "Generic (PLEG): container finished" podID="41c7c636-6457-4a7d-b0da-30c590b643c7" containerID="ae84c5b4d9bf124138f331d12ea865041ecf356b53cae9f743c7d06f073a5270" exitCode=0 Nov 25 09:48:07 crc kubenswrapper[4776]: I1125 09:48:07.222214 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvl56" event={"ID":"41c7c636-6457-4a7d-b0da-30c590b643c7","Type":"ContainerDied","Data":"ae84c5b4d9bf124138f331d12ea865041ecf356b53cae9f743c7d06f073a5270"} Nov 25 09:48:07 crc kubenswrapper[4776]: I1125 09:48:07.222243 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvl56" event={"ID":"41c7c636-6457-4a7d-b0da-30c590b643c7","Type":"ContainerStarted","Data":"fa9a325b0b1fb67af9753d6386d3c0fdbbdb14c12ade7e3f72ada430e72baf58"} Nov 25 09:48:07 crc kubenswrapper[4776]: I1125 09:48:07.252762 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="09f0463c-f091-495e-9773-c07fe902c200" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.0.173:8081/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 25 09:48:07 crc kubenswrapper[4776]: I1125 09:48:07.685306 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83c080b8-d976-4e81-b103-2442bb2eafe1" path="/var/lib/kubelet/pods/83c080b8-d976-4e81-b103-2442bb2eafe1/volumes" Nov 25 09:48:07 crc kubenswrapper[4776]: I1125 09:48:07.686687 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9266e4bf-3e2e-4ed5-b799-02e56af27091" path="/var/lib/kubelet/pods/9266e4bf-3e2e-4ed5-b799-02e56af27091/volumes" Nov 25 09:48:07 crc kubenswrapper[4776]: I1125 09:48:07.687721 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aab7d356-ec67-4e5d-9c25-f26e39786f79" path="/var/lib/kubelet/pods/aab7d356-ec67-4e5d-9c25-f26e39786f79/volumes" Nov 25 09:48:07 crc kubenswrapper[4776]: I1125 09:48:07.690692 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c59c3d1e-c5c3-4036-ae28-436585e303ab" path="/var/lib/kubelet/pods/c59c3d1e-c5c3-4036-ae28-436585e303ab/volumes" Nov 25 09:48:07 crc kubenswrapper[4776]: I1125 09:48:07.691522 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e51292cd-00f5-4e0a-ba2b-6319bd8a5e45" path="/var/lib/kubelet/pods/e51292cd-00f5-4e0a-ba2b-6319bd8a5e45/volumes" Nov 25 09:48:07 crc kubenswrapper[4776]: I1125 09:48:07.692222 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f182a7e0-ebd8-4258-9269-43a662e39af8" path="/var/lib/kubelet/pods/f182a7e0-ebd8-4258-9269-43a662e39af8/volumes" Nov 25 09:48:07 crc kubenswrapper[4776]: I1125 09:48:07.948313 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-54dbffbb57-kmqfb" Nov 25 09:48:07 crc kubenswrapper[4776]: I1125 09:48:07.974952 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-ovndb-tls-certs\") pod \"fe3ff67b-01ff-480f-be91-7c8235593c97\" (UID: \"fe3ff67b-01ff-480f-be91-7c8235593c97\") " Nov 25 09:48:07 crc kubenswrapper[4776]: I1125 09:48:07.975107 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-combined-ca-bundle\") pod \"fe3ff67b-01ff-480f-be91-7c8235593c97\" (UID: \"fe3ff67b-01ff-480f-be91-7c8235593c97\") " Nov 25 09:48:07 crc kubenswrapper[4776]: I1125 09:48:07.975139 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-internal-tls-certs\") pod \"fe3ff67b-01ff-480f-be91-7c8235593c97\" (UID: \"fe3ff67b-01ff-480f-be91-7c8235593c97\") " Nov 25 09:48:07 crc kubenswrapper[4776]: I1125 09:48:07.975176 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-config\") pod \"fe3ff67b-01ff-480f-be91-7c8235593c97\" (UID: \"fe3ff67b-01ff-480f-be91-7c8235593c97\") " Nov 25 09:48:07 crc kubenswrapper[4776]: I1125 09:48:07.975220 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-httpd-config\") pod \"fe3ff67b-01ff-480f-be91-7c8235593c97\" (UID: \"fe3ff67b-01ff-480f-be91-7c8235593c97\") " Nov 25 09:48:07 crc kubenswrapper[4776]: I1125 09:48:07.975267 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f78rw\" (UniqueName: \"kubernetes.io/projected/fe3ff67b-01ff-480f-be91-7c8235593c97-kube-api-access-f78rw\") pod \"fe3ff67b-01ff-480f-be91-7c8235593c97\" (UID: \"fe3ff67b-01ff-480f-be91-7c8235593c97\") " Nov 25 09:48:07 crc kubenswrapper[4776]: I1125 09:48:07.975304 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-public-tls-certs\") pod \"fe3ff67b-01ff-480f-be91-7c8235593c97\" (UID: \"fe3ff67b-01ff-480f-be91-7c8235593c97\") " Nov 25 09:48:07 crc kubenswrapper[4776]: I1125 09:48:07.981760 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "fe3ff67b-01ff-480f-be91-7c8235593c97" (UID: "fe3ff67b-01ff-480f-be91-7c8235593c97"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:07 crc kubenswrapper[4776]: I1125 09:48:07.992596 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe3ff67b-01ff-480f-be91-7c8235593c97-kube-api-access-f78rw" (OuterVolumeSpecName: "kube-api-access-f78rw") pod "fe3ff67b-01ff-480f-be91-7c8235593c97" (UID: "fe3ff67b-01ff-480f-be91-7c8235593c97"). InnerVolumeSpecName "kube-api-access-f78rw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:08 crc kubenswrapper[4776]: I1125 09:48:08.021854 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-config" (OuterVolumeSpecName: "config") pod "fe3ff67b-01ff-480f-be91-7c8235593c97" (UID: "fe3ff67b-01ff-480f-be91-7c8235593c97"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:08 crc kubenswrapper[4776]: I1125 09:48:08.029838 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "fe3ff67b-01ff-480f-be91-7c8235593c97" (UID: "fe3ff67b-01ff-480f-be91-7c8235593c97"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:08 crc kubenswrapper[4776]: I1125 09:48:08.033994 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fe3ff67b-01ff-480f-be91-7c8235593c97" (UID: "fe3ff67b-01ff-480f-be91-7c8235593c97"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:08 crc kubenswrapper[4776]: I1125 09:48:08.041345 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "fe3ff67b-01ff-480f-be91-7c8235593c97" (UID: "fe3ff67b-01ff-480f-be91-7c8235593c97"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:08 crc kubenswrapper[4776]: I1125 09:48:08.043451 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "fe3ff67b-01ff-480f-be91-7c8235593c97" (UID: "fe3ff67b-01ff-480f-be91-7c8235593c97"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:08 crc kubenswrapper[4776]: I1125 09:48:08.076593 4776 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:08 crc kubenswrapper[4776]: I1125 09:48:08.076643 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:08 crc kubenswrapper[4776]: I1125 09:48:08.076684 4776 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:08 crc kubenswrapper[4776]: I1125 09:48:08.076697 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:08 crc kubenswrapper[4776]: I1125 09:48:08.076710 4776 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:08 crc kubenswrapper[4776]: I1125 09:48:08.076721 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f78rw\" (UniqueName: \"kubernetes.io/projected/fe3ff67b-01ff-480f-be91-7c8235593c97-kube-api-access-f78rw\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:08 crc kubenswrapper[4776]: I1125 09:48:08.076734 4776 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe3ff67b-01ff-480f-be91-7c8235593c97-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:08 crc kubenswrapper[4776]: I1125 09:48:08.248368 4776 generic.go:334] "Generic (PLEG): container finished" podID="fe3ff67b-01ff-480f-be91-7c8235593c97" containerID="94e2c5021a1689b990cd8fd474841522443a046e3012b95cb3f5f1f9b90efdd1" exitCode=0 Nov 25 09:48:08 crc kubenswrapper[4776]: I1125 09:48:08.248444 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54dbffbb57-kmqfb" event={"ID":"fe3ff67b-01ff-480f-be91-7c8235593c97","Type":"ContainerDied","Data":"94e2c5021a1689b990cd8fd474841522443a046e3012b95cb3f5f1f9b90efdd1"} Nov 25 09:48:08 crc kubenswrapper[4776]: I1125 09:48:08.248477 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54dbffbb57-kmqfb" event={"ID":"fe3ff67b-01ff-480f-be91-7c8235593c97","Type":"ContainerDied","Data":"7d83e210b3c97dae8119bdb60c3219ec6037f6359c74ccc6a38954fcea697f4c"} Nov 25 09:48:08 crc kubenswrapper[4776]: I1125 09:48:08.248497 4776 scope.go:117] "RemoveContainer" containerID="5c1129db350942a00317a1332a9a8db16c63fbd966ba3d63890450c144964ddb" Nov 25 09:48:08 crc kubenswrapper[4776]: I1125 09:48:08.248658 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-54dbffbb57-kmqfb" Nov 25 09:48:08 crc kubenswrapper[4776]: I1125 09:48:08.251752 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvl56" event={"ID":"41c7c636-6457-4a7d-b0da-30c590b643c7","Type":"ContainerStarted","Data":"b0a376848e32ca571b595940521d8008deb7cfaab3ad42b2be71790ec63e8a01"} Nov 25 09:48:08 crc kubenswrapper[4776]: I1125 09:48:08.273640 4776 scope.go:117] "RemoveContainer" containerID="94e2c5021a1689b990cd8fd474841522443a046e3012b95cb3f5f1f9b90efdd1" Nov 25 09:48:08 crc kubenswrapper[4776]: I1125 09:48:08.291929 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-54dbffbb57-kmqfb"] Nov 25 09:48:08 crc kubenswrapper[4776]: I1125 09:48:08.297587 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-54dbffbb57-kmqfb"] Nov 25 09:48:08 crc kubenswrapper[4776]: I1125 09:48:08.298503 4776 scope.go:117] "RemoveContainer" containerID="5c1129db350942a00317a1332a9a8db16c63fbd966ba3d63890450c144964ddb" Nov 25 09:48:08 crc kubenswrapper[4776]: E1125 09:48:08.299059 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c1129db350942a00317a1332a9a8db16c63fbd966ba3d63890450c144964ddb\": container with ID starting with 5c1129db350942a00317a1332a9a8db16c63fbd966ba3d63890450c144964ddb not found: ID does not exist" containerID="5c1129db350942a00317a1332a9a8db16c63fbd966ba3d63890450c144964ddb" Nov 25 09:48:08 crc kubenswrapper[4776]: I1125 09:48:08.299142 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c1129db350942a00317a1332a9a8db16c63fbd966ba3d63890450c144964ddb"} err="failed to get container status \"5c1129db350942a00317a1332a9a8db16c63fbd966ba3d63890450c144964ddb\": rpc error: code = NotFound desc = could not find container \"5c1129db350942a00317a1332a9a8db16c63fbd966ba3d63890450c144964ddb\": container with ID starting with 5c1129db350942a00317a1332a9a8db16c63fbd966ba3d63890450c144964ddb not found: ID does not exist" Nov 25 09:48:08 crc kubenswrapper[4776]: I1125 09:48:08.299165 4776 scope.go:117] "RemoveContainer" containerID="94e2c5021a1689b990cd8fd474841522443a046e3012b95cb3f5f1f9b90efdd1" Nov 25 09:48:08 crc kubenswrapper[4776]: E1125 09:48:08.299438 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94e2c5021a1689b990cd8fd474841522443a046e3012b95cb3f5f1f9b90efdd1\": container with ID starting with 94e2c5021a1689b990cd8fd474841522443a046e3012b95cb3f5f1f9b90efdd1 not found: ID does not exist" containerID="94e2c5021a1689b990cd8fd474841522443a046e3012b95cb3f5f1f9b90efdd1" Nov 25 09:48:08 crc kubenswrapper[4776]: I1125 09:48:08.299474 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94e2c5021a1689b990cd8fd474841522443a046e3012b95cb3f5f1f9b90efdd1"} err="failed to get container status \"94e2c5021a1689b990cd8fd474841522443a046e3012b95cb3f5f1f9b90efdd1\": rpc error: code = NotFound desc = could not find container \"94e2c5021a1689b990cd8fd474841522443a046e3012b95cb3f5f1f9b90efdd1\": container with ID starting with 94e2c5021a1689b990cd8fd474841522443a046e3012b95cb3f5f1f9b90efdd1 not found: ID does not exist" Nov 25 09:48:08 crc kubenswrapper[4776]: E1125 09:48:08.583901 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:08 crc kubenswrapper[4776]: E1125 09:48:08.583982 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/14b51bba-0b15-469b-bdea-90fae8d9b976-operator-scripts podName:14b51bba-0b15-469b-bdea-90fae8d9b976 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:16.58396172 +0000 UTC m=+1441.625021273 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/14b51bba-0b15-469b-bdea-90fae8d9b976-operator-scripts") pod "novaapi3b46-account-delete-9w5r9" (UID: "14b51bba-0b15-469b-bdea-90fae8d9b976") : configmap "openstack-scripts" not found Nov 25 09:48:08 crc kubenswrapper[4776]: E1125 09:48:08.584031 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:08 crc kubenswrapper[4776]: E1125 09:48:08.584112 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ef85f90f-c377-4273-9fff-451420d6008a-operator-scripts podName:ef85f90f-c377-4273-9fff-451420d6008a nodeName:}" failed. No retries permitted until 2025-11-25 09:48:16.584096073 +0000 UTC m=+1441.625155626 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/ef85f90f-c377-4273-9fff-451420d6008a-operator-scripts") pod "cinder774d-account-delete-b29n5" (UID: "ef85f90f-c377-4273-9fff-451420d6008a") : configmap "openstack-scripts" not found Nov 25 09:48:08 crc kubenswrapper[4776]: E1125 09:48:08.584142 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:08 crc kubenswrapper[4776]: E1125 09:48:08.584162 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/73f2c2ab-6cb6-4544-a3f7-20722e5b6400-operator-scripts podName:73f2c2ab-6cb6-4544-a3f7-20722e5b6400 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:16.584156315 +0000 UTC m=+1441.625215858 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/73f2c2ab-6cb6-4544-a3f7-20722e5b6400-operator-scripts") pod "neutron5151-account-delete-8vbjg" (UID: "73f2c2ab-6cb6-4544-a3f7-20722e5b6400") : configmap "openstack-scripts" not found Nov 25 09:48:08 crc kubenswrapper[4776]: E1125 09:48:08.584207 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:08 crc kubenswrapper[4776]: E1125 09:48:08.584227 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/946ea713-bfba-449f-82ab-28e915469938-operator-scripts podName:946ea713-bfba-449f-82ab-28e915469938 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:16.584221107 +0000 UTC m=+1441.625280660 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/946ea713-bfba-449f-82ab-28e915469938-operator-scripts") pod "novacell09dc4-account-delete-w5xcj" (UID: "946ea713-bfba-449f-82ab-28e915469938") : configmap "openstack-scripts" not found Nov 25 09:48:08 crc kubenswrapper[4776]: E1125 09:48:08.584609 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:08 crc kubenswrapper[4776]: E1125 09:48:08.584763 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6be099e6-89d6-45a0-9d0f-4dca4d91701a-operator-scripts podName:6be099e6-89d6-45a0-9d0f-4dca4d91701a nodeName:}" failed. No retries permitted until 2025-11-25 09:48:16.58474078 +0000 UTC m=+1441.625800413 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/6be099e6-89d6-45a0-9d0f-4dca4d91701a-operator-scripts") pod "placemente176-account-delete-wbgz5" (UID: "6be099e6-89d6-45a0-9d0f-4dca4d91701a") : configmap "openstack-scripts" not found Nov 25 09:48:08 crc kubenswrapper[4776]: E1125 09:48:08.685032 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:08 crc kubenswrapper[4776]: E1125 09:48:08.685117 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/bcaec0f3-13af-4f73-8e7f-7e575b8e2008-operator-scripts podName:bcaec0f3-13af-4f73-8e7f-7e575b8e2008 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:16.685096603 +0000 UTC m=+1441.726156156 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/bcaec0f3-13af-4f73-8e7f-7e575b8e2008-operator-scripts") pod "glanceb9b5-account-delete-tjrnr" (UID: "bcaec0f3-13af-4f73-8e7f-7e575b8e2008") : configmap "openstack-scripts" not found Nov 25 09:48:09 crc kubenswrapper[4776]: I1125 09:48:09.262216 4776 generic.go:334] "Generic (PLEG): container finished" podID="41c7c636-6457-4a7d-b0da-30c590b643c7" containerID="b0a376848e32ca571b595940521d8008deb7cfaab3ad42b2be71790ec63e8a01" exitCode=0 Nov 25 09:48:09 crc kubenswrapper[4776]: I1125 09:48:09.262272 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvl56" event={"ID":"41c7c636-6457-4a7d-b0da-30c590b643c7","Type":"ContainerDied","Data":"b0a376848e32ca571b595940521d8008deb7cfaab3ad42b2be71790ec63e8a01"} Nov 25 09:48:09 crc kubenswrapper[4776]: I1125 09:48:09.673918 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe3ff67b-01ff-480f-be91-7c8235593c97" path="/var/lib/kubelet/pods/fe3ff67b-01ff-480f-be91-7c8235593c97/volumes" Nov 25 09:48:10 crc kubenswrapper[4776]: I1125 09:48:10.284970 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvl56" event={"ID":"41c7c636-6457-4a7d-b0da-30c590b643c7","Type":"ContainerStarted","Data":"f47954c53fabd913eea349a21f67f1dd6e34eacf30d0b802dec2287bfa0ff4fe"} Nov 25 09:48:10 crc kubenswrapper[4776]: I1125 09:48:10.311594 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wvl56" podStartSLOduration=1.8998179579999999 podStartE2EDuration="4.311563979s" podCreationTimestamp="2025-11-25 09:48:06 +0000 UTC" firstStartedPulling="2025-11-25 09:48:07.223653787 +0000 UTC m=+1432.264713340" lastFinishedPulling="2025-11-25 09:48:09.635399808 +0000 UTC m=+1434.676459361" observedRunningTime="2025-11-25 09:48:10.303805784 +0000 UTC m=+1435.344865347" watchObservedRunningTime="2025-11-25 09:48:10.311563979 +0000 UTC m=+1435.352623532" Nov 25 09:48:10 crc kubenswrapper[4776]: E1125 09:48:10.628007 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06 is running failed: container process not found" containerID="eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 25 09:48:10 crc kubenswrapper[4776]: E1125 09:48:10.628399 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06 is running failed: container process not found" containerID="eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 25 09:48:10 crc kubenswrapper[4776]: E1125 09:48:10.628600 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06 is running failed: container process not found" containerID="eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 25 09:48:10 crc kubenswrapper[4776]: E1125 09:48:10.628625 4776 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-7fgjf" podUID="10fe3dc2-2bde-4380-8ad3-d71832e31095" containerName="ovsdb-server" Nov 25 09:48:10 crc kubenswrapper[4776]: E1125 09:48:10.629304 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5002006f906f0b17afbf25b23648688a842975cfd3b385d57e3191040607fb57" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 25 09:48:10 crc kubenswrapper[4776]: E1125 09:48:10.631223 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5002006f906f0b17afbf25b23648688a842975cfd3b385d57e3191040607fb57" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 25 09:48:10 crc kubenswrapper[4776]: E1125 09:48:10.634228 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5002006f906f0b17afbf25b23648688a842975cfd3b385d57e3191040607fb57" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 25 09:48:10 crc kubenswrapper[4776]: E1125 09:48:10.634434 4776 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-7fgjf" podUID="10fe3dc2-2bde-4380-8ad3-d71832e31095" containerName="ovs-vswitchd" Nov 25 09:48:15 crc kubenswrapper[4776]: E1125 09:48:15.628527 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06 is running failed: container process not found" containerID="eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 25 09:48:15 crc kubenswrapper[4776]: E1125 09:48:15.629711 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06 is running failed: container process not found" containerID="eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 25 09:48:15 crc kubenswrapper[4776]: E1125 09:48:15.630653 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06 is running failed: container process not found" containerID="eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 25 09:48:15 crc kubenswrapper[4776]: E1125 09:48:15.630693 4776 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-7fgjf" podUID="10fe3dc2-2bde-4380-8ad3-d71832e31095" containerName="ovsdb-server" Nov 25 09:48:15 crc kubenswrapper[4776]: E1125 09:48:15.631192 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5002006f906f0b17afbf25b23648688a842975cfd3b385d57e3191040607fb57" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 25 09:48:15 crc kubenswrapper[4776]: E1125 09:48:15.632618 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5002006f906f0b17afbf25b23648688a842975cfd3b385d57e3191040607fb57" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 25 09:48:15 crc kubenswrapper[4776]: E1125 09:48:15.634864 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5002006f906f0b17afbf25b23648688a842975cfd3b385d57e3191040607fb57" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 25 09:48:15 crc kubenswrapper[4776]: E1125 09:48:15.634917 4776 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-7fgjf" podUID="10fe3dc2-2bde-4380-8ad3-d71832e31095" containerName="ovs-vswitchd" Nov 25 09:48:16 crc kubenswrapper[4776]: I1125 09:48:16.495237 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wvl56" Nov 25 09:48:16 crc kubenswrapper[4776]: I1125 09:48:16.495753 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wvl56" Nov 25 09:48:16 crc kubenswrapper[4776]: I1125 09:48:16.551980 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wvl56" Nov 25 09:48:16 crc kubenswrapper[4776]: E1125 09:48:16.604154 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:16 crc kubenswrapper[4776]: E1125 09:48:16.604269 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:16 crc kubenswrapper[4776]: E1125 09:48:16.604316 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:16 crc kubenswrapper[4776]: E1125 09:48:16.604327 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/14b51bba-0b15-469b-bdea-90fae8d9b976-operator-scripts podName:14b51bba-0b15-469b-bdea-90fae8d9b976 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:32.604306603 +0000 UTC m=+1457.645366156 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/14b51bba-0b15-469b-bdea-90fae8d9b976-operator-scripts") pod "novaapi3b46-account-delete-9w5r9" (UID: "14b51bba-0b15-469b-bdea-90fae8d9b976") : configmap "openstack-scripts" not found Nov 25 09:48:16 crc kubenswrapper[4776]: E1125 09:48:16.604381 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:16 crc kubenswrapper[4776]: E1125 09:48:16.604422 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ef85f90f-c377-4273-9fff-451420d6008a-operator-scripts podName:ef85f90f-c377-4273-9fff-451420d6008a nodeName:}" failed. No retries permitted until 2025-11-25 09:48:32.604399805 +0000 UTC m=+1457.645459398 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/ef85f90f-c377-4273-9fff-451420d6008a-operator-scripts") pod "cinder774d-account-delete-b29n5" (UID: "ef85f90f-c377-4273-9fff-451420d6008a") : configmap "openstack-scripts" not found Nov 25 09:48:16 crc kubenswrapper[4776]: E1125 09:48:16.604497 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6be099e6-89d6-45a0-9d0f-4dca4d91701a-operator-scripts podName:6be099e6-89d6-45a0-9d0f-4dca4d91701a nodeName:}" failed. No retries permitted until 2025-11-25 09:48:32.604465157 +0000 UTC m=+1457.645524900 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/6be099e6-89d6-45a0-9d0f-4dca4d91701a-operator-scripts") pod "placemente176-account-delete-wbgz5" (UID: "6be099e6-89d6-45a0-9d0f-4dca4d91701a") : configmap "openstack-scripts" not found Nov 25 09:48:16 crc kubenswrapper[4776]: E1125 09:48:16.604695 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:16 crc kubenswrapper[4776]: E1125 09:48:16.604719 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/73f2c2ab-6cb6-4544-a3f7-20722e5b6400-operator-scripts podName:73f2c2ab-6cb6-4544-a3f7-20722e5b6400 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:32.604708133 +0000 UTC m=+1457.645767936 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/73f2c2ab-6cb6-4544-a3f7-20722e5b6400-operator-scripts") pod "neutron5151-account-delete-8vbjg" (UID: "73f2c2ab-6cb6-4544-a3f7-20722e5b6400") : configmap "openstack-scripts" not found Nov 25 09:48:16 crc kubenswrapper[4776]: E1125 09:48:16.604910 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/946ea713-bfba-449f-82ab-28e915469938-operator-scripts podName:946ea713-bfba-449f-82ab-28e915469938 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:32.604886298 +0000 UTC m=+1457.645945851 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/946ea713-bfba-449f-82ab-28e915469938-operator-scripts") pod "novacell09dc4-account-delete-w5xcj" (UID: "946ea713-bfba-449f-82ab-28e915469938") : configmap "openstack-scripts" not found Nov 25 09:48:16 crc kubenswrapper[4776]: E1125 09:48:16.705259 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:16 crc kubenswrapper[4776]: E1125 09:48:16.706359 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/bcaec0f3-13af-4f73-8e7f-7e575b8e2008-operator-scripts podName:bcaec0f3-13af-4f73-8e7f-7e575b8e2008 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:32.706339238 +0000 UTC m=+1457.747398791 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/bcaec0f3-13af-4f73-8e7f-7e575b8e2008-operator-scripts") pod "glanceb9b5-account-delete-tjrnr" (UID: "bcaec0f3-13af-4f73-8e7f-7e575b8e2008") : configmap "openstack-scripts" not found Nov 25 09:48:17 crc kubenswrapper[4776]: I1125 09:48:17.393847 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wvl56" Nov 25 09:48:17 crc kubenswrapper[4776]: I1125 09:48:17.441325 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wvl56"] Nov 25 09:48:17 crc kubenswrapper[4776]: I1125 09:48:17.818686 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:48:17 crc kubenswrapper[4776]: I1125 09:48:17.818748 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:48:19 crc kubenswrapper[4776]: I1125 09:48:19.369928 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wvl56" podUID="41c7c636-6457-4a7d-b0da-30c590b643c7" containerName="registry-server" containerID="cri-o://f47954c53fabd913eea349a21f67f1dd6e34eacf30d0b802dec2287bfa0ff4fe" gracePeriod=2 Nov 25 09:48:19 crc kubenswrapper[4776]: I1125 09:48:19.746141 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wvl56" Nov 25 09:48:19 crc kubenswrapper[4776]: I1125 09:48:19.853875 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41c7c636-6457-4a7d-b0da-30c590b643c7-catalog-content\") pod \"41c7c636-6457-4a7d-b0da-30c590b643c7\" (UID: \"41c7c636-6457-4a7d-b0da-30c590b643c7\") " Nov 25 09:48:19 crc kubenswrapper[4776]: I1125 09:48:19.853952 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkkz5\" (UniqueName: \"kubernetes.io/projected/41c7c636-6457-4a7d-b0da-30c590b643c7-kube-api-access-hkkz5\") pod \"41c7c636-6457-4a7d-b0da-30c590b643c7\" (UID: \"41c7c636-6457-4a7d-b0da-30c590b643c7\") " Nov 25 09:48:19 crc kubenswrapper[4776]: I1125 09:48:19.854037 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41c7c636-6457-4a7d-b0da-30c590b643c7-utilities\") pod \"41c7c636-6457-4a7d-b0da-30c590b643c7\" (UID: \"41c7c636-6457-4a7d-b0da-30c590b643c7\") " Nov 25 09:48:19 crc kubenswrapper[4776]: I1125 09:48:19.856825 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41c7c636-6457-4a7d-b0da-30c590b643c7-utilities" (OuterVolumeSpecName: "utilities") pod "41c7c636-6457-4a7d-b0da-30c590b643c7" (UID: "41c7c636-6457-4a7d-b0da-30c590b643c7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:19 crc kubenswrapper[4776]: I1125 09:48:19.860943 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41c7c636-6457-4a7d-b0da-30c590b643c7-kube-api-access-hkkz5" (OuterVolumeSpecName: "kube-api-access-hkkz5") pod "41c7c636-6457-4a7d-b0da-30c590b643c7" (UID: "41c7c636-6457-4a7d-b0da-30c590b643c7"). InnerVolumeSpecName "kube-api-access-hkkz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:19 crc kubenswrapper[4776]: I1125 09:48:19.874629 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41c7c636-6457-4a7d-b0da-30c590b643c7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "41c7c636-6457-4a7d-b0da-30c590b643c7" (UID: "41c7c636-6457-4a7d-b0da-30c590b643c7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:19 crc kubenswrapper[4776]: I1125 09:48:19.955850 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41c7c636-6457-4a7d-b0da-30c590b643c7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:19 crc kubenswrapper[4776]: I1125 09:48:19.955883 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkkz5\" (UniqueName: \"kubernetes.io/projected/41c7c636-6457-4a7d-b0da-30c590b643c7-kube-api-access-hkkz5\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:19 crc kubenswrapper[4776]: I1125 09:48:19.955904 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41c7c636-6457-4a7d-b0da-30c590b643c7-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:20 crc kubenswrapper[4776]: I1125 09:48:20.379096 4776 generic.go:334] "Generic (PLEG): container finished" podID="41c7c636-6457-4a7d-b0da-30c590b643c7" containerID="f47954c53fabd913eea349a21f67f1dd6e34eacf30d0b802dec2287bfa0ff4fe" exitCode=0 Nov 25 09:48:20 crc kubenswrapper[4776]: I1125 09:48:20.379139 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvl56" event={"ID":"41c7c636-6457-4a7d-b0da-30c590b643c7","Type":"ContainerDied","Data":"f47954c53fabd913eea349a21f67f1dd6e34eacf30d0b802dec2287bfa0ff4fe"} Nov 25 09:48:20 crc kubenswrapper[4776]: I1125 09:48:20.379149 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wvl56" Nov 25 09:48:20 crc kubenswrapper[4776]: I1125 09:48:20.379164 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvl56" event={"ID":"41c7c636-6457-4a7d-b0da-30c590b643c7","Type":"ContainerDied","Data":"fa9a325b0b1fb67af9753d6386d3c0fdbbdb14c12ade7e3f72ada430e72baf58"} Nov 25 09:48:20 crc kubenswrapper[4776]: I1125 09:48:20.379182 4776 scope.go:117] "RemoveContainer" containerID="f47954c53fabd913eea349a21f67f1dd6e34eacf30d0b802dec2287bfa0ff4fe" Nov 25 09:48:20 crc kubenswrapper[4776]: I1125 09:48:20.406253 4776 scope.go:117] "RemoveContainer" containerID="b0a376848e32ca571b595940521d8008deb7cfaab3ad42b2be71790ec63e8a01" Nov 25 09:48:20 crc kubenswrapper[4776]: I1125 09:48:20.417346 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wvl56"] Nov 25 09:48:20 crc kubenswrapper[4776]: I1125 09:48:20.424786 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wvl56"] Nov 25 09:48:20 crc kubenswrapper[4776]: I1125 09:48:20.431517 4776 scope.go:117] "RemoveContainer" containerID="ae84c5b4d9bf124138f331d12ea865041ecf356b53cae9f743c7d06f073a5270" Nov 25 09:48:20 crc kubenswrapper[4776]: I1125 09:48:20.460941 4776 scope.go:117] "RemoveContainer" containerID="f47954c53fabd913eea349a21f67f1dd6e34eacf30d0b802dec2287bfa0ff4fe" Nov 25 09:48:20 crc kubenswrapper[4776]: E1125 09:48:20.461502 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f47954c53fabd913eea349a21f67f1dd6e34eacf30d0b802dec2287bfa0ff4fe\": container with ID starting with f47954c53fabd913eea349a21f67f1dd6e34eacf30d0b802dec2287bfa0ff4fe not found: ID does not exist" containerID="f47954c53fabd913eea349a21f67f1dd6e34eacf30d0b802dec2287bfa0ff4fe" Nov 25 09:48:20 crc kubenswrapper[4776]: I1125 09:48:20.461541 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f47954c53fabd913eea349a21f67f1dd6e34eacf30d0b802dec2287bfa0ff4fe"} err="failed to get container status \"f47954c53fabd913eea349a21f67f1dd6e34eacf30d0b802dec2287bfa0ff4fe\": rpc error: code = NotFound desc = could not find container \"f47954c53fabd913eea349a21f67f1dd6e34eacf30d0b802dec2287bfa0ff4fe\": container with ID starting with f47954c53fabd913eea349a21f67f1dd6e34eacf30d0b802dec2287bfa0ff4fe not found: ID does not exist" Nov 25 09:48:20 crc kubenswrapper[4776]: I1125 09:48:20.461593 4776 scope.go:117] "RemoveContainer" containerID="b0a376848e32ca571b595940521d8008deb7cfaab3ad42b2be71790ec63e8a01" Nov 25 09:48:20 crc kubenswrapper[4776]: E1125 09:48:20.462023 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0a376848e32ca571b595940521d8008deb7cfaab3ad42b2be71790ec63e8a01\": container with ID starting with b0a376848e32ca571b595940521d8008deb7cfaab3ad42b2be71790ec63e8a01 not found: ID does not exist" containerID="b0a376848e32ca571b595940521d8008deb7cfaab3ad42b2be71790ec63e8a01" Nov 25 09:48:20 crc kubenswrapper[4776]: I1125 09:48:20.462097 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0a376848e32ca571b595940521d8008deb7cfaab3ad42b2be71790ec63e8a01"} err="failed to get container status \"b0a376848e32ca571b595940521d8008deb7cfaab3ad42b2be71790ec63e8a01\": rpc error: code = NotFound desc = could not find container \"b0a376848e32ca571b595940521d8008deb7cfaab3ad42b2be71790ec63e8a01\": container with ID starting with b0a376848e32ca571b595940521d8008deb7cfaab3ad42b2be71790ec63e8a01 not found: ID does not exist" Nov 25 09:48:20 crc kubenswrapper[4776]: I1125 09:48:20.462124 4776 scope.go:117] "RemoveContainer" containerID="ae84c5b4d9bf124138f331d12ea865041ecf356b53cae9f743c7d06f073a5270" Nov 25 09:48:20 crc kubenswrapper[4776]: E1125 09:48:20.462498 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae84c5b4d9bf124138f331d12ea865041ecf356b53cae9f743c7d06f073a5270\": container with ID starting with ae84c5b4d9bf124138f331d12ea865041ecf356b53cae9f743c7d06f073a5270 not found: ID does not exist" containerID="ae84c5b4d9bf124138f331d12ea865041ecf356b53cae9f743c7d06f073a5270" Nov 25 09:48:20 crc kubenswrapper[4776]: I1125 09:48:20.462533 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae84c5b4d9bf124138f331d12ea865041ecf356b53cae9f743c7d06f073a5270"} err="failed to get container status \"ae84c5b4d9bf124138f331d12ea865041ecf356b53cae9f743c7d06f073a5270\": rpc error: code = NotFound desc = could not find container \"ae84c5b4d9bf124138f331d12ea865041ecf356b53cae9f743c7d06f073a5270\": container with ID starting with ae84c5b4d9bf124138f331d12ea865041ecf356b53cae9f743c7d06f073a5270 not found: ID does not exist" Nov 25 09:48:20 crc kubenswrapper[4776]: E1125 09:48:20.628437 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06 is running failed: container process not found" containerID="eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 25 09:48:20 crc kubenswrapper[4776]: E1125 09:48:20.629031 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06 is running failed: container process not found" containerID="eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 25 09:48:20 crc kubenswrapper[4776]: E1125 09:48:20.629403 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06 is running failed: container process not found" containerID="eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 25 09:48:20 crc kubenswrapper[4776]: E1125 09:48:20.629472 4776 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-7fgjf" podUID="10fe3dc2-2bde-4380-8ad3-d71832e31095" containerName="ovsdb-server" Nov 25 09:48:20 crc kubenswrapper[4776]: E1125 09:48:20.630644 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5002006f906f0b17afbf25b23648688a842975cfd3b385d57e3191040607fb57" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 25 09:48:20 crc kubenswrapper[4776]: E1125 09:48:20.632053 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5002006f906f0b17afbf25b23648688a842975cfd3b385d57e3191040607fb57" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 25 09:48:20 crc kubenswrapper[4776]: E1125 09:48:20.633374 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5002006f906f0b17afbf25b23648688a842975cfd3b385d57e3191040607fb57" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 25 09:48:20 crc kubenswrapper[4776]: E1125 09:48:20.633456 4776 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-7fgjf" podUID="10fe3dc2-2bde-4380-8ad3-d71832e31095" containerName="ovs-vswitchd" Nov 25 09:48:21 crc kubenswrapper[4776]: I1125 09:48:21.672583 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41c7c636-6457-4a7d-b0da-30c590b643c7" path="/var/lib/kubelet/pods/41c7c636-6457-4a7d-b0da-30c590b643c7/volumes" Nov 25 09:48:25 crc kubenswrapper[4776]: E1125 09:48:25.627991 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06 is running failed: container process not found" containerID="eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 25 09:48:25 crc kubenswrapper[4776]: E1125 09:48:25.628912 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06 is running failed: container process not found" containerID="eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 25 09:48:25 crc kubenswrapper[4776]: E1125 09:48:25.629378 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06 is running failed: container process not found" containerID="eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Nov 25 09:48:25 crc kubenswrapper[4776]: E1125 09:48:25.629426 4776 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-7fgjf" podUID="10fe3dc2-2bde-4380-8ad3-d71832e31095" containerName="ovsdb-server" Nov 25 09:48:25 crc kubenswrapper[4776]: E1125 09:48:25.629821 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5002006f906f0b17afbf25b23648688a842975cfd3b385d57e3191040607fb57" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 25 09:48:25 crc kubenswrapper[4776]: E1125 09:48:25.632531 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5002006f906f0b17afbf25b23648688a842975cfd3b385d57e3191040607fb57" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 25 09:48:25 crc kubenswrapper[4776]: E1125 09:48:25.635170 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="5002006f906f0b17afbf25b23648688a842975cfd3b385d57e3191040607fb57" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Nov 25 09:48:25 crc kubenswrapper[4776]: E1125 09:48:25.635375 4776 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-7fgjf" podUID="10fe3dc2-2bde-4380-8ad3-d71832e31095" containerName="ovs-vswitchd" Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.454818 4776 generic.go:334] "Generic (PLEG): container finished" podID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerID="faa52779c1f67996100c644db4f91149b383bf2ef821c114ac1e92b3b697987f" exitCode=137 Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.454889 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerDied","Data":"faa52779c1f67996100c644db4f91149b383bf2ef821c114ac1e92b3b697987f"} Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.457787 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7fgjf_10fe3dc2-2bde-4380-8ad3-d71832e31095/ovs-vswitchd/0.log" Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.458592 4776 generic.go:334] "Generic (PLEG): container finished" podID="10fe3dc2-2bde-4380-8ad3-d71832e31095" containerID="5002006f906f0b17afbf25b23648688a842975cfd3b385d57e3191040607fb57" exitCode=137 Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.458638 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7fgjf" event={"ID":"10fe3dc2-2bde-4380-8ad3-d71832e31095","Type":"ContainerDied","Data":"5002006f906f0b17afbf25b23648688a842975cfd3b385d57e3191040607fb57"} Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.642110 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7fgjf_10fe3dc2-2bde-4380-8ad3-d71832e31095/ovs-vswitchd/0.log" Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.643476 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-7fgjf" Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.696378 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.759638 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/10fe3dc2-2bde-4380-8ad3-d71832e31095-var-log\") pod \"10fe3dc2-2bde-4380-8ad3-d71832e31095\" (UID: \"10fe3dc2-2bde-4380-8ad3-d71832e31095\") " Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.759742 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/10fe3dc2-2bde-4380-8ad3-d71832e31095-var-run\") pod \"10fe3dc2-2bde-4380-8ad3-d71832e31095\" (UID: \"10fe3dc2-2bde-4380-8ad3-d71832e31095\") " Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.759791 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10fe3dc2-2bde-4380-8ad3-d71832e31095-scripts\") pod \"10fe3dc2-2bde-4380-8ad3-d71832e31095\" (UID: \"10fe3dc2-2bde-4380-8ad3-d71832e31095\") " Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.759894 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/10fe3dc2-2bde-4380-8ad3-d71832e31095-var-lib\") pod \"10fe3dc2-2bde-4380-8ad3-d71832e31095\" (UID: \"10fe3dc2-2bde-4380-8ad3-d71832e31095\") " Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.759960 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwjwq\" (UniqueName: \"kubernetes.io/projected/10fe3dc2-2bde-4380-8ad3-d71832e31095-kube-api-access-cwjwq\") pod \"10fe3dc2-2bde-4380-8ad3-d71832e31095\" (UID: \"10fe3dc2-2bde-4380-8ad3-d71832e31095\") " Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.759996 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/10fe3dc2-2bde-4380-8ad3-d71832e31095-etc-ovs\") pod \"10fe3dc2-2bde-4380-8ad3-d71832e31095\" (UID: \"10fe3dc2-2bde-4380-8ad3-d71832e31095\") " Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.760096 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10fe3dc2-2bde-4380-8ad3-d71832e31095-var-run" (OuterVolumeSpecName: "var-run") pod "10fe3dc2-2bde-4380-8ad3-d71832e31095" (UID: "10fe3dc2-2bde-4380-8ad3-d71832e31095"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.760137 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10fe3dc2-2bde-4380-8ad3-d71832e31095-var-lib" (OuterVolumeSpecName: "var-lib") pod "10fe3dc2-2bde-4380-8ad3-d71832e31095" (UID: "10fe3dc2-2bde-4380-8ad3-d71832e31095"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.760189 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10fe3dc2-2bde-4380-8ad3-d71832e31095-var-log" (OuterVolumeSpecName: "var-log") pod "10fe3dc2-2bde-4380-8ad3-d71832e31095" (UID: "10fe3dc2-2bde-4380-8ad3-d71832e31095"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.760295 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10fe3dc2-2bde-4380-8ad3-d71832e31095-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "10fe3dc2-2bde-4380-8ad3-d71832e31095" (UID: "10fe3dc2-2bde-4380-8ad3-d71832e31095"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.760518 4776 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/10fe3dc2-2bde-4380-8ad3-d71832e31095-var-lib\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.760555 4776 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/10fe3dc2-2bde-4380-8ad3-d71832e31095-etc-ovs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.760568 4776 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/10fe3dc2-2bde-4380-8ad3-d71832e31095-var-log\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.760576 4776 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/10fe3dc2-2bde-4380-8ad3-d71832e31095-var-run\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.761514 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10fe3dc2-2bde-4380-8ad3-d71832e31095-scripts" (OuterVolumeSpecName: "scripts") pod "10fe3dc2-2bde-4380-8ad3-d71832e31095" (UID: "10fe3dc2-2bde-4380-8ad3-d71832e31095"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.766043 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10fe3dc2-2bde-4380-8ad3-d71832e31095-kube-api-access-cwjwq" (OuterVolumeSpecName: "kube-api-access-cwjwq") pod "10fe3dc2-2bde-4380-8ad3-d71832e31095" (UID: "10fe3dc2-2bde-4380-8ad3-d71832e31095"). InnerVolumeSpecName "kube-api-access-cwjwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.861871 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6c49z\" (UniqueName: \"kubernetes.io/projected/ebd843d5-4a0b-4660-9636-fc19d8fddec6-kube-api-access-6c49z\") pod \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\" (UID: \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\") " Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.862011 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ebd843d5-4a0b-4660-9636-fc19d8fddec6-etc-swift\") pod \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\" (UID: \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\") " Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.862100 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/ebd843d5-4a0b-4660-9636-fc19d8fddec6-lock\") pod \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\" (UID: \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\") " Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.862141 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\" (UID: \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\") " Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.862174 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/ebd843d5-4a0b-4660-9636-fc19d8fddec6-cache\") pod \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\" (UID: \"ebd843d5-4a0b-4660-9636-fc19d8fddec6\") " Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.862551 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebd843d5-4a0b-4660-9636-fc19d8fddec6-lock" (OuterVolumeSpecName: "lock") pod "ebd843d5-4a0b-4660-9636-fc19d8fddec6" (UID: "ebd843d5-4a0b-4660-9636-fc19d8fddec6"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.863231 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwjwq\" (UniqueName: \"kubernetes.io/projected/10fe3dc2-2bde-4380-8ad3-d71832e31095-kube-api-access-cwjwq\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.863252 4776 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/ebd843d5-4a0b-4660-9636-fc19d8fddec6-lock\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.863263 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10fe3dc2-2bde-4380-8ad3-d71832e31095-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.863443 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebd843d5-4a0b-4660-9636-fc19d8fddec6-cache" (OuterVolumeSpecName: "cache") pod "ebd843d5-4a0b-4660-9636-fc19d8fddec6" (UID: "ebd843d5-4a0b-4660-9636-fc19d8fddec6"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.865552 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "swift") pod "ebd843d5-4a0b-4660-9636-fc19d8fddec6" (UID: "ebd843d5-4a0b-4660-9636-fc19d8fddec6"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.865743 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebd843d5-4a0b-4660-9636-fc19d8fddec6-kube-api-access-6c49z" (OuterVolumeSpecName: "kube-api-access-6c49z") pod "ebd843d5-4a0b-4660-9636-fc19d8fddec6" (UID: "ebd843d5-4a0b-4660-9636-fc19d8fddec6"). InnerVolumeSpecName "kube-api-access-6c49z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.865794 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebd843d5-4a0b-4660-9636-fc19d8fddec6-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "ebd843d5-4a0b-4660-9636-fc19d8fddec6" (UID: "ebd843d5-4a0b-4660-9636-fc19d8fddec6"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.965345 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6c49z\" (UniqueName: \"kubernetes.io/projected/ebd843d5-4a0b-4660-9636-fc19d8fddec6-kube-api-access-6c49z\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.965385 4776 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ebd843d5-4a0b-4660-9636-fc19d8fddec6-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.965420 4776 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.965432 4776 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/ebd843d5-4a0b-4660-9636-fc19d8fddec6-cache\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:26 crc kubenswrapper[4776]: I1125 09:48:26.982306 4776 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.066895 4776 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.475190 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"ebd843d5-4a0b-4660-9636-fc19d8fddec6","Type":"ContainerDied","Data":"8d9f1bd809faa88090af0396f6842af5c4e67d29000a1f49bbc2554cee1a30ea"} Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.475272 4776 scope.go:117] "RemoveContainer" containerID="faa52779c1f67996100c644db4f91149b383bf2ef821c114ac1e92b3b697987f" Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.475279 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.479106 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7fgjf" event={"ID":"10fe3dc2-2bde-4380-8ad3-d71832e31095","Type":"ContainerDied","Data":"39cdf7b2812d7e93af5c98ff938ad95f0c62191b3c2d0b2154912df748de511b"} Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.479165 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-7fgjf" Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.496658 4776 scope.go:117] "RemoveContainer" containerID="152026f7bf460a5258df8a4110443a6e694c32ee892ab2dcb9469e4391545baa" Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.523080 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.523584 4776 scope.go:117] "RemoveContainer" containerID="8e84b91c7719175c8908ad8fa6987a175f46744663fe91972c8bbb602ae871d8" Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.531332 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.537420 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-7fgjf"] Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.543484 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-7fgjf"] Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.545944 4776 scope.go:117] "RemoveContainer" containerID="49bb08566118dc9e19698c8156598960f4778c7e09f832e3eeeaa9b6d4fc9a62" Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.562752 4776 scope.go:117] "RemoveContainer" containerID="670e0e074022b9b3b944ddaeb78954a059e4c4b8e2f7af4e2f91000daddc1263" Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.582665 4776 scope.go:117] "RemoveContainer" containerID="a5e5c6ad114abc906cc7868aa09dc3d1a6a83ffb494d14de45deba5107b01b02" Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.600436 4776 scope.go:117] "RemoveContainer" containerID="2c713a54dd52d79687c4437832031c9afbb02bd373076897a51654ae360ff22a" Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.621582 4776 scope.go:117] "RemoveContainer" containerID="f0cdb71ed47fdaaa7eabd7f81c93a293616c6c022e1c4d8cef28b88de0dfb8d5" Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.639173 4776 scope.go:117] "RemoveContainer" containerID="434eabc5a25e34093e38beb759273c9fac1b8758c3a03273e4ddb5f564fcdc08" Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.654960 4776 scope.go:117] "RemoveContainer" containerID="a4e815cd4ca77d4505a242b624a83fcda104ea6e85674eb4c15ddc064562ffd6" Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.671485 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10fe3dc2-2bde-4380-8ad3-d71832e31095" path="/var/lib/kubelet/pods/10fe3dc2-2bde-4380-8ad3-d71832e31095/volumes" Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.672384 4776 scope.go:117] "RemoveContainer" containerID="e2072d3b3f2b295702755794df1a3f5a2a130db157825a15a28d4ac1c3a1c674" Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.672576 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" path="/var/lib/kubelet/pods/ebd843d5-4a0b-4660-9636-fc19d8fddec6/volumes" Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.691130 4776 scope.go:117] "RemoveContainer" containerID="cd073bb8b2d7b566c99b6e569d6dbfa6b16b08d5db2e84e47a8d2d85eb563f1b" Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.708159 4776 scope.go:117] "RemoveContainer" containerID="6ccd151880ed401c0372e024b6b953cdc0eba4d9d9e8cdd47435d059001cbc30" Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.725128 4776 scope.go:117] "RemoveContainer" containerID="05b65f4bed8c7914cf657df0f540c0d19593a2c84c93c9ae6c4f106530470ec4" Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.741320 4776 scope.go:117] "RemoveContainer" containerID="9e7cf1b3a65acbcaaedff95080720317c88059b1b6b759f4434c871973a12bc3" Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.759623 4776 scope.go:117] "RemoveContainer" containerID="5002006f906f0b17afbf25b23648688a842975cfd3b385d57e3191040607fb57" Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.775765 4776 scope.go:117] "RemoveContainer" containerID="eb1dc820fa1786b91dbf97123d997e7441420bfb2fb9a786cae8b336a0be1d06" Nov 25 09:48:27 crc kubenswrapper[4776]: I1125 09:48:27.792363 4776 scope.go:117] "RemoveContainer" containerID="b43aaf9ab18cf2548b29395d03bbfa83bdd75486872c6906cf0d6deac19222b4" Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.418524 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi3b46-account-delete-9w5r9" Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.519825 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder774d-account-delete-b29n5" Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.525664 4776 generic.go:334] "Generic (PLEG): container finished" podID="14b51bba-0b15-469b-bdea-90fae8d9b976" containerID="d3de2925044530f7fd248214ce715db43c6f7a123e71e4f5fd11e8af0b5cda0e" exitCode=137 Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.525721 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi3b46-account-delete-9w5r9" event={"ID":"14b51bba-0b15-469b-bdea-90fae8d9b976","Type":"ContainerDied","Data":"d3de2925044530f7fd248214ce715db43c6f7a123e71e4f5fd11e8af0b5cda0e"} Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.525750 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapi3b46-account-delete-9w5r9" event={"ID":"14b51bba-0b15-469b-bdea-90fae8d9b976","Type":"ContainerDied","Data":"5d66e9e059cc4ecbbbdd54624c88073c9a95bd315ce5454ae88b21b09a0588b0"} Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.525768 4776 scope.go:117] "RemoveContainer" containerID="d3de2925044530f7fd248214ce715db43c6f7a123e71e4f5fd11e8af0b5cda0e" Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.525879 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapi3b46-account-delete-9w5r9" Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.528017 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron5151-account-delete-8vbjg" Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.528182 4776 generic.go:334] "Generic (PLEG): container finished" podID="ef85f90f-c377-4273-9fff-451420d6008a" containerID="02f077bb80ad2e06d150c6b44d4bfbe70a57d6d488294c69cf9fd29d74fce9a8" exitCode=137 Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.528237 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder774d-account-delete-b29n5" event={"ID":"ef85f90f-c377-4273-9fff-451420d6008a","Type":"ContainerDied","Data":"02f077bb80ad2e06d150c6b44d4bfbe70a57d6d488294c69cf9fd29d74fce9a8"} Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.528258 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder774d-account-delete-b29n5" event={"ID":"ef85f90f-c377-4273-9fff-451420d6008a","Type":"ContainerDied","Data":"5e4df52a7d69593f81f35d33034b747cbf372fe6bdfff6a79400c8d265865396"} Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.528297 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder774d-account-delete-b29n5" Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.537250 4776 generic.go:334] "Generic (PLEG): container finished" podID="73f2c2ab-6cb6-4544-a3f7-20722e5b6400" containerID="f44277236f6a634e8f611e825fd05aabebb686cac5979f168a7f50210ff4ef17" exitCode=137 Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.537315 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron5151-account-delete-8vbjg" event={"ID":"73f2c2ab-6cb6-4544-a3f7-20722e5b6400","Type":"ContainerDied","Data":"f44277236f6a634e8f611e825fd05aabebb686cac5979f168a7f50210ff4ef17"} Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.537348 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron5151-account-delete-8vbjg" event={"ID":"73f2c2ab-6cb6-4544-a3f7-20722e5b6400","Type":"ContainerDied","Data":"b773b52fed0b22bef9171161cc75302f0462620c1fb47cb91658f91785551dc1"} Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.537419 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron5151-account-delete-8vbjg" Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.554085 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/14b51bba-0b15-469b-bdea-90fae8d9b976-operator-scripts\") pod \"14b51bba-0b15-469b-bdea-90fae8d9b976\" (UID: \"14b51bba-0b15-469b-bdea-90fae8d9b976\") " Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.555027 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcs9z\" (UniqueName: \"kubernetes.io/projected/14b51bba-0b15-469b-bdea-90fae8d9b976-kube-api-access-gcs9z\") pod \"14b51bba-0b15-469b-bdea-90fae8d9b976\" (UID: \"14b51bba-0b15-469b-bdea-90fae8d9b976\") " Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.555325 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14b51bba-0b15-469b-bdea-90fae8d9b976-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "14b51bba-0b15-469b-bdea-90fae8d9b976" (UID: "14b51bba-0b15-469b-bdea-90fae8d9b976"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.556088 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/14b51bba-0b15-469b-bdea-90fae8d9b976-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.559784 4776 scope.go:117] "RemoveContainer" containerID="d3de2925044530f7fd248214ce715db43c6f7a123e71e4f5fd11e8af0b5cda0e" Nov 25 09:48:32 crc kubenswrapper[4776]: E1125 09:48:32.560215 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3de2925044530f7fd248214ce715db43c6f7a123e71e4f5fd11e8af0b5cda0e\": container with ID starting with d3de2925044530f7fd248214ce715db43c6f7a123e71e4f5fd11e8af0b5cda0e not found: ID does not exist" containerID="d3de2925044530f7fd248214ce715db43c6f7a123e71e4f5fd11e8af0b5cda0e" Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.560254 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3de2925044530f7fd248214ce715db43c6f7a123e71e4f5fd11e8af0b5cda0e"} err="failed to get container status \"d3de2925044530f7fd248214ce715db43c6f7a123e71e4f5fd11e8af0b5cda0e\": rpc error: code = NotFound desc = could not find container \"d3de2925044530f7fd248214ce715db43c6f7a123e71e4f5fd11e8af0b5cda0e\": container with ID starting with d3de2925044530f7fd248214ce715db43c6f7a123e71e4f5fd11e8af0b5cda0e not found: ID does not exist" Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.560278 4776 scope.go:117] "RemoveContainer" containerID="02f077bb80ad2e06d150c6b44d4bfbe70a57d6d488294c69cf9fd29d74fce9a8" Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.560384 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14b51bba-0b15-469b-bdea-90fae8d9b976-kube-api-access-gcs9z" (OuterVolumeSpecName: "kube-api-access-gcs9z") pod "14b51bba-0b15-469b-bdea-90fae8d9b976" (UID: "14b51bba-0b15-469b-bdea-90fae8d9b976"). InnerVolumeSpecName "kube-api-access-gcs9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.583290 4776 scope.go:117] "RemoveContainer" containerID="02f077bb80ad2e06d150c6b44d4bfbe70a57d6d488294c69cf9fd29d74fce9a8" Nov 25 09:48:32 crc kubenswrapper[4776]: E1125 09:48:32.583718 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02f077bb80ad2e06d150c6b44d4bfbe70a57d6d488294c69cf9fd29d74fce9a8\": container with ID starting with 02f077bb80ad2e06d150c6b44d4bfbe70a57d6d488294c69cf9fd29d74fce9a8 not found: ID does not exist" containerID="02f077bb80ad2e06d150c6b44d4bfbe70a57d6d488294c69cf9fd29d74fce9a8" Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.583830 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02f077bb80ad2e06d150c6b44d4bfbe70a57d6d488294c69cf9fd29d74fce9a8"} err="failed to get container status \"02f077bb80ad2e06d150c6b44d4bfbe70a57d6d488294c69cf9fd29d74fce9a8\": rpc error: code = NotFound desc = could not find container \"02f077bb80ad2e06d150c6b44d4bfbe70a57d6d488294c69cf9fd29d74fce9a8\": container with ID starting with 02f077bb80ad2e06d150c6b44d4bfbe70a57d6d488294c69cf9fd29d74fce9a8 not found: ID does not exist" Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.583905 4776 scope.go:117] "RemoveContainer" containerID="f44277236f6a634e8f611e825fd05aabebb686cac5979f168a7f50210ff4ef17" Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.597909 4776 scope.go:117] "RemoveContainer" containerID="f44277236f6a634e8f611e825fd05aabebb686cac5979f168a7f50210ff4ef17" Nov 25 09:48:32 crc kubenswrapper[4776]: E1125 09:48:32.598382 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f44277236f6a634e8f611e825fd05aabebb686cac5979f168a7f50210ff4ef17\": container with ID starting with f44277236f6a634e8f611e825fd05aabebb686cac5979f168a7f50210ff4ef17 not found: ID does not exist" containerID="f44277236f6a634e8f611e825fd05aabebb686cac5979f168a7f50210ff4ef17" Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.598415 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f44277236f6a634e8f611e825fd05aabebb686cac5979f168a7f50210ff4ef17"} err="failed to get container status \"f44277236f6a634e8f611e825fd05aabebb686cac5979f168a7f50210ff4ef17\": rpc error: code = NotFound desc = could not find container \"f44277236f6a634e8f611e825fd05aabebb686cac5979f168a7f50210ff4ef17\": container with ID starting with f44277236f6a634e8f611e825fd05aabebb686cac5979f168a7f50210ff4ef17 not found: ID does not exist" Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.657184 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqcbj\" (UniqueName: \"kubernetes.io/projected/ef85f90f-c377-4273-9fff-451420d6008a-kube-api-access-gqcbj\") pod \"ef85f90f-c377-4273-9fff-451420d6008a\" (UID: \"ef85f90f-c377-4273-9fff-451420d6008a\") " Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.657281 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef85f90f-c377-4273-9fff-451420d6008a-operator-scripts\") pod \"ef85f90f-c377-4273-9fff-451420d6008a\" (UID: \"ef85f90f-c377-4273-9fff-451420d6008a\") " Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.657345 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73f2c2ab-6cb6-4544-a3f7-20722e5b6400-operator-scripts\") pod \"73f2c2ab-6cb6-4544-a3f7-20722e5b6400\" (UID: \"73f2c2ab-6cb6-4544-a3f7-20722e5b6400\") " Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.657975 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef85f90f-c377-4273-9fff-451420d6008a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ef85f90f-c377-4273-9fff-451420d6008a" (UID: "ef85f90f-c377-4273-9fff-451420d6008a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.658098 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92frh\" (UniqueName: \"kubernetes.io/projected/73f2c2ab-6cb6-4544-a3f7-20722e5b6400-kube-api-access-92frh\") pod \"73f2c2ab-6cb6-4544-a3f7-20722e5b6400\" (UID: \"73f2c2ab-6cb6-4544-a3f7-20722e5b6400\") " Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.658095 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73f2c2ab-6cb6-4544-a3f7-20722e5b6400-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "73f2c2ab-6cb6-4544-a3f7-20722e5b6400" (UID: "73f2c2ab-6cb6-4544-a3f7-20722e5b6400"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:32 crc kubenswrapper[4776]: E1125 09:48:32.658478 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:32 crc kubenswrapper[4776]: E1125 09:48:32.658548 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6be099e6-89d6-45a0-9d0f-4dca4d91701a-operator-scripts podName:6be099e6-89d6-45a0-9d0f-4dca4d91701a nodeName:}" failed. No retries permitted until 2025-11-25 09:49:04.658530368 +0000 UTC m=+1489.699589981 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/6be099e6-89d6-45a0-9d0f-4dca4d91701a-operator-scripts") pod "placemente176-account-delete-wbgz5" (UID: "6be099e6-89d6-45a0-9d0f-4dca4d91701a") : configmap "openstack-scripts" not found Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.658570 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73f2c2ab-6cb6-4544-a3f7-20722e5b6400-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:32 crc kubenswrapper[4776]: E1125 09:48:32.658575 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.658586 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcs9z\" (UniqueName: \"kubernetes.io/projected/14b51bba-0b15-469b-bdea-90fae8d9b976-kube-api-access-gcs9z\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.658601 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ef85f90f-c377-4273-9fff-451420d6008a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:32 crc kubenswrapper[4776]: E1125 09:48:32.658610 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/946ea713-bfba-449f-82ab-28e915469938-operator-scripts podName:946ea713-bfba-449f-82ab-28e915469938 nodeName:}" failed. No retries permitted until 2025-11-25 09:49:04.658599209 +0000 UTC m=+1489.699658762 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/946ea713-bfba-449f-82ab-28e915469938-operator-scripts") pod "novacell09dc4-account-delete-w5xcj" (UID: "946ea713-bfba-449f-82ab-28e915469938") : configmap "openstack-scripts" not found Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.660551 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef85f90f-c377-4273-9fff-451420d6008a-kube-api-access-gqcbj" (OuterVolumeSpecName: "kube-api-access-gqcbj") pod "ef85f90f-c377-4273-9fff-451420d6008a" (UID: "ef85f90f-c377-4273-9fff-451420d6008a"). InnerVolumeSpecName "kube-api-access-gqcbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.660689 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73f2c2ab-6cb6-4544-a3f7-20722e5b6400-kube-api-access-92frh" (OuterVolumeSpecName: "kube-api-access-92frh") pod "73f2c2ab-6cb6-4544-a3f7-20722e5b6400" (UID: "73f2c2ab-6cb6-4544-a3f7-20722e5b6400"). InnerVolumeSpecName "kube-api-access-92frh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.760039 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqcbj\" (UniqueName: \"kubernetes.io/projected/ef85f90f-c377-4273-9fff-451420d6008a-kube-api-access-gqcbj\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.760089 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92frh\" (UniqueName: \"kubernetes.io/projected/73f2c2ab-6cb6-4544-a3f7-20722e5b6400-kube-api-access-92frh\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:32 crc kubenswrapper[4776]: E1125 09:48:32.760171 4776 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Nov 25 09:48:32 crc kubenswrapper[4776]: E1125 09:48:32.760270 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/bcaec0f3-13af-4f73-8e7f-7e575b8e2008-operator-scripts podName:bcaec0f3-13af-4f73-8e7f-7e575b8e2008 nodeName:}" failed. No retries permitted until 2025-11-25 09:49:04.760247345 +0000 UTC m=+1489.801306948 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/bcaec0f3-13af-4f73-8e7f-7e575b8e2008-operator-scripts") pod "glanceb9b5-account-delete-tjrnr" (UID: "bcaec0f3-13af-4f73-8e7f-7e575b8e2008") : configmap "openstack-scripts" not found Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.881008 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapi3b46-account-delete-9w5r9"] Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.892245 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapi3b46-account-delete-9w5r9"] Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.902155 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder774d-account-delete-b29n5"] Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.907429 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder774d-account-delete-b29n5"] Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.912764 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron5151-account-delete-8vbjg"] Nov 25 09:48:32 crc kubenswrapper[4776]: I1125 09:48:32.918294 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron5151-account-delete-8vbjg"] Nov 25 09:48:33 crc kubenswrapper[4776]: I1125 09:48:33.547036 4776 generic.go:334] "Generic (PLEG): container finished" podID="946ea713-bfba-449f-82ab-28e915469938" containerID="ac0deb2776fd44dcbddf8b4d4bae71247d519e589f78423415504e15ea5a4a0d" exitCode=137 Nov 25 09:48:33 crc kubenswrapper[4776]: I1125 09:48:33.547141 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell09dc4-account-delete-w5xcj" event={"ID":"946ea713-bfba-449f-82ab-28e915469938","Type":"ContainerDied","Data":"ac0deb2776fd44dcbddf8b4d4bae71247d519e589f78423415504e15ea5a4a0d"} Nov 25 09:48:33 crc kubenswrapper[4776]: I1125 09:48:33.675629 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14b51bba-0b15-469b-bdea-90fae8d9b976" path="/var/lib/kubelet/pods/14b51bba-0b15-469b-bdea-90fae8d9b976/volumes" Nov 25 09:48:33 crc kubenswrapper[4776]: I1125 09:48:33.676452 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73f2c2ab-6cb6-4544-a3f7-20722e5b6400" path="/var/lib/kubelet/pods/73f2c2ab-6cb6-4544-a3f7-20722e5b6400/volumes" Nov 25 09:48:33 crc kubenswrapper[4776]: I1125 09:48:33.677126 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef85f90f-c377-4273-9fff-451420d6008a" path="/var/lib/kubelet/pods/ef85f90f-c377-4273-9fff-451420d6008a/volumes" Nov 25 09:48:33 crc kubenswrapper[4776]: I1125 09:48:33.882716 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell09dc4-account-delete-w5xcj" Nov 25 09:48:33 crc kubenswrapper[4776]: I1125 09:48:33.981923 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59dkv\" (UniqueName: \"kubernetes.io/projected/946ea713-bfba-449f-82ab-28e915469938-kube-api-access-59dkv\") pod \"946ea713-bfba-449f-82ab-28e915469938\" (UID: \"946ea713-bfba-449f-82ab-28e915469938\") " Nov 25 09:48:33 crc kubenswrapper[4776]: I1125 09:48:33.981979 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/946ea713-bfba-449f-82ab-28e915469938-operator-scripts\") pod \"946ea713-bfba-449f-82ab-28e915469938\" (UID: \"946ea713-bfba-449f-82ab-28e915469938\") " Nov 25 09:48:33 crc kubenswrapper[4776]: I1125 09:48:33.983116 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/946ea713-bfba-449f-82ab-28e915469938-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "946ea713-bfba-449f-82ab-28e915469938" (UID: "946ea713-bfba-449f-82ab-28e915469938"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:33 crc kubenswrapper[4776]: I1125 09:48:33.986491 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/946ea713-bfba-449f-82ab-28e915469938-kube-api-access-59dkv" (OuterVolumeSpecName: "kube-api-access-59dkv") pod "946ea713-bfba-449f-82ab-28e915469938" (UID: "946ea713-bfba-449f-82ab-28e915469938"). InnerVolumeSpecName "kube-api-access-59dkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:34 crc kubenswrapper[4776]: I1125 09:48:34.083145 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59dkv\" (UniqueName: \"kubernetes.io/projected/946ea713-bfba-449f-82ab-28e915469938-kube-api-access-59dkv\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:34 crc kubenswrapper[4776]: I1125 09:48:34.083186 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/946ea713-bfba-449f-82ab-28e915469938-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:34 crc kubenswrapper[4776]: I1125 09:48:34.560105 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell09dc4-account-delete-w5xcj" event={"ID":"946ea713-bfba-449f-82ab-28e915469938","Type":"ContainerDied","Data":"3426e4cb115f29a90f341756fdee32ff5ebae07b8b55d943e28a698f73f1eb3d"} Nov 25 09:48:34 crc kubenswrapper[4776]: I1125 09:48:34.560179 4776 scope.go:117] "RemoveContainer" containerID="ac0deb2776fd44dcbddf8b4d4bae71247d519e589f78423415504e15ea5a4a0d" Nov 25 09:48:34 crc kubenswrapper[4776]: I1125 09:48:34.560311 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell09dc4-account-delete-w5xcj" Nov 25 09:48:34 crc kubenswrapper[4776]: I1125 09:48:34.589512 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell09dc4-account-delete-w5xcj"] Nov 25 09:48:34 crc kubenswrapper[4776]: I1125 09:48:34.594880 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell09dc4-account-delete-w5xcj"] Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.147754 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placemente176-account-delete-wbgz5" Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.203472 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glanceb9b5-account-delete-tjrnr" Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.300080 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wb29h\" (UniqueName: \"kubernetes.io/projected/6be099e6-89d6-45a0-9d0f-4dca4d91701a-kube-api-access-wb29h\") pod \"6be099e6-89d6-45a0-9d0f-4dca4d91701a\" (UID: \"6be099e6-89d6-45a0-9d0f-4dca4d91701a\") " Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.300242 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6be099e6-89d6-45a0-9d0f-4dca4d91701a-operator-scripts\") pod \"6be099e6-89d6-45a0-9d0f-4dca4d91701a\" (UID: \"6be099e6-89d6-45a0-9d0f-4dca4d91701a\") " Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.300967 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6be099e6-89d6-45a0-9d0f-4dca4d91701a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6be099e6-89d6-45a0-9d0f-4dca4d91701a" (UID: "6be099e6-89d6-45a0-9d0f-4dca4d91701a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.305140 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6be099e6-89d6-45a0-9d0f-4dca4d91701a-kube-api-access-wb29h" (OuterVolumeSpecName: "kube-api-access-wb29h") pod "6be099e6-89d6-45a0-9d0f-4dca4d91701a" (UID: "6be099e6-89d6-45a0-9d0f-4dca4d91701a"). InnerVolumeSpecName "kube-api-access-wb29h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.402602 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfmhc\" (UniqueName: \"kubernetes.io/projected/bcaec0f3-13af-4f73-8e7f-7e575b8e2008-kube-api-access-vfmhc\") pod \"bcaec0f3-13af-4f73-8e7f-7e575b8e2008\" (UID: \"bcaec0f3-13af-4f73-8e7f-7e575b8e2008\") " Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.403006 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcaec0f3-13af-4f73-8e7f-7e575b8e2008-operator-scripts\") pod \"bcaec0f3-13af-4f73-8e7f-7e575b8e2008\" (UID: \"bcaec0f3-13af-4f73-8e7f-7e575b8e2008\") " Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.403479 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6be099e6-89d6-45a0-9d0f-4dca4d91701a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.403599 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wb29h\" (UniqueName: \"kubernetes.io/projected/6be099e6-89d6-45a0-9d0f-4dca4d91701a-kube-api-access-wb29h\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.403673 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcaec0f3-13af-4f73-8e7f-7e575b8e2008-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bcaec0f3-13af-4f73-8e7f-7e575b8e2008" (UID: "bcaec0f3-13af-4f73-8e7f-7e575b8e2008"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.405511 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcaec0f3-13af-4f73-8e7f-7e575b8e2008-kube-api-access-vfmhc" (OuterVolumeSpecName: "kube-api-access-vfmhc") pod "bcaec0f3-13af-4f73-8e7f-7e575b8e2008" (UID: "bcaec0f3-13af-4f73-8e7f-7e575b8e2008"). InnerVolumeSpecName "kube-api-access-vfmhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.504954 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfmhc\" (UniqueName: \"kubernetes.io/projected/bcaec0f3-13af-4f73-8e7f-7e575b8e2008-kube-api-access-vfmhc\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.505001 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcaec0f3-13af-4f73-8e7f-7e575b8e2008-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.571995 4776 generic.go:334] "Generic (PLEG): container finished" podID="bcaec0f3-13af-4f73-8e7f-7e575b8e2008" containerID="4a42c4c2fcc9fc1a6adb53fc048a0c22c503282432d6521db1b327054c0ea919" exitCode=137 Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.572084 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glanceb9b5-account-delete-tjrnr" Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.572130 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glanceb9b5-account-delete-tjrnr" event={"ID":"bcaec0f3-13af-4f73-8e7f-7e575b8e2008","Type":"ContainerDied","Data":"4a42c4c2fcc9fc1a6adb53fc048a0c22c503282432d6521db1b327054c0ea919"} Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.572218 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glanceb9b5-account-delete-tjrnr" event={"ID":"bcaec0f3-13af-4f73-8e7f-7e575b8e2008","Type":"ContainerDied","Data":"abe37879060260786a380d4efa643725c294e581dc2808c91e62da1b2dd37969"} Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.572247 4776 scope.go:117] "RemoveContainer" containerID="4a42c4c2fcc9fc1a6adb53fc048a0c22c503282432d6521db1b327054c0ea919" Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.576349 4776 generic.go:334] "Generic (PLEG): container finished" podID="6be099e6-89d6-45a0-9d0f-4dca4d91701a" containerID="52647c9b509f4a2e0420df93981105d258326ab2d8003e1be47968ae44e82e11" exitCode=137 Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.576425 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placemente176-account-delete-wbgz5" event={"ID":"6be099e6-89d6-45a0-9d0f-4dca4d91701a","Type":"ContainerDied","Data":"52647c9b509f4a2e0420df93981105d258326ab2d8003e1be47968ae44e82e11"} Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.576465 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placemente176-account-delete-wbgz5" event={"ID":"6be099e6-89d6-45a0-9d0f-4dca4d91701a","Type":"ContainerDied","Data":"d7ddbba5ee21f93a892892dc7ab2ec9efca17632ed5b4c898a4b773f4b5b6b2c"} Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.576370 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placemente176-account-delete-wbgz5" Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.595679 4776 scope.go:117] "RemoveContainer" containerID="4a42c4c2fcc9fc1a6adb53fc048a0c22c503282432d6521db1b327054c0ea919" Nov 25 09:48:35 crc kubenswrapper[4776]: E1125 09:48:35.596156 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a42c4c2fcc9fc1a6adb53fc048a0c22c503282432d6521db1b327054c0ea919\": container with ID starting with 4a42c4c2fcc9fc1a6adb53fc048a0c22c503282432d6521db1b327054c0ea919 not found: ID does not exist" containerID="4a42c4c2fcc9fc1a6adb53fc048a0c22c503282432d6521db1b327054c0ea919" Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.596190 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a42c4c2fcc9fc1a6adb53fc048a0c22c503282432d6521db1b327054c0ea919"} err="failed to get container status \"4a42c4c2fcc9fc1a6adb53fc048a0c22c503282432d6521db1b327054c0ea919\": rpc error: code = NotFound desc = could not find container \"4a42c4c2fcc9fc1a6adb53fc048a0c22c503282432d6521db1b327054c0ea919\": container with ID starting with 4a42c4c2fcc9fc1a6adb53fc048a0c22c503282432d6521db1b327054c0ea919 not found: ID does not exist" Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.596215 4776 scope.go:117] "RemoveContainer" containerID="52647c9b509f4a2e0420df93981105d258326ab2d8003e1be47968ae44e82e11" Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.609609 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placemente176-account-delete-wbgz5"] Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.617613 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placemente176-account-delete-wbgz5"] Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.621390 4776 scope.go:117] "RemoveContainer" containerID="52647c9b509f4a2e0420df93981105d258326ab2d8003e1be47968ae44e82e11" Nov 25 09:48:35 crc kubenswrapper[4776]: E1125 09:48:35.621967 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52647c9b509f4a2e0420df93981105d258326ab2d8003e1be47968ae44e82e11\": container with ID starting with 52647c9b509f4a2e0420df93981105d258326ab2d8003e1be47968ae44e82e11 not found: ID does not exist" containerID="52647c9b509f4a2e0420df93981105d258326ab2d8003e1be47968ae44e82e11" Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.622004 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52647c9b509f4a2e0420df93981105d258326ab2d8003e1be47968ae44e82e11"} err="failed to get container status \"52647c9b509f4a2e0420df93981105d258326ab2d8003e1be47968ae44e82e11\": rpc error: code = NotFound desc = could not find container \"52647c9b509f4a2e0420df93981105d258326ab2d8003e1be47968ae44e82e11\": container with ID starting with 52647c9b509f4a2e0420df93981105d258326ab2d8003e1be47968ae44e82e11 not found: ID does not exist" Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.622718 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glanceb9b5-account-delete-tjrnr"] Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.627006 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glanceb9b5-account-delete-tjrnr"] Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.669610 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6be099e6-89d6-45a0-9d0f-4dca4d91701a" path="/var/lib/kubelet/pods/6be099e6-89d6-45a0-9d0f-4dca4d91701a/volumes" Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.670093 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="946ea713-bfba-449f-82ab-28e915469938" path="/var/lib/kubelet/pods/946ea713-bfba-449f-82ab-28e915469938/volumes" Nov 25 09:48:35 crc kubenswrapper[4776]: I1125 09:48:35.670997 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcaec0f3-13af-4f73-8e7f-7e575b8e2008" path="/var/lib/kubelet/pods/bcaec0f3-13af-4f73-8e7f-7e575b8e2008/volumes" Nov 25 09:48:47 crc kubenswrapper[4776]: I1125 09:48:47.818699 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:48:47 crc kubenswrapper[4776]: I1125 09:48:47.819298 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:48:47 crc kubenswrapper[4776]: I1125 09:48:47.819698 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 09:48:47 crc kubenswrapper[4776]: I1125 09:48:47.820465 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"41bf411c5d6ef6679dd0fb3b86628f86f08b8ccef2b894d89b25148005eb3847"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 09:48:47 crc kubenswrapper[4776]: I1125 09:48:47.820534 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://41bf411c5d6ef6679dd0fb3b86628f86f08b8ccef2b894d89b25148005eb3847" gracePeriod=600 Nov 25 09:48:48 crc kubenswrapper[4776]: I1125 09:48:48.682262 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="41bf411c5d6ef6679dd0fb3b86628f86f08b8ccef2b894d89b25148005eb3847" exitCode=0 Nov 25 09:48:48 crc kubenswrapper[4776]: I1125 09:48:48.682356 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"41bf411c5d6ef6679dd0fb3b86628f86f08b8ccef2b894d89b25148005eb3847"} Nov 25 09:48:48 crc kubenswrapper[4776]: I1125 09:48:48.682997 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722"} Nov 25 09:48:48 crc kubenswrapper[4776]: I1125 09:48:48.683036 4776 scope.go:117] "RemoveContainer" containerID="bbf70cb3f31b21dd9b2d356f05e12d08a7b4f65263789554ac47b2758bc077be" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.096108 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-c9bhx"] Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.097892 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="account-auditor" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.097965 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="account-auditor" Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.098040 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe3ff67b-01ff-480f-be91-7c8235593c97" containerName="neutron-api" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.098116 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe3ff67b-01ff-480f-be91-7c8235593c97" containerName="neutron-api" Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.098187 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10fe3dc2-2bde-4380-8ad3-d71832e31095" containerName="ovs-vswitchd" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.098242 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="10fe3dc2-2bde-4380-8ad3-d71832e31095" containerName="ovs-vswitchd" Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.098299 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef85f90f-c377-4273-9fff-451420d6008a" containerName="mariadb-account-delete" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.098352 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef85f90f-c377-4273-9fff-451420d6008a" containerName="mariadb-account-delete" Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.098409 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10fe3dc2-2bde-4380-8ad3-d71832e31095" containerName="ovsdb-server" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.098457 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="10fe3dc2-2bde-4380-8ad3-d71832e31095" containerName="ovsdb-server" Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.098516 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="account-server" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.098572 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="account-server" Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.098659 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="object-updater" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.098731 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="object-updater" Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.098791 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="container-replicator" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.098875 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="container-replicator" Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.098943 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10fe3dc2-2bde-4380-8ad3-d71832e31095" containerName="ovsdb-server-init" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.098997 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="10fe3dc2-2bde-4380-8ad3-d71832e31095" containerName="ovsdb-server-init" Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.099051 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41c7c636-6457-4a7d-b0da-30c590b643c7" containerName="registry-server" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.099147 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="41c7c636-6457-4a7d-b0da-30c590b643c7" containerName="registry-server" Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.099268 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="object-replicator" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.099337 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="object-replicator" Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.099403 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41c7c636-6457-4a7d-b0da-30c590b643c7" containerName="extract-utilities" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.099478 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="41c7c636-6457-4a7d-b0da-30c590b643c7" containerName="extract-utilities" Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.099548 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="object-auditor" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.099616 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="object-auditor" Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.099689 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="account-replicator" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.099756 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="account-replicator" Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.099829 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="container-auditor" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.099909 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="container-auditor" Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.099997 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="container-updater" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.100089 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="container-updater" Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.100178 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="object-server" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.100249 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="object-server" Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.100326 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe3ff67b-01ff-480f-be91-7c8235593c97" containerName="neutron-httpd" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.100400 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe3ff67b-01ff-480f-be91-7c8235593c97" containerName="neutron-httpd" Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.100469 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="object-expirer" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.100535 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="object-expirer" Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.100610 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="rsync" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.100676 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="rsync" Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.100750 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73f2c2ab-6cb6-4544-a3f7-20722e5b6400" containerName="mariadb-account-delete" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.100817 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="73f2c2ab-6cb6-4544-a3f7-20722e5b6400" containerName="mariadb-account-delete" Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.100910 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="account-reaper" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.100976 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="account-reaper" Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.101042 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6be099e6-89d6-45a0-9d0f-4dca4d91701a" containerName="mariadb-account-delete" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.101127 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="6be099e6-89d6-45a0-9d0f-4dca4d91701a" containerName="mariadb-account-delete" Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.101202 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="swift-recon-cron" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.101266 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="swift-recon-cron" Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.101340 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcaec0f3-13af-4f73-8e7f-7e575b8e2008" containerName="mariadb-account-delete" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.101403 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcaec0f3-13af-4f73-8e7f-7e575b8e2008" containerName="mariadb-account-delete" Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.101469 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="946ea713-bfba-449f-82ab-28e915469938" containerName="mariadb-account-delete" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.101532 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="946ea713-bfba-449f-82ab-28e915469938" containerName="mariadb-account-delete" Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.101602 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41c7c636-6457-4a7d-b0da-30c590b643c7" containerName="extract-content" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.101682 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="41c7c636-6457-4a7d-b0da-30c590b643c7" containerName="extract-content" Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.101756 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="container-server" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.101820 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="container-server" Nov 25 09:48:55 crc kubenswrapper[4776]: E1125 09:48:55.101892 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14b51bba-0b15-469b-bdea-90fae8d9b976" containerName="mariadb-account-delete" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.101955 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="14b51bba-0b15-469b-bdea-90fae8d9b976" containerName="mariadb-account-delete" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.102225 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="object-auditor" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.102309 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe3ff67b-01ff-480f-be91-7c8235593c97" containerName="neutron-api" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.102382 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="73f2c2ab-6cb6-4544-a3f7-20722e5b6400" containerName="mariadb-account-delete" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.102445 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="container-auditor" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.102510 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="object-expirer" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.102575 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="14b51bba-0b15-469b-bdea-90fae8d9b976" containerName="mariadb-account-delete" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.102638 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef85f90f-c377-4273-9fff-451420d6008a" containerName="mariadb-account-delete" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.102708 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="swift-recon-cron" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.103907 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="rsync" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.104013 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="object-updater" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.104099 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="container-replicator" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.104184 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="container-server" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.104255 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="10fe3dc2-2bde-4380-8ad3-d71832e31095" containerName="ovs-vswitchd" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.104326 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="account-auditor" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.104393 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="6be099e6-89d6-45a0-9d0f-4dca4d91701a" containerName="mariadb-account-delete" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.104475 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="container-updater" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.104545 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="account-server" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.104613 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="account-replicator" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.104684 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe3ff67b-01ff-480f-be91-7c8235593c97" containerName="neutron-httpd" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.104751 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcaec0f3-13af-4f73-8e7f-7e575b8e2008" containerName="mariadb-account-delete" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.104822 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="account-reaper" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.104896 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="41c7c636-6457-4a7d-b0da-30c590b643c7" containerName="registry-server" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.104965 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="object-replicator" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.105034 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebd843d5-4a0b-4660-9636-fc19d8fddec6" containerName="object-server" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.105132 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="10fe3dc2-2bde-4380-8ad3-d71832e31095" containerName="ovsdb-server" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.105210 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="946ea713-bfba-449f-82ab-28e915469938" containerName="mariadb-account-delete" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.106612 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c9bhx" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.114408 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c9bhx"] Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.194553 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084-catalog-content\") pod \"certified-operators-c9bhx\" (UID: \"d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084\") " pod="openshift-marketplace/certified-operators-c9bhx" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.194618 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084-utilities\") pod \"certified-operators-c9bhx\" (UID: \"d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084\") " pod="openshift-marketplace/certified-operators-c9bhx" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.194722 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k5lj\" (UniqueName: \"kubernetes.io/projected/d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084-kube-api-access-7k5lj\") pod \"certified-operators-c9bhx\" (UID: \"d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084\") " pod="openshift-marketplace/certified-operators-c9bhx" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.295754 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084-catalog-content\") pod \"certified-operators-c9bhx\" (UID: \"d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084\") " pod="openshift-marketplace/certified-operators-c9bhx" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.295820 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084-utilities\") pod \"certified-operators-c9bhx\" (UID: \"d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084\") " pod="openshift-marketplace/certified-operators-c9bhx" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.295889 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k5lj\" (UniqueName: \"kubernetes.io/projected/d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084-kube-api-access-7k5lj\") pod \"certified-operators-c9bhx\" (UID: \"d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084\") " pod="openshift-marketplace/certified-operators-c9bhx" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.296415 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084-catalog-content\") pod \"certified-operators-c9bhx\" (UID: \"d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084\") " pod="openshift-marketplace/certified-operators-c9bhx" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.296466 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084-utilities\") pod \"certified-operators-c9bhx\" (UID: \"d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084\") " pod="openshift-marketplace/certified-operators-c9bhx" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.326734 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k5lj\" (UniqueName: \"kubernetes.io/projected/d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084-kube-api-access-7k5lj\") pod \"certified-operators-c9bhx\" (UID: \"d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084\") " pod="openshift-marketplace/certified-operators-c9bhx" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.429605 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c9bhx" Nov 25 09:48:55 crc kubenswrapper[4776]: I1125 09:48:55.900723 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c9bhx"] Nov 25 09:48:56 crc kubenswrapper[4776]: I1125 09:48:56.756991 4776 generic.go:334] "Generic (PLEG): container finished" podID="d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084" containerID="2c75bb7b99d80b4a7281402dc4d81d6b75cc447066f674251e617617e0d1f347" exitCode=0 Nov 25 09:48:56 crc kubenswrapper[4776]: I1125 09:48:56.757041 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c9bhx" event={"ID":"d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084","Type":"ContainerDied","Data":"2c75bb7b99d80b4a7281402dc4d81d6b75cc447066f674251e617617e0d1f347"} Nov 25 09:48:56 crc kubenswrapper[4776]: I1125 09:48:56.757092 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c9bhx" event={"ID":"d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084","Type":"ContainerStarted","Data":"e3bdf786717dadb79a78a96b42a347a2211646791c30d887ee52ffc43a41e9c3"} Nov 25 09:48:57 crc kubenswrapper[4776]: I1125 09:48:57.768757 4776 generic.go:334] "Generic (PLEG): container finished" podID="d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084" containerID="ccf56fbb35bfb3a1943ea55a80f21eeaa61925a00cf964146fb30315c6dc6a4e" exitCode=0 Nov 25 09:48:57 crc kubenswrapper[4776]: I1125 09:48:57.768821 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c9bhx" event={"ID":"d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084","Type":"ContainerDied","Data":"ccf56fbb35bfb3a1943ea55a80f21eeaa61925a00cf964146fb30315c6dc6a4e"} Nov 25 09:48:58 crc kubenswrapper[4776]: I1125 09:48:58.778208 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c9bhx" event={"ID":"d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084","Type":"ContainerStarted","Data":"cda5a83dd03b22e9a697fc50dba2231fc1f446c9d0d747c7014a1ae380fc6cd6"} Nov 25 09:48:58 crc kubenswrapper[4776]: I1125 09:48:58.797020 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-c9bhx" podStartSLOduration=2.301199351 podStartE2EDuration="3.797003501s" podCreationTimestamp="2025-11-25 09:48:55 +0000 UTC" firstStartedPulling="2025-11-25 09:48:56.760941616 +0000 UTC m=+1481.802001169" lastFinishedPulling="2025-11-25 09:48:58.256745766 +0000 UTC m=+1483.297805319" observedRunningTime="2025-11-25 09:48:58.79220855 +0000 UTC m=+1483.833268123" watchObservedRunningTime="2025-11-25 09:48:58.797003501 +0000 UTC m=+1483.838063054" Nov 25 09:49:05 crc kubenswrapper[4776]: I1125 09:49:05.430187 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-c9bhx" Nov 25 09:49:05 crc kubenswrapper[4776]: I1125 09:49:05.430689 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-c9bhx" Nov 25 09:49:05 crc kubenswrapper[4776]: I1125 09:49:05.472058 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-c9bhx" Nov 25 09:49:05 crc kubenswrapper[4776]: I1125 09:49:05.879783 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-c9bhx" Nov 25 09:49:05 crc kubenswrapper[4776]: I1125 09:49:05.935755 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c9bhx"] Nov 25 09:49:07 crc kubenswrapper[4776]: I1125 09:49:07.850413 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-c9bhx" podUID="d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084" containerName="registry-server" containerID="cri-o://cda5a83dd03b22e9a697fc50dba2231fc1f446c9d0d747c7014a1ae380fc6cd6" gracePeriod=2 Nov 25 09:49:08 crc kubenswrapper[4776]: I1125 09:49:08.892806 4776 generic.go:334] "Generic (PLEG): container finished" podID="d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084" containerID="cda5a83dd03b22e9a697fc50dba2231fc1f446c9d0d747c7014a1ae380fc6cd6" exitCode=0 Nov 25 09:49:08 crc kubenswrapper[4776]: I1125 09:49:08.893168 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c9bhx" event={"ID":"d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084","Type":"ContainerDied","Data":"cda5a83dd03b22e9a697fc50dba2231fc1f446c9d0d747c7014a1ae380fc6cd6"} Nov 25 09:49:08 crc kubenswrapper[4776]: I1125 09:49:08.952801 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c9bhx" Nov 25 09:49:08 crc kubenswrapper[4776]: I1125 09:49:08.990343 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084-utilities\") pod \"d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084\" (UID: \"d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084\") " Nov 25 09:49:08 crc kubenswrapper[4776]: I1125 09:49:08.990422 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7k5lj\" (UniqueName: \"kubernetes.io/projected/d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084-kube-api-access-7k5lj\") pod \"d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084\" (UID: \"d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084\") " Nov 25 09:49:08 crc kubenswrapper[4776]: I1125 09:49:08.990491 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084-catalog-content\") pod \"d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084\" (UID: \"d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084\") " Nov 25 09:49:08 crc kubenswrapper[4776]: I1125 09:49:08.999869 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084-utilities" (OuterVolumeSpecName: "utilities") pod "d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084" (UID: "d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:49:09 crc kubenswrapper[4776]: I1125 09:49:09.005276 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084-kube-api-access-7k5lj" (OuterVolumeSpecName: "kube-api-access-7k5lj") pod "d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084" (UID: "d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084"). InnerVolumeSpecName "kube-api-access-7k5lj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:49:09 crc kubenswrapper[4776]: I1125 09:49:09.042277 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084" (UID: "d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:49:09 crc kubenswrapper[4776]: I1125 09:49:09.092330 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:49:09 crc kubenswrapper[4776]: I1125 09:49:09.092365 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7k5lj\" (UniqueName: \"kubernetes.io/projected/d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084-kube-api-access-7k5lj\") on node \"crc\" DevicePath \"\"" Nov 25 09:49:09 crc kubenswrapper[4776]: I1125 09:49:09.092377 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:49:09 crc kubenswrapper[4776]: I1125 09:49:09.904883 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c9bhx" event={"ID":"d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084","Type":"ContainerDied","Data":"e3bdf786717dadb79a78a96b42a347a2211646791c30d887ee52ffc43a41e9c3"} Nov 25 09:49:09 crc kubenswrapper[4776]: I1125 09:49:09.905315 4776 scope.go:117] "RemoveContainer" containerID="cda5a83dd03b22e9a697fc50dba2231fc1f446c9d0d747c7014a1ae380fc6cd6" Nov 25 09:49:09 crc kubenswrapper[4776]: I1125 09:49:09.907277 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c9bhx" Nov 25 09:49:09 crc kubenswrapper[4776]: I1125 09:49:09.934020 4776 scope.go:117] "RemoveContainer" containerID="ccf56fbb35bfb3a1943ea55a80f21eeaa61925a00cf964146fb30315c6dc6a4e" Nov 25 09:49:09 crc kubenswrapper[4776]: I1125 09:49:09.940421 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c9bhx"] Nov 25 09:49:09 crc kubenswrapper[4776]: I1125 09:49:09.947369 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-c9bhx"] Nov 25 09:49:09 crc kubenswrapper[4776]: I1125 09:49:09.961442 4776 scope.go:117] "RemoveContainer" containerID="2c75bb7b99d80b4a7281402dc4d81d6b75cc447066f674251e617617e0d1f347" Nov 25 09:49:11 crc kubenswrapper[4776]: I1125 09:49:11.675306 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084" path="/var/lib/kubelet/pods/d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084/volumes" Nov 25 09:49:17 crc kubenswrapper[4776]: I1125 09:49:17.748297 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2zdtf"] Nov 25 09:49:17 crc kubenswrapper[4776]: E1125 09:49:17.750693 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084" containerName="registry-server" Nov 25 09:49:17 crc kubenswrapper[4776]: I1125 09:49:17.750864 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084" containerName="registry-server" Nov 25 09:49:17 crc kubenswrapper[4776]: E1125 09:49:17.751014 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084" containerName="extract-content" Nov 25 09:49:17 crc kubenswrapper[4776]: I1125 09:49:17.751152 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084" containerName="extract-content" Nov 25 09:49:17 crc kubenswrapper[4776]: E1125 09:49:17.751277 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084" containerName="extract-utilities" Nov 25 09:49:17 crc kubenswrapper[4776]: I1125 09:49:17.751385 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084" containerName="extract-utilities" Nov 25 09:49:17 crc kubenswrapper[4776]: I1125 09:49:17.751700 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6dcc4d8-2ff9-428d-a5b5-ff734c9f5084" containerName="registry-server" Nov 25 09:49:17 crc kubenswrapper[4776]: I1125 09:49:17.753473 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2zdtf" Nov 25 09:49:17 crc kubenswrapper[4776]: I1125 09:49:17.764281 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2zdtf"] Nov 25 09:49:17 crc kubenswrapper[4776]: I1125 09:49:17.809461 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d21a892-5265-4d39-9608-f2df1d7a6efe-catalog-content\") pod \"community-operators-2zdtf\" (UID: \"6d21a892-5265-4d39-9608-f2df1d7a6efe\") " pod="openshift-marketplace/community-operators-2zdtf" Nov 25 09:49:17 crc kubenswrapper[4776]: I1125 09:49:17.809843 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgf4r\" (UniqueName: \"kubernetes.io/projected/6d21a892-5265-4d39-9608-f2df1d7a6efe-kube-api-access-mgf4r\") pod \"community-operators-2zdtf\" (UID: \"6d21a892-5265-4d39-9608-f2df1d7a6efe\") " pod="openshift-marketplace/community-operators-2zdtf" Nov 25 09:49:17 crc kubenswrapper[4776]: I1125 09:49:17.809889 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d21a892-5265-4d39-9608-f2df1d7a6efe-utilities\") pod \"community-operators-2zdtf\" (UID: \"6d21a892-5265-4d39-9608-f2df1d7a6efe\") " pod="openshift-marketplace/community-operators-2zdtf" Nov 25 09:49:17 crc kubenswrapper[4776]: I1125 09:49:17.911571 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d21a892-5265-4d39-9608-f2df1d7a6efe-utilities\") pod \"community-operators-2zdtf\" (UID: \"6d21a892-5265-4d39-9608-f2df1d7a6efe\") " pod="openshift-marketplace/community-operators-2zdtf" Nov 25 09:49:17 crc kubenswrapper[4776]: I1125 09:49:17.911678 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d21a892-5265-4d39-9608-f2df1d7a6efe-catalog-content\") pod \"community-operators-2zdtf\" (UID: \"6d21a892-5265-4d39-9608-f2df1d7a6efe\") " pod="openshift-marketplace/community-operators-2zdtf" Nov 25 09:49:17 crc kubenswrapper[4776]: I1125 09:49:17.911713 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgf4r\" (UniqueName: \"kubernetes.io/projected/6d21a892-5265-4d39-9608-f2df1d7a6efe-kube-api-access-mgf4r\") pod \"community-operators-2zdtf\" (UID: \"6d21a892-5265-4d39-9608-f2df1d7a6efe\") " pod="openshift-marketplace/community-operators-2zdtf" Nov 25 09:49:17 crc kubenswrapper[4776]: I1125 09:49:17.912094 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d21a892-5265-4d39-9608-f2df1d7a6efe-utilities\") pod \"community-operators-2zdtf\" (UID: \"6d21a892-5265-4d39-9608-f2df1d7a6efe\") " pod="openshift-marketplace/community-operators-2zdtf" Nov 25 09:49:17 crc kubenswrapper[4776]: I1125 09:49:17.912181 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d21a892-5265-4d39-9608-f2df1d7a6efe-catalog-content\") pod \"community-operators-2zdtf\" (UID: \"6d21a892-5265-4d39-9608-f2df1d7a6efe\") " pod="openshift-marketplace/community-operators-2zdtf" Nov 25 09:49:17 crc kubenswrapper[4776]: I1125 09:49:17.937247 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgf4r\" (UniqueName: \"kubernetes.io/projected/6d21a892-5265-4d39-9608-f2df1d7a6efe-kube-api-access-mgf4r\") pod \"community-operators-2zdtf\" (UID: \"6d21a892-5265-4d39-9608-f2df1d7a6efe\") " pod="openshift-marketplace/community-operators-2zdtf" Nov 25 09:49:18 crc kubenswrapper[4776]: I1125 09:49:18.071654 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2zdtf" Nov 25 09:49:18 crc kubenswrapper[4776]: I1125 09:49:18.528481 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2zdtf"] Nov 25 09:49:18 crc kubenswrapper[4776]: I1125 09:49:18.986508 4776 generic.go:334] "Generic (PLEG): container finished" podID="6d21a892-5265-4d39-9608-f2df1d7a6efe" containerID="c46ed205d91f4101cfadf2963a35c1ec890662a76dea6eb709a3f027da281ede" exitCode=0 Nov 25 09:49:18 crc kubenswrapper[4776]: I1125 09:49:18.986562 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2zdtf" event={"ID":"6d21a892-5265-4d39-9608-f2df1d7a6efe","Type":"ContainerDied","Data":"c46ed205d91f4101cfadf2963a35c1ec890662a76dea6eb709a3f027da281ede"} Nov 25 09:49:18 crc kubenswrapper[4776]: I1125 09:49:18.986595 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2zdtf" event={"ID":"6d21a892-5265-4d39-9608-f2df1d7a6efe","Type":"ContainerStarted","Data":"5cad6fa0e47ca2e66eb0763eabaef276e07d721bb9e733cc4d48bda8f227848b"} Nov 25 09:49:20 crc kubenswrapper[4776]: I1125 09:49:19.999671 4776 generic.go:334] "Generic (PLEG): container finished" podID="6d21a892-5265-4d39-9608-f2df1d7a6efe" containerID="593748590e7d653aaf5dff8d58a132040532fb1b6d924e9bd02a1b4e6ad0db1f" exitCode=0 Nov 25 09:49:20 crc kubenswrapper[4776]: I1125 09:49:19.999769 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2zdtf" event={"ID":"6d21a892-5265-4d39-9608-f2df1d7a6efe","Type":"ContainerDied","Data":"593748590e7d653aaf5dff8d58a132040532fb1b6d924e9bd02a1b4e6ad0db1f"} Nov 25 09:49:21 crc kubenswrapper[4776]: I1125 09:49:21.013215 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2zdtf" event={"ID":"6d21a892-5265-4d39-9608-f2df1d7a6efe","Type":"ContainerStarted","Data":"5a437d7735baa7b64b54abf1bb274516f93636fca4b44ba303b2f6995d7ba14b"} Nov 25 09:49:21 crc kubenswrapper[4776]: I1125 09:49:21.031193 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2zdtf" podStartSLOduration=2.30854866 podStartE2EDuration="4.031172404s" podCreationTimestamp="2025-11-25 09:49:17 +0000 UTC" firstStartedPulling="2025-11-25 09:49:18.98959504 +0000 UTC m=+1504.030654593" lastFinishedPulling="2025-11-25 09:49:20.712218784 +0000 UTC m=+1505.753278337" observedRunningTime="2025-11-25 09:49:21.029862851 +0000 UTC m=+1506.070922404" watchObservedRunningTime="2025-11-25 09:49:21.031172404 +0000 UTC m=+1506.072231957" Nov 25 09:49:27 crc kubenswrapper[4776]: I1125 09:49:27.441960 4776 scope.go:117] "RemoveContainer" containerID="92b79dbb4ed078714aec0c9d66fe7c33b71b68f0384cad8a81c0c460830f785d" Nov 25 09:49:27 crc kubenswrapper[4776]: I1125 09:49:27.465163 4776 scope.go:117] "RemoveContainer" containerID="f098fe0a1fe35cd7b0310e84ef175144d5d65e426c5ec2e06ed70195987f5c40" Nov 25 09:49:27 crc kubenswrapper[4776]: I1125 09:49:27.492955 4776 scope.go:117] "RemoveContainer" containerID="2bd361fc4287baa9805ecbc3dbe86ea5dc2b96fd42699533192f2b89fd8fd73c" Nov 25 09:49:27 crc kubenswrapper[4776]: I1125 09:49:27.523603 4776 scope.go:117] "RemoveContainer" containerID="3114febb3411856186b21a212016a8980b4f4ed26617cd0c55aadc1ea4fb8615" Nov 25 09:49:27 crc kubenswrapper[4776]: I1125 09:49:27.550636 4776 scope.go:117] "RemoveContainer" containerID="3516bb91cfe2308e98f8eae2924c9343a1bcfca0ad8e712dfbdb8076f26a23b2" Nov 25 09:49:27 crc kubenswrapper[4776]: I1125 09:49:27.584596 4776 scope.go:117] "RemoveContainer" containerID="fb166b129931ff672d247f77f4c3b881346f25b0410d683eb6bb28f72d3af0a4" Nov 25 09:49:27 crc kubenswrapper[4776]: I1125 09:49:27.622682 4776 scope.go:117] "RemoveContainer" containerID="f87ecf56de084525eec7915cd66a47ade4adc2c2476c6ecb8bae1e7236795cc6" Nov 25 09:49:27 crc kubenswrapper[4776]: I1125 09:49:27.642903 4776 scope.go:117] "RemoveContainer" containerID="061b5a0bcf2039c372d93606dd4f4d9c15e175c9e5f3706f9ed15162829b3321" Nov 25 09:49:27 crc kubenswrapper[4776]: I1125 09:49:27.673789 4776 scope.go:117] "RemoveContainer" containerID="356e591a2a72a0074a6ee72d621e09ac39b9f3d69f4d225fdbc0fdceddce7b5a" Nov 25 09:49:27 crc kubenswrapper[4776]: I1125 09:49:27.703219 4776 scope.go:117] "RemoveContainer" containerID="99459662f37648298f22a0f72639de05480bb2132930a4572102745915daeb31" Nov 25 09:49:27 crc kubenswrapper[4776]: I1125 09:49:27.723927 4776 scope.go:117] "RemoveContainer" containerID="c1a79c25e3fcf070aa4efe449ff4827dd403b86266f770d2f2b4441e527d7adf" Nov 25 09:49:27 crc kubenswrapper[4776]: I1125 09:49:27.740654 4776 scope.go:117] "RemoveContainer" containerID="6048c18060333e95c2e0be12fac988647d61e6f042d149f41998c255e6566c91" Nov 25 09:49:27 crc kubenswrapper[4776]: I1125 09:49:27.762641 4776 scope.go:117] "RemoveContainer" containerID="1ee992a9feeeafbc3dad19eff013d51a325f2d13b85873bd8bf629934e586f03" Nov 25 09:49:27 crc kubenswrapper[4776]: I1125 09:49:27.779258 4776 scope.go:117] "RemoveContainer" containerID="0013160d2c55a26010d58e0144d8a89639c5b92e40762472567d0dae6bd635e9" Nov 25 09:49:27 crc kubenswrapper[4776]: I1125 09:49:27.798953 4776 scope.go:117] "RemoveContainer" containerID="aeee9511209e1b516d5a4fa252549d9534596d3569e1ff114ec890eef5d6d49d" Nov 25 09:49:27 crc kubenswrapper[4776]: I1125 09:49:27.817274 4776 scope.go:117] "RemoveContainer" containerID="fe29f80314d191348c18f899e06138d8db881e961c44e2efa613f33717c12d39" Nov 25 09:49:27 crc kubenswrapper[4776]: I1125 09:49:27.835517 4776 scope.go:117] "RemoveContainer" containerID="bbe973c26582167ba2a33fb740db5cdcd63e979ff3085ac9bf82ee71344b4aff" Nov 25 09:49:27 crc kubenswrapper[4776]: I1125 09:49:27.862752 4776 scope.go:117] "RemoveContainer" containerID="fcc2df6898ae37ac42d535f5a9578c9c98363157deb46bb5d8844b5320228fca" Nov 25 09:49:28 crc kubenswrapper[4776]: I1125 09:49:28.072386 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2zdtf" Nov 25 09:49:28 crc kubenswrapper[4776]: I1125 09:49:28.072443 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2zdtf" Nov 25 09:49:28 crc kubenswrapper[4776]: I1125 09:49:28.120718 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2zdtf" Nov 25 09:49:29 crc kubenswrapper[4776]: I1125 09:49:29.129399 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2zdtf" Nov 25 09:49:29 crc kubenswrapper[4776]: I1125 09:49:29.189368 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2zdtf"] Nov 25 09:49:31 crc kubenswrapper[4776]: I1125 09:49:31.104197 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2zdtf" podUID="6d21a892-5265-4d39-9608-f2df1d7a6efe" containerName="registry-server" containerID="cri-o://5a437d7735baa7b64b54abf1bb274516f93636fca4b44ba303b2f6995d7ba14b" gracePeriod=2 Nov 25 09:49:31 crc kubenswrapper[4776]: I1125 09:49:31.488813 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2zdtf" Nov 25 09:49:31 crc kubenswrapper[4776]: I1125 09:49:31.505518 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d21a892-5265-4d39-9608-f2df1d7a6efe-catalog-content\") pod \"6d21a892-5265-4d39-9608-f2df1d7a6efe\" (UID: \"6d21a892-5265-4d39-9608-f2df1d7a6efe\") " Nov 25 09:49:31 crc kubenswrapper[4776]: I1125 09:49:31.505590 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d21a892-5265-4d39-9608-f2df1d7a6efe-utilities\") pod \"6d21a892-5265-4d39-9608-f2df1d7a6efe\" (UID: \"6d21a892-5265-4d39-9608-f2df1d7a6efe\") " Nov 25 09:49:31 crc kubenswrapper[4776]: I1125 09:49:31.505677 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgf4r\" (UniqueName: \"kubernetes.io/projected/6d21a892-5265-4d39-9608-f2df1d7a6efe-kube-api-access-mgf4r\") pod \"6d21a892-5265-4d39-9608-f2df1d7a6efe\" (UID: \"6d21a892-5265-4d39-9608-f2df1d7a6efe\") " Nov 25 09:49:31 crc kubenswrapper[4776]: I1125 09:49:31.506627 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d21a892-5265-4d39-9608-f2df1d7a6efe-utilities" (OuterVolumeSpecName: "utilities") pod "6d21a892-5265-4d39-9608-f2df1d7a6efe" (UID: "6d21a892-5265-4d39-9608-f2df1d7a6efe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:49:31 crc kubenswrapper[4776]: I1125 09:49:31.526263 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d21a892-5265-4d39-9608-f2df1d7a6efe-kube-api-access-mgf4r" (OuterVolumeSpecName: "kube-api-access-mgf4r") pod "6d21a892-5265-4d39-9608-f2df1d7a6efe" (UID: "6d21a892-5265-4d39-9608-f2df1d7a6efe"). InnerVolumeSpecName "kube-api-access-mgf4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:49:31 crc kubenswrapper[4776]: I1125 09:49:31.607373 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d21a892-5265-4d39-9608-f2df1d7a6efe-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:49:31 crc kubenswrapper[4776]: I1125 09:49:31.607428 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgf4r\" (UniqueName: \"kubernetes.io/projected/6d21a892-5265-4d39-9608-f2df1d7a6efe-kube-api-access-mgf4r\") on node \"crc\" DevicePath \"\"" Nov 25 09:49:31 crc kubenswrapper[4776]: I1125 09:49:31.608221 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d21a892-5265-4d39-9608-f2df1d7a6efe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6d21a892-5265-4d39-9608-f2df1d7a6efe" (UID: "6d21a892-5265-4d39-9608-f2df1d7a6efe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:49:31 crc kubenswrapper[4776]: I1125 09:49:31.708413 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d21a892-5265-4d39-9608-f2df1d7a6efe-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:49:32 crc kubenswrapper[4776]: I1125 09:49:32.113926 4776 generic.go:334] "Generic (PLEG): container finished" podID="6d21a892-5265-4d39-9608-f2df1d7a6efe" containerID="5a437d7735baa7b64b54abf1bb274516f93636fca4b44ba303b2f6995d7ba14b" exitCode=0 Nov 25 09:49:32 crc kubenswrapper[4776]: I1125 09:49:32.113966 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2zdtf" event={"ID":"6d21a892-5265-4d39-9608-f2df1d7a6efe","Type":"ContainerDied","Data":"5a437d7735baa7b64b54abf1bb274516f93636fca4b44ba303b2f6995d7ba14b"} Nov 25 09:49:32 crc kubenswrapper[4776]: I1125 09:49:32.113984 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2zdtf" Nov 25 09:49:32 crc kubenswrapper[4776]: I1125 09:49:32.113993 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2zdtf" event={"ID":"6d21a892-5265-4d39-9608-f2df1d7a6efe","Type":"ContainerDied","Data":"5cad6fa0e47ca2e66eb0763eabaef276e07d721bb9e733cc4d48bda8f227848b"} Nov 25 09:49:32 crc kubenswrapper[4776]: I1125 09:49:32.114007 4776 scope.go:117] "RemoveContainer" containerID="5a437d7735baa7b64b54abf1bb274516f93636fca4b44ba303b2f6995d7ba14b" Nov 25 09:49:32 crc kubenswrapper[4776]: I1125 09:49:32.138320 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2zdtf"] Nov 25 09:49:32 crc kubenswrapper[4776]: I1125 09:49:32.138478 4776 scope.go:117] "RemoveContainer" containerID="593748590e7d653aaf5dff8d58a132040532fb1b6d924e9bd02a1b4e6ad0db1f" Nov 25 09:49:32 crc kubenswrapper[4776]: I1125 09:49:32.144169 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2zdtf"] Nov 25 09:49:32 crc kubenswrapper[4776]: I1125 09:49:32.158654 4776 scope.go:117] "RemoveContainer" containerID="c46ed205d91f4101cfadf2963a35c1ec890662a76dea6eb709a3f027da281ede" Nov 25 09:49:32 crc kubenswrapper[4776]: I1125 09:49:32.179514 4776 scope.go:117] "RemoveContainer" containerID="5a437d7735baa7b64b54abf1bb274516f93636fca4b44ba303b2f6995d7ba14b" Nov 25 09:49:32 crc kubenswrapper[4776]: E1125 09:49:32.180056 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a437d7735baa7b64b54abf1bb274516f93636fca4b44ba303b2f6995d7ba14b\": container with ID starting with 5a437d7735baa7b64b54abf1bb274516f93636fca4b44ba303b2f6995d7ba14b not found: ID does not exist" containerID="5a437d7735baa7b64b54abf1bb274516f93636fca4b44ba303b2f6995d7ba14b" Nov 25 09:49:32 crc kubenswrapper[4776]: I1125 09:49:32.180311 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a437d7735baa7b64b54abf1bb274516f93636fca4b44ba303b2f6995d7ba14b"} err="failed to get container status \"5a437d7735baa7b64b54abf1bb274516f93636fca4b44ba303b2f6995d7ba14b\": rpc error: code = NotFound desc = could not find container \"5a437d7735baa7b64b54abf1bb274516f93636fca4b44ba303b2f6995d7ba14b\": container with ID starting with 5a437d7735baa7b64b54abf1bb274516f93636fca4b44ba303b2f6995d7ba14b not found: ID does not exist" Nov 25 09:49:32 crc kubenswrapper[4776]: I1125 09:49:32.180338 4776 scope.go:117] "RemoveContainer" containerID="593748590e7d653aaf5dff8d58a132040532fb1b6d924e9bd02a1b4e6ad0db1f" Nov 25 09:49:32 crc kubenswrapper[4776]: E1125 09:49:32.180595 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"593748590e7d653aaf5dff8d58a132040532fb1b6d924e9bd02a1b4e6ad0db1f\": container with ID starting with 593748590e7d653aaf5dff8d58a132040532fb1b6d924e9bd02a1b4e6ad0db1f not found: ID does not exist" containerID="593748590e7d653aaf5dff8d58a132040532fb1b6d924e9bd02a1b4e6ad0db1f" Nov 25 09:49:32 crc kubenswrapper[4776]: I1125 09:49:32.180625 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"593748590e7d653aaf5dff8d58a132040532fb1b6d924e9bd02a1b4e6ad0db1f"} err="failed to get container status \"593748590e7d653aaf5dff8d58a132040532fb1b6d924e9bd02a1b4e6ad0db1f\": rpc error: code = NotFound desc = could not find container \"593748590e7d653aaf5dff8d58a132040532fb1b6d924e9bd02a1b4e6ad0db1f\": container with ID starting with 593748590e7d653aaf5dff8d58a132040532fb1b6d924e9bd02a1b4e6ad0db1f not found: ID does not exist" Nov 25 09:49:32 crc kubenswrapper[4776]: I1125 09:49:32.180699 4776 scope.go:117] "RemoveContainer" containerID="c46ed205d91f4101cfadf2963a35c1ec890662a76dea6eb709a3f027da281ede" Nov 25 09:49:32 crc kubenswrapper[4776]: E1125 09:49:32.181117 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c46ed205d91f4101cfadf2963a35c1ec890662a76dea6eb709a3f027da281ede\": container with ID starting with c46ed205d91f4101cfadf2963a35c1ec890662a76dea6eb709a3f027da281ede not found: ID does not exist" containerID="c46ed205d91f4101cfadf2963a35c1ec890662a76dea6eb709a3f027da281ede" Nov 25 09:49:32 crc kubenswrapper[4776]: I1125 09:49:32.181170 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c46ed205d91f4101cfadf2963a35c1ec890662a76dea6eb709a3f027da281ede"} err="failed to get container status \"c46ed205d91f4101cfadf2963a35c1ec890662a76dea6eb709a3f027da281ede\": rpc error: code = NotFound desc = could not find container \"c46ed205d91f4101cfadf2963a35c1ec890662a76dea6eb709a3f027da281ede\": container with ID starting with c46ed205d91f4101cfadf2963a35c1ec890662a76dea6eb709a3f027da281ede not found: ID does not exist" Nov 25 09:49:33 crc kubenswrapper[4776]: I1125 09:49:33.673828 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d21a892-5265-4d39-9608-f2df1d7a6efe" path="/var/lib/kubelet/pods/6d21a892-5265-4d39-9608-f2df1d7a6efe/volumes" Nov 25 09:50:28 crc kubenswrapper[4776]: I1125 09:50:28.262670 4776 scope.go:117] "RemoveContainer" containerID="4b1650f8058627a45e7f620b18a860f042c250a0cb8ce2b2c139288d15aaa398" Nov 25 09:50:28 crc kubenswrapper[4776]: I1125 09:50:28.299872 4776 scope.go:117] "RemoveContainer" containerID="d35530a8d2f45153bec6fcc975ede0bc99a55fb6d60c6edac19a05f0939799d7" Nov 25 09:50:28 crc kubenswrapper[4776]: I1125 09:50:28.331986 4776 scope.go:117] "RemoveContainer" containerID="5cb968840c90f1277c3bd3ba5789c34ed207afcd5657d28a1ee8b953a59bcbe3" Nov 25 09:50:28 crc kubenswrapper[4776]: I1125 09:50:28.369799 4776 scope.go:117] "RemoveContainer" containerID="1edff0bfa155e1f89f986bf9450cc1b87e57f55c9434668ff95b152144e5b031" Nov 25 09:50:28 crc kubenswrapper[4776]: I1125 09:50:28.396619 4776 scope.go:117] "RemoveContainer" containerID="820a46f0a557e2d6d0e23522585d4acb1b79eaa6244a8621f29bee90ba4a6713" Nov 25 09:50:28 crc kubenswrapper[4776]: I1125 09:50:28.434997 4776 scope.go:117] "RemoveContainer" containerID="f6f2a58bc0a49e61eaaf2dd6aaa0193a12cd24b220d82c3d3115c644b6cac73b" Nov 25 09:50:28 crc kubenswrapper[4776]: I1125 09:50:28.458006 4776 scope.go:117] "RemoveContainer" containerID="a0cb3a2dd3f2cee8e943eca96c9c023eed24dd0a9a9aafe2008ae9f59ba5d6ee" Nov 25 09:50:28 crc kubenswrapper[4776]: I1125 09:50:28.492081 4776 scope.go:117] "RemoveContainer" containerID="32677d2e2acbdc045ea2ec0ca221f8ca335f3a14b393cf46c3db835819bda093" Nov 25 09:50:28 crc kubenswrapper[4776]: I1125 09:50:28.524308 4776 scope.go:117] "RemoveContainer" containerID="89ea5f6d57e939e9a504efd0696cae5f58d4950592a81b46f72127d3204ecdab" Nov 25 09:50:28 crc kubenswrapper[4776]: I1125 09:50:28.551992 4776 scope.go:117] "RemoveContainer" containerID="f6caa82b641d6b347437cba0f2afb987e0ce5625f8944257e0067c00afe42059" Nov 25 09:51:17 crc kubenswrapper[4776]: I1125 09:51:17.818406 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:51:17 crc kubenswrapper[4776]: I1125 09:51:17.820047 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:51:28 crc kubenswrapper[4776]: I1125 09:51:28.704119 4776 scope.go:117] "RemoveContainer" containerID="53eb409078616814ade62b3d6be13d09644f9faf11f7f4738e36db9a3a38eefb" Nov 25 09:51:28 crc kubenswrapper[4776]: I1125 09:51:28.727829 4776 scope.go:117] "RemoveContainer" containerID="02c4c64a3ee9a1d43c07e566240a85097d6186935e3986969648ee476f3e015f" Nov 25 09:51:28 crc kubenswrapper[4776]: I1125 09:51:28.752692 4776 scope.go:117] "RemoveContainer" containerID="1505fc4e03c873c1557f6ddb2bb810c54cb98dd95eb6841cce7981791552b151" Nov 25 09:51:28 crc kubenswrapper[4776]: I1125 09:51:28.786855 4776 scope.go:117] "RemoveContainer" containerID="ef86e26a3c64bb3fa983b0e4cb1ba97b5d01472982d3395d5f9382b30981a3d2" Nov 25 09:51:28 crc kubenswrapper[4776]: I1125 09:51:28.803515 4776 scope.go:117] "RemoveContainer" containerID="c2a85ebe5755f4af658dd25fd030364579d92bdc461690f7c2025585b4779bd8" Nov 25 09:51:28 crc kubenswrapper[4776]: I1125 09:51:28.828671 4776 scope.go:117] "RemoveContainer" containerID="94b41979aa9c10d3bd3737ef33206925359824fb82b712cb4748abcf10f1e055" Nov 25 09:51:28 crc kubenswrapper[4776]: I1125 09:51:28.855964 4776 scope.go:117] "RemoveContainer" containerID="9779df52088b267add1650137fc3fc3f92ad6d3ea2e75a34659668d2ef3813bb" Nov 25 09:51:28 crc kubenswrapper[4776]: I1125 09:51:28.875441 4776 scope.go:117] "RemoveContainer" containerID="8c4015a72b16a36f9de326c38c3b6c1efa071b368ac1bacb90a14ea00d371022" Nov 25 09:51:28 crc kubenswrapper[4776]: I1125 09:51:28.893426 4776 scope.go:117] "RemoveContainer" containerID="87f8cbf578187a46d6e2d2c1082d1ed105b039c2788db71db6340e4c1b8457d4" Nov 25 09:51:28 crc kubenswrapper[4776]: I1125 09:51:28.916452 4776 scope.go:117] "RemoveContainer" containerID="ea0e61b5432e05c0fedaf2f966d0d05c1d4707abe42eef391fb793278d77f5c7" Nov 25 09:51:47 crc kubenswrapper[4776]: I1125 09:51:47.818397 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:51:47 crc kubenswrapper[4776]: I1125 09:51:47.818929 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:52:17 crc kubenswrapper[4776]: I1125 09:52:17.818939 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:52:17 crc kubenswrapper[4776]: I1125 09:52:17.819529 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:52:17 crc kubenswrapper[4776]: I1125 09:52:17.819577 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 09:52:17 crc kubenswrapper[4776]: I1125 09:52:17.820244 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 09:52:17 crc kubenswrapper[4776]: I1125 09:52:17.820303 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722" gracePeriod=600 Nov 25 09:52:17 crc kubenswrapper[4776]: E1125 09:52:17.953534 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 09:52:18 crc kubenswrapper[4776]: I1125 09:52:18.896304 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722" exitCode=0 Nov 25 09:52:18 crc kubenswrapper[4776]: I1125 09:52:18.896358 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722"} Nov 25 09:52:18 crc kubenswrapper[4776]: I1125 09:52:18.896404 4776 scope.go:117] "RemoveContainer" containerID="41bf411c5d6ef6679dd0fb3b86628f86f08b8ccef2b894d89b25148005eb3847" Nov 25 09:52:18 crc kubenswrapper[4776]: I1125 09:52:18.896854 4776 scope.go:117] "RemoveContainer" containerID="670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722" Nov 25 09:52:18 crc kubenswrapper[4776]: E1125 09:52:18.897089 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 09:52:29 crc kubenswrapper[4776]: I1125 09:52:29.074642 4776 scope.go:117] "RemoveContainer" containerID="5f45f2b73f855ed2de26c3204f7878eea5330ecac2bd8412a8721a9a5a319cfb" Nov 25 09:52:29 crc kubenswrapper[4776]: I1125 09:52:29.115861 4776 scope.go:117] "RemoveContainer" containerID="e2585e863a2a360dafd403d35d262a6c229963ec204b453fbc68f98d8cde8db2" Nov 25 09:52:29 crc kubenswrapper[4776]: I1125 09:52:29.142364 4776 scope.go:117] "RemoveContainer" containerID="de673d55efb03d21b4bc57b9296a8b59f87945fbe34c20d80c4980cad983590b" Nov 25 09:52:29 crc kubenswrapper[4776]: I1125 09:52:29.178309 4776 scope.go:117] "RemoveContainer" containerID="f7079fb542df2e880a7d48696479a3e07c7298972160f08d1c04ea9542b544ab" Nov 25 09:52:29 crc kubenswrapper[4776]: I1125 09:52:29.197373 4776 scope.go:117] "RemoveContainer" containerID="51468cb0acece8ce1a386bdf49793e6322b08cddac5cf23fc46abef1f5330e22" Nov 25 09:52:29 crc kubenswrapper[4776]: I1125 09:52:29.213905 4776 scope.go:117] "RemoveContainer" containerID="d4105354d99c01f581fbadfbf25a6148558ca3890a4fe12519c12245b6fa0449" Nov 25 09:52:32 crc kubenswrapper[4776]: I1125 09:52:32.662570 4776 scope.go:117] "RemoveContainer" containerID="670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722" Nov 25 09:52:32 crc kubenswrapper[4776]: E1125 09:52:32.663149 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 09:52:44 crc kubenswrapper[4776]: I1125 09:52:44.663036 4776 scope.go:117] "RemoveContainer" containerID="670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722" Nov 25 09:52:44 crc kubenswrapper[4776]: E1125 09:52:44.663899 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 09:52:55 crc kubenswrapper[4776]: I1125 09:52:55.666714 4776 scope.go:117] "RemoveContainer" containerID="670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722" Nov 25 09:52:55 crc kubenswrapper[4776]: E1125 09:52:55.667638 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 09:53:08 crc kubenswrapper[4776]: I1125 09:53:08.667555 4776 scope.go:117] "RemoveContainer" containerID="670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722" Nov 25 09:53:08 crc kubenswrapper[4776]: E1125 09:53:08.668230 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 09:53:21 crc kubenswrapper[4776]: I1125 09:53:21.661985 4776 scope.go:117] "RemoveContainer" containerID="670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722" Nov 25 09:53:21 crc kubenswrapper[4776]: E1125 09:53:21.662828 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 09:53:29 crc kubenswrapper[4776]: I1125 09:53:29.315341 4776 scope.go:117] "RemoveContainer" containerID="79cbebf3e9c8cb7e6ea34200c99cf5ad70ddabde13fac5d58d25f07b09c5e055" Nov 25 09:53:29 crc kubenswrapper[4776]: I1125 09:53:29.342874 4776 scope.go:117] "RemoveContainer" containerID="5a25e29978f0284d518a88b255bee88fa125d11b08d1eaeb950230e38ea74769" Nov 25 09:53:29 crc kubenswrapper[4776]: I1125 09:53:29.357712 4776 scope.go:117] "RemoveContainer" containerID="716b38eadb9be8e7ee2b70b2dd192c0196f35d29a4682d29db4b4d8f9899ba52" Nov 25 09:53:35 crc kubenswrapper[4776]: I1125 09:53:35.669439 4776 scope.go:117] "RemoveContainer" containerID="670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722" Nov 25 09:53:35 crc kubenswrapper[4776]: E1125 09:53:35.670364 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 09:53:50 crc kubenswrapper[4776]: I1125 09:53:50.662669 4776 scope.go:117] "RemoveContainer" containerID="670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722" Nov 25 09:53:50 crc kubenswrapper[4776]: E1125 09:53:50.664371 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 09:54:04 crc kubenswrapper[4776]: I1125 09:54:04.662385 4776 scope.go:117] "RemoveContainer" containerID="670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722" Nov 25 09:54:04 crc kubenswrapper[4776]: E1125 09:54:04.663171 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 09:54:18 crc kubenswrapper[4776]: I1125 09:54:18.662832 4776 scope.go:117] "RemoveContainer" containerID="670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722" Nov 25 09:54:18 crc kubenswrapper[4776]: E1125 09:54:18.663542 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 09:54:29 crc kubenswrapper[4776]: I1125 09:54:29.435237 4776 scope.go:117] "RemoveContainer" containerID="4a672773256a5735e511e0d75fb11f455144c24c1ad5bbd0829cbe34523b873d" Nov 25 09:54:29 crc kubenswrapper[4776]: I1125 09:54:29.466833 4776 scope.go:117] "RemoveContainer" containerID="2a2229b7cd7d8596847c255b5a632079e86a284eb29c3cd724e9e64edc2fcf8b" Nov 25 09:54:29 crc kubenswrapper[4776]: I1125 09:54:29.487989 4776 scope.go:117] "RemoveContainer" containerID="5006a19b68d700e697cd2f465b73452fcf31deed84559cb5659fa9a712fb7a12" Nov 25 09:54:29 crc kubenswrapper[4776]: I1125 09:54:29.507715 4776 scope.go:117] "RemoveContainer" containerID="77f3f2bd4f6583ba506502ce3622fd8e1b8dd903ba70835674336a5c58d9241b" Nov 25 09:54:30 crc kubenswrapper[4776]: I1125 09:54:30.662687 4776 scope.go:117] "RemoveContainer" containerID="670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722" Nov 25 09:54:30 crc kubenswrapper[4776]: E1125 09:54:30.663627 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 09:54:45 crc kubenswrapper[4776]: I1125 09:54:45.668173 4776 scope.go:117] "RemoveContainer" containerID="670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722" Nov 25 09:54:45 crc kubenswrapper[4776]: E1125 09:54:45.668949 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 09:54:59 crc kubenswrapper[4776]: I1125 09:54:59.662461 4776 scope.go:117] "RemoveContainer" containerID="670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722" Nov 25 09:54:59 crc kubenswrapper[4776]: E1125 09:54:59.663323 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 09:55:13 crc kubenswrapper[4776]: I1125 09:55:13.662976 4776 scope.go:117] "RemoveContainer" containerID="670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722" Nov 25 09:55:13 crc kubenswrapper[4776]: E1125 09:55:13.663628 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 09:55:24 crc kubenswrapper[4776]: I1125 09:55:24.661722 4776 scope.go:117] "RemoveContainer" containerID="670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722" Nov 25 09:55:24 crc kubenswrapper[4776]: E1125 09:55:24.662597 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 09:55:38 crc kubenswrapper[4776]: I1125 09:55:38.662932 4776 scope.go:117] "RemoveContainer" containerID="670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722" Nov 25 09:55:38 crc kubenswrapper[4776]: E1125 09:55:38.664019 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 09:55:53 crc kubenswrapper[4776]: I1125 09:55:53.662567 4776 scope.go:117] "RemoveContainer" containerID="670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722" Nov 25 09:55:53 crc kubenswrapper[4776]: E1125 09:55:53.663349 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 09:56:04 crc kubenswrapper[4776]: I1125 09:56:04.662985 4776 scope.go:117] "RemoveContainer" containerID="670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722" Nov 25 09:56:04 crc kubenswrapper[4776]: E1125 09:56:04.664202 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 09:56:19 crc kubenswrapper[4776]: I1125 09:56:19.663032 4776 scope.go:117] "RemoveContainer" containerID="670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722" Nov 25 09:56:19 crc kubenswrapper[4776]: E1125 09:56:19.664003 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 09:56:32 crc kubenswrapper[4776]: I1125 09:56:32.662422 4776 scope.go:117] "RemoveContainer" containerID="670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722" Nov 25 09:56:32 crc kubenswrapper[4776]: E1125 09:56:32.663250 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 09:56:44 crc kubenswrapper[4776]: I1125 09:56:44.662986 4776 scope.go:117] "RemoveContainer" containerID="670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722" Nov 25 09:56:44 crc kubenswrapper[4776]: E1125 09:56:44.664265 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 09:56:57 crc kubenswrapper[4776]: I1125 09:56:57.663135 4776 scope.go:117] "RemoveContainer" containerID="670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722" Nov 25 09:56:57 crc kubenswrapper[4776]: E1125 09:56:57.664106 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 09:57:12 crc kubenswrapper[4776]: I1125 09:57:12.662521 4776 scope.go:117] "RemoveContainer" containerID="670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722" Nov 25 09:57:12 crc kubenswrapper[4776]: E1125 09:57:12.663566 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 09:57:27 crc kubenswrapper[4776]: I1125 09:57:27.662516 4776 scope.go:117] "RemoveContainer" containerID="670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722" Nov 25 09:57:28 crc kubenswrapper[4776]: I1125 09:57:28.266839 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"7c8b8b54f516cb84cd523fbee04ee9767e3fadd57ce4fa6310ba05067cb37d20"} Nov 25 09:58:04 crc kubenswrapper[4776]: I1125 09:58:04.198409 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2j7sc"] Nov 25 09:58:04 crc kubenswrapper[4776]: E1125 09:58:04.199232 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d21a892-5265-4d39-9608-f2df1d7a6efe" containerName="registry-server" Nov 25 09:58:04 crc kubenswrapper[4776]: I1125 09:58:04.199248 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d21a892-5265-4d39-9608-f2df1d7a6efe" containerName="registry-server" Nov 25 09:58:04 crc kubenswrapper[4776]: E1125 09:58:04.199264 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d21a892-5265-4d39-9608-f2df1d7a6efe" containerName="extract-content" Nov 25 09:58:04 crc kubenswrapper[4776]: I1125 09:58:04.199271 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d21a892-5265-4d39-9608-f2df1d7a6efe" containerName="extract-content" Nov 25 09:58:04 crc kubenswrapper[4776]: E1125 09:58:04.199290 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d21a892-5265-4d39-9608-f2df1d7a6efe" containerName="extract-utilities" Nov 25 09:58:04 crc kubenswrapper[4776]: I1125 09:58:04.199298 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d21a892-5265-4d39-9608-f2df1d7a6efe" containerName="extract-utilities" Nov 25 09:58:04 crc kubenswrapper[4776]: I1125 09:58:04.199434 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d21a892-5265-4d39-9608-f2df1d7a6efe" containerName="registry-server" Nov 25 09:58:04 crc kubenswrapper[4776]: I1125 09:58:04.200373 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2j7sc" Nov 25 09:58:04 crc kubenswrapper[4776]: I1125 09:58:04.210756 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2j7sc"] Nov 25 09:58:04 crc kubenswrapper[4776]: I1125 09:58:04.388368 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3632a088-45f2-44a9-99d9-6f5070426236-catalog-content\") pod \"redhat-operators-2j7sc\" (UID: \"3632a088-45f2-44a9-99d9-6f5070426236\") " pod="openshift-marketplace/redhat-operators-2j7sc" Nov 25 09:58:04 crc kubenswrapper[4776]: I1125 09:58:04.388451 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxmbg\" (UniqueName: \"kubernetes.io/projected/3632a088-45f2-44a9-99d9-6f5070426236-kube-api-access-bxmbg\") pod \"redhat-operators-2j7sc\" (UID: \"3632a088-45f2-44a9-99d9-6f5070426236\") " pod="openshift-marketplace/redhat-operators-2j7sc" Nov 25 09:58:04 crc kubenswrapper[4776]: I1125 09:58:04.388586 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3632a088-45f2-44a9-99d9-6f5070426236-utilities\") pod \"redhat-operators-2j7sc\" (UID: \"3632a088-45f2-44a9-99d9-6f5070426236\") " pod="openshift-marketplace/redhat-operators-2j7sc" Nov 25 09:58:04 crc kubenswrapper[4776]: I1125 09:58:04.490159 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3632a088-45f2-44a9-99d9-6f5070426236-catalog-content\") pod \"redhat-operators-2j7sc\" (UID: \"3632a088-45f2-44a9-99d9-6f5070426236\") " pod="openshift-marketplace/redhat-operators-2j7sc" Nov 25 09:58:04 crc kubenswrapper[4776]: I1125 09:58:04.490222 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxmbg\" (UniqueName: \"kubernetes.io/projected/3632a088-45f2-44a9-99d9-6f5070426236-kube-api-access-bxmbg\") pod \"redhat-operators-2j7sc\" (UID: \"3632a088-45f2-44a9-99d9-6f5070426236\") " pod="openshift-marketplace/redhat-operators-2j7sc" Nov 25 09:58:04 crc kubenswrapper[4776]: I1125 09:58:04.490250 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3632a088-45f2-44a9-99d9-6f5070426236-utilities\") pod \"redhat-operators-2j7sc\" (UID: \"3632a088-45f2-44a9-99d9-6f5070426236\") " pod="openshift-marketplace/redhat-operators-2j7sc" Nov 25 09:58:04 crc kubenswrapper[4776]: I1125 09:58:04.490725 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3632a088-45f2-44a9-99d9-6f5070426236-catalog-content\") pod \"redhat-operators-2j7sc\" (UID: \"3632a088-45f2-44a9-99d9-6f5070426236\") " pod="openshift-marketplace/redhat-operators-2j7sc" Nov 25 09:58:04 crc kubenswrapper[4776]: I1125 09:58:04.490786 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3632a088-45f2-44a9-99d9-6f5070426236-utilities\") pod \"redhat-operators-2j7sc\" (UID: \"3632a088-45f2-44a9-99d9-6f5070426236\") " pod="openshift-marketplace/redhat-operators-2j7sc" Nov 25 09:58:04 crc kubenswrapper[4776]: I1125 09:58:04.510730 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxmbg\" (UniqueName: \"kubernetes.io/projected/3632a088-45f2-44a9-99d9-6f5070426236-kube-api-access-bxmbg\") pod \"redhat-operators-2j7sc\" (UID: \"3632a088-45f2-44a9-99d9-6f5070426236\") " pod="openshift-marketplace/redhat-operators-2j7sc" Nov 25 09:58:04 crc kubenswrapper[4776]: I1125 09:58:04.565584 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2j7sc" Nov 25 09:58:05 crc kubenswrapper[4776]: I1125 09:58:05.041025 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2j7sc"] Nov 25 09:58:05 crc kubenswrapper[4776]: I1125 09:58:05.585610 4776 generic.go:334] "Generic (PLEG): container finished" podID="3632a088-45f2-44a9-99d9-6f5070426236" containerID="3cb0e224904e52ba8bf13878d7b9c45a59c0434ac63f5b153e4eefd7eaebaf62" exitCode=0 Nov 25 09:58:05 crc kubenswrapper[4776]: I1125 09:58:05.585687 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2j7sc" event={"ID":"3632a088-45f2-44a9-99d9-6f5070426236","Type":"ContainerDied","Data":"3cb0e224904e52ba8bf13878d7b9c45a59c0434ac63f5b153e4eefd7eaebaf62"} Nov 25 09:58:05 crc kubenswrapper[4776]: I1125 09:58:05.586170 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2j7sc" event={"ID":"3632a088-45f2-44a9-99d9-6f5070426236","Type":"ContainerStarted","Data":"6eda476ce08c5a0a9459395df7fa2bd5b22c3b27d7c5d0d98e69e8e7b3516bca"} Nov 25 09:58:05 crc kubenswrapper[4776]: I1125 09:58:05.587756 4776 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 09:58:07 crc kubenswrapper[4776]: I1125 09:58:07.600860 4776 generic.go:334] "Generic (PLEG): container finished" podID="3632a088-45f2-44a9-99d9-6f5070426236" containerID="edc5314bf557b9a6b8a9e231428cdf5f671b091a13d0a49dbb9cf13050698a14" exitCode=0 Nov 25 09:58:07 crc kubenswrapper[4776]: I1125 09:58:07.600977 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2j7sc" event={"ID":"3632a088-45f2-44a9-99d9-6f5070426236","Type":"ContainerDied","Data":"edc5314bf557b9a6b8a9e231428cdf5f671b091a13d0a49dbb9cf13050698a14"} Nov 25 09:58:08 crc kubenswrapper[4776]: I1125 09:58:08.609862 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2j7sc" event={"ID":"3632a088-45f2-44a9-99d9-6f5070426236","Type":"ContainerStarted","Data":"8a46d09e89c721d68b30c32b253250b032053be3e6cb1d17f4f370a0d8dbc5ce"} Nov 25 09:58:08 crc kubenswrapper[4776]: I1125 09:58:08.628624 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2j7sc" podStartSLOduration=2.126333115 podStartE2EDuration="4.628609228s" podCreationTimestamp="2025-11-25 09:58:04 +0000 UTC" firstStartedPulling="2025-11-25 09:58:05.587475328 +0000 UTC m=+2030.628534891" lastFinishedPulling="2025-11-25 09:58:08.089751451 +0000 UTC m=+2033.130811004" observedRunningTime="2025-11-25 09:58:08.62588628 +0000 UTC m=+2033.666945833" watchObservedRunningTime="2025-11-25 09:58:08.628609228 +0000 UTC m=+2033.669668781" Nov 25 09:58:14 crc kubenswrapper[4776]: I1125 09:58:14.566044 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2j7sc" Nov 25 09:58:14 crc kubenswrapper[4776]: I1125 09:58:14.566521 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2j7sc" Nov 25 09:58:14 crc kubenswrapper[4776]: I1125 09:58:14.613256 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2j7sc" Nov 25 09:58:14 crc kubenswrapper[4776]: I1125 09:58:14.705260 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2j7sc" Nov 25 09:58:14 crc kubenswrapper[4776]: I1125 09:58:14.844605 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2j7sc"] Nov 25 09:58:16 crc kubenswrapper[4776]: I1125 09:58:16.672933 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2j7sc" podUID="3632a088-45f2-44a9-99d9-6f5070426236" containerName="registry-server" containerID="cri-o://8a46d09e89c721d68b30c32b253250b032053be3e6cb1d17f4f370a0d8dbc5ce" gracePeriod=2 Nov 25 09:58:17 crc kubenswrapper[4776]: I1125 09:58:17.083855 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2j7sc" Nov 25 09:58:17 crc kubenswrapper[4776]: I1125 09:58:17.276955 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3632a088-45f2-44a9-99d9-6f5070426236-catalog-content\") pod \"3632a088-45f2-44a9-99d9-6f5070426236\" (UID: \"3632a088-45f2-44a9-99d9-6f5070426236\") " Nov 25 09:58:17 crc kubenswrapper[4776]: I1125 09:58:17.277316 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3632a088-45f2-44a9-99d9-6f5070426236-utilities\") pod \"3632a088-45f2-44a9-99d9-6f5070426236\" (UID: \"3632a088-45f2-44a9-99d9-6f5070426236\") " Nov 25 09:58:17 crc kubenswrapper[4776]: I1125 09:58:17.277377 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxmbg\" (UniqueName: \"kubernetes.io/projected/3632a088-45f2-44a9-99d9-6f5070426236-kube-api-access-bxmbg\") pod \"3632a088-45f2-44a9-99d9-6f5070426236\" (UID: \"3632a088-45f2-44a9-99d9-6f5070426236\") " Nov 25 09:58:17 crc kubenswrapper[4776]: I1125 09:58:17.278831 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3632a088-45f2-44a9-99d9-6f5070426236-utilities" (OuterVolumeSpecName: "utilities") pod "3632a088-45f2-44a9-99d9-6f5070426236" (UID: "3632a088-45f2-44a9-99d9-6f5070426236"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:58:17 crc kubenswrapper[4776]: I1125 09:58:17.293451 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3632a088-45f2-44a9-99d9-6f5070426236-kube-api-access-bxmbg" (OuterVolumeSpecName: "kube-api-access-bxmbg") pod "3632a088-45f2-44a9-99d9-6f5070426236" (UID: "3632a088-45f2-44a9-99d9-6f5070426236"). InnerVolumeSpecName "kube-api-access-bxmbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:58:17 crc kubenswrapper[4776]: I1125 09:58:17.379259 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3632a088-45f2-44a9-99d9-6f5070426236-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:58:17 crc kubenswrapper[4776]: I1125 09:58:17.379304 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxmbg\" (UniqueName: \"kubernetes.io/projected/3632a088-45f2-44a9-99d9-6f5070426236-kube-api-access-bxmbg\") on node \"crc\" DevicePath \"\"" Nov 25 09:58:17 crc kubenswrapper[4776]: I1125 09:58:17.468598 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3632a088-45f2-44a9-99d9-6f5070426236-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3632a088-45f2-44a9-99d9-6f5070426236" (UID: "3632a088-45f2-44a9-99d9-6f5070426236"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:58:17 crc kubenswrapper[4776]: I1125 09:58:17.480054 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3632a088-45f2-44a9-99d9-6f5070426236-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:58:17 crc kubenswrapper[4776]: I1125 09:58:17.681531 4776 generic.go:334] "Generic (PLEG): container finished" podID="3632a088-45f2-44a9-99d9-6f5070426236" containerID="8a46d09e89c721d68b30c32b253250b032053be3e6cb1d17f4f370a0d8dbc5ce" exitCode=0 Nov 25 09:58:17 crc kubenswrapper[4776]: I1125 09:58:17.681602 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2j7sc" event={"ID":"3632a088-45f2-44a9-99d9-6f5070426236","Type":"ContainerDied","Data":"8a46d09e89c721d68b30c32b253250b032053be3e6cb1d17f4f370a0d8dbc5ce"} Nov 25 09:58:17 crc kubenswrapper[4776]: I1125 09:58:17.681649 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2j7sc" event={"ID":"3632a088-45f2-44a9-99d9-6f5070426236","Type":"ContainerDied","Data":"6eda476ce08c5a0a9459395df7fa2bd5b22c3b27d7c5d0d98e69e8e7b3516bca"} Nov 25 09:58:17 crc kubenswrapper[4776]: I1125 09:58:17.681678 4776 scope.go:117] "RemoveContainer" containerID="8a46d09e89c721d68b30c32b253250b032053be3e6cb1d17f4f370a0d8dbc5ce" Nov 25 09:58:17 crc kubenswrapper[4776]: I1125 09:58:17.682881 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2j7sc" Nov 25 09:58:17 crc kubenswrapper[4776]: I1125 09:58:17.702307 4776 scope.go:117] "RemoveContainer" containerID="edc5314bf557b9a6b8a9e231428cdf5f671b091a13d0a49dbb9cf13050698a14" Nov 25 09:58:17 crc kubenswrapper[4776]: I1125 09:58:17.712924 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2j7sc"] Nov 25 09:58:17 crc kubenswrapper[4776]: I1125 09:58:17.721447 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2j7sc"] Nov 25 09:58:17 crc kubenswrapper[4776]: I1125 09:58:17.741225 4776 scope.go:117] "RemoveContainer" containerID="3cb0e224904e52ba8bf13878d7b9c45a59c0434ac63f5b153e4eefd7eaebaf62" Nov 25 09:58:17 crc kubenswrapper[4776]: I1125 09:58:17.760153 4776 scope.go:117] "RemoveContainer" containerID="8a46d09e89c721d68b30c32b253250b032053be3e6cb1d17f4f370a0d8dbc5ce" Nov 25 09:58:17 crc kubenswrapper[4776]: E1125 09:58:17.760547 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a46d09e89c721d68b30c32b253250b032053be3e6cb1d17f4f370a0d8dbc5ce\": container with ID starting with 8a46d09e89c721d68b30c32b253250b032053be3e6cb1d17f4f370a0d8dbc5ce not found: ID does not exist" containerID="8a46d09e89c721d68b30c32b253250b032053be3e6cb1d17f4f370a0d8dbc5ce" Nov 25 09:58:17 crc kubenswrapper[4776]: I1125 09:58:17.760579 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a46d09e89c721d68b30c32b253250b032053be3e6cb1d17f4f370a0d8dbc5ce"} err="failed to get container status \"8a46d09e89c721d68b30c32b253250b032053be3e6cb1d17f4f370a0d8dbc5ce\": rpc error: code = NotFound desc = could not find container \"8a46d09e89c721d68b30c32b253250b032053be3e6cb1d17f4f370a0d8dbc5ce\": container with ID starting with 8a46d09e89c721d68b30c32b253250b032053be3e6cb1d17f4f370a0d8dbc5ce not found: ID does not exist" Nov 25 09:58:17 crc kubenswrapper[4776]: I1125 09:58:17.760599 4776 scope.go:117] "RemoveContainer" containerID="edc5314bf557b9a6b8a9e231428cdf5f671b091a13d0a49dbb9cf13050698a14" Nov 25 09:58:17 crc kubenswrapper[4776]: E1125 09:58:17.760983 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edc5314bf557b9a6b8a9e231428cdf5f671b091a13d0a49dbb9cf13050698a14\": container with ID starting with edc5314bf557b9a6b8a9e231428cdf5f671b091a13d0a49dbb9cf13050698a14 not found: ID does not exist" containerID="edc5314bf557b9a6b8a9e231428cdf5f671b091a13d0a49dbb9cf13050698a14" Nov 25 09:58:17 crc kubenswrapper[4776]: I1125 09:58:17.761007 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edc5314bf557b9a6b8a9e231428cdf5f671b091a13d0a49dbb9cf13050698a14"} err="failed to get container status \"edc5314bf557b9a6b8a9e231428cdf5f671b091a13d0a49dbb9cf13050698a14\": rpc error: code = NotFound desc = could not find container \"edc5314bf557b9a6b8a9e231428cdf5f671b091a13d0a49dbb9cf13050698a14\": container with ID starting with edc5314bf557b9a6b8a9e231428cdf5f671b091a13d0a49dbb9cf13050698a14 not found: ID does not exist" Nov 25 09:58:17 crc kubenswrapper[4776]: I1125 09:58:17.761019 4776 scope.go:117] "RemoveContainer" containerID="3cb0e224904e52ba8bf13878d7b9c45a59c0434ac63f5b153e4eefd7eaebaf62" Nov 25 09:58:17 crc kubenswrapper[4776]: E1125 09:58:17.761303 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cb0e224904e52ba8bf13878d7b9c45a59c0434ac63f5b153e4eefd7eaebaf62\": container with ID starting with 3cb0e224904e52ba8bf13878d7b9c45a59c0434ac63f5b153e4eefd7eaebaf62 not found: ID does not exist" containerID="3cb0e224904e52ba8bf13878d7b9c45a59c0434ac63f5b153e4eefd7eaebaf62" Nov 25 09:58:17 crc kubenswrapper[4776]: I1125 09:58:17.761339 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cb0e224904e52ba8bf13878d7b9c45a59c0434ac63f5b153e4eefd7eaebaf62"} err="failed to get container status \"3cb0e224904e52ba8bf13878d7b9c45a59c0434ac63f5b153e4eefd7eaebaf62\": rpc error: code = NotFound desc = could not find container \"3cb0e224904e52ba8bf13878d7b9c45a59c0434ac63f5b153e4eefd7eaebaf62\": container with ID starting with 3cb0e224904e52ba8bf13878d7b9c45a59c0434ac63f5b153e4eefd7eaebaf62 not found: ID does not exist" Nov 25 09:58:19 crc kubenswrapper[4776]: I1125 09:58:19.670766 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3632a088-45f2-44a9-99d9-6f5070426236" path="/var/lib/kubelet/pods/3632a088-45f2-44a9-99d9-6f5070426236/volumes" Nov 25 09:58:52 crc kubenswrapper[4776]: I1125 09:58:52.948827 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nfr57"] Nov 25 09:58:52 crc kubenswrapper[4776]: E1125 09:58:52.949908 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3632a088-45f2-44a9-99d9-6f5070426236" containerName="extract-utilities" Nov 25 09:58:52 crc kubenswrapper[4776]: I1125 09:58:52.949928 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="3632a088-45f2-44a9-99d9-6f5070426236" containerName="extract-utilities" Nov 25 09:58:52 crc kubenswrapper[4776]: E1125 09:58:52.949943 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3632a088-45f2-44a9-99d9-6f5070426236" containerName="extract-content" Nov 25 09:58:52 crc kubenswrapper[4776]: I1125 09:58:52.949952 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="3632a088-45f2-44a9-99d9-6f5070426236" containerName="extract-content" Nov 25 09:58:52 crc kubenswrapper[4776]: E1125 09:58:52.949983 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3632a088-45f2-44a9-99d9-6f5070426236" containerName="registry-server" Nov 25 09:58:52 crc kubenswrapper[4776]: I1125 09:58:52.949994 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="3632a088-45f2-44a9-99d9-6f5070426236" containerName="registry-server" Nov 25 09:58:52 crc kubenswrapper[4776]: I1125 09:58:52.950231 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="3632a088-45f2-44a9-99d9-6f5070426236" containerName="registry-server" Nov 25 09:58:52 crc kubenswrapper[4776]: I1125 09:58:52.951890 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nfr57" Nov 25 09:58:52 crc kubenswrapper[4776]: I1125 09:58:52.961249 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nfr57"] Nov 25 09:58:53 crc kubenswrapper[4776]: I1125 09:58:53.124718 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf85v\" (UniqueName: \"kubernetes.io/projected/15b68276-4800-40ac-a834-53be7071aeba-kube-api-access-tf85v\") pod \"redhat-marketplace-nfr57\" (UID: \"15b68276-4800-40ac-a834-53be7071aeba\") " pod="openshift-marketplace/redhat-marketplace-nfr57" Nov 25 09:58:53 crc kubenswrapper[4776]: I1125 09:58:53.124801 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15b68276-4800-40ac-a834-53be7071aeba-catalog-content\") pod \"redhat-marketplace-nfr57\" (UID: \"15b68276-4800-40ac-a834-53be7071aeba\") " pod="openshift-marketplace/redhat-marketplace-nfr57" Nov 25 09:58:53 crc kubenswrapper[4776]: I1125 09:58:53.124850 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15b68276-4800-40ac-a834-53be7071aeba-utilities\") pod \"redhat-marketplace-nfr57\" (UID: \"15b68276-4800-40ac-a834-53be7071aeba\") " pod="openshift-marketplace/redhat-marketplace-nfr57" Nov 25 09:58:53 crc kubenswrapper[4776]: I1125 09:58:53.225674 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15b68276-4800-40ac-a834-53be7071aeba-catalog-content\") pod \"redhat-marketplace-nfr57\" (UID: \"15b68276-4800-40ac-a834-53be7071aeba\") " pod="openshift-marketplace/redhat-marketplace-nfr57" Nov 25 09:58:53 crc kubenswrapper[4776]: I1125 09:58:53.225761 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15b68276-4800-40ac-a834-53be7071aeba-utilities\") pod \"redhat-marketplace-nfr57\" (UID: \"15b68276-4800-40ac-a834-53be7071aeba\") " pod="openshift-marketplace/redhat-marketplace-nfr57" Nov 25 09:58:53 crc kubenswrapper[4776]: I1125 09:58:53.225834 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf85v\" (UniqueName: \"kubernetes.io/projected/15b68276-4800-40ac-a834-53be7071aeba-kube-api-access-tf85v\") pod \"redhat-marketplace-nfr57\" (UID: \"15b68276-4800-40ac-a834-53be7071aeba\") " pod="openshift-marketplace/redhat-marketplace-nfr57" Nov 25 09:58:53 crc kubenswrapper[4776]: I1125 09:58:53.226338 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15b68276-4800-40ac-a834-53be7071aeba-utilities\") pod \"redhat-marketplace-nfr57\" (UID: \"15b68276-4800-40ac-a834-53be7071aeba\") " pod="openshift-marketplace/redhat-marketplace-nfr57" Nov 25 09:58:53 crc kubenswrapper[4776]: I1125 09:58:53.226442 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15b68276-4800-40ac-a834-53be7071aeba-catalog-content\") pod \"redhat-marketplace-nfr57\" (UID: \"15b68276-4800-40ac-a834-53be7071aeba\") " pod="openshift-marketplace/redhat-marketplace-nfr57" Nov 25 09:58:53 crc kubenswrapper[4776]: I1125 09:58:53.246811 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf85v\" (UniqueName: \"kubernetes.io/projected/15b68276-4800-40ac-a834-53be7071aeba-kube-api-access-tf85v\") pod \"redhat-marketplace-nfr57\" (UID: \"15b68276-4800-40ac-a834-53be7071aeba\") " pod="openshift-marketplace/redhat-marketplace-nfr57" Nov 25 09:58:53 crc kubenswrapper[4776]: I1125 09:58:53.283403 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nfr57" Nov 25 09:58:53 crc kubenswrapper[4776]: I1125 09:58:53.694992 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nfr57"] Nov 25 09:58:53 crc kubenswrapper[4776]: I1125 09:58:53.984423 4776 generic.go:334] "Generic (PLEG): container finished" podID="15b68276-4800-40ac-a834-53be7071aeba" containerID="847568010fa4fd0889d187ad362ae479702cafd9c73cc26c49f34946367e24f2" exitCode=0 Nov 25 09:58:53 crc kubenswrapper[4776]: I1125 09:58:53.985363 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nfr57" event={"ID":"15b68276-4800-40ac-a834-53be7071aeba","Type":"ContainerDied","Data":"847568010fa4fd0889d187ad362ae479702cafd9c73cc26c49f34946367e24f2"} Nov 25 09:58:53 crc kubenswrapper[4776]: I1125 09:58:53.985751 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nfr57" event={"ID":"15b68276-4800-40ac-a834-53be7071aeba","Type":"ContainerStarted","Data":"d0ba043ad876b39a968c0b343de66e73f802ea5ec09dfd6d279c9985b19e5813"} Nov 25 09:58:54 crc kubenswrapper[4776]: I1125 09:58:54.995454 4776 generic.go:334] "Generic (PLEG): container finished" podID="15b68276-4800-40ac-a834-53be7071aeba" containerID="ef896494ba4c9a29ebee6dffeaccb6b1b3ddfc0525ddf9235abab0e34e3bd653" exitCode=0 Nov 25 09:58:54 crc kubenswrapper[4776]: I1125 09:58:54.995553 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nfr57" event={"ID":"15b68276-4800-40ac-a834-53be7071aeba","Type":"ContainerDied","Data":"ef896494ba4c9a29ebee6dffeaccb6b1b3ddfc0525ddf9235abab0e34e3bd653"} Nov 25 09:58:56 crc kubenswrapper[4776]: I1125 09:58:56.006059 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nfr57" event={"ID":"15b68276-4800-40ac-a834-53be7071aeba","Type":"ContainerStarted","Data":"e4360ec20ca03e80c94c522e4fa3efc2b801d2a1565d585511b274ce21be34b2"} Nov 25 09:58:56 crc kubenswrapper[4776]: I1125 09:58:56.030012 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nfr57" podStartSLOduration=2.663800422 podStartE2EDuration="4.029996723s" podCreationTimestamp="2025-11-25 09:58:52 +0000 UTC" firstStartedPulling="2025-11-25 09:58:53.987273482 +0000 UTC m=+2079.028333055" lastFinishedPulling="2025-11-25 09:58:55.353469803 +0000 UTC m=+2080.394529356" observedRunningTime="2025-11-25 09:58:56.024980467 +0000 UTC m=+2081.066040020" watchObservedRunningTime="2025-11-25 09:58:56.029996723 +0000 UTC m=+2081.071056276" Nov 25 09:59:03 crc kubenswrapper[4776]: I1125 09:59:03.284016 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nfr57" Nov 25 09:59:03 crc kubenswrapper[4776]: I1125 09:59:03.285331 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nfr57" Nov 25 09:59:03 crc kubenswrapper[4776]: I1125 09:59:03.363095 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nfr57" Nov 25 09:59:03 crc kubenswrapper[4776]: I1125 09:59:03.427971 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nfr57" Nov 25 09:59:03 crc kubenswrapper[4776]: I1125 09:59:03.596554 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nfr57"] Nov 25 09:59:05 crc kubenswrapper[4776]: I1125 09:59:05.316257 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nfr57" podUID="15b68276-4800-40ac-a834-53be7071aeba" containerName="registry-server" containerID="cri-o://e4360ec20ca03e80c94c522e4fa3efc2b801d2a1565d585511b274ce21be34b2" gracePeriod=2 Nov 25 09:59:05 crc kubenswrapper[4776]: I1125 09:59:05.737141 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nfr57" Nov 25 09:59:05 crc kubenswrapper[4776]: I1125 09:59:05.821639 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15b68276-4800-40ac-a834-53be7071aeba-catalog-content\") pod \"15b68276-4800-40ac-a834-53be7071aeba\" (UID: \"15b68276-4800-40ac-a834-53be7071aeba\") " Nov 25 09:59:05 crc kubenswrapper[4776]: I1125 09:59:05.821778 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15b68276-4800-40ac-a834-53be7071aeba-utilities\") pod \"15b68276-4800-40ac-a834-53be7071aeba\" (UID: \"15b68276-4800-40ac-a834-53be7071aeba\") " Nov 25 09:59:05 crc kubenswrapper[4776]: I1125 09:59:05.821813 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tf85v\" (UniqueName: \"kubernetes.io/projected/15b68276-4800-40ac-a834-53be7071aeba-kube-api-access-tf85v\") pod \"15b68276-4800-40ac-a834-53be7071aeba\" (UID: \"15b68276-4800-40ac-a834-53be7071aeba\") " Nov 25 09:59:05 crc kubenswrapper[4776]: I1125 09:59:05.822929 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15b68276-4800-40ac-a834-53be7071aeba-utilities" (OuterVolumeSpecName: "utilities") pod "15b68276-4800-40ac-a834-53be7071aeba" (UID: "15b68276-4800-40ac-a834-53be7071aeba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:59:05 crc kubenswrapper[4776]: I1125 09:59:05.827651 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15b68276-4800-40ac-a834-53be7071aeba-kube-api-access-tf85v" (OuterVolumeSpecName: "kube-api-access-tf85v") pod "15b68276-4800-40ac-a834-53be7071aeba" (UID: "15b68276-4800-40ac-a834-53be7071aeba"). InnerVolumeSpecName "kube-api-access-tf85v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:59:05 crc kubenswrapper[4776]: I1125 09:59:05.844876 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15b68276-4800-40ac-a834-53be7071aeba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "15b68276-4800-40ac-a834-53be7071aeba" (UID: "15b68276-4800-40ac-a834-53be7071aeba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:59:05 crc kubenswrapper[4776]: I1125 09:59:05.923390 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15b68276-4800-40ac-a834-53be7071aeba-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:05 crc kubenswrapper[4776]: I1125 09:59:05.923439 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tf85v\" (UniqueName: \"kubernetes.io/projected/15b68276-4800-40ac-a834-53be7071aeba-kube-api-access-tf85v\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:05 crc kubenswrapper[4776]: I1125 09:59:05.923459 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15b68276-4800-40ac-a834-53be7071aeba-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:06 crc kubenswrapper[4776]: I1125 09:59:06.327000 4776 generic.go:334] "Generic (PLEG): container finished" podID="15b68276-4800-40ac-a834-53be7071aeba" containerID="e4360ec20ca03e80c94c522e4fa3efc2b801d2a1565d585511b274ce21be34b2" exitCode=0 Nov 25 09:59:06 crc kubenswrapper[4776]: I1125 09:59:06.327047 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nfr57" event={"ID":"15b68276-4800-40ac-a834-53be7071aeba","Type":"ContainerDied","Data":"e4360ec20ca03e80c94c522e4fa3efc2b801d2a1565d585511b274ce21be34b2"} Nov 25 09:59:06 crc kubenswrapper[4776]: I1125 09:59:06.327078 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nfr57" Nov 25 09:59:06 crc kubenswrapper[4776]: I1125 09:59:06.327137 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nfr57" event={"ID":"15b68276-4800-40ac-a834-53be7071aeba","Type":"ContainerDied","Data":"d0ba043ad876b39a968c0b343de66e73f802ea5ec09dfd6d279c9985b19e5813"} Nov 25 09:59:06 crc kubenswrapper[4776]: I1125 09:59:06.327155 4776 scope.go:117] "RemoveContainer" containerID="e4360ec20ca03e80c94c522e4fa3efc2b801d2a1565d585511b274ce21be34b2" Nov 25 09:59:06 crc kubenswrapper[4776]: I1125 09:59:06.357313 4776 scope.go:117] "RemoveContainer" containerID="ef896494ba4c9a29ebee6dffeaccb6b1b3ddfc0525ddf9235abab0e34e3bd653" Nov 25 09:59:06 crc kubenswrapper[4776]: I1125 09:59:06.371877 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nfr57"] Nov 25 09:59:06 crc kubenswrapper[4776]: I1125 09:59:06.378061 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nfr57"] Nov 25 09:59:06 crc kubenswrapper[4776]: I1125 09:59:06.397185 4776 scope.go:117] "RemoveContainer" containerID="847568010fa4fd0889d187ad362ae479702cafd9c73cc26c49f34946367e24f2" Nov 25 09:59:06 crc kubenswrapper[4776]: I1125 09:59:06.413927 4776 scope.go:117] "RemoveContainer" containerID="e4360ec20ca03e80c94c522e4fa3efc2b801d2a1565d585511b274ce21be34b2" Nov 25 09:59:06 crc kubenswrapper[4776]: E1125 09:59:06.414357 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4360ec20ca03e80c94c522e4fa3efc2b801d2a1565d585511b274ce21be34b2\": container with ID starting with e4360ec20ca03e80c94c522e4fa3efc2b801d2a1565d585511b274ce21be34b2 not found: ID does not exist" containerID="e4360ec20ca03e80c94c522e4fa3efc2b801d2a1565d585511b274ce21be34b2" Nov 25 09:59:06 crc kubenswrapper[4776]: I1125 09:59:06.414387 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4360ec20ca03e80c94c522e4fa3efc2b801d2a1565d585511b274ce21be34b2"} err="failed to get container status \"e4360ec20ca03e80c94c522e4fa3efc2b801d2a1565d585511b274ce21be34b2\": rpc error: code = NotFound desc = could not find container \"e4360ec20ca03e80c94c522e4fa3efc2b801d2a1565d585511b274ce21be34b2\": container with ID starting with e4360ec20ca03e80c94c522e4fa3efc2b801d2a1565d585511b274ce21be34b2 not found: ID does not exist" Nov 25 09:59:06 crc kubenswrapper[4776]: I1125 09:59:06.414410 4776 scope.go:117] "RemoveContainer" containerID="ef896494ba4c9a29ebee6dffeaccb6b1b3ddfc0525ddf9235abab0e34e3bd653" Nov 25 09:59:06 crc kubenswrapper[4776]: E1125 09:59:06.414784 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef896494ba4c9a29ebee6dffeaccb6b1b3ddfc0525ddf9235abab0e34e3bd653\": container with ID starting with ef896494ba4c9a29ebee6dffeaccb6b1b3ddfc0525ddf9235abab0e34e3bd653 not found: ID does not exist" containerID="ef896494ba4c9a29ebee6dffeaccb6b1b3ddfc0525ddf9235abab0e34e3bd653" Nov 25 09:59:06 crc kubenswrapper[4776]: I1125 09:59:06.414826 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef896494ba4c9a29ebee6dffeaccb6b1b3ddfc0525ddf9235abab0e34e3bd653"} err="failed to get container status \"ef896494ba4c9a29ebee6dffeaccb6b1b3ddfc0525ddf9235abab0e34e3bd653\": rpc error: code = NotFound desc = could not find container \"ef896494ba4c9a29ebee6dffeaccb6b1b3ddfc0525ddf9235abab0e34e3bd653\": container with ID starting with ef896494ba4c9a29ebee6dffeaccb6b1b3ddfc0525ddf9235abab0e34e3bd653 not found: ID does not exist" Nov 25 09:59:06 crc kubenswrapper[4776]: I1125 09:59:06.414851 4776 scope.go:117] "RemoveContainer" containerID="847568010fa4fd0889d187ad362ae479702cafd9c73cc26c49f34946367e24f2" Nov 25 09:59:06 crc kubenswrapper[4776]: E1125 09:59:06.415288 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"847568010fa4fd0889d187ad362ae479702cafd9c73cc26c49f34946367e24f2\": container with ID starting with 847568010fa4fd0889d187ad362ae479702cafd9c73cc26c49f34946367e24f2 not found: ID does not exist" containerID="847568010fa4fd0889d187ad362ae479702cafd9c73cc26c49f34946367e24f2" Nov 25 09:59:06 crc kubenswrapper[4776]: I1125 09:59:06.415399 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"847568010fa4fd0889d187ad362ae479702cafd9c73cc26c49f34946367e24f2"} err="failed to get container status \"847568010fa4fd0889d187ad362ae479702cafd9c73cc26c49f34946367e24f2\": rpc error: code = NotFound desc = could not find container \"847568010fa4fd0889d187ad362ae479702cafd9c73cc26c49f34946367e24f2\": container with ID starting with 847568010fa4fd0889d187ad362ae479702cafd9c73cc26c49f34946367e24f2 not found: ID does not exist" Nov 25 09:59:07 crc kubenswrapper[4776]: I1125 09:59:07.673527 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15b68276-4800-40ac-a834-53be7071aeba" path="/var/lib/kubelet/pods/15b68276-4800-40ac-a834-53be7071aeba/volumes" Nov 25 09:59:29 crc kubenswrapper[4776]: I1125 09:59:29.584232 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2684k"] Nov 25 09:59:29 crc kubenswrapper[4776]: E1125 09:59:29.585092 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15b68276-4800-40ac-a834-53be7071aeba" containerName="extract-utilities" Nov 25 09:59:29 crc kubenswrapper[4776]: I1125 09:59:29.585108 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="15b68276-4800-40ac-a834-53be7071aeba" containerName="extract-utilities" Nov 25 09:59:29 crc kubenswrapper[4776]: E1125 09:59:29.585123 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15b68276-4800-40ac-a834-53be7071aeba" containerName="extract-content" Nov 25 09:59:29 crc kubenswrapper[4776]: I1125 09:59:29.585130 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="15b68276-4800-40ac-a834-53be7071aeba" containerName="extract-content" Nov 25 09:59:29 crc kubenswrapper[4776]: E1125 09:59:29.585142 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15b68276-4800-40ac-a834-53be7071aeba" containerName="registry-server" Nov 25 09:59:29 crc kubenswrapper[4776]: I1125 09:59:29.585148 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="15b68276-4800-40ac-a834-53be7071aeba" containerName="registry-server" Nov 25 09:59:29 crc kubenswrapper[4776]: I1125 09:59:29.585297 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="15b68276-4800-40ac-a834-53be7071aeba" containerName="registry-server" Nov 25 09:59:29 crc kubenswrapper[4776]: I1125 09:59:29.586340 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2684k" Nov 25 09:59:29 crc kubenswrapper[4776]: I1125 09:59:29.611331 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2684k"] Nov 25 09:59:29 crc kubenswrapper[4776]: I1125 09:59:29.677932 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcm4v\" (UniqueName: \"kubernetes.io/projected/ffe818dc-9e81-48f9-91de-e770bffeea36-kube-api-access-fcm4v\") pod \"certified-operators-2684k\" (UID: \"ffe818dc-9e81-48f9-91de-e770bffeea36\") " pod="openshift-marketplace/certified-operators-2684k" Nov 25 09:59:29 crc kubenswrapper[4776]: I1125 09:59:29.677997 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ffe818dc-9e81-48f9-91de-e770bffeea36-catalog-content\") pod \"certified-operators-2684k\" (UID: \"ffe818dc-9e81-48f9-91de-e770bffeea36\") " pod="openshift-marketplace/certified-operators-2684k" Nov 25 09:59:29 crc kubenswrapper[4776]: I1125 09:59:29.678118 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ffe818dc-9e81-48f9-91de-e770bffeea36-utilities\") pod \"certified-operators-2684k\" (UID: \"ffe818dc-9e81-48f9-91de-e770bffeea36\") " pod="openshift-marketplace/certified-operators-2684k" Nov 25 09:59:29 crc kubenswrapper[4776]: I1125 09:59:29.778978 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ffe818dc-9e81-48f9-91de-e770bffeea36-utilities\") pod \"certified-operators-2684k\" (UID: \"ffe818dc-9e81-48f9-91de-e770bffeea36\") " pod="openshift-marketplace/certified-operators-2684k" Nov 25 09:59:29 crc kubenswrapper[4776]: I1125 09:59:29.779283 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcm4v\" (UniqueName: \"kubernetes.io/projected/ffe818dc-9e81-48f9-91de-e770bffeea36-kube-api-access-fcm4v\") pod \"certified-operators-2684k\" (UID: \"ffe818dc-9e81-48f9-91de-e770bffeea36\") " pod="openshift-marketplace/certified-operators-2684k" Nov 25 09:59:29 crc kubenswrapper[4776]: I1125 09:59:29.779405 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ffe818dc-9e81-48f9-91de-e770bffeea36-catalog-content\") pod \"certified-operators-2684k\" (UID: \"ffe818dc-9e81-48f9-91de-e770bffeea36\") " pod="openshift-marketplace/certified-operators-2684k" Nov 25 09:59:29 crc kubenswrapper[4776]: I1125 09:59:29.779615 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ffe818dc-9e81-48f9-91de-e770bffeea36-utilities\") pod \"certified-operators-2684k\" (UID: \"ffe818dc-9e81-48f9-91de-e770bffeea36\") " pod="openshift-marketplace/certified-operators-2684k" Nov 25 09:59:29 crc kubenswrapper[4776]: I1125 09:59:29.779787 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ffe818dc-9e81-48f9-91de-e770bffeea36-catalog-content\") pod \"certified-operators-2684k\" (UID: \"ffe818dc-9e81-48f9-91de-e770bffeea36\") " pod="openshift-marketplace/certified-operators-2684k" Nov 25 09:59:29 crc kubenswrapper[4776]: I1125 09:59:29.808058 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcm4v\" (UniqueName: \"kubernetes.io/projected/ffe818dc-9e81-48f9-91de-e770bffeea36-kube-api-access-fcm4v\") pod \"certified-operators-2684k\" (UID: \"ffe818dc-9e81-48f9-91de-e770bffeea36\") " pod="openshift-marketplace/certified-operators-2684k" Nov 25 09:59:29 crc kubenswrapper[4776]: I1125 09:59:29.909497 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2684k" Nov 25 09:59:30 crc kubenswrapper[4776]: I1125 09:59:30.190828 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2684k"] Nov 25 09:59:30 crc kubenswrapper[4776]: I1125 09:59:30.528557 4776 generic.go:334] "Generic (PLEG): container finished" podID="ffe818dc-9e81-48f9-91de-e770bffeea36" containerID="88b37ab6121985811351671fe074323fe322fb3e2b86c0edde2809dbc5075f30" exitCode=0 Nov 25 09:59:30 crc kubenswrapper[4776]: I1125 09:59:30.528596 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2684k" event={"ID":"ffe818dc-9e81-48f9-91de-e770bffeea36","Type":"ContainerDied","Data":"88b37ab6121985811351671fe074323fe322fb3e2b86c0edde2809dbc5075f30"} Nov 25 09:59:30 crc kubenswrapper[4776]: I1125 09:59:30.528620 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2684k" event={"ID":"ffe818dc-9e81-48f9-91de-e770bffeea36","Type":"ContainerStarted","Data":"7916a5d796cb587d46b7fb32baad2922416746833a389f1210a61441378fbc9c"} Nov 25 09:59:31 crc kubenswrapper[4776]: I1125 09:59:31.539125 4776 generic.go:334] "Generic (PLEG): container finished" podID="ffe818dc-9e81-48f9-91de-e770bffeea36" containerID="2ef76e679f60d4a286f7e5697ab72bd71b410265dd97a30d6d4a05bcdd825b49" exitCode=0 Nov 25 09:59:31 crc kubenswrapper[4776]: I1125 09:59:31.539191 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2684k" event={"ID":"ffe818dc-9e81-48f9-91de-e770bffeea36","Type":"ContainerDied","Data":"2ef76e679f60d4a286f7e5697ab72bd71b410265dd97a30d6d4a05bcdd825b49"} Nov 25 09:59:32 crc kubenswrapper[4776]: I1125 09:59:32.548526 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2684k" event={"ID":"ffe818dc-9e81-48f9-91de-e770bffeea36","Type":"ContainerStarted","Data":"958d77c8936a761b364150eb0c346df6057efc009fe5c607f0598413ec26032e"} Nov 25 09:59:32 crc kubenswrapper[4776]: I1125 09:59:32.569553 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2684k" podStartSLOduration=1.8872868729999999 podStartE2EDuration="3.569527985s" podCreationTimestamp="2025-11-25 09:59:29 +0000 UTC" firstStartedPulling="2025-11-25 09:59:30.530271901 +0000 UTC m=+2115.571331454" lastFinishedPulling="2025-11-25 09:59:32.212513013 +0000 UTC m=+2117.253572566" observedRunningTime="2025-11-25 09:59:32.567231787 +0000 UTC m=+2117.608291340" watchObservedRunningTime="2025-11-25 09:59:32.569527985 +0000 UTC m=+2117.610587548" Nov 25 09:59:39 crc kubenswrapper[4776]: I1125 09:59:39.910442 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2684k" Nov 25 09:59:39 crc kubenswrapper[4776]: I1125 09:59:39.911295 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2684k" Nov 25 09:59:39 crc kubenswrapper[4776]: I1125 09:59:39.983509 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2684k" Nov 25 09:59:40 crc kubenswrapper[4776]: I1125 09:59:40.681522 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2684k" Nov 25 09:59:40 crc kubenswrapper[4776]: I1125 09:59:40.730334 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2684k"] Nov 25 09:59:42 crc kubenswrapper[4776]: I1125 09:59:42.639703 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2684k" podUID="ffe818dc-9e81-48f9-91de-e770bffeea36" containerName="registry-server" containerID="cri-o://958d77c8936a761b364150eb0c346df6057efc009fe5c607f0598413ec26032e" gracePeriod=2 Nov 25 09:59:43 crc kubenswrapper[4776]: I1125 09:59:43.010056 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2684k" Nov 25 09:59:43 crc kubenswrapper[4776]: I1125 09:59:43.069388 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ffe818dc-9e81-48f9-91de-e770bffeea36-utilities\") pod \"ffe818dc-9e81-48f9-91de-e770bffeea36\" (UID: \"ffe818dc-9e81-48f9-91de-e770bffeea36\") " Nov 25 09:59:43 crc kubenswrapper[4776]: I1125 09:59:43.069440 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcm4v\" (UniqueName: \"kubernetes.io/projected/ffe818dc-9e81-48f9-91de-e770bffeea36-kube-api-access-fcm4v\") pod \"ffe818dc-9e81-48f9-91de-e770bffeea36\" (UID: \"ffe818dc-9e81-48f9-91de-e770bffeea36\") " Nov 25 09:59:43 crc kubenswrapper[4776]: I1125 09:59:43.069474 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ffe818dc-9e81-48f9-91de-e770bffeea36-catalog-content\") pod \"ffe818dc-9e81-48f9-91de-e770bffeea36\" (UID: \"ffe818dc-9e81-48f9-91de-e770bffeea36\") " Nov 25 09:59:43 crc kubenswrapper[4776]: I1125 09:59:43.071185 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ffe818dc-9e81-48f9-91de-e770bffeea36-utilities" (OuterVolumeSpecName: "utilities") pod "ffe818dc-9e81-48f9-91de-e770bffeea36" (UID: "ffe818dc-9e81-48f9-91de-e770bffeea36"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:59:43 crc kubenswrapper[4776]: I1125 09:59:43.077226 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffe818dc-9e81-48f9-91de-e770bffeea36-kube-api-access-fcm4v" (OuterVolumeSpecName: "kube-api-access-fcm4v") pod "ffe818dc-9e81-48f9-91de-e770bffeea36" (UID: "ffe818dc-9e81-48f9-91de-e770bffeea36"). InnerVolumeSpecName "kube-api-access-fcm4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:59:43 crc kubenswrapper[4776]: I1125 09:59:43.127459 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ffe818dc-9e81-48f9-91de-e770bffeea36-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ffe818dc-9e81-48f9-91de-e770bffeea36" (UID: "ffe818dc-9e81-48f9-91de-e770bffeea36"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:59:43 crc kubenswrapper[4776]: I1125 09:59:43.170635 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ffe818dc-9e81-48f9-91de-e770bffeea36-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:43 crc kubenswrapper[4776]: I1125 09:59:43.170681 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcm4v\" (UniqueName: \"kubernetes.io/projected/ffe818dc-9e81-48f9-91de-e770bffeea36-kube-api-access-fcm4v\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:43 crc kubenswrapper[4776]: I1125 09:59:43.170693 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ffe818dc-9e81-48f9-91de-e770bffeea36-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:43 crc kubenswrapper[4776]: I1125 09:59:43.653117 4776 generic.go:334] "Generic (PLEG): container finished" podID="ffe818dc-9e81-48f9-91de-e770bffeea36" containerID="958d77c8936a761b364150eb0c346df6057efc009fe5c607f0598413ec26032e" exitCode=0 Nov 25 09:59:43 crc kubenswrapper[4776]: I1125 09:59:43.653283 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2684k" Nov 25 09:59:43 crc kubenswrapper[4776]: I1125 09:59:43.654044 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2684k" event={"ID":"ffe818dc-9e81-48f9-91de-e770bffeea36","Type":"ContainerDied","Data":"958d77c8936a761b364150eb0c346df6057efc009fe5c607f0598413ec26032e"} Nov 25 09:59:43 crc kubenswrapper[4776]: I1125 09:59:43.654430 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2684k" event={"ID":"ffe818dc-9e81-48f9-91de-e770bffeea36","Type":"ContainerDied","Data":"7916a5d796cb587d46b7fb32baad2922416746833a389f1210a61441378fbc9c"} Nov 25 09:59:43 crc kubenswrapper[4776]: I1125 09:59:43.654508 4776 scope.go:117] "RemoveContainer" containerID="958d77c8936a761b364150eb0c346df6057efc009fe5c607f0598413ec26032e" Nov 25 09:59:43 crc kubenswrapper[4776]: I1125 09:59:43.689383 4776 scope.go:117] "RemoveContainer" containerID="2ef76e679f60d4a286f7e5697ab72bd71b410265dd97a30d6d4a05bcdd825b49" Nov 25 09:59:43 crc kubenswrapper[4776]: I1125 09:59:43.722242 4776 scope.go:117] "RemoveContainer" containerID="88b37ab6121985811351671fe074323fe322fb3e2b86c0edde2809dbc5075f30" Nov 25 09:59:43 crc kubenswrapper[4776]: I1125 09:59:43.725105 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2684k"] Nov 25 09:59:43 crc kubenswrapper[4776]: I1125 09:59:43.734654 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2684k"] Nov 25 09:59:43 crc kubenswrapper[4776]: I1125 09:59:43.769367 4776 scope.go:117] "RemoveContainer" containerID="958d77c8936a761b364150eb0c346df6057efc009fe5c607f0598413ec26032e" Nov 25 09:59:43 crc kubenswrapper[4776]: E1125 09:59:43.769974 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"958d77c8936a761b364150eb0c346df6057efc009fe5c607f0598413ec26032e\": container with ID starting with 958d77c8936a761b364150eb0c346df6057efc009fe5c607f0598413ec26032e not found: ID does not exist" containerID="958d77c8936a761b364150eb0c346df6057efc009fe5c607f0598413ec26032e" Nov 25 09:59:43 crc kubenswrapper[4776]: I1125 09:59:43.770053 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"958d77c8936a761b364150eb0c346df6057efc009fe5c607f0598413ec26032e"} err="failed to get container status \"958d77c8936a761b364150eb0c346df6057efc009fe5c607f0598413ec26032e\": rpc error: code = NotFound desc = could not find container \"958d77c8936a761b364150eb0c346df6057efc009fe5c607f0598413ec26032e\": container with ID starting with 958d77c8936a761b364150eb0c346df6057efc009fe5c607f0598413ec26032e not found: ID does not exist" Nov 25 09:59:43 crc kubenswrapper[4776]: I1125 09:59:43.770117 4776 scope.go:117] "RemoveContainer" containerID="2ef76e679f60d4a286f7e5697ab72bd71b410265dd97a30d6d4a05bcdd825b49" Nov 25 09:59:43 crc kubenswrapper[4776]: E1125 09:59:43.770658 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ef76e679f60d4a286f7e5697ab72bd71b410265dd97a30d6d4a05bcdd825b49\": container with ID starting with 2ef76e679f60d4a286f7e5697ab72bd71b410265dd97a30d6d4a05bcdd825b49 not found: ID does not exist" containerID="2ef76e679f60d4a286f7e5697ab72bd71b410265dd97a30d6d4a05bcdd825b49" Nov 25 09:59:43 crc kubenswrapper[4776]: I1125 09:59:43.770723 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ef76e679f60d4a286f7e5697ab72bd71b410265dd97a30d6d4a05bcdd825b49"} err="failed to get container status \"2ef76e679f60d4a286f7e5697ab72bd71b410265dd97a30d6d4a05bcdd825b49\": rpc error: code = NotFound desc = could not find container \"2ef76e679f60d4a286f7e5697ab72bd71b410265dd97a30d6d4a05bcdd825b49\": container with ID starting with 2ef76e679f60d4a286f7e5697ab72bd71b410265dd97a30d6d4a05bcdd825b49 not found: ID does not exist" Nov 25 09:59:43 crc kubenswrapper[4776]: I1125 09:59:43.770767 4776 scope.go:117] "RemoveContainer" containerID="88b37ab6121985811351671fe074323fe322fb3e2b86c0edde2809dbc5075f30" Nov 25 09:59:43 crc kubenswrapper[4776]: E1125 09:59:43.771155 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88b37ab6121985811351671fe074323fe322fb3e2b86c0edde2809dbc5075f30\": container with ID starting with 88b37ab6121985811351671fe074323fe322fb3e2b86c0edde2809dbc5075f30 not found: ID does not exist" containerID="88b37ab6121985811351671fe074323fe322fb3e2b86c0edde2809dbc5075f30" Nov 25 09:59:43 crc kubenswrapper[4776]: I1125 09:59:43.771222 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88b37ab6121985811351671fe074323fe322fb3e2b86c0edde2809dbc5075f30"} err="failed to get container status \"88b37ab6121985811351671fe074323fe322fb3e2b86c0edde2809dbc5075f30\": rpc error: code = NotFound desc = could not find container \"88b37ab6121985811351671fe074323fe322fb3e2b86c0edde2809dbc5075f30\": container with ID starting with 88b37ab6121985811351671fe074323fe322fb3e2b86c0edde2809dbc5075f30 not found: ID does not exist" Nov 25 09:59:45 crc kubenswrapper[4776]: I1125 09:59:45.678863 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffe818dc-9e81-48f9-91de-e770bffeea36" path="/var/lib/kubelet/pods/ffe818dc-9e81-48f9-91de-e770bffeea36/volumes" Nov 25 09:59:46 crc kubenswrapper[4776]: I1125 09:59:46.092779 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7srcr"] Nov 25 09:59:46 crc kubenswrapper[4776]: E1125 09:59:46.093337 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffe818dc-9e81-48f9-91de-e770bffeea36" containerName="extract-utilities" Nov 25 09:59:46 crc kubenswrapper[4776]: I1125 09:59:46.093360 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffe818dc-9e81-48f9-91de-e770bffeea36" containerName="extract-utilities" Nov 25 09:59:46 crc kubenswrapper[4776]: E1125 09:59:46.093382 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffe818dc-9e81-48f9-91de-e770bffeea36" containerName="registry-server" Nov 25 09:59:46 crc kubenswrapper[4776]: I1125 09:59:46.093392 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffe818dc-9e81-48f9-91de-e770bffeea36" containerName="registry-server" Nov 25 09:59:46 crc kubenswrapper[4776]: E1125 09:59:46.093426 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffe818dc-9e81-48f9-91de-e770bffeea36" containerName="extract-content" Nov 25 09:59:46 crc kubenswrapper[4776]: I1125 09:59:46.093436 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffe818dc-9e81-48f9-91de-e770bffeea36" containerName="extract-content" Nov 25 09:59:46 crc kubenswrapper[4776]: I1125 09:59:46.093698 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffe818dc-9e81-48f9-91de-e770bffeea36" containerName="registry-server" Nov 25 09:59:46 crc kubenswrapper[4776]: I1125 09:59:46.101856 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7srcr" Nov 25 09:59:46 crc kubenswrapper[4776]: I1125 09:59:46.112362 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7srcr"] Nov 25 09:59:46 crc kubenswrapper[4776]: I1125 09:59:46.115636 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35c6cb73-51b4-4258-a6a3-43cba2007a49-utilities\") pod \"community-operators-7srcr\" (UID: \"35c6cb73-51b4-4258-a6a3-43cba2007a49\") " pod="openshift-marketplace/community-operators-7srcr" Nov 25 09:59:46 crc kubenswrapper[4776]: I1125 09:59:46.115701 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ptr4\" (UniqueName: \"kubernetes.io/projected/35c6cb73-51b4-4258-a6a3-43cba2007a49-kube-api-access-9ptr4\") pod \"community-operators-7srcr\" (UID: \"35c6cb73-51b4-4258-a6a3-43cba2007a49\") " pod="openshift-marketplace/community-operators-7srcr" Nov 25 09:59:46 crc kubenswrapper[4776]: I1125 09:59:46.115734 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35c6cb73-51b4-4258-a6a3-43cba2007a49-catalog-content\") pod \"community-operators-7srcr\" (UID: \"35c6cb73-51b4-4258-a6a3-43cba2007a49\") " pod="openshift-marketplace/community-operators-7srcr" Nov 25 09:59:46 crc kubenswrapper[4776]: I1125 09:59:46.217557 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35c6cb73-51b4-4258-a6a3-43cba2007a49-utilities\") pod \"community-operators-7srcr\" (UID: \"35c6cb73-51b4-4258-a6a3-43cba2007a49\") " pod="openshift-marketplace/community-operators-7srcr" Nov 25 09:59:46 crc kubenswrapper[4776]: I1125 09:59:46.217654 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ptr4\" (UniqueName: \"kubernetes.io/projected/35c6cb73-51b4-4258-a6a3-43cba2007a49-kube-api-access-9ptr4\") pod \"community-operators-7srcr\" (UID: \"35c6cb73-51b4-4258-a6a3-43cba2007a49\") " pod="openshift-marketplace/community-operators-7srcr" Nov 25 09:59:46 crc kubenswrapper[4776]: I1125 09:59:46.217700 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35c6cb73-51b4-4258-a6a3-43cba2007a49-catalog-content\") pod \"community-operators-7srcr\" (UID: \"35c6cb73-51b4-4258-a6a3-43cba2007a49\") " pod="openshift-marketplace/community-operators-7srcr" Nov 25 09:59:46 crc kubenswrapper[4776]: I1125 09:59:46.218171 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35c6cb73-51b4-4258-a6a3-43cba2007a49-utilities\") pod \"community-operators-7srcr\" (UID: \"35c6cb73-51b4-4258-a6a3-43cba2007a49\") " pod="openshift-marketplace/community-operators-7srcr" Nov 25 09:59:46 crc kubenswrapper[4776]: I1125 09:59:46.218310 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35c6cb73-51b4-4258-a6a3-43cba2007a49-catalog-content\") pod \"community-operators-7srcr\" (UID: \"35c6cb73-51b4-4258-a6a3-43cba2007a49\") " pod="openshift-marketplace/community-operators-7srcr" Nov 25 09:59:46 crc kubenswrapper[4776]: I1125 09:59:46.242032 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ptr4\" (UniqueName: \"kubernetes.io/projected/35c6cb73-51b4-4258-a6a3-43cba2007a49-kube-api-access-9ptr4\") pod \"community-operators-7srcr\" (UID: \"35c6cb73-51b4-4258-a6a3-43cba2007a49\") " pod="openshift-marketplace/community-operators-7srcr" Nov 25 09:59:46 crc kubenswrapper[4776]: I1125 09:59:46.424109 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7srcr" Nov 25 09:59:46 crc kubenswrapper[4776]: I1125 09:59:46.954011 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7srcr"] Nov 25 09:59:47 crc kubenswrapper[4776]: I1125 09:59:47.689374 4776 generic.go:334] "Generic (PLEG): container finished" podID="35c6cb73-51b4-4258-a6a3-43cba2007a49" containerID="82af4ae638d838069f48ed767c8475e1b6b134d5ec63edd53731e67e29115215" exitCode=0 Nov 25 09:59:47 crc kubenswrapper[4776]: I1125 09:59:47.689546 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7srcr" event={"ID":"35c6cb73-51b4-4258-a6a3-43cba2007a49","Type":"ContainerDied","Data":"82af4ae638d838069f48ed767c8475e1b6b134d5ec63edd53731e67e29115215"} Nov 25 09:59:47 crc kubenswrapper[4776]: I1125 09:59:47.689697 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7srcr" event={"ID":"35c6cb73-51b4-4258-a6a3-43cba2007a49","Type":"ContainerStarted","Data":"3b6da74ac206f11a576fe07c446e4ea2b759aaec0fc311ec21c7dfa092e3dd2e"} Nov 25 09:59:47 crc kubenswrapper[4776]: I1125 09:59:47.818777 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:59:47 crc kubenswrapper[4776]: I1125 09:59:47.818857 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:59:48 crc kubenswrapper[4776]: I1125 09:59:48.699619 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7srcr" event={"ID":"35c6cb73-51b4-4258-a6a3-43cba2007a49","Type":"ContainerStarted","Data":"5eb0a8cc5f5c972934bbe6ea04907c09bd15725c2cbd70044b4fd45cadaa36d5"} Nov 25 09:59:49 crc kubenswrapper[4776]: I1125 09:59:49.712718 4776 generic.go:334] "Generic (PLEG): container finished" podID="35c6cb73-51b4-4258-a6a3-43cba2007a49" containerID="5eb0a8cc5f5c972934bbe6ea04907c09bd15725c2cbd70044b4fd45cadaa36d5" exitCode=0 Nov 25 09:59:49 crc kubenswrapper[4776]: I1125 09:59:49.712807 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7srcr" event={"ID":"35c6cb73-51b4-4258-a6a3-43cba2007a49","Type":"ContainerDied","Data":"5eb0a8cc5f5c972934bbe6ea04907c09bd15725c2cbd70044b4fd45cadaa36d5"} Nov 25 09:59:50 crc kubenswrapper[4776]: I1125 09:59:50.725755 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7srcr" event={"ID":"35c6cb73-51b4-4258-a6a3-43cba2007a49","Type":"ContainerStarted","Data":"7e69a75e4bafd241ea1a92a72e22bebaec449e1cc9df49f447302ede65e0c6a1"} Nov 25 09:59:50 crc kubenswrapper[4776]: I1125 09:59:50.752316 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7srcr" podStartSLOduration=2.178588161 podStartE2EDuration="4.752294661s" podCreationTimestamp="2025-11-25 09:59:46 +0000 UTC" firstStartedPulling="2025-11-25 09:59:47.691035055 +0000 UTC m=+2132.732094608" lastFinishedPulling="2025-11-25 09:59:50.264741545 +0000 UTC m=+2135.305801108" observedRunningTime="2025-11-25 09:59:50.748520166 +0000 UTC m=+2135.789579749" watchObservedRunningTime="2025-11-25 09:59:50.752294661 +0000 UTC m=+2135.793354224" Nov 25 09:59:56 crc kubenswrapper[4776]: I1125 09:59:56.425053 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7srcr" Nov 25 09:59:56 crc kubenswrapper[4776]: I1125 09:59:56.425535 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7srcr" Nov 25 09:59:56 crc kubenswrapper[4776]: I1125 09:59:56.493347 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7srcr" Nov 25 09:59:56 crc kubenswrapper[4776]: I1125 09:59:56.828279 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7srcr" Nov 25 09:59:56 crc kubenswrapper[4776]: I1125 09:59:56.869585 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7srcr"] Nov 25 09:59:58 crc kubenswrapper[4776]: I1125 09:59:58.794593 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7srcr" podUID="35c6cb73-51b4-4258-a6a3-43cba2007a49" containerName="registry-server" containerID="cri-o://7e69a75e4bafd241ea1a92a72e22bebaec449e1cc9df49f447302ede65e0c6a1" gracePeriod=2 Nov 25 09:59:59 crc kubenswrapper[4776]: I1125 09:59:59.223211 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7srcr" Nov 25 09:59:59 crc kubenswrapper[4776]: I1125 09:59:59.327454 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ptr4\" (UniqueName: \"kubernetes.io/projected/35c6cb73-51b4-4258-a6a3-43cba2007a49-kube-api-access-9ptr4\") pod \"35c6cb73-51b4-4258-a6a3-43cba2007a49\" (UID: \"35c6cb73-51b4-4258-a6a3-43cba2007a49\") " Nov 25 09:59:59 crc kubenswrapper[4776]: I1125 09:59:59.327554 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35c6cb73-51b4-4258-a6a3-43cba2007a49-catalog-content\") pod \"35c6cb73-51b4-4258-a6a3-43cba2007a49\" (UID: \"35c6cb73-51b4-4258-a6a3-43cba2007a49\") " Nov 25 09:59:59 crc kubenswrapper[4776]: I1125 09:59:59.327616 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35c6cb73-51b4-4258-a6a3-43cba2007a49-utilities\") pod \"35c6cb73-51b4-4258-a6a3-43cba2007a49\" (UID: \"35c6cb73-51b4-4258-a6a3-43cba2007a49\") " Nov 25 09:59:59 crc kubenswrapper[4776]: I1125 09:59:59.329656 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35c6cb73-51b4-4258-a6a3-43cba2007a49-utilities" (OuterVolumeSpecName: "utilities") pod "35c6cb73-51b4-4258-a6a3-43cba2007a49" (UID: "35c6cb73-51b4-4258-a6a3-43cba2007a49"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:59:59 crc kubenswrapper[4776]: I1125 09:59:59.334737 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35c6cb73-51b4-4258-a6a3-43cba2007a49-kube-api-access-9ptr4" (OuterVolumeSpecName: "kube-api-access-9ptr4") pod "35c6cb73-51b4-4258-a6a3-43cba2007a49" (UID: "35c6cb73-51b4-4258-a6a3-43cba2007a49"). InnerVolumeSpecName "kube-api-access-9ptr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:59:59 crc kubenswrapper[4776]: I1125 09:59:59.430133 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ptr4\" (UniqueName: \"kubernetes.io/projected/35c6cb73-51b4-4258-a6a3-43cba2007a49-kube-api-access-9ptr4\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:59 crc kubenswrapper[4776]: I1125 09:59:59.430202 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35c6cb73-51b4-4258-a6a3-43cba2007a49-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:59 crc kubenswrapper[4776]: I1125 09:59:59.731577 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35c6cb73-51b4-4258-a6a3-43cba2007a49-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "35c6cb73-51b4-4258-a6a3-43cba2007a49" (UID: "35c6cb73-51b4-4258-a6a3-43cba2007a49"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:59:59 crc kubenswrapper[4776]: I1125 09:59:59.735762 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35c6cb73-51b4-4258-a6a3-43cba2007a49-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:59 crc kubenswrapper[4776]: I1125 09:59:59.803388 4776 generic.go:334] "Generic (PLEG): container finished" podID="35c6cb73-51b4-4258-a6a3-43cba2007a49" containerID="7e69a75e4bafd241ea1a92a72e22bebaec449e1cc9df49f447302ede65e0c6a1" exitCode=0 Nov 25 09:59:59 crc kubenswrapper[4776]: I1125 09:59:59.803426 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7srcr" event={"ID":"35c6cb73-51b4-4258-a6a3-43cba2007a49","Type":"ContainerDied","Data":"7e69a75e4bafd241ea1a92a72e22bebaec449e1cc9df49f447302ede65e0c6a1"} Nov 25 09:59:59 crc kubenswrapper[4776]: I1125 09:59:59.803452 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7srcr" event={"ID":"35c6cb73-51b4-4258-a6a3-43cba2007a49","Type":"ContainerDied","Data":"3b6da74ac206f11a576fe07c446e4ea2b759aaec0fc311ec21c7dfa092e3dd2e"} Nov 25 09:59:59 crc kubenswrapper[4776]: I1125 09:59:59.803455 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7srcr" Nov 25 09:59:59 crc kubenswrapper[4776]: I1125 09:59:59.803469 4776 scope.go:117] "RemoveContainer" containerID="7e69a75e4bafd241ea1a92a72e22bebaec449e1cc9df49f447302ede65e0c6a1" Nov 25 09:59:59 crc kubenswrapper[4776]: I1125 09:59:59.822704 4776 scope.go:117] "RemoveContainer" containerID="5eb0a8cc5f5c972934bbe6ea04907c09bd15725c2cbd70044b4fd45cadaa36d5" Nov 25 09:59:59 crc kubenswrapper[4776]: I1125 09:59:59.849177 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7srcr"] Nov 25 09:59:59 crc kubenswrapper[4776]: I1125 09:59:59.849834 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7srcr"] Nov 25 09:59:59 crc kubenswrapper[4776]: I1125 09:59:59.851696 4776 scope.go:117] "RemoveContainer" containerID="82af4ae638d838069f48ed767c8475e1b6b134d5ec63edd53731e67e29115215" Nov 25 09:59:59 crc kubenswrapper[4776]: I1125 09:59:59.871647 4776 scope.go:117] "RemoveContainer" containerID="7e69a75e4bafd241ea1a92a72e22bebaec449e1cc9df49f447302ede65e0c6a1" Nov 25 09:59:59 crc kubenswrapper[4776]: E1125 09:59:59.872160 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e69a75e4bafd241ea1a92a72e22bebaec449e1cc9df49f447302ede65e0c6a1\": container with ID starting with 7e69a75e4bafd241ea1a92a72e22bebaec449e1cc9df49f447302ede65e0c6a1 not found: ID does not exist" containerID="7e69a75e4bafd241ea1a92a72e22bebaec449e1cc9df49f447302ede65e0c6a1" Nov 25 09:59:59 crc kubenswrapper[4776]: I1125 09:59:59.872242 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e69a75e4bafd241ea1a92a72e22bebaec449e1cc9df49f447302ede65e0c6a1"} err="failed to get container status \"7e69a75e4bafd241ea1a92a72e22bebaec449e1cc9df49f447302ede65e0c6a1\": rpc error: code = NotFound desc = could not find container \"7e69a75e4bafd241ea1a92a72e22bebaec449e1cc9df49f447302ede65e0c6a1\": container with ID starting with 7e69a75e4bafd241ea1a92a72e22bebaec449e1cc9df49f447302ede65e0c6a1 not found: ID does not exist" Nov 25 09:59:59 crc kubenswrapper[4776]: I1125 09:59:59.872293 4776 scope.go:117] "RemoveContainer" containerID="5eb0a8cc5f5c972934bbe6ea04907c09bd15725c2cbd70044b4fd45cadaa36d5" Nov 25 09:59:59 crc kubenswrapper[4776]: E1125 09:59:59.872660 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5eb0a8cc5f5c972934bbe6ea04907c09bd15725c2cbd70044b4fd45cadaa36d5\": container with ID starting with 5eb0a8cc5f5c972934bbe6ea04907c09bd15725c2cbd70044b4fd45cadaa36d5 not found: ID does not exist" containerID="5eb0a8cc5f5c972934bbe6ea04907c09bd15725c2cbd70044b4fd45cadaa36d5" Nov 25 09:59:59 crc kubenswrapper[4776]: I1125 09:59:59.872726 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5eb0a8cc5f5c972934bbe6ea04907c09bd15725c2cbd70044b4fd45cadaa36d5"} err="failed to get container status \"5eb0a8cc5f5c972934bbe6ea04907c09bd15725c2cbd70044b4fd45cadaa36d5\": rpc error: code = NotFound desc = could not find container \"5eb0a8cc5f5c972934bbe6ea04907c09bd15725c2cbd70044b4fd45cadaa36d5\": container with ID starting with 5eb0a8cc5f5c972934bbe6ea04907c09bd15725c2cbd70044b4fd45cadaa36d5 not found: ID does not exist" Nov 25 09:59:59 crc kubenswrapper[4776]: I1125 09:59:59.872757 4776 scope.go:117] "RemoveContainer" containerID="82af4ae638d838069f48ed767c8475e1b6b134d5ec63edd53731e67e29115215" Nov 25 09:59:59 crc kubenswrapper[4776]: E1125 09:59:59.873031 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82af4ae638d838069f48ed767c8475e1b6b134d5ec63edd53731e67e29115215\": container with ID starting with 82af4ae638d838069f48ed767c8475e1b6b134d5ec63edd53731e67e29115215 not found: ID does not exist" containerID="82af4ae638d838069f48ed767c8475e1b6b134d5ec63edd53731e67e29115215" Nov 25 09:59:59 crc kubenswrapper[4776]: I1125 09:59:59.873056 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82af4ae638d838069f48ed767c8475e1b6b134d5ec63edd53731e67e29115215"} err="failed to get container status \"82af4ae638d838069f48ed767c8475e1b6b134d5ec63edd53731e67e29115215\": rpc error: code = NotFound desc = could not find container \"82af4ae638d838069f48ed767c8475e1b6b134d5ec63edd53731e67e29115215\": container with ID starting with 82af4ae638d838069f48ed767c8475e1b6b134d5ec63edd53731e67e29115215 not found: ID does not exist" Nov 25 10:00:00 crc kubenswrapper[4776]: I1125 10:00:00.168678 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401080-m5xwz"] Nov 25 10:00:00 crc kubenswrapper[4776]: E1125 10:00:00.169016 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35c6cb73-51b4-4258-a6a3-43cba2007a49" containerName="registry-server" Nov 25 10:00:00 crc kubenswrapper[4776]: I1125 10:00:00.169030 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="35c6cb73-51b4-4258-a6a3-43cba2007a49" containerName="registry-server" Nov 25 10:00:00 crc kubenswrapper[4776]: E1125 10:00:00.169055 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35c6cb73-51b4-4258-a6a3-43cba2007a49" containerName="extract-utilities" Nov 25 10:00:00 crc kubenswrapper[4776]: I1125 10:00:00.169079 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="35c6cb73-51b4-4258-a6a3-43cba2007a49" containerName="extract-utilities" Nov 25 10:00:00 crc kubenswrapper[4776]: E1125 10:00:00.169090 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35c6cb73-51b4-4258-a6a3-43cba2007a49" containerName="extract-content" Nov 25 10:00:00 crc kubenswrapper[4776]: I1125 10:00:00.169096 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="35c6cb73-51b4-4258-a6a3-43cba2007a49" containerName="extract-content" Nov 25 10:00:00 crc kubenswrapper[4776]: I1125 10:00:00.169285 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="35c6cb73-51b4-4258-a6a3-43cba2007a49" containerName="registry-server" Nov 25 10:00:00 crc kubenswrapper[4776]: I1125 10:00:00.169759 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-m5xwz" Nov 25 10:00:00 crc kubenswrapper[4776]: I1125 10:00:00.172530 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 10:00:00 crc kubenswrapper[4776]: I1125 10:00:00.172854 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 10:00:00 crc kubenswrapper[4776]: I1125 10:00:00.177734 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401080-m5xwz"] Nov 25 10:00:00 crc kubenswrapper[4776]: I1125 10:00:00.244970 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ad146178-73be-4f2b-a2d1-87905a70e2f0-config-volume\") pod \"collect-profiles-29401080-m5xwz\" (UID: \"ad146178-73be-4f2b-a2d1-87905a70e2f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-m5xwz" Nov 25 10:00:00 crc kubenswrapper[4776]: I1125 10:00:00.245111 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spswp\" (UniqueName: \"kubernetes.io/projected/ad146178-73be-4f2b-a2d1-87905a70e2f0-kube-api-access-spswp\") pod \"collect-profiles-29401080-m5xwz\" (UID: \"ad146178-73be-4f2b-a2d1-87905a70e2f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-m5xwz" Nov 25 10:00:00 crc kubenswrapper[4776]: I1125 10:00:00.245150 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ad146178-73be-4f2b-a2d1-87905a70e2f0-secret-volume\") pod \"collect-profiles-29401080-m5xwz\" (UID: \"ad146178-73be-4f2b-a2d1-87905a70e2f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-m5xwz" Nov 25 10:00:00 crc kubenswrapper[4776]: I1125 10:00:00.346251 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ad146178-73be-4f2b-a2d1-87905a70e2f0-config-volume\") pod \"collect-profiles-29401080-m5xwz\" (UID: \"ad146178-73be-4f2b-a2d1-87905a70e2f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-m5xwz" Nov 25 10:00:00 crc kubenswrapper[4776]: I1125 10:00:00.346369 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spswp\" (UniqueName: \"kubernetes.io/projected/ad146178-73be-4f2b-a2d1-87905a70e2f0-kube-api-access-spswp\") pod \"collect-profiles-29401080-m5xwz\" (UID: \"ad146178-73be-4f2b-a2d1-87905a70e2f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-m5xwz" Nov 25 10:00:00 crc kubenswrapper[4776]: I1125 10:00:00.346415 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ad146178-73be-4f2b-a2d1-87905a70e2f0-secret-volume\") pod \"collect-profiles-29401080-m5xwz\" (UID: \"ad146178-73be-4f2b-a2d1-87905a70e2f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-m5xwz" Nov 25 10:00:00 crc kubenswrapper[4776]: I1125 10:00:00.347292 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ad146178-73be-4f2b-a2d1-87905a70e2f0-config-volume\") pod \"collect-profiles-29401080-m5xwz\" (UID: \"ad146178-73be-4f2b-a2d1-87905a70e2f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-m5xwz" Nov 25 10:00:00 crc kubenswrapper[4776]: I1125 10:00:00.353966 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ad146178-73be-4f2b-a2d1-87905a70e2f0-secret-volume\") pod \"collect-profiles-29401080-m5xwz\" (UID: \"ad146178-73be-4f2b-a2d1-87905a70e2f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-m5xwz" Nov 25 10:00:00 crc kubenswrapper[4776]: I1125 10:00:00.368232 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spswp\" (UniqueName: \"kubernetes.io/projected/ad146178-73be-4f2b-a2d1-87905a70e2f0-kube-api-access-spswp\") pod \"collect-profiles-29401080-m5xwz\" (UID: \"ad146178-73be-4f2b-a2d1-87905a70e2f0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-m5xwz" Nov 25 10:00:00 crc kubenswrapper[4776]: I1125 10:00:00.493033 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-m5xwz" Nov 25 10:00:00 crc kubenswrapper[4776]: I1125 10:00:00.965723 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401080-m5xwz"] Nov 25 10:00:01 crc kubenswrapper[4776]: I1125 10:00:01.679249 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35c6cb73-51b4-4258-a6a3-43cba2007a49" path="/var/lib/kubelet/pods/35c6cb73-51b4-4258-a6a3-43cba2007a49/volumes" Nov 25 10:00:01 crc kubenswrapper[4776]: I1125 10:00:01.823647 4776 generic.go:334] "Generic (PLEG): container finished" podID="ad146178-73be-4f2b-a2d1-87905a70e2f0" containerID="5976e99e2d279f9b29baf268695259602ba8e471cdba2c32bd0654801018f07d" exitCode=0 Nov 25 10:00:01 crc kubenswrapper[4776]: I1125 10:00:01.823693 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-m5xwz" event={"ID":"ad146178-73be-4f2b-a2d1-87905a70e2f0","Type":"ContainerDied","Data":"5976e99e2d279f9b29baf268695259602ba8e471cdba2c32bd0654801018f07d"} Nov 25 10:00:01 crc kubenswrapper[4776]: I1125 10:00:01.823719 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-m5xwz" event={"ID":"ad146178-73be-4f2b-a2d1-87905a70e2f0","Type":"ContainerStarted","Data":"b5bc10d9c041cb5c784e3f018ae558c2f8d236cad2d527754bafa6dba5d7840a"} Nov 25 10:00:03 crc kubenswrapper[4776]: I1125 10:00:03.160782 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-m5xwz" Nov 25 10:00:03 crc kubenswrapper[4776]: I1125 10:00:03.187650 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ad146178-73be-4f2b-a2d1-87905a70e2f0-config-volume\") pod \"ad146178-73be-4f2b-a2d1-87905a70e2f0\" (UID: \"ad146178-73be-4f2b-a2d1-87905a70e2f0\") " Nov 25 10:00:03 crc kubenswrapper[4776]: I1125 10:00:03.187721 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spswp\" (UniqueName: \"kubernetes.io/projected/ad146178-73be-4f2b-a2d1-87905a70e2f0-kube-api-access-spswp\") pod \"ad146178-73be-4f2b-a2d1-87905a70e2f0\" (UID: \"ad146178-73be-4f2b-a2d1-87905a70e2f0\") " Nov 25 10:00:03 crc kubenswrapper[4776]: I1125 10:00:03.187755 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ad146178-73be-4f2b-a2d1-87905a70e2f0-secret-volume\") pod \"ad146178-73be-4f2b-a2d1-87905a70e2f0\" (UID: \"ad146178-73be-4f2b-a2d1-87905a70e2f0\") " Nov 25 10:00:03 crc kubenswrapper[4776]: I1125 10:00:03.188554 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad146178-73be-4f2b-a2d1-87905a70e2f0-config-volume" (OuterVolumeSpecName: "config-volume") pod "ad146178-73be-4f2b-a2d1-87905a70e2f0" (UID: "ad146178-73be-4f2b-a2d1-87905a70e2f0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:00:03 crc kubenswrapper[4776]: I1125 10:00:03.194168 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad146178-73be-4f2b-a2d1-87905a70e2f0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ad146178-73be-4f2b-a2d1-87905a70e2f0" (UID: "ad146178-73be-4f2b-a2d1-87905a70e2f0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:00:03 crc kubenswrapper[4776]: I1125 10:00:03.195691 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad146178-73be-4f2b-a2d1-87905a70e2f0-kube-api-access-spswp" (OuterVolumeSpecName: "kube-api-access-spswp") pod "ad146178-73be-4f2b-a2d1-87905a70e2f0" (UID: "ad146178-73be-4f2b-a2d1-87905a70e2f0"). InnerVolumeSpecName "kube-api-access-spswp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:00:03 crc kubenswrapper[4776]: I1125 10:00:03.289875 4776 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ad146178-73be-4f2b-a2d1-87905a70e2f0-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 10:00:03 crc kubenswrapper[4776]: I1125 10:00:03.289916 4776 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ad146178-73be-4f2b-a2d1-87905a70e2f0-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 10:00:03 crc kubenswrapper[4776]: I1125 10:00:03.289930 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spswp\" (UniqueName: \"kubernetes.io/projected/ad146178-73be-4f2b-a2d1-87905a70e2f0-kube-api-access-spswp\") on node \"crc\" DevicePath \"\"" Nov 25 10:00:03 crc kubenswrapper[4776]: I1125 10:00:03.844286 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-m5xwz" event={"ID":"ad146178-73be-4f2b-a2d1-87905a70e2f0","Type":"ContainerDied","Data":"b5bc10d9c041cb5c784e3f018ae558c2f8d236cad2d527754bafa6dba5d7840a"} Nov 25 10:00:03 crc kubenswrapper[4776]: I1125 10:00:03.844722 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5bc10d9c041cb5c784e3f018ae558c2f8d236cad2d527754bafa6dba5d7840a" Nov 25 10:00:03 crc kubenswrapper[4776]: I1125 10:00:03.844534 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-m5xwz" Nov 25 10:00:04 crc kubenswrapper[4776]: I1125 10:00:04.255432 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401035-dht2z"] Nov 25 10:00:04 crc kubenswrapper[4776]: I1125 10:00:04.262474 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401035-dht2z"] Nov 25 10:00:05 crc kubenswrapper[4776]: I1125 10:00:05.675044 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b51fc9b-c461-4e80-9a56-4f0e43e02399" path="/var/lib/kubelet/pods/4b51fc9b-c461-4e80-9a56-4f0e43e02399/volumes" Nov 25 10:00:17 crc kubenswrapper[4776]: I1125 10:00:17.818710 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:00:17 crc kubenswrapper[4776]: I1125 10:00:17.819559 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:00:29 crc kubenswrapper[4776]: I1125 10:00:29.692765 4776 scope.go:117] "RemoveContainer" containerID="3f95652774809f6ee88562122ca9c3788000edd112955de44fc0382bad0e8c84" Nov 25 10:00:47 crc kubenswrapper[4776]: I1125 10:00:47.818825 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:00:47 crc kubenswrapper[4776]: I1125 10:00:47.819567 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:00:47 crc kubenswrapper[4776]: I1125 10:00:47.819635 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 10:00:47 crc kubenswrapper[4776]: I1125 10:00:47.820545 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7c8b8b54f516cb84cd523fbee04ee9767e3fadd57ce4fa6310ba05067cb37d20"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:00:47 crc kubenswrapper[4776]: I1125 10:00:47.820649 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://7c8b8b54f516cb84cd523fbee04ee9767e3fadd57ce4fa6310ba05067cb37d20" gracePeriod=600 Nov 25 10:00:48 crc kubenswrapper[4776]: I1125 10:00:48.218735 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="7c8b8b54f516cb84cd523fbee04ee9767e3fadd57ce4fa6310ba05067cb37d20" exitCode=0 Nov 25 10:00:48 crc kubenswrapper[4776]: I1125 10:00:48.218819 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"7c8b8b54f516cb84cd523fbee04ee9767e3fadd57ce4fa6310ba05067cb37d20"} Nov 25 10:00:48 crc kubenswrapper[4776]: I1125 10:00:48.219237 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288"} Nov 25 10:00:48 crc kubenswrapper[4776]: I1125 10:00:48.219307 4776 scope.go:117] "RemoveContainer" containerID="670b4a02206417b8340fd14867ce56f08800a3b8c3928b901c8621339231c722" Nov 25 10:03:17 crc kubenswrapper[4776]: I1125 10:03:17.818605 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:03:17 crc kubenswrapper[4776]: I1125 10:03:17.819621 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:03:47 crc kubenswrapper[4776]: I1125 10:03:47.818680 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:03:47 crc kubenswrapper[4776]: I1125 10:03:47.819480 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:04:17 crc kubenswrapper[4776]: I1125 10:04:17.818126 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:04:17 crc kubenswrapper[4776]: I1125 10:04:17.818635 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:04:17 crc kubenswrapper[4776]: I1125 10:04:17.818683 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 10:04:17 crc kubenswrapper[4776]: I1125 10:04:17.819441 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:04:17 crc kubenswrapper[4776]: I1125 10:04:17.819509 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288" gracePeriod=600 Nov 25 10:04:17 crc kubenswrapper[4776]: E1125 10:04:17.939563 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:04:18 crc kubenswrapper[4776]: I1125 10:04:18.224934 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288" exitCode=0 Nov 25 10:04:18 crc kubenswrapper[4776]: I1125 10:04:18.225028 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288"} Nov 25 10:04:18 crc kubenswrapper[4776]: I1125 10:04:18.225303 4776 scope.go:117] "RemoveContainer" containerID="7c8b8b54f516cb84cd523fbee04ee9767e3fadd57ce4fa6310ba05067cb37d20" Nov 25 10:04:18 crc kubenswrapper[4776]: I1125 10:04:18.225857 4776 scope.go:117] "RemoveContainer" containerID="4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288" Nov 25 10:04:18 crc kubenswrapper[4776]: E1125 10:04:18.226108 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:04:31 crc kubenswrapper[4776]: I1125 10:04:31.662723 4776 scope.go:117] "RemoveContainer" containerID="4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288" Nov 25 10:04:31 crc kubenswrapper[4776]: E1125 10:04:31.663431 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:04:43 crc kubenswrapper[4776]: I1125 10:04:43.662956 4776 scope.go:117] "RemoveContainer" containerID="4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288" Nov 25 10:04:43 crc kubenswrapper[4776]: E1125 10:04:43.663771 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:04:55 crc kubenswrapper[4776]: I1125 10:04:55.667148 4776 scope.go:117] "RemoveContainer" containerID="4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288" Nov 25 10:04:55 crc kubenswrapper[4776]: E1125 10:04:55.669479 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:05:10 crc kubenswrapper[4776]: I1125 10:05:10.662303 4776 scope.go:117] "RemoveContainer" containerID="4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288" Nov 25 10:05:10 crc kubenswrapper[4776]: E1125 10:05:10.663039 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:05:23 crc kubenswrapper[4776]: I1125 10:05:23.664206 4776 scope.go:117] "RemoveContainer" containerID="4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288" Nov 25 10:05:23 crc kubenswrapper[4776]: E1125 10:05:23.665221 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:05:37 crc kubenswrapper[4776]: I1125 10:05:37.663759 4776 scope.go:117] "RemoveContainer" containerID="4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288" Nov 25 10:05:37 crc kubenswrapper[4776]: E1125 10:05:37.665633 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:05:50 crc kubenswrapper[4776]: I1125 10:05:50.662310 4776 scope.go:117] "RemoveContainer" containerID="4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288" Nov 25 10:05:50 crc kubenswrapper[4776]: E1125 10:05:50.663238 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:06:05 crc kubenswrapper[4776]: I1125 10:06:05.667269 4776 scope.go:117] "RemoveContainer" containerID="4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288" Nov 25 10:06:05 crc kubenswrapper[4776]: E1125 10:06:05.668187 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:06:20 crc kubenswrapper[4776]: I1125 10:06:20.662392 4776 scope.go:117] "RemoveContainer" containerID="4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288" Nov 25 10:06:20 crc kubenswrapper[4776]: E1125 10:06:20.663152 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:06:31 crc kubenswrapper[4776]: I1125 10:06:31.662484 4776 scope.go:117] "RemoveContainer" containerID="4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288" Nov 25 10:06:31 crc kubenswrapper[4776]: E1125 10:06:31.663607 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:06:46 crc kubenswrapper[4776]: I1125 10:06:46.662770 4776 scope.go:117] "RemoveContainer" containerID="4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288" Nov 25 10:06:46 crc kubenswrapper[4776]: E1125 10:06:46.663834 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:07:00 crc kubenswrapper[4776]: I1125 10:07:00.662161 4776 scope.go:117] "RemoveContainer" containerID="4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288" Nov 25 10:07:00 crc kubenswrapper[4776]: E1125 10:07:00.662919 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:07:13 crc kubenswrapper[4776]: I1125 10:07:13.663014 4776 scope.go:117] "RemoveContainer" containerID="4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288" Nov 25 10:07:13 crc kubenswrapper[4776]: E1125 10:07:13.663913 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:07:25 crc kubenswrapper[4776]: I1125 10:07:25.669261 4776 scope.go:117] "RemoveContainer" containerID="4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288" Nov 25 10:07:25 crc kubenswrapper[4776]: E1125 10:07:25.670392 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:07:39 crc kubenswrapper[4776]: I1125 10:07:39.663392 4776 scope.go:117] "RemoveContainer" containerID="4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288" Nov 25 10:07:39 crc kubenswrapper[4776]: E1125 10:07:39.664577 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:07:54 crc kubenswrapper[4776]: I1125 10:07:54.663302 4776 scope.go:117] "RemoveContainer" containerID="4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288" Nov 25 10:07:54 crc kubenswrapper[4776]: E1125 10:07:54.664206 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:08:07 crc kubenswrapper[4776]: I1125 10:08:07.662275 4776 scope.go:117] "RemoveContainer" containerID="4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288" Nov 25 10:08:07 crc kubenswrapper[4776]: E1125 10:08:07.663123 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:08:21 crc kubenswrapper[4776]: I1125 10:08:21.662719 4776 scope.go:117] "RemoveContainer" containerID="4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288" Nov 25 10:08:21 crc kubenswrapper[4776]: E1125 10:08:21.663842 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:08:34 crc kubenswrapper[4776]: I1125 10:08:34.662655 4776 scope.go:117] "RemoveContainer" containerID="4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288" Nov 25 10:08:34 crc kubenswrapper[4776]: E1125 10:08:34.663504 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:08:46 crc kubenswrapper[4776]: I1125 10:08:46.661738 4776 scope.go:117] "RemoveContainer" containerID="4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288" Nov 25 10:08:46 crc kubenswrapper[4776]: E1125 10:08:46.662488 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:08:58 crc kubenswrapper[4776]: I1125 10:08:58.661940 4776 scope.go:117] "RemoveContainer" containerID="4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288" Nov 25 10:08:58 crc kubenswrapper[4776]: E1125 10:08:58.662537 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:08:59 crc kubenswrapper[4776]: I1125 10:08:59.897481 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w8mp6"] Nov 25 10:08:59 crc kubenswrapper[4776]: E1125 10:08:59.897789 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad146178-73be-4f2b-a2d1-87905a70e2f0" containerName="collect-profiles" Nov 25 10:08:59 crc kubenswrapper[4776]: I1125 10:08:59.897802 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad146178-73be-4f2b-a2d1-87905a70e2f0" containerName="collect-profiles" Nov 25 10:08:59 crc kubenswrapper[4776]: I1125 10:08:59.897958 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad146178-73be-4f2b-a2d1-87905a70e2f0" containerName="collect-profiles" Nov 25 10:08:59 crc kubenswrapper[4776]: I1125 10:08:59.899180 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w8mp6" Nov 25 10:08:59 crc kubenswrapper[4776]: I1125 10:08:59.909839 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w8mp6"] Nov 25 10:09:00 crc kubenswrapper[4776]: I1125 10:09:00.014702 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhfnl\" (UniqueName: \"kubernetes.io/projected/0d809729-2b3c-43ce-b84b-0319ee628fbd-kube-api-access-hhfnl\") pod \"redhat-operators-w8mp6\" (UID: \"0d809729-2b3c-43ce-b84b-0319ee628fbd\") " pod="openshift-marketplace/redhat-operators-w8mp6" Nov 25 10:09:00 crc kubenswrapper[4776]: I1125 10:09:00.014763 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d809729-2b3c-43ce-b84b-0319ee628fbd-catalog-content\") pod \"redhat-operators-w8mp6\" (UID: \"0d809729-2b3c-43ce-b84b-0319ee628fbd\") " pod="openshift-marketplace/redhat-operators-w8mp6" Nov 25 10:09:00 crc kubenswrapper[4776]: I1125 10:09:00.014875 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d809729-2b3c-43ce-b84b-0319ee628fbd-utilities\") pod \"redhat-operators-w8mp6\" (UID: \"0d809729-2b3c-43ce-b84b-0319ee628fbd\") " pod="openshift-marketplace/redhat-operators-w8mp6" Nov 25 10:09:00 crc kubenswrapper[4776]: I1125 10:09:00.116051 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d809729-2b3c-43ce-b84b-0319ee628fbd-catalog-content\") pod \"redhat-operators-w8mp6\" (UID: \"0d809729-2b3c-43ce-b84b-0319ee628fbd\") " pod="openshift-marketplace/redhat-operators-w8mp6" Nov 25 10:09:00 crc kubenswrapper[4776]: I1125 10:09:00.116145 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d809729-2b3c-43ce-b84b-0319ee628fbd-utilities\") pod \"redhat-operators-w8mp6\" (UID: \"0d809729-2b3c-43ce-b84b-0319ee628fbd\") " pod="openshift-marketplace/redhat-operators-w8mp6" Nov 25 10:09:00 crc kubenswrapper[4776]: I1125 10:09:00.116215 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhfnl\" (UniqueName: \"kubernetes.io/projected/0d809729-2b3c-43ce-b84b-0319ee628fbd-kube-api-access-hhfnl\") pod \"redhat-operators-w8mp6\" (UID: \"0d809729-2b3c-43ce-b84b-0319ee628fbd\") " pod="openshift-marketplace/redhat-operators-w8mp6" Nov 25 10:09:00 crc kubenswrapper[4776]: I1125 10:09:00.116671 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d809729-2b3c-43ce-b84b-0319ee628fbd-catalog-content\") pod \"redhat-operators-w8mp6\" (UID: \"0d809729-2b3c-43ce-b84b-0319ee628fbd\") " pod="openshift-marketplace/redhat-operators-w8mp6" Nov 25 10:09:00 crc kubenswrapper[4776]: I1125 10:09:00.116709 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d809729-2b3c-43ce-b84b-0319ee628fbd-utilities\") pod \"redhat-operators-w8mp6\" (UID: \"0d809729-2b3c-43ce-b84b-0319ee628fbd\") " pod="openshift-marketplace/redhat-operators-w8mp6" Nov 25 10:09:00 crc kubenswrapper[4776]: I1125 10:09:00.146139 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhfnl\" (UniqueName: \"kubernetes.io/projected/0d809729-2b3c-43ce-b84b-0319ee628fbd-kube-api-access-hhfnl\") pod \"redhat-operators-w8mp6\" (UID: \"0d809729-2b3c-43ce-b84b-0319ee628fbd\") " pod="openshift-marketplace/redhat-operators-w8mp6" Nov 25 10:09:00 crc kubenswrapper[4776]: I1125 10:09:00.216264 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w8mp6" Nov 25 10:09:00 crc kubenswrapper[4776]: I1125 10:09:00.452946 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w8mp6"] Nov 25 10:09:00 crc kubenswrapper[4776]: I1125 10:09:00.568025 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8mp6" event={"ID":"0d809729-2b3c-43ce-b84b-0319ee628fbd","Type":"ContainerStarted","Data":"b097f6de3edcea92f9770d6354905d4a57238f3bdb99c7f740f4697361f03660"} Nov 25 10:09:01 crc kubenswrapper[4776]: I1125 10:09:01.576359 4776 generic.go:334] "Generic (PLEG): container finished" podID="0d809729-2b3c-43ce-b84b-0319ee628fbd" containerID="edd7afd0c0fd9c81f0a3196054f7667a0438e07020cf5ce9670e7fe3b3584eda" exitCode=0 Nov 25 10:09:01 crc kubenswrapper[4776]: I1125 10:09:01.576404 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8mp6" event={"ID":"0d809729-2b3c-43ce-b84b-0319ee628fbd","Type":"ContainerDied","Data":"edd7afd0c0fd9c81f0a3196054f7667a0438e07020cf5ce9670e7fe3b3584eda"} Nov 25 10:09:01 crc kubenswrapper[4776]: I1125 10:09:01.578574 4776 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 10:09:02 crc kubenswrapper[4776]: I1125 10:09:02.584061 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8mp6" event={"ID":"0d809729-2b3c-43ce-b84b-0319ee628fbd","Type":"ContainerStarted","Data":"63875fc6484dbc3583129fab686d149054e80d705ce3cdcd7658dea799861266"} Nov 25 10:09:03 crc kubenswrapper[4776]: I1125 10:09:03.595641 4776 generic.go:334] "Generic (PLEG): container finished" podID="0d809729-2b3c-43ce-b84b-0319ee628fbd" containerID="63875fc6484dbc3583129fab686d149054e80d705ce3cdcd7658dea799861266" exitCode=0 Nov 25 10:09:03 crc kubenswrapper[4776]: I1125 10:09:03.595788 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8mp6" event={"ID":"0d809729-2b3c-43ce-b84b-0319ee628fbd","Type":"ContainerDied","Data":"63875fc6484dbc3583129fab686d149054e80d705ce3cdcd7658dea799861266"} Nov 25 10:09:04 crc kubenswrapper[4776]: I1125 10:09:04.604254 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8mp6" event={"ID":"0d809729-2b3c-43ce-b84b-0319ee628fbd","Type":"ContainerStarted","Data":"ef776026e31b88d2d21aaea4c0710752f6d83730a32e71d021c72470e563d345"} Nov 25 10:09:10 crc kubenswrapper[4776]: I1125 10:09:10.217253 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-w8mp6" Nov 25 10:09:10 crc kubenswrapper[4776]: I1125 10:09:10.218390 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-w8mp6" Nov 25 10:09:10 crc kubenswrapper[4776]: I1125 10:09:10.268781 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-w8mp6" Nov 25 10:09:10 crc kubenswrapper[4776]: I1125 10:09:10.284432 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w8mp6" podStartSLOduration=8.829400373 podStartE2EDuration="11.28441585s" podCreationTimestamp="2025-11-25 10:08:59 +0000 UTC" firstStartedPulling="2025-11-25 10:09:01.578299511 +0000 UTC m=+2686.619359064" lastFinishedPulling="2025-11-25 10:09:04.033314978 +0000 UTC m=+2689.074374541" observedRunningTime="2025-11-25 10:09:04.631010977 +0000 UTC m=+2689.672070530" watchObservedRunningTime="2025-11-25 10:09:10.28441585 +0000 UTC m=+2695.325475413" Nov 25 10:09:10 crc kubenswrapper[4776]: I1125 10:09:10.682505 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-w8mp6" Nov 25 10:09:10 crc kubenswrapper[4776]: I1125 10:09:10.739926 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w8mp6"] Nov 25 10:09:11 crc kubenswrapper[4776]: I1125 10:09:11.663238 4776 scope.go:117] "RemoveContainer" containerID="4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288" Nov 25 10:09:11 crc kubenswrapper[4776]: E1125 10:09:11.663678 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:09:12 crc kubenswrapper[4776]: I1125 10:09:12.659919 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-w8mp6" podUID="0d809729-2b3c-43ce-b84b-0319ee628fbd" containerName="registry-server" containerID="cri-o://ef776026e31b88d2d21aaea4c0710752f6d83730a32e71d021c72470e563d345" gracePeriod=2 Nov 25 10:09:13 crc kubenswrapper[4776]: I1125 10:09:13.106491 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w8mp6" Nov 25 10:09:13 crc kubenswrapper[4776]: I1125 10:09:13.213413 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d809729-2b3c-43ce-b84b-0319ee628fbd-utilities\") pod \"0d809729-2b3c-43ce-b84b-0319ee628fbd\" (UID: \"0d809729-2b3c-43ce-b84b-0319ee628fbd\") " Nov 25 10:09:13 crc kubenswrapper[4776]: I1125 10:09:13.213556 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d809729-2b3c-43ce-b84b-0319ee628fbd-catalog-content\") pod \"0d809729-2b3c-43ce-b84b-0319ee628fbd\" (UID: \"0d809729-2b3c-43ce-b84b-0319ee628fbd\") " Nov 25 10:09:13 crc kubenswrapper[4776]: I1125 10:09:13.213653 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhfnl\" (UniqueName: \"kubernetes.io/projected/0d809729-2b3c-43ce-b84b-0319ee628fbd-kube-api-access-hhfnl\") pod \"0d809729-2b3c-43ce-b84b-0319ee628fbd\" (UID: \"0d809729-2b3c-43ce-b84b-0319ee628fbd\") " Nov 25 10:09:13 crc kubenswrapper[4776]: I1125 10:09:13.214757 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d809729-2b3c-43ce-b84b-0319ee628fbd-utilities" (OuterVolumeSpecName: "utilities") pod "0d809729-2b3c-43ce-b84b-0319ee628fbd" (UID: "0d809729-2b3c-43ce-b84b-0319ee628fbd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:09:13 crc kubenswrapper[4776]: I1125 10:09:13.215264 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d809729-2b3c-43ce-b84b-0319ee628fbd-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:13 crc kubenswrapper[4776]: I1125 10:09:13.218835 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d809729-2b3c-43ce-b84b-0319ee628fbd-kube-api-access-hhfnl" (OuterVolumeSpecName: "kube-api-access-hhfnl") pod "0d809729-2b3c-43ce-b84b-0319ee628fbd" (UID: "0d809729-2b3c-43ce-b84b-0319ee628fbd"). InnerVolumeSpecName "kube-api-access-hhfnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:09:13 crc kubenswrapper[4776]: I1125 10:09:13.317008 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhfnl\" (UniqueName: \"kubernetes.io/projected/0d809729-2b3c-43ce-b84b-0319ee628fbd-kube-api-access-hhfnl\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:13 crc kubenswrapper[4776]: I1125 10:09:13.668395 4776 generic.go:334] "Generic (PLEG): container finished" podID="0d809729-2b3c-43ce-b84b-0319ee628fbd" containerID="ef776026e31b88d2d21aaea4c0710752f6d83730a32e71d021c72470e563d345" exitCode=0 Nov 25 10:09:13 crc kubenswrapper[4776]: I1125 10:09:13.668515 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w8mp6" Nov 25 10:09:13 crc kubenswrapper[4776]: I1125 10:09:13.683092 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8mp6" event={"ID":"0d809729-2b3c-43ce-b84b-0319ee628fbd","Type":"ContainerDied","Data":"ef776026e31b88d2d21aaea4c0710752f6d83730a32e71d021c72470e563d345"} Nov 25 10:09:13 crc kubenswrapper[4776]: I1125 10:09:13.683136 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w8mp6" event={"ID":"0d809729-2b3c-43ce-b84b-0319ee628fbd","Type":"ContainerDied","Data":"b097f6de3edcea92f9770d6354905d4a57238f3bdb99c7f740f4697361f03660"} Nov 25 10:09:13 crc kubenswrapper[4776]: I1125 10:09:13.683155 4776 scope.go:117] "RemoveContainer" containerID="ef776026e31b88d2d21aaea4c0710752f6d83730a32e71d021c72470e563d345" Nov 25 10:09:13 crc kubenswrapper[4776]: I1125 10:09:13.714546 4776 scope.go:117] "RemoveContainer" containerID="63875fc6484dbc3583129fab686d149054e80d705ce3cdcd7658dea799861266" Nov 25 10:09:13 crc kubenswrapper[4776]: I1125 10:09:13.733690 4776 scope.go:117] "RemoveContainer" containerID="edd7afd0c0fd9c81f0a3196054f7667a0438e07020cf5ce9670e7fe3b3584eda" Nov 25 10:09:13 crc kubenswrapper[4776]: I1125 10:09:13.763338 4776 scope.go:117] "RemoveContainer" containerID="ef776026e31b88d2d21aaea4c0710752f6d83730a32e71d021c72470e563d345" Nov 25 10:09:13 crc kubenswrapper[4776]: E1125 10:09:13.763828 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef776026e31b88d2d21aaea4c0710752f6d83730a32e71d021c72470e563d345\": container with ID starting with ef776026e31b88d2d21aaea4c0710752f6d83730a32e71d021c72470e563d345 not found: ID does not exist" containerID="ef776026e31b88d2d21aaea4c0710752f6d83730a32e71d021c72470e563d345" Nov 25 10:09:13 crc kubenswrapper[4776]: I1125 10:09:13.763871 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef776026e31b88d2d21aaea4c0710752f6d83730a32e71d021c72470e563d345"} err="failed to get container status \"ef776026e31b88d2d21aaea4c0710752f6d83730a32e71d021c72470e563d345\": rpc error: code = NotFound desc = could not find container \"ef776026e31b88d2d21aaea4c0710752f6d83730a32e71d021c72470e563d345\": container with ID starting with ef776026e31b88d2d21aaea4c0710752f6d83730a32e71d021c72470e563d345 not found: ID does not exist" Nov 25 10:09:13 crc kubenswrapper[4776]: I1125 10:09:13.763895 4776 scope.go:117] "RemoveContainer" containerID="63875fc6484dbc3583129fab686d149054e80d705ce3cdcd7658dea799861266" Nov 25 10:09:13 crc kubenswrapper[4776]: E1125 10:09:13.764281 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63875fc6484dbc3583129fab686d149054e80d705ce3cdcd7658dea799861266\": container with ID starting with 63875fc6484dbc3583129fab686d149054e80d705ce3cdcd7658dea799861266 not found: ID does not exist" containerID="63875fc6484dbc3583129fab686d149054e80d705ce3cdcd7658dea799861266" Nov 25 10:09:13 crc kubenswrapper[4776]: I1125 10:09:13.764318 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63875fc6484dbc3583129fab686d149054e80d705ce3cdcd7658dea799861266"} err="failed to get container status \"63875fc6484dbc3583129fab686d149054e80d705ce3cdcd7658dea799861266\": rpc error: code = NotFound desc = could not find container \"63875fc6484dbc3583129fab686d149054e80d705ce3cdcd7658dea799861266\": container with ID starting with 63875fc6484dbc3583129fab686d149054e80d705ce3cdcd7658dea799861266 not found: ID does not exist" Nov 25 10:09:13 crc kubenswrapper[4776]: I1125 10:09:13.764342 4776 scope.go:117] "RemoveContainer" containerID="edd7afd0c0fd9c81f0a3196054f7667a0438e07020cf5ce9670e7fe3b3584eda" Nov 25 10:09:13 crc kubenswrapper[4776]: E1125 10:09:13.764625 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edd7afd0c0fd9c81f0a3196054f7667a0438e07020cf5ce9670e7fe3b3584eda\": container with ID starting with edd7afd0c0fd9c81f0a3196054f7667a0438e07020cf5ce9670e7fe3b3584eda not found: ID does not exist" containerID="edd7afd0c0fd9c81f0a3196054f7667a0438e07020cf5ce9670e7fe3b3584eda" Nov 25 10:09:13 crc kubenswrapper[4776]: I1125 10:09:13.764653 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edd7afd0c0fd9c81f0a3196054f7667a0438e07020cf5ce9670e7fe3b3584eda"} err="failed to get container status \"edd7afd0c0fd9c81f0a3196054f7667a0438e07020cf5ce9670e7fe3b3584eda\": rpc error: code = NotFound desc = could not find container \"edd7afd0c0fd9c81f0a3196054f7667a0438e07020cf5ce9670e7fe3b3584eda\": container with ID starting with edd7afd0c0fd9c81f0a3196054f7667a0438e07020cf5ce9670e7fe3b3584eda not found: ID does not exist" Nov 25 10:09:13 crc kubenswrapper[4776]: I1125 10:09:13.901706 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d809729-2b3c-43ce-b84b-0319ee628fbd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0d809729-2b3c-43ce-b84b-0319ee628fbd" (UID: "0d809729-2b3c-43ce-b84b-0319ee628fbd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:09:13 crc kubenswrapper[4776]: I1125 10:09:13.926222 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d809729-2b3c-43ce-b84b-0319ee628fbd-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:14 crc kubenswrapper[4776]: I1125 10:09:14.002089 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w8mp6"] Nov 25 10:09:14 crc kubenswrapper[4776]: I1125 10:09:14.012279 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-w8mp6"] Nov 25 10:09:15 crc kubenswrapper[4776]: I1125 10:09:15.678307 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d809729-2b3c-43ce-b84b-0319ee628fbd" path="/var/lib/kubelet/pods/0d809729-2b3c-43ce-b84b-0319ee628fbd/volumes" Nov 25 10:09:26 crc kubenswrapper[4776]: I1125 10:09:26.662345 4776 scope.go:117] "RemoveContainer" containerID="4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288" Nov 25 10:09:27 crc kubenswrapper[4776]: I1125 10:09:27.785493 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"41c62a00c2902589260314ef41fa8e27ca25f86c7f781c0b248f7cc291cf5f53"} Nov 25 10:09:33 crc kubenswrapper[4776]: I1125 10:09:33.749366 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2llhj"] Nov 25 10:09:33 crc kubenswrapper[4776]: E1125 10:09:33.750171 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d809729-2b3c-43ce-b84b-0319ee628fbd" containerName="extract-utilities" Nov 25 10:09:33 crc kubenswrapper[4776]: I1125 10:09:33.750183 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d809729-2b3c-43ce-b84b-0319ee628fbd" containerName="extract-utilities" Nov 25 10:09:33 crc kubenswrapper[4776]: E1125 10:09:33.750196 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d809729-2b3c-43ce-b84b-0319ee628fbd" containerName="extract-content" Nov 25 10:09:33 crc kubenswrapper[4776]: I1125 10:09:33.750202 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d809729-2b3c-43ce-b84b-0319ee628fbd" containerName="extract-content" Nov 25 10:09:33 crc kubenswrapper[4776]: E1125 10:09:33.750212 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d809729-2b3c-43ce-b84b-0319ee628fbd" containerName="registry-server" Nov 25 10:09:33 crc kubenswrapper[4776]: I1125 10:09:33.750218 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d809729-2b3c-43ce-b84b-0319ee628fbd" containerName="registry-server" Nov 25 10:09:33 crc kubenswrapper[4776]: I1125 10:09:33.750347 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d809729-2b3c-43ce-b84b-0319ee628fbd" containerName="registry-server" Nov 25 10:09:33 crc kubenswrapper[4776]: I1125 10:09:33.751290 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2llhj" Nov 25 10:09:33 crc kubenswrapper[4776]: I1125 10:09:33.763119 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2llhj"] Nov 25 10:09:33 crc kubenswrapper[4776]: I1125 10:09:33.926535 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4fda217-d8e5-49bd-802a-f0588ade20a4-catalog-content\") pod \"redhat-marketplace-2llhj\" (UID: \"f4fda217-d8e5-49bd-802a-f0588ade20a4\") " pod="openshift-marketplace/redhat-marketplace-2llhj" Nov 25 10:09:33 crc kubenswrapper[4776]: I1125 10:09:33.926589 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hrqn\" (UniqueName: \"kubernetes.io/projected/f4fda217-d8e5-49bd-802a-f0588ade20a4-kube-api-access-2hrqn\") pod \"redhat-marketplace-2llhj\" (UID: \"f4fda217-d8e5-49bd-802a-f0588ade20a4\") " pod="openshift-marketplace/redhat-marketplace-2llhj" Nov 25 10:09:33 crc kubenswrapper[4776]: I1125 10:09:33.926738 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4fda217-d8e5-49bd-802a-f0588ade20a4-utilities\") pod \"redhat-marketplace-2llhj\" (UID: \"f4fda217-d8e5-49bd-802a-f0588ade20a4\") " pod="openshift-marketplace/redhat-marketplace-2llhj" Nov 25 10:09:34 crc kubenswrapper[4776]: I1125 10:09:34.028228 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4fda217-d8e5-49bd-802a-f0588ade20a4-utilities\") pod \"redhat-marketplace-2llhj\" (UID: \"f4fda217-d8e5-49bd-802a-f0588ade20a4\") " pod="openshift-marketplace/redhat-marketplace-2llhj" Nov 25 10:09:34 crc kubenswrapper[4776]: I1125 10:09:34.028569 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4fda217-d8e5-49bd-802a-f0588ade20a4-catalog-content\") pod \"redhat-marketplace-2llhj\" (UID: \"f4fda217-d8e5-49bd-802a-f0588ade20a4\") " pod="openshift-marketplace/redhat-marketplace-2llhj" Nov 25 10:09:34 crc kubenswrapper[4776]: I1125 10:09:34.028593 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hrqn\" (UniqueName: \"kubernetes.io/projected/f4fda217-d8e5-49bd-802a-f0588ade20a4-kube-api-access-2hrqn\") pod \"redhat-marketplace-2llhj\" (UID: \"f4fda217-d8e5-49bd-802a-f0588ade20a4\") " pod="openshift-marketplace/redhat-marketplace-2llhj" Nov 25 10:09:34 crc kubenswrapper[4776]: I1125 10:09:34.028677 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4fda217-d8e5-49bd-802a-f0588ade20a4-utilities\") pod \"redhat-marketplace-2llhj\" (UID: \"f4fda217-d8e5-49bd-802a-f0588ade20a4\") " pod="openshift-marketplace/redhat-marketplace-2llhj" Nov 25 10:09:34 crc kubenswrapper[4776]: I1125 10:09:34.029170 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4fda217-d8e5-49bd-802a-f0588ade20a4-catalog-content\") pod \"redhat-marketplace-2llhj\" (UID: \"f4fda217-d8e5-49bd-802a-f0588ade20a4\") " pod="openshift-marketplace/redhat-marketplace-2llhj" Nov 25 10:09:34 crc kubenswrapper[4776]: I1125 10:09:34.055178 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hrqn\" (UniqueName: \"kubernetes.io/projected/f4fda217-d8e5-49bd-802a-f0588ade20a4-kube-api-access-2hrqn\") pod \"redhat-marketplace-2llhj\" (UID: \"f4fda217-d8e5-49bd-802a-f0588ade20a4\") " pod="openshift-marketplace/redhat-marketplace-2llhj" Nov 25 10:09:34 crc kubenswrapper[4776]: I1125 10:09:34.067615 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2llhj" Nov 25 10:09:34 crc kubenswrapper[4776]: I1125 10:09:34.516744 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2llhj"] Nov 25 10:09:34 crc kubenswrapper[4776]: I1125 10:09:34.842217 4776 generic.go:334] "Generic (PLEG): container finished" podID="f4fda217-d8e5-49bd-802a-f0588ade20a4" containerID="59ccf84bb212af568b15a093b07a95d19b9e07a5d83880fc1e8a3f4060fb5afb" exitCode=0 Nov 25 10:09:34 crc kubenswrapper[4776]: I1125 10:09:34.842342 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2llhj" event={"ID":"f4fda217-d8e5-49bd-802a-f0588ade20a4","Type":"ContainerDied","Data":"59ccf84bb212af568b15a093b07a95d19b9e07a5d83880fc1e8a3f4060fb5afb"} Nov 25 10:09:34 crc kubenswrapper[4776]: I1125 10:09:34.842542 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2llhj" event={"ID":"f4fda217-d8e5-49bd-802a-f0588ade20a4","Type":"ContainerStarted","Data":"1c870fc227c26e66509b38df1edcf593ffeb37521befe2f3f0fff61293eab11f"} Nov 25 10:09:35 crc kubenswrapper[4776]: I1125 10:09:35.860691 4776 generic.go:334] "Generic (PLEG): container finished" podID="f4fda217-d8e5-49bd-802a-f0588ade20a4" containerID="0c925c7a71238a65aab798c23d8708f77d12d9a79ad3bb8c5ce6702bfe227cc8" exitCode=0 Nov 25 10:09:35 crc kubenswrapper[4776]: I1125 10:09:35.861187 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2llhj" event={"ID":"f4fda217-d8e5-49bd-802a-f0588ade20a4","Type":"ContainerDied","Data":"0c925c7a71238a65aab798c23d8708f77d12d9a79ad3bb8c5ce6702bfe227cc8"} Nov 25 10:09:36 crc kubenswrapper[4776]: I1125 10:09:36.872977 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2llhj" event={"ID":"f4fda217-d8e5-49bd-802a-f0588ade20a4","Type":"ContainerStarted","Data":"9d3c05ec1aa079c004adf21e080afaa4e7fc6ab73e067f31eab5fdcd762358f2"} Nov 25 10:09:36 crc kubenswrapper[4776]: I1125 10:09:36.904809 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2llhj" podStartSLOduration=2.244079973 podStartE2EDuration="3.904788164s" podCreationTimestamp="2025-11-25 10:09:33 +0000 UTC" firstStartedPulling="2025-11-25 10:09:34.843464173 +0000 UTC m=+2719.884523736" lastFinishedPulling="2025-11-25 10:09:36.504172374 +0000 UTC m=+2721.545231927" observedRunningTime="2025-11-25 10:09:36.895021319 +0000 UTC m=+2721.936080912" watchObservedRunningTime="2025-11-25 10:09:36.904788164 +0000 UTC m=+2721.945847727" Nov 25 10:09:44 crc kubenswrapper[4776]: I1125 10:09:44.067917 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2llhj" Nov 25 10:09:44 crc kubenswrapper[4776]: I1125 10:09:44.068527 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2llhj" Nov 25 10:09:44 crc kubenswrapper[4776]: I1125 10:09:44.114045 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2llhj" Nov 25 10:09:44 crc kubenswrapper[4776]: I1125 10:09:44.999599 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2llhj" Nov 25 10:09:45 crc kubenswrapper[4776]: I1125 10:09:45.052663 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2llhj"] Nov 25 10:09:46 crc kubenswrapper[4776]: I1125 10:09:46.974664 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2llhj" podUID="f4fda217-d8e5-49bd-802a-f0588ade20a4" containerName="registry-server" containerID="cri-o://9d3c05ec1aa079c004adf21e080afaa4e7fc6ab73e067f31eab5fdcd762358f2" gracePeriod=2 Nov 25 10:09:47 crc kubenswrapper[4776]: I1125 10:09:47.445387 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2llhj" Nov 25 10:09:47 crc kubenswrapper[4776]: I1125 10:09:47.618224 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4fda217-d8e5-49bd-802a-f0588ade20a4-catalog-content\") pod \"f4fda217-d8e5-49bd-802a-f0588ade20a4\" (UID: \"f4fda217-d8e5-49bd-802a-f0588ade20a4\") " Nov 25 10:09:47 crc kubenswrapper[4776]: I1125 10:09:47.618289 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4fda217-d8e5-49bd-802a-f0588ade20a4-utilities\") pod \"f4fda217-d8e5-49bd-802a-f0588ade20a4\" (UID: \"f4fda217-d8e5-49bd-802a-f0588ade20a4\") " Nov 25 10:09:47 crc kubenswrapper[4776]: I1125 10:09:47.618334 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hrqn\" (UniqueName: \"kubernetes.io/projected/f4fda217-d8e5-49bd-802a-f0588ade20a4-kube-api-access-2hrqn\") pod \"f4fda217-d8e5-49bd-802a-f0588ade20a4\" (UID: \"f4fda217-d8e5-49bd-802a-f0588ade20a4\") " Nov 25 10:09:47 crc kubenswrapper[4776]: I1125 10:09:47.619538 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4fda217-d8e5-49bd-802a-f0588ade20a4-utilities" (OuterVolumeSpecName: "utilities") pod "f4fda217-d8e5-49bd-802a-f0588ade20a4" (UID: "f4fda217-d8e5-49bd-802a-f0588ade20a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:09:47 crc kubenswrapper[4776]: I1125 10:09:47.630274 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4fda217-d8e5-49bd-802a-f0588ade20a4-kube-api-access-2hrqn" (OuterVolumeSpecName: "kube-api-access-2hrqn") pod "f4fda217-d8e5-49bd-802a-f0588ade20a4" (UID: "f4fda217-d8e5-49bd-802a-f0588ade20a4"). InnerVolumeSpecName "kube-api-access-2hrqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:09:47 crc kubenswrapper[4776]: I1125 10:09:47.636146 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4fda217-d8e5-49bd-802a-f0588ade20a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f4fda217-d8e5-49bd-802a-f0588ade20a4" (UID: "f4fda217-d8e5-49bd-802a-f0588ade20a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:09:47 crc kubenswrapper[4776]: I1125 10:09:47.720522 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4fda217-d8e5-49bd-802a-f0588ade20a4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:47 crc kubenswrapper[4776]: I1125 10:09:47.720565 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4fda217-d8e5-49bd-802a-f0588ade20a4-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:47 crc kubenswrapper[4776]: I1125 10:09:47.720581 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hrqn\" (UniqueName: \"kubernetes.io/projected/f4fda217-d8e5-49bd-802a-f0588ade20a4-kube-api-access-2hrqn\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:47 crc kubenswrapper[4776]: I1125 10:09:47.986253 4776 generic.go:334] "Generic (PLEG): container finished" podID="f4fda217-d8e5-49bd-802a-f0588ade20a4" containerID="9d3c05ec1aa079c004adf21e080afaa4e7fc6ab73e067f31eab5fdcd762358f2" exitCode=0 Nov 25 10:09:47 crc kubenswrapper[4776]: I1125 10:09:47.986302 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2llhj" event={"ID":"f4fda217-d8e5-49bd-802a-f0588ade20a4","Type":"ContainerDied","Data":"9d3c05ec1aa079c004adf21e080afaa4e7fc6ab73e067f31eab5fdcd762358f2"} Nov 25 10:09:47 crc kubenswrapper[4776]: I1125 10:09:47.986322 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2llhj" Nov 25 10:09:47 crc kubenswrapper[4776]: I1125 10:09:47.986354 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2llhj" event={"ID":"f4fda217-d8e5-49bd-802a-f0588ade20a4","Type":"ContainerDied","Data":"1c870fc227c26e66509b38df1edcf593ffeb37521befe2f3f0fff61293eab11f"} Nov 25 10:09:47 crc kubenswrapper[4776]: I1125 10:09:47.986381 4776 scope.go:117] "RemoveContainer" containerID="9d3c05ec1aa079c004adf21e080afaa4e7fc6ab73e067f31eab5fdcd762358f2" Nov 25 10:09:48 crc kubenswrapper[4776]: I1125 10:09:48.011009 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2llhj"] Nov 25 10:09:48 crc kubenswrapper[4776]: I1125 10:09:48.019716 4776 scope.go:117] "RemoveContainer" containerID="0c925c7a71238a65aab798c23d8708f77d12d9a79ad3bb8c5ce6702bfe227cc8" Nov 25 10:09:48 crc kubenswrapper[4776]: I1125 10:09:48.021543 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2llhj"] Nov 25 10:09:48 crc kubenswrapper[4776]: I1125 10:09:48.034895 4776 scope.go:117] "RemoveContainer" containerID="59ccf84bb212af568b15a093b07a95d19b9e07a5d83880fc1e8a3f4060fb5afb" Nov 25 10:09:48 crc kubenswrapper[4776]: I1125 10:09:48.070041 4776 scope.go:117] "RemoveContainer" containerID="9d3c05ec1aa079c004adf21e080afaa4e7fc6ab73e067f31eab5fdcd762358f2" Nov 25 10:09:48 crc kubenswrapper[4776]: E1125 10:09:48.070408 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d3c05ec1aa079c004adf21e080afaa4e7fc6ab73e067f31eab5fdcd762358f2\": container with ID starting with 9d3c05ec1aa079c004adf21e080afaa4e7fc6ab73e067f31eab5fdcd762358f2 not found: ID does not exist" containerID="9d3c05ec1aa079c004adf21e080afaa4e7fc6ab73e067f31eab5fdcd762358f2" Nov 25 10:09:48 crc kubenswrapper[4776]: I1125 10:09:48.070443 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d3c05ec1aa079c004adf21e080afaa4e7fc6ab73e067f31eab5fdcd762358f2"} err="failed to get container status \"9d3c05ec1aa079c004adf21e080afaa4e7fc6ab73e067f31eab5fdcd762358f2\": rpc error: code = NotFound desc = could not find container \"9d3c05ec1aa079c004adf21e080afaa4e7fc6ab73e067f31eab5fdcd762358f2\": container with ID starting with 9d3c05ec1aa079c004adf21e080afaa4e7fc6ab73e067f31eab5fdcd762358f2 not found: ID does not exist" Nov 25 10:09:48 crc kubenswrapper[4776]: I1125 10:09:48.070467 4776 scope.go:117] "RemoveContainer" containerID="0c925c7a71238a65aab798c23d8708f77d12d9a79ad3bb8c5ce6702bfe227cc8" Nov 25 10:09:48 crc kubenswrapper[4776]: E1125 10:09:48.070707 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c925c7a71238a65aab798c23d8708f77d12d9a79ad3bb8c5ce6702bfe227cc8\": container with ID starting with 0c925c7a71238a65aab798c23d8708f77d12d9a79ad3bb8c5ce6702bfe227cc8 not found: ID does not exist" containerID="0c925c7a71238a65aab798c23d8708f77d12d9a79ad3bb8c5ce6702bfe227cc8" Nov 25 10:09:48 crc kubenswrapper[4776]: I1125 10:09:48.070732 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c925c7a71238a65aab798c23d8708f77d12d9a79ad3bb8c5ce6702bfe227cc8"} err="failed to get container status \"0c925c7a71238a65aab798c23d8708f77d12d9a79ad3bb8c5ce6702bfe227cc8\": rpc error: code = NotFound desc = could not find container \"0c925c7a71238a65aab798c23d8708f77d12d9a79ad3bb8c5ce6702bfe227cc8\": container with ID starting with 0c925c7a71238a65aab798c23d8708f77d12d9a79ad3bb8c5ce6702bfe227cc8 not found: ID does not exist" Nov 25 10:09:48 crc kubenswrapper[4776]: I1125 10:09:48.070748 4776 scope.go:117] "RemoveContainer" containerID="59ccf84bb212af568b15a093b07a95d19b9e07a5d83880fc1e8a3f4060fb5afb" Nov 25 10:09:48 crc kubenswrapper[4776]: E1125 10:09:48.070987 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59ccf84bb212af568b15a093b07a95d19b9e07a5d83880fc1e8a3f4060fb5afb\": container with ID starting with 59ccf84bb212af568b15a093b07a95d19b9e07a5d83880fc1e8a3f4060fb5afb not found: ID does not exist" containerID="59ccf84bb212af568b15a093b07a95d19b9e07a5d83880fc1e8a3f4060fb5afb" Nov 25 10:09:48 crc kubenswrapper[4776]: I1125 10:09:48.071026 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59ccf84bb212af568b15a093b07a95d19b9e07a5d83880fc1e8a3f4060fb5afb"} err="failed to get container status \"59ccf84bb212af568b15a093b07a95d19b9e07a5d83880fc1e8a3f4060fb5afb\": rpc error: code = NotFound desc = could not find container \"59ccf84bb212af568b15a093b07a95d19b9e07a5d83880fc1e8a3f4060fb5afb\": container with ID starting with 59ccf84bb212af568b15a093b07a95d19b9e07a5d83880fc1e8a3f4060fb5afb not found: ID does not exist" Nov 25 10:09:49 crc kubenswrapper[4776]: I1125 10:09:49.672469 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4fda217-d8e5-49bd-802a-f0588ade20a4" path="/var/lib/kubelet/pods/f4fda217-d8e5-49bd-802a-f0588ade20a4/volumes" Nov 25 10:09:53 crc kubenswrapper[4776]: I1125 10:09:53.901149 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rqc7d"] Nov 25 10:09:53 crc kubenswrapper[4776]: E1125 10:09:53.901715 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4fda217-d8e5-49bd-802a-f0588ade20a4" containerName="extract-utilities" Nov 25 10:09:53 crc kubenswrapper[4776]: I1125 10:09:53.901727 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4fda217-d8e5-49bd-802a-f0588ade20a4" containerName="extract-utilities" Nov 25 10:09:53 crc kubenswrapper[4776]: E1125 10:09:53.901744 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4fda217-d8e5-49bd-802a-f0588ade20a4" containerName="extract-content" Nov 25 10:09:53 crc kubenswrapper[4776]: I1125 10:09:53.901751 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4fda217-d8e5-49bd-802a-f0588ade20a4" containerName="extract-content" Nov 25 10:09:53 crc kubenswrapper[4776]: E1125 10:09:53.901767 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4fda217-d8e5-49bd-802a-f0588ade20a4" containerName="registry-server" Nov 25 10:09:53 crc kubenswrapper[4776]: I1125 10:09:53.901776 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4fda217-d8e5-49bd-802a-f0588ade20a4" containerName="registry-server" Nov 25 10:09:53 crc kubenswrapper[4776]: I1125 10:09:53.901917 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4fda217-d8e5-49bd-802a-f0588ade20a4" containerName="registry-server" Nov 25 10:09:53 crc kubenswrapper[4776]: I1125 10:09:53.903591 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rqc7d" Nov 25 10:09:53 crc kubenswrapper[4776]: I1125 10:09:53.916312 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rqc7d"] Nov 25 10:09:53 crc kubenswrapper[4776]: I1125 10:09:53.928960 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9e73a37-f483-4d21-9497-963e2b93f03d-catalog-content\") pod \"community-operators-rqc7d\" (UID: \"c9e73a37-f483-4d21-9497-963e2b93f03d\") " pod="openshift-marketplace/community-operators-rqc7d" Nov 25 10:09:53 crc kubenswrapper[4776]: I1125 10:09:53.929044 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9e73a37-f483-4d21-9497-963e2b93f03d-utilities\") pod \"community-operators-rqc7d\" (UID: \"c9e73a37-f483-4d21-9497-963e2b93f03d\") " pod="openshift-marketplace/community-operators-rqc7d" Nov 25 10:09:53 crc kubenswrapper[4776]: I1125 10:09:53.929110 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tggdb\" (UniqueName: \"kubernetes.io/projected/c9e73a37-f483-4d21-9497-963e2b93f03d-kube-api-access-tggdb\") pod \"community-operators-rqc7d\" (UID: \"c9e73a37-f483-4d21-9497-963e2b93f03d\") " pod="openshift-marketplace/community-operators-rqc7d" Nov 25 10:09:54 crc kubenswrapper[4776]: I1125 10:09:54.029940 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9e73a37-f483-4d21-9497-963e2b93f03d-utilities\") pod \"community-operators-rqc7d\" (UID: \"c9e73a37-f483-4d21-9497-963e2b93f03d\") " pod="openshift-marketplace/community-operators-rqc7d" Nov 25 10:09:54 crc kubenswrapper[4776]: I1125 10:09:54.030032 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tggdb\" (UniqueName: \"kubernetes.io/projected/c9e73a37-f483-4d21-9497-963e2b93f03d-kube-api-access-tggdb\") pod \"community-operators-rqc7d\" (UID: \"c9e73a37-f483-4d21-9497-963e2b93f03d\") " pod="openshift-marketplace/community-operators-rqc7d" Nov 25 10:09:54 crc kubenswrapper[4776]: I1125 10:09:54.030127 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9e73a37-f483-4d21-9497-963e2b93f03d-catalog-content\") pod \"community-operators-rqc7d\" (UID: \"c9e73a37-f483-4d21-9497-963e2b93f03d\") " pod="openshift-marketplace/community-operators-rqc7d" Nov 25 10:09:54 crc kubenswrapper[4776]: I1125 10:09:54.030675 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9e73a37-f483-4d21-9497-963e2b93f03d-utilities\") pod \"community-operators-rqc7d\" (UID: \"c9e73a37-f483-4d21-9497-963e2b93f03d\") " pod="openshift-marketplace/community-operators-rqc7d" Nov 25 10:09:54 crc kubenswrapper[4776]: I1125 10:09:54.032473 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9e73a37-f483-4d21-9497-963e2b93f03d-catalog-content\") pod \"community-operators-rqc7d\" (UID: \"c9e73a37-f483-4d21-9497-963e2b93f03d\") " pod="openshift-marketplace/community-operators-rqc7d" Nov 25 10:09:54 crc kubenswrapper[4776]: I1125 10:09:54.054900 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tggdb\" (UniqueName: \"kubernetes.io/projected/c9e73a37-f483-4d21-9497-963e2b93f03d-kube-api-access-tggdb\") pod \"community-operators-rqc7d\" (UID: \"c9e73a37-f483-4d21-9497-963e2b93f03d\") " pod="openshift-marketplace/community-operators-rqc7d" Nov 25 10:09:54 crc kubenswrapper[4776]: I1125 10:09:54.221308 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rqc7d" Nov 25 10:09:54 crc kubenswrapper[4776]: I1125 10:09:54.708959 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rqc7d"] Nov 25 10:09:55 crc kubenswrapper[4776]: I1125 10:09:55.040777 4776 generic.go:334] "Generic (PLEG): container finished" podID="c9e73a37-f483-4d21-9497-963e2b93f03d" containerID="bb763832602fa43f17077a9d0cd67df2c84ad34e407405e7e5f92d405c5af901" exitCode=0 Nov 25 10:09:55 crc kubenswrapper[4776]: I1125 10:09:55.040828 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqc7d" event={"ID":"c9e73a37-f483-4d21-9497-963e2b93f03d","Type":"ContainerDied","Data":"bb763832602fa43f17077a9d0cd67df2c84ad34e407405e7e5f92d405c5af901"} Nov 25 10:09:55 crc kubenswrapper[4776]: I1125 10:09:55.040887 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqc7d" event={"ID":"c9e73a37-f483-4d21-9497-963e2b93f03d","Type":"ContainerStarted","Data":"b541b21dee67a5c346e7822788819662fcbc3f0a501e250d427f4d968eacc01f"} Nov 25 10:09:56 crc kubenswrapper[4776]: I1125 10:09:56.049589 4776 generic.go:334] "Generic (PLEG): container finished" podID="c9e73a37-f483-4d21-9497-963e2b93f03d" containerID="79f6612ab9fd6217ff254cc2a4a8b657bc5b24c5b871fddbb96419a9e470b688" exitCode=0 Nov 25 10:09:56 crc kubenswrapper[4776]: I1125 10:09:56.049679 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqc7d" event={"ID":"c9e73a37-f483-4d21-9497-963e2b93f03d","Type":"ContainerDied","Data":"79f6612ab9fd6217ff254cc2a4a8b657bc5b24c5b871fddbb96419a9e470b688"} Nov 25 10:09:57 crc kubenswrapper[4776]: I1125 10:09:57.059847 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqc7d" event={"ID":"c9e73a37-f483-4d21-9497-963e2b93f03d","Type":"ContainerStarted","Data":"af2d17c21541b948f5b77c2b38f8513190587db2a624a3b3c41f9f9cfff4268e"} Nov 25 10:09:57 crc kubenswrapper[4776]: I1125 10:09:57.081570 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rqc7d" podStartSLOduration=2.643994493 podStartE2EDuration="4.081552082s" podCreationTimestamp="2025-11-25 10:09:53 +0000 UTC" firstStartedPulling="2025-11-25 10:09:55.04255474 +0000 UTC m=+2740.083614303" lastFinishedPulling="2025-11-25 10:09:56.480112339 +0000 UTC m=+2741.521171892" observedRunningTime="2025-11-25 10:09:57.079538732 +0000 UTC m=+2742.120598285" watchObservedRunningTime="2025-11-25 10:09:57.081552082 +0000 UTC m=+2742.122611645" Nov 25 10:10:04 crc kubenswrapper[4776]: I1125 10:10:04.222414 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rqc7d" Nov 25 10:10:04 crc kubenswrapper[4776]: I1125 10:10:04.223006 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rqc7d" Nov 25 10:10:04 crc kubenswrapper[4776]: I1125 10:10:04.287077 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rqc7d" Nov 25 10:10:05 crc kubenswrapper[4776]: I1125 10:10:05.156529 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rqc7d" Nov 25 10:10:05 crc kubenswrapper[4776]: I1125 10:10:05.196962 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rqc7d"] Nov 25 10:10:07 crc kubenswrapper[4776]: I1125 10:10:07.138100 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rqc7d" podUID="c9e73a37-f483-4d21-9497-963e2b93f03d" containerName="registry-server" containerID="cri-o://af2d17c21541b948f5b77c2b38f8513190587db2a624a3b3c41f9f9cfff4268e" gracePeriod=2 Nov 25 10:10:07 crc kubenswrapper[4776]: I1125 10:10:07.503027 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rqc7d" Nov 25 10:10:07 crc kubenswrapper[4776]: I1125 10:10:07.616691 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9e73a37-f483-4d21-9497-963e2b93f03d-utilities\") pod \"c9e73a37-f483-4d21-9497-963e2b93f03d\" (UID: \"c9e73a37-f483-4d21-9497-963e2b93f03d\") " Nov 25 10:10:07 crc kubenswrapper[4776]: I1125 10:10:07.616842 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tggdb\" (UniqueName: \"kubernetes.io/projected/c9e73a37-f483-4d21-9497-963e2b93f03d-kube-api-access-tggdb\") pod \"c9e73a37-f483-4d21-9497-963e2b93f03d\" (UID: \"c9e73a37-f483-4d21-9497-963e2b93f03d\") " Nov 25 10:10:07 crc kubenswrapper[4776]: I1125 10:10:07.616944 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9e73a37-f483-4d21-9497-963e2b93f03d-catalog-content\") pod \"c9e73a37-f483-4d21-9497-963e2b93f03d\" (UID: \"c9e73a37-f483-4d21-9497-963e2b93f03d\") " Nov 25 10:10:07 crc kubenswrapper[4776]: I1125 10:10:07.617572 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9e73a37-f483-4d21-9497-963e2b93f03d-utilities" (OuterVolumeSpecName: "utilities") pod "c9e73a37-f483-4d21-9497-963e2b93f03d" (UID: "c9e73a37-f483-4d21-9497-963e2b93f03d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:10:07 crc kubenswrapper[4776]: I1125 10:10:07.626392 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9e73a37-f483-4d21-9497-963e2b93f03d-kube-api-access-tggdb" (OuterVolumeSpecName: "kube-api-access-tggdb") pod "c9e73a37-f483-4d21-9497-963e2b93f03d" (UID: "c9e73a37-f483-4d21-9497-963e2b93f03d"). InnerVolumeSpecName "kube-api-access-tggdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:10:07 crc kubenswrapper[4776]: I1125 10:10:07.673942 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9e73a37-f483-4d21-9497-963e2b93f03d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c9e73a37-f483-4d21-9497-963e2b93f03d" (UID: "c9e73a37-f483-4d21-9497-963e2b93f03d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:10:07 crc kubenswrapper[4776]: I1125 10:10:07.719170 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9e73a37-f483-4d21-9497-963e2b93f03d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:10:07 crc kubenswrapper[4776]: I1125 10:10:07.719215 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9e73a37-f483-4d21-9497-963e2b93f03d-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:10:07 crc kubenswrapper[4776]: I1125 10:10:07.719227 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tggdb\" (UniqueName: \"kubernetes.io/projected/c9e73a37-f483-4d21-9497-963e2b93f03d-kube-api-access-tggdb\") on node \"crc\" DevicePath \"\"" Nov 25 10:10:08 crc kubenswrapper[4776]: I1125 10:10:08.146220 4776 generic.go:334] "Generic (PLEG): container finished" podID="c9e73a37-f483-4d21-9497-963e2b93f03d" containerID="af2d17c21541b948f5b77c2b38f8513190587db2a624a3b3c41f9f9cfff4268e" exitCode=0 Nov 25 10:10:08 crc kubenswrapper[4776]: I1125 10:10:08.146259 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rqc7d" Nov 25 10:10:08 crc kubenswrapper[4776]: I1125 10:10:08.146269 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqc7d" event={"ID":"c9e73a37-f483-4d21-9497-963e2b93f03d","Type":"ContainerDied","Data":"af2d17c21541b948f5b77c2b38f8513190587db2a624a3b3c41f9f9cfff4268e"} Nov 25 10:10:08 crc kubenswrapper[4776]: I1125 10:10:08.146306 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqc7d" event={"ID":"c9e73a37-f483-4d21-9497-963e2b93f03d","Type":"ContainerDied","Data":"b541b21dee67a5c346e7822788819662fcbc3f0a501e250d427f4d968eacc01f"} Nov 25 10:10:08 crc kubenswrapper[4776]: I1125 10:10:08.146325 4776 scope.go:117] "RemoveContainer" containerID="af2d17c21541b948f5b77c2b38f8513190587db2a624a3b3c41f9f9cfff4268e" Nov 25 10:10:08 crc kubenswrapper[4776]: I1125 10:10:08.176951 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rqc7d"] Nov 25 10:10:08 crc kubenswrapper[4776]: I1125 10:10:08.178000 4776 scope.go:117] "RemoveContainer" containerID="79f6612ab9fd6217ff254cc2a4a8b657bc5b24c5b871fddbb96419a9e470b688" Nov 25 10:10:08 crc kubenswrapper[4776]: I1125 10:10:08.187540 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rqc7d"] Nov 25 10:10:08 crc kubenswrapper[4776]: I1125 10:10:08.197219 4776 scope.go:117] "RemoveContainer" containerID="bb763832602fa43f17077a9d0cd67df2c84ad34e407405e7e5f92d405c5af901" Nov 25 10:10:08 crc kubenswrapper[4776]: I1125 10:10:08.229030 4776 scope.go:117] "RemoveContainer" containerID="af2d17c21541b948f5b77c2b38f8513190587db2a624a3b3c41f9f9cfff4268e" Nov 25 10:10:08 crc kubenswrapper[4776]: E1125 10:10:08.229592 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af2d17c21541b948f5b77c2b38f8513190587db2a624a3b3c41f9f9cfff4268e\": container with ID starting with af2d17c21541b948f5b77c2b38f8513190587db2a624a3b3c41f9f9cfff4268e not found: ID does not exist" containerID="af2d17c21541b948f5b77c2b38f8513190587db2a624a3b3c41f9f9cfff4268e" Nov 25 10:10:08 crc kubenswrapper[4776]: I1125 10:10:08.229636 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af2d17c21541b948f5b77c2b38f8513190587db2a624a3b3c41f9f9cfff4268e"} err="failed to get container status \"af2d17c21541b948f5b77c2b38f8513190587db2a624a3b3c41f9f9cfff4268e\": rpc error: code = NotFound desc = could not find container \"af2d17c21541b948f5b77c2b38f8513190587db2a624a3b3c41f9f9cfff4268e\": container with ID starting with af2d17c21541b948f5b77c2b38f8513190587db2a624a3b3c41f9f9cfff4268e not found: ID does not exist" Nov 25 10:10:08 crc kubenswrapper[4776]: I1125 10:10:08.229663 4776 scope.go:117] "RemoveContainer" containerID="79f6612ab9fd6217ff254cc2a4a8b657bc5b24c5b871fddbb96419a9e470b688" Nov 25 10:10:08 crc kubenswrapper[4776]: E1125 10:10:08.229998 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79f6612ab9fd6217ff254cc2a4a8b657bc5b24c5b871fddbb96419a9e470b688\": container with ID starting with 79f6612ab9fd6217ff254cc2a4a8b657bc5b24c5b871fddbb96419a9e470b688 not found: ID does not exist" containerID="79f6612ab9fd6217ff254cc2a4a8b657bc5b24c5b871fddbb96419a9e470b688" Nov 25 10:10:08 crc kubenswrapper[4776]: I1125 10:10:08.230019 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79f6612ab9fd6217ff254cc2a4a8b657bc5b24c5b871fddbb96419a9e470b688"} err="failed to get container status \"79f6612ab9fd6217ff254cc2a4a8b657bc5b24c5b871fddbb96419a9e470b688\": rpc error: code = NotFound desc = could not find container \"79f6612ab9fd6217ff254cc2a4a8b657bc5b24c5b871fddbb96419a9e470b688\": container with ID starting with 79f6612ab9fd6217ff254cc2a4a8b657bc5b24c5b871fddbb96419a9e470b688 not found: ID does not exist" Nov 25 10:10:08 crc kubenswrapper[4776]: I1125 10:10:08.230032 4776 scope.go:117] "RemoveContainer" containerID="bb763832602fa43f17077a9d0cd67df2c84ad34e407405e7e5f92d405c5af901" Nov 25 10:10:08 crc kubenswrapper[4776]: E1125 10:10:08.230219 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb763832602fa43f17077a9d0cd67df2c84ad34e407405e7e5f92d405c5af901\": container with ID starting with bb763832602fa43f17077a9d0cd67df2c84ad34e407405e7e5f92d405c5af901 not found: ID does not exist" containerID="bb763832602fa43f17077a9d0cd67df2c84ad34e407405e7e5f92d405c5af901" Nov 25 10:10:08 crc kubenswrapper[4776]: I1125 10:10:08.230243 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb763832602fa43f17077a9d0cd67df2c84ad34e407405e7e5f92d405c5af901"} err="failed to get container status \"bb763832602fa43f17077a9d0cd67df2c84ad34e407405e7e5f92d405c5af901\": rpc error: code = NotFound desc = could not find container \"bb763832602fa43f17077a9d0cd67df2c84ad34e407405e7e5f92d405c5af901\": container with ID starting with bb763832602fa43f17077a9d0cd67df2c84ad34e407405e7e5f92d405c5af901 not found: ID does not exist" Nov 25 10:10:09 crc kubenswrapper[4776]: I1125 10:10:09.675003 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9e73a37-f483-4d21-9497-963e2b93f03d" path="/var/lib/kubelet/pods/c9e73a37-f483-4d21-9497-963e2b93f03d/volumes" Nov 25 10:10:18 crc kubenswrapper[4776]: I1125 10:10:18.649621 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2h5gs"] Nov 25 10:10:18 crc kubenswrapper[4776]: E1125 10:10:18.650630 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9e73a37-f483-4d21-9497-963e2b93f03d" containerName="registry-server" Nov 25 10:10:18 crc kubenswrapper[4776]: I1125 10:10:18.650648 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9e73a37-f483-4d21-9497-963e2b93f03d" containerName="registry-server" Nov 25 10:10:18 crc kubenswrapper[4776]: E1125 10:10:18.650680 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9e73a37-f483-4d21-9497-963e2b93f03d" containerName="extract-content" Nov 25 10:10:18 crc kubenswrapper[4776]: I1125 10:10:18.650687 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9e73a37-f483-4d21-9497-963e2b93f03d" containerName="extract-content" Nov 25 10:10:18 crc kubenswrapper[4776]: E1125 10:10:18.650708 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9e73a37-f483-4d21-9497-963e2b93f03d" containerName="extract-utilities" Nov 25 10:10:18 crc kubenswrapper[4776]: I1125 10:10:18.650715 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9e73a37-f483-4d21-9497-963e2b93f03d" containerName="extract-utilities" Nov 25 10:10:18 crc kubenswrapper[4776]: I1125 10:10:18.651106 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9e73a37-f483-4d21-9497-963e2b93f03d" containerName="registry-server" Nov 25 10:10:18 crc kubenswrapper[4776]: I1125 10:10:18.652414 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2h5gs" Nov 25 10:10:18 crc kubenswrapper[4776]: I1125 10:10:18.659382 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2h5gs"] Nov 25 10:10:18 crc kubenswrapper[4776]: I1125 10:10:18.678887 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d6cc563-6d0c-4050-8d05-3e395605dafe-catalog-content\") pod \"certified-operators-2h5gs\" (UID: \"1d6cc563-6d0c-4050-8d05-3e395605dafe\") " pod="openshift-marketplace/certified-operators-2h5gs" Nov 25 10:10:18 crc kubenswrapper[4776]: I1125 10:10:18.679027 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzxpm\" (UniqueName: \"kubernetes.io/projected/1d6cc563-6d0c-4050-8d05-3e395605dafe-kube-api-access-hzxpm\") pod \"certified-operators-2h5gs\" (UID: \"1d6cc563-6d0c-4050-8d05-3e395605dafe\") " pod="openshift-marketplace/certified-operators-2h5gs" Nov 25 10:10:18 crc kubenswrapper[4776]: I1125 10:10:18.679202 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d6cc563-6d0c-4050-8d05-3e395605dafe-utilities\") pod \"certified-operators-2h5gs\" (UID: \"1d6cc563-6d0c-4050-8d05-3e395605dafe\") " pod="openshift-marketplace/certified-operators-2h5gs" Nov 25 10:10:18 crc kubenswrapper[4776]: I1125 10:10:18.780138 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzxpm\" (UniqueName: \"kubernetes.io/projected/1d6cc563-6d0c-4050-8d05-3e395605dafe-kube-api-access-hzxpm\") pod \"certified-operators-2h5gs\" (UID: \"1d6cc563-6d0c-4050-8d05-3e395605dafe\") " pod="openshift-marketplace/certified-operators-2h5gs" Nov 25 10:10:18 crc kubenswrapper[4776]: I1125 10:10:18.780241 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d6cc563-6d0c-4050-8d05-3e395605dafe-utilities\") pod \"certified-operators-2h5gs\" (UID: \"1d6cc563-6d0c-4050-8d05-3e395605dafe\") " pod="openshift-marketplace/certified-operators-2h5gs" Nov 25 10:10:18 crc kubenswrapper[4776]: I1125 10:10:18.780274 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d6cc563-6d0c-4050-8d05-3e395605dafe-catalog-content\") pod \"certified-operators-2h5gs\" (UID: \"1d6cc563-6d0c-4050-8d05-3e395605dafe\") " pod="openshift-marketplace/certified-operators-2h5gs" Nov 25 10:10:18 crc kubenswrapper[4776]: I1125 10:10:18.780794 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d6cc563-6d0c-4050-8d05-3e395605dafe-utilities\") pod \"certified-operators-2h5gs\" (UID: \"1d6cc563-6d0c-4050-8d05-3e395605dafe\") " pod="openshift-marketplace/certified-operators-2h5gs" Nov 25 10:10:18 crc kubenswrapper[4776]: I1125 10:10:18.780827 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d6cc563-6d0c-4050-8d05-3e395605dafe-catalog-content\") pod \"certified-operators-2h5gs\" (UID: \"1d6cc563-6d0c-4050-8d05-3e395605dafe\") " pod="openshift-marketplace/certified-operators-2h5gs" Nov 25 10:10:18 crc kubenswrapper[4776]: I1125 10:10:18.800855 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzxpm\" (UniqueName: \"kubernetes.io/projected/1d6cc563-6d0c-4050-8d05-3e395605dafe-kube-api-access-hzxpm\") pod \"certified-operators-2h5gs\" (UID: \"1d6cc563-6d0c-4050-8d05-3e395605dafe\") " pod="openshift-marketplace/certified-operators-2h5gs" Nov 25 10:10:18 crc kubenswrapper[4776]: I1125 10:10:18.977649 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2h5gs" Nov 25 10:10:19 crc kubenswrapper[4776]: I1125 10:10:19.231867 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2h5gs"] Nov 25 10:10:19 crc kubenswrapper[4776]: I1125 10:10:19.256251 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2h5gs" event={"ID":"1d6cc563-6d0c-4050-8d05-3e395605dafe","Type":"ContainerStarted","Data":"2ef1d2e3a45585705ac34bcd62fb3b3426aeef6d98f8d28caab8dfa5407f17a1"} Nov 25 10:10:20 crc kubenswrapper[4776]: I1125 10:10:20.264475 4776 generic.go:334] "Generic (PLEG): container finished" podID="1d6cc563-6d0c-4050-8d05-3e395605dafe" containerID="cf27fefc8aadea14579986700e5da29d26fed64a8089d70d879f41948539aec5" exitCode=0 Nov 25 10:10:20 crc kubenswrapper[4776]: I1125 10:10:20.264510 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2h5gs" event={"ID":"1d6cc563-6d0c-4050-8d05-3e395605dafe","Type":"ContainerDied","Data":"cf27fefc8aadea14579986700e5da29d26fed64a8089d70d879f41948539aec5"} Nov 25 10:10:21 crc kubenswrapper[4776]: I1125 10:10:21.275917 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2h5gs" event={"ID":"1d6cc563-6d0c-4050-8d05-3e395605dafe","Type":"ContainerStarted","Data":"59a9e2377147760b8e39a1e971ba8dcc495244908d6277f83b03eea51bbea02a"} Nov 25 10:10:22 crc kubenswrapper[4776]: I1125 10:10:22.287147 4776 generic.go:334] "Generic (PLEG): container finished" podID="1d6cc563-6d0c-4050-8d05-3e395605dafe" containerID="59a9e2377147760b8e39a1e971ba8dcc495244908d6277f83b03eea51bbea02a" exitCode=0 Nov 25 10:10:22 crc kubenswrapper[4776]: I1125 10:10:22.287224 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2h5gs" event={"ID":"1d6cc563-6d0c-4050-8d05-3e395605dafe","Type":"ContainerDied","Data":"59a9e2377147760b8e39a1e971ba8dcc495244908d6277f83b03eea51bbea02a"} Nov 25 10:10:23 crc kubenswrapper[4776]: I1125 10:10:23.296340 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2h5gs" event={"ID":"1d6cc563-6d0c-4050-8d05-3e395605dafe","Type":"ContainerStarted","Data":"242319c933f830f0a42d8c0945c8f2a70e0c19c64e644df60a916315ecbe04d4"} Nov 25 10:10:23 crc kubenswrapper[4776]: I1125 10:10:23.319086 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2h5gs" podStartSLOduration=2.897675739 podStartE2EDuration="5.319038461s" podCreationTimestamp="2025-11-25 10:10:18 +0000 UTC" firstStartedPulling="2025-11-25 10:10:20.266150861 +0000 UTC m=+2765.307210414" lastFinishedPulling="2025-11-25 10:10:22.687513583 +0000 UTC m=+2767.728573136" observedRunningTime="2025-11-25 10:10:23.315966954 +0000 UTC m=+2768.357026527" watchObservedRunningTime="2025-11-25 10:10:23.319038461 +0000 UTC m=+2768.360098014" Nov 25 10:10:28 crc kubenswrapper[4776]: I1125 10:10:28.978642 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2h5gs" Nov 25 10:10:28 crc kubenswrapper[4776]: I1125 10:10:28.979024 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2h5gs" Nov 25 10:10:29 crc kubenswrapper[4776]: I1125 10:10:29.016976 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2h5gs" Nov 25 10:10:29 crc kubenswrapper[4776]: I1125 10:10:29.406664 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2h5gs" Nov 25 10:10:29 crc kubenswrapper[4776]: I1125 10:10:29.463765 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2h5gs"] Nov 25 10:10:31 crc kubenswrapper[4776]: I1125 10:10:31.361322 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2h5gs" podUID="1d6cc563-6d0c-4050-8d05-3e395605dafe" containerName="registry-server" containerID="cri-o://242319c933f830f0a42d8c0945c8f2a70e0c19c64e644df60a916315ecbe04d4" gracePeriod=2 Nov 25 10:10:32 crc kubenswrapper[4776]: I1125 10:10:32.284402 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2h5gs" Nov 25 10:10:32 crc kubenswrapper[4776]: I1125 10:10:32.370401 4776 generic.go:334] "Generic (PLEG): container finished" podID="1d6cc563-6d0c-4050-8d05-3e395605dafe" containerID="242319c933f830f0a42d8c0945c8f2a70e0c19c64e644df60a916315ecbe04d4" exitCode=0 Nov 25 10:10:32 crc kubenswrapper[4776]: I1125 10:10:32.370450 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2h5gs" event={"ID":"1d6cc563-6d0c-4050-8d05-3e395605dafe","Type":"ContainerDied","Data":"242319c933f830f0a42d8c0945c8f2a70e0c19c64e644df60a916315ecbe04d4"} Nov 25 10:10:32 crc kubenswrapper[4776]: I1125 10:10:32.370480 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2h5gs" event={"ID":"1d6cc563-6d0c-4050-8d05-3e395605dafe","Type":"ContainerDied","Data":"2ef1d2e3a45585705ac34bcd62fb3b3426aeef6d98f8d28caab8dfa5407f17a1"} Nov 25 10:10:32 crc kubenswrapper[4776]: I1125 10:10:32.370502 4776 scope.go:117] "RemoveContainer" containerID="242319c933f830f0a42d8c0945c8f2a70e0c19c64e644df60a916315ecbe04d4" Nov 25 10:10:32 crc kubenswrapper[4776]: I1125 10:10:32.370635 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2h5gs" Nov 25 10:10:32 crc kubenswrapper[4776]: I1125 10:10:32.399137 4776 scope.go:117] "RemoveContainer" containerID="59a9e2377147760b8e39a1e971ba8dcc495244908d6277f83b03eea51bbea02a" Nov 25 10:10:32 crc kubenswrapper[4776]: I1125 10:10:32.418765 4776 scope.go:117] "RemoveContainer" containerID="cf27fefc8aadea14579986700e5da29d26fed64a8089d70d879f41948539aec5" Nov 25 10:10:32 crc kubenswrapper[4776]: I1125 10:10:32.442246 4776 scope.go:117] "RemoveContainer" containerID="242319c933f830f0a42d8c0945c8f2a70e0c19c64e644df60a916315ecbe04d4" Nov 25 10:10:32 crc kubenswrapper[4776]: E1125 10:10:32.442715 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"242319c933f830f0a42d8c0945c8f2a70e0c19c64e644df60a916315ecbe04d4\": container with ID starting with 242319c933f830f0a42d8c0945c8f2a70e0c19c64e644df60a916315ecbe04d4 not found: ID does not exist" containerID="242319c933f830f0a42d8c0945c8f2a70e0c19c64e644df60a916315ecbe04d4" Nov 25 10:10:32 crc kubenswrapper[4776]: I1125 10:10:32.442751 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"242319c933f830f0a42d8c0945c8f2a70e0c19c64e644df60a916315ecbe04d4"} err="failed to get container status \"242319c933f830f0a42d8c0945c8f2a70e0c19c64e644df60a916315ecbe04d4\": rpc error: code = NotFound desc = could not find container \"242319c933f830f0a42d8c0945c8f2a70e0c19c64e644df60a916315ecbe04d4\": container with ID starting with 242319c933f830f0a42d8c0945c8f2a70e0c19c64e644df60a916315ecbe04d4 not found: ID does not exist" Nov 25 10:10:32 crc kubenswrapper[4776]: I1125 10:10:32.442771 4776 scope.go:117] "RemoveContainer" containerID="59a9e2377147760b8e39a1e971ba8dcc495244908d6277f83b03eea51bbea02a" Nov 25 10:10:32 crc kubenswrapper[4776]: E1125 10:10:32.443193 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59a9e2377147760b8e39a1e971ba8dcc495244908d6277f83b03eea51bbea02a\": container with ID starting with 59a9e2377147760b8e39a1e971ba8dcc495244908d6277f83b03eea51bbea02a not found: ID does not exist" containerID="59a9e2377147760b8e39a1e971ba8dcc495244908d6277f83b03eea51bbea02a" Nov 25 10:10:32 crc kubenswrapper[4776]: I1125 10:10:32.443242 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59a9e2377147760b8e39a1e971ba8dcc495244908d6277f83b03eea51bbea02a"} err="failed to get container status \"59a9e2377147760b8e39a1e971ba8dcc495244908d6277f83b03eea51bbea02a\": rpc error: code = NotFound desc = could not find container \"59a9e2377147760b8e39a1e971ba8dcc495244908d6277f83b03eea51bbea02a\": container with ID starting with 59a9e2377147760b8e39a1e971ba8dcc495244908d6277f83b03eea51bbea02a not found: ID does not exist" Nov 25 10:10:32 crc kubenswrapper[4776]: I1125 10:10:32.443265 4776 scope.go:117] "RemoveContainer" containerID="cf27fefc8aadea14579986700e5da29d26fed64a8089d70d879f41948539aec5" Nov 25 10:10:32 crc kubenswrapper[4776]: E1125 10:10:32.443672 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf27fefc8aadea14579986700e5da29d26fed64a8089d70d879f41948539aec5\": container with ID starting with cf27fefc8aadea14579986700e5da29d26fed64a8089d70d879f41948539aec5 not found: ID does not exist" containerID="cf27fefc8aadea14579986700e5da29d26fed64a8089d70d879f41948539aec5" Nov 25 10:10:32 crc kubenswrapper[4776]: I1125 10:10:32.443696 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf27fefc8aadea14579986700e5da29d26fed64a8089d70d879f41948539aec5"} err="failed to get container status \"cf27fefc8aadea14579986700e5da29d26fed64a8089d70d879f41948539aec5\": rpc error: code = NotFound desc = could not find container \"cf27fefc8aadea14579986700e5da29d26fed64a8089d70d879f41948539aec5\": container with ID starting with cf27fefc8aadea14579986700e5da29d26fed64a8089d70d879f41948539aec5 not found: ID does not exist" Nov 25 10:10:32 crc kubenswrapper[4776]: I1125 10:10:32.477375 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d6cc563-6d0c-4050-8d05-3e395605dafe-utilities\") pod \"1d6cc563-6d0c-4050-8d05-3e395605dafe\" (UID: \"1d6cc563-6d0c-4050-8d05-3e395605dafe\") " Nov 25 10:10:32 crc kubenswrapper[4776]: I1125 10:10:32.477454 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d6cc563-6d0c-4050-8d05-3e395605dafe-catalog-content\") pod \"1d6cc563-6d0c-4050-8d05-3e395605dafe\" (UID: \"1d6cc563-6d0c-4050-8d05-3e395605dafe\") " Nov 25 10:10:32 crc kubenswrapper[4776]: I1125 10:10:32.477537 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzxpm\" (UniqueName: \"kubernetes.io/projected/1d6cc563-6d0c-4050-8d05-3e395605dafe-kube-api-access-hzxpm\") pod \"1d6cc563-6d0c-4050-8d05-3e395605dafe\" (UID: \"1d6cc563-6d0c-4050-8d05-3e395605dafe\") " Nov 25 10:10:32 crc kubenswrapper[4776]: I1125 10:10:32.478681 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d6cc563-6d0c-4050-8d05-3e395605dafe-utilities" (OuterVolumeSpecName: "utilities") pod "1d6cc563-6d0c-4050-8d05-3e395605dafe" (UID: "1d6cc563-6d0c-4050-8d05-3e395605dafe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:10:32 crc kubenswrapper[4776]: I1125 10:10:32.486069 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d6cc563-6d0c-4050-8d05-3e395605dafe-kube-api-access-hzxpm" (OuterVolumeSpecName: "kube-api-access-hzxpm") pod "1d6cc563-6d0c-4050-8d05-3e395605dafe" (UID: "1d6cc563-6d0c-4050-8d05-3e395605dafe"). InnerVolumeSpecName "kube-api-access-hzxpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:10:32 crc kubenswrapper[4776]: I1125 10:10:32.527973 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d6cc563-6d0c-4050-8d05-3e395605dafe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d6cc563-6d0c-4050-8d05-3e395605dafe" (UID: "1d6cc563-6d0c-4050-8d05-3e395605dafe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:10:32 crc kubenswrapper[4776]: I1125 10:10:32.578785 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzxpm\" (UniqueName: \"kubernetes.io/projected/1d6cc563-6d0c-4050-8d05-3e395605dafe-kube-api-access-hzxpm\") on node \"crc\" DevicePath \"\"" Nov 25 10:10:32 crc kubenswrapper[4776]: I1125 10:10:32.578819 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d6cc563-6d0c-4050-8d05-3e395605dafe-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:10:32 crc kubenswrapper[4776]: I1125 10:10:32.578829 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d6cc563-6d0c-4050-8d05-3e395605dafe-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:10:32 crc kubenswrapper[4776]: I1125 10:10:32.715683 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2h5gs"] Nov 25 10:10:32 crc kubenswrapper[4776]: I1125 10:10:32.722934 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2h5gs"] Nov 25 10:10:33 crc kubenswrapper[4776]: I1125 10:10:33.679917 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d6cc563-6d0c-4050-8d05-3e395605dafe" path="/var/lib/kubelet/pods/1d6cc563-6d0c-4050-8d05-3e395605dafe/volumes" Nov 25 10:11:47 crc kubenswrapper[4776]: I1125 10:11:47.818311 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:11:47 crc kubenswrapper[4776]: I1125 10:11:47.818904 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:12:17 crc kubenswrapper[4776]: I1125 10:12:17.818373 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:12:17 crc kubenswrapper[4776]: I1125 10:12:17.818994 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:12:47 crc kubenswrapper[4776]: I1125 10:12:47.818310 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:12:47 crc kubenswrapper[4776]: I1125 10:12:47.818939 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:12:47 crc kubenswrapper[4776]: I1125 10:12:47.818990 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 10:12:47 crc kubenswrapper[4776]: I1125 10:12:47.819522 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"41c62a00c2902589260314ef41fa8e27ca25f86c7f781c0b248f7cc291cf5f53"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:12:47 crc kubenswrapper[4776]: I1125 10:12:47.819572 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://41c62a00c2902589260314ef41fa8e27ca25f86c7f781c0b248f7cc291cf5f53" gracePeriod=600 Nov 25 10:12:48 crc kubenswrapper[4776]: I1125 10:12:48.761827 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="41c62a00c2902589260314ef41fa8e27ca25f86c7f781c0b248f7cc291cf5f53" exitCode=0 Nov 25 10:12:48 crc kubenswrapper[4776]: I1125 10:12:48.761878 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"41c62a00c2902589260314ef41fa8e27ca25f86c7f781c0b248f7cc291cf5f53"} Nov 25 10:12:48 crc kubenswrapper[4776]: I1125 10:12:48.762348 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255"} Nov 25 10:12:48 crc kubenswrapper[4776]: I1125 10:12:48.762381 4776 scope.go:117] "RemoveContainer" containerID="4588971ba22ce79f8fe6f426ac8ff87d99cf7b7212337a8cbdb2129a888b9288" Nov 25 10:15:00 crc kubenswrapper[4776]: I1125 10:15:00.186164 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401095-rnclm"] Nov 25 10:15:00 crc kubenswrapper[4776]: E1125 10:15:00.187189 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d6cc563-6d0c-4050-8d05-3e395605dafe" containerName="registry-server" Nov 25 10:15:00 crc kubenswrapper[4776]: I1125 10:15:00.187209 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d6cc563-6d0c-4050-8d05-3e395605dafe" containerName="registry-server" Nov 25 10:15:00 crc kubenswrapper[4776]: E1125 10:15:00.187234 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d6cc563-6d0c-4050-8d05-3e395605dafe" containerName="extract-utilities" Nov 25 10:15:00 crc kubenswrapper[4776]: I1125 10:15:00.187243 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d6cc563-6d0c-4050-8d05-3e395605dafe" containerName="extract-utilities" Nov 25 10:15:00 crc kubenswrapper[4776]: E1125 10:15:00.187254 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d6cc563-6d0c-4050-8d05-3e395605dafe" containerName="extract-content" Nov 25 10:15:00 crc kubenswrapper[4776]: I1125 10:15:00.187262 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d6cc563-6d0c-4050-8d05-3e395605dafe" containerName="extract-content" Nov 25 10:15:00 crc kubenswrapper[4776]: I1125 10:15:00.187452 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d6cc563-6d0c-4050-8d05-3e395605dafe" containerName="registry-server" Nov 25 10:15:00 crc kubenswrapper[4776]: I1125 10:15:00.187984 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-rnclm" Nov 25 10:15:00 crc kubenswrapper[4776]: I1125 10:15:00.196960 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 10:15:00 crc kubenswrapper[4776]: I1125 10:15:00.197363 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401095-rnclm"] Nov 25 10:15:00 crc kubenswrapper[4776]: I1125 10:15:00.198400 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 10:15:00 crc kubenswrapper[4776]: I1125 10:15:00.223116 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/625731b7-d813-461a-b2b3-8b7e64622ff4-config-volume\") pod \"collect-profiles-29401095-rnclm\" (UID: \"625731b7-d813-461a-b2b3-8b7e64622ff4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-rnclm" Nov 25 10:15:00 crc kubenswrapper[4776]: I1125 10:15:00.223177 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5nlc\" (UniqueName: \"kubernetes.io/projected/625731b7-d813-461a-b2b3-8b7e64622ff4-kube-api-access-c5nlc\") pod \"collect-profiles-29401095-rnclm\" (UID: \"625731b7-d813-461a-b2b3-8b7e64622ff4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-rnclm" Nov 25 10:15:00 crc kubenswrapper[4776]: I1125 10:15:00.223337 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/625731b7-d813-461a-b2b3-8b7e64622ff4-secret-volume\") pod \"collect-profiles-29401095-rnclm\" (UID: \"625731b7-d813-461a-b2b3-8b7e64622ff4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-rnclm" Nov 25 10:15:00 crc kubenswrapper[4776]: I1125 10:15:00.324535 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/625731b7-d813-461a-b2b3-8b7e64622ff4-config-volume\") pod \"collect-profiles-29401095-rnclm\" (UID: \"625731b7-d813-461a-b2b3-8b7e64622ff4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-rnclm" Nov 25 10:15:00 crc kubenswrapper[4776]: I1125 10:15:00.324790 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5nlc\" (UniqueName: \"kubernetes.io/projected/625731b7-d813-461a-b2b3-8b7e64622ff4-kube-api-access-c5nlc\") pod \"collect-profiles-29401095-rnclm\" (UID: \"625731b7-d813-461a-b2b3-8b7e64622ff4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-rnclm" Nov 25 10:15:00 crc kubenswrapper[4776]: I1125 10:15:00.324879 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/625731b7-d813-461a-b2b3-8b7e64622ff4-secret-volume\") pod \"collect-profiles-29401095-rnclm\" (UID: \"625731b7-d813-461a-b2b3-8b7e64622ff4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-rnclm" Nov 25 10:15:00 crc kubenswrapper[4776]: I1125 10:15:00.325423 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/625731b7-d813-461a-b2b3-8b7e64622ff4-config-volume\") pod \"collect-profiles-29401095-rnclm\" (UID: \"625731b7-d813-461a-b2b3-8b7e64622ff4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-rnclm" Nov 25 10:15:00 crc kubenswrapper[4776]: I1125 10:15:00.331553 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/625731b7-d813-461a-b2b3-8b7e64622ff4-secret-volume\") pod \"collect-profiles-29401095-rnclm\" (UID: \"625731b7-d813-461a-b2b3-8b7e64622ff4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-rnclm" Nov 25 10:15:00 crc kubenswrapper[4776]: I1125 10:15:00.342879 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5nlc\" (UniqueName: \"kubernetes.io/projected/625731b7-d813-461a-b2b3-8b7e64622ff4-kube-api-access-c5nlc\") pod \"collect-profiles-29401095-rnclm\" (UID: \"625731b7-d813-461a-b2b3-8b7e64622ff4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-rnclm" Nov 25 10:15:00 crc kubenswrapper[4776]: I1125 10:15:00.527018 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-rnclm" Nov 25 10:15:00 crc kubenswrapper[4776]: I1125 10:15:00.949572 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401095-rnclm"] Nov 25 10:15:01 crc kubenswrapper[4776]: I1125 10:15:01.876733 4776 generic.go:334] "Generic (PLEG): container finished" podID="625731b7-d813-461a-b2b3-8b7e64622ff4" containerID="c189a2b86eca91bc5be33fbc2bafdcdda468548845cf5f1546dd2249836d2c54" exitCode=0 Nov 25 10:15:01 crc kubenswrapper[4776]: I1125 10:15:01.876812 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-rnclm" event={"ID":"625731b7-d813-461a-b2b3-8b7e64622ff4","Type":"ContainerDied","Data":"c189a2b86eca91bc5be33fbc2bafdcdda468548845cf5f1546dd2249836d2c54"} Nov 25 10:15:01 crc kubenswrapper[4776]: I1125 10:15:01.876865 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-rnclm" event={"ID":"625731b7-d813-461a-b2b3-8b7e64622ff4","Type":"ContainerStarted","Data":"ceca01a65cde73482b27e51bf379688f34292e2a84a88cfab9db0b38a00dbf80"} Nov 25 10:15:03 crc kubenswrapper[4776]: I1125 10:15:03.181636 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-rnclm" Nov 25 10:15:03 crc kubenswrapper[4776]: I1125 10:15:03.364532 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/625731b7-d813-461a-b2b3-8b7e64622ff4-secret-volume\") pod \"625731b7-d813-461a-b2b3-8b7e64622ff4\" (UID: \"625731b7-d813-461a-b2b3-8b7e64622ff4\") " Nov 25 10:15:03 crc kubenswrapper[4776]: I1125 10:15:03.364589 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5nlc\" (UniqueName: \"kubernetes.io/projected/625731b7-d813-461a-b2b3-8b7e64622ff4-kube-api-access-c5nlc\") pod \"625731b7-d813-461a-b2b3-8b7e64622ff4\" (UID: \"625731b7-d813-461a-b2b3-8b7e64622ff4\") " Nov 25 10:15:03 crc kubenswrapper[4776]: I1125 10:15:03.364678 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/625731b7-d813-461a-b2b3-8b7e64622ff4-config-volume\") pod \"625731b7-d813-461a-b2b3-8b7e64622ff4\" (UID: \"625731b7-d813-461a-b2b3-8b7e64622ff4\") " Nov 25 10:15:03 crc kubenswrapper[4776]: I1125 10:15:03.365739 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/625731b7-d813-461a-b2b3-8b7e64622ff4-config-volume" (OuterVolumeSpecName: "config-volume") pod "625731b7-d813-461a-b2b3-8b7e64622ff4" (UID: "625731b7-d813-461a-b2b3-8b7e64622ff4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:15:03 crc kubenswrapper[4776]: I1125 10:15:03.370694 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/625731b7-d813-461a-b2b3-8b7e64622ff4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "625731b7-d813-461a-b2b3-8b7e64622ff4" (UID: "625731b7-d813-461a-b2b3-8b7e64622ff4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:15:03 crc kubenswrapper[4776]: I1125 10:15:03.371349 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/625731b7-d813-461a-b2b3-8b7e64622ff4-kube-api-access-c5nlc" (OuterVolumeSpecName: "kube-api-access-c5nlc") pod "625731b7-d813-461a-b2b3-8b7e64622ff4" (UID: "625731b7-d813-461a-b2b3-8b7e64622ff4"). InnerVolumeSpecName "kube-api-access-c5nlc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:15:03 crc kubenswrapper[4776]: I1125 10:15:03.466456 4776 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/625731b7-d813-461a-b2b3-8b7e64622ff4-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 10:15:03 crc kubenswrapper[4776]: I1125 10:15:03.466487 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5nlc\" (UniqueName: \"kubernetes.io/projected/625731b7-d813-461a-b2b3-8b7e64622ff4-kube-api-access-c5nlc\") on node \"crc\" DevicePath \"\"" Nov 25 10:15:03 crc kubenswrapper[4776]: I1125 10:15:03.466497 4776 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/625731b7-d813-461a-b2b3-8b7e64622ff4-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 10:15:03 crc kubenswrapper[4776]: I1125 10:15:03.896573 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-rnclm" event={"ID":"625731b7-d813-461a-b2b3-8b7e64622ff4","Type":"ContainerDied","Data":"ceca01a65cde73482b27e51bf379688f34292e2a84a88cfab9db0b38a00dbf80"} Nov 25 10:15:03 crc kubenswrapper[4776]: I1125 10:15:03.896613 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ceca01a65cde73482b27e51bf379688f34292e2a84a88cfab9db0b38a00dbf80" Nov 25 10:15:03 crc kubenswrapper[4776]: I1125 10:15:03.896661 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-rnclm" Nov 25 10:15:04 crc kubenswrapper[4776]: I1125 10:15:04.265384 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401050-2dn8b"] Nov 25 10:15:04 crc kubenswrapper[4776]: I1125 10:15:04.269976 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401050-2dn8b"] Nov 25 10:15:05 crc kubenswrapper[4776]: I1125 10:15:05.686149 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e37acd8f-6b60-44d4-9d6f-21fd887ed739" path="/var/lib/kubelet/pods/e37acd8f-6b60-44d4-9d6f-21fd887ed739/volumes" Nov 25 10:15:17 crc kubenswrapper[4776]: I1125 10:15:17.818506 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:15:17 crc kubenswrapper[4776]: I1125 10:15:17.819132 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:15:30 crc kubenswrapper[4776]: I1125 10:15:30.079002 4776 scope.go:117] "RemoveContainer" containerID="252bc9b9dbc3c6b6f3ef21f4a2d7fa9d52b302ae1da040029a791e6056f1b8e5" Nov 25 10:15:47 crc kubenswrapper[4776]: I1125 10:15:47.818612 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:15:47 crc kubenswrapper[4776]: I1125 10:15:47.819248 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:16:17 crc kubenswrapper[4776]: I1125 10:16:17.818497 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:16:17 crc kubenswrapper[4776]: I1125 10:16:17.819662 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:16:17 crc kubenswrapper[4776]: I1125 10:16:17.819723 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 10:16:17 crc kubenswrapper[4776]: I1125 10:16:17.820270 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:16:17 crc kubenswrapper[4776]: I1125 10:16:17.820340 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255" gracePeriod=600 Nov 25 10:16:17 crc kubenswrapper[4776]: E1125 10:16:17.939301 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:16:18 crc kubenswrapper[4776]: I1125 10:16:18.430966 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255" exitCode=0 Nov 25 10:16:18 crc kubenswrapper[4776]: I1125 10:16:18.431094 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255"} Nov 25 10:16:18 crc kubenswrapper[4776]: I1125 10:16:18.431360 4776 scope.go:117] "RemoveContainer" containerID="41c62a00c2902589260314ef41fa8e27ca25f86c7f781c0b248f7cc291cf5f53" Nov 25 10:16:18 crc kubenswrapper[4776]: I1125 10:16:18.432309 4776 scope.go:117] "RemoveContainer" containerID="81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255" Nov 25 10:16:18 crc kubenswrapper[4776]: E1125 10:16:18.432659 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:16:33 crc kubenswrapper[4776]: I1125 10:16:33.662655 4776 scope.go:117] "RemoveContainer" containerID="81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255" Nov 25 10:16:33 crc kubenswrapper[4776]: E1125 10:16:33.663460 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:16:46 crc kubenswrapper[4776]: I1125 10:16:46.662155 4776 scope.go:117] "RemoveContainer" containerID="81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255" Nov 25 10:16:46 crc kubenswrapper[4776]: E1125 10:16:46.663096 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:17:00 crc kubenswrapper[4776]: I1125 10:17:00.663954 4776 scope.go:117] "RemoveContainer" containerID="81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255" Nov 25 10:17:00 crc kubenswrapper[4776]: E1125 10:17:00.664999 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:17:11 crc kubenswrapper[4776]: I1125 10:17:11.662859 4776 scope.go:117] "RemoveContainer" containerID="81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255" Nov 25 10:17:11 crc kubenswrapper[4776]: E1125 10:17:11.663526 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:17:24 crc kubenswrapper[4776]: I1125 10:17:24.662709 4776 scope.go:117] "RemoveContainer" containerID="81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255" Nov 25 10:17:24 crc kubenswrapper[4776]: E1125 10:17:24.663852 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:17:36 crc kubenswrapper[4776]: I1125 10:17:36.662561 4776 scope.go:117] "RemoveContainer" containerID="81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255" Nov 25 10:17:36 crc kubenswrapper[4776]: E1125 10:17:36.663630 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:17:48 crc kubenswrapper[4776]: I1125 10:17:48.662459 4776 scope.go:117] "RemoveContainer" containerID="81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255" Nov 25 10:17:48 crc kubenswrapper[4776]: E1125 10:17:48.663236 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:17:59 crc kubenswrapper[4776]: I1125 10:17:59.662788 4776 scope.go:117] "RemoveContainer" containerID="81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255" Nov 25 10:17:59 crc kubenswrapper[4776]: E1125 10:17:59.663599 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:18:10 crc kubenswrapper[4776]: I1125 10:18:10.663056 4776 scope.go:117] "RemoveContainer" containerID="81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255" Nov 25 10:18:10 crc kubenswrapper[4776]: E1125 10:18:10.664649 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:18:21 crc kubenswrapper[4776]: I1125 10:18:21.662376 4776 scope.go:117] "RemoveContainer" containerID="81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255" Nov 25 10:18:21 crc kubenswrapper[4776]: E1125 10:18:21.663267 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:18:33 crc kubenswrapper[4776]: I1125 10:18:33.662806 4776 scope.go:117] "RemoveContainer" containerID="81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255" Nov 25 10:18:33 crc kubenswrapper[4776]: E1125 10:18:33.663638 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:18:46 crc kubenswrapper[4776]: I1125 10:18:46.662220 4776 scope.go:117] "RemoveContainer" containerID="81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255" Nov 25 10:18:46 crc kubenswrapper[4776]: E1125 10:18:46.663081 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:18:58 crc kubenswrapper[4776]: I1125 10:18:58.662753 4776 scope.go:117] "RemoveContainer" containerID="81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255" Nov 25 10:18:58 crc kubenswrapper[4776]: E1125 10:18:58.664129 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:19:10 crc kubenswrapper[4776]: I1125 10:19:10.942641 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4xwlm"] Nov 25 10:19:10 crc kubenswrapper[4776]: E1125 10:19:10.943576 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="625731b7-d813-461a-b2b3-8b7e64622ff4" containerName="collect-profiles" Nov 25 10:19:10 crc kubenswrapper[4776]: I1125 10:19:10.943590 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="625731b7-d813-461a-b2b3-8b7e64622ff4" containerName="collect-profiles" Nov 25 10:19:10 crc kubenswrapper[4776]: I1125 10:19:10.943805 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="625731b7-d813-461a-b2b3-8b7e64622ff4" containerName="collect-profiles" Nov 25 10:19:10 crc kubenswrapper[4776]: I1125 10:19:10.945031 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4xwlm" Nov 25 10:19:10 crc kubenswrapper[4776]: I1125 10:19:10.949430 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4xwlm"] Nov 25 10:19:11 crc kubenswrapper[4776]: I1125 10:19:11.133160 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkbg2\" (UniqueName: \"kubernetes.io/projected/a4429813-b873-4168-b913-b2f3e9f92567-kube-api-access-bkbg2\") pod \"redhat-operators-4xwlm\" (UID: \"a4429813-b873-4168-b913-b2f3e9f92567\") " pod="openshift-marketplace/redhat-operators-4xwlm" Nov 25 10:19:11 crc kubenswrapper[4776]: I1125 10:19:11.133284 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4429813-b873-4168-b913-b2f3e9f92567-catalog-content\") pod \"redhat-operators-4xwlm\" (UID: \"a4429813-b873-4168-b913-b2f3e9f92567\") " pod="openshift-marketplace/redhat-operators-4xwlm" Nov 25 10:19:11 crc kubenswrapper[4776]: I1125 10:19:11.133321 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4429813-b873-4168-b913-b2f3e9f92567-utilities\") pod \"redhat-operators-4xwlm\" (UID: \"a4429813-b873-4168-b913-b2f3e9f92567\") " pod="openshift-marketplace/redhat-operators-4xwlm" Nov 25 10:19:11 crc kubenswrapper[4776]: I1125 10:19:11.243214 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4429813-b873-4168-b913-b2f3e9f92567-utilities\") pod \"redhat-operators-4xwlm\" (UID: \"a4429813-b873-4168-b913-b2f3e9f92567\") " pod="openshift-marketplace/redhat-operators-4xwlm" Nov 25 10:19:11 crc kubenswrapper[4776]: I1125 10:19:11.243398 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkbg2\" (UniqueName: \"kubernetes.io/projected/a4429813-b873-4168-b913-b2f3e9f92567-kube-api-access-bkbg2\") pod \"redhat-operators-4xwlm\" (UID: \"a4429813-b873-4168-b913-b2f3e9f92567\") " pod="openshift-marketplace/redhat-operators-4xwlm" Nov 25 10:19:11 crc kubenswrapper[4776]: I1125 10:19:11.243514 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4429813-b873-4168-b913-b2f3e9f92567-catalog-content\") pod \"redhat-operators-4xwlm\" (UID: \"a4429813-b873-4168-b913-b2f3e9f92567\") " pod="openshift-marketplace/redhat-operators-4xwlm" Nov 25 10:19:11 crc kubenswrapper[4776]: I1125 10:19:11.244188 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4429813-b873-4168-b913-b2f3e9f92567-catalog-content\") pod \"redhat-operators-4xwlm\" (UID: \"a4429813-b873-4168-b913-b2f3e9f92567\") " pod="openshift-marketplace/redhat-operators-4xwlm" Nov 25 10:19:11 crc kubenswrapper[4776]: I1125 10:19:11.244482 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4429813-b873-4168-b913-b2f3e9f92567-utilities\") pod \"redhat-operators-4xwlm\" (UID: \"a4429813-b873-4168-b913-b2f3e9f92567\") " pod="openshift-marketplace/redhat-operators-4xwlm" Nov 25 10:19:11 crc kubenswrapper[4776]: I1125 10:19:11.275262 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkbg2\" (UniqueName: \"kubernetes.io/projected/a4429813-b873-4168-b913-b2f3e9f92567-kube-api-access-bkbg2\") pod \"redhat-operators-4xwlm\" (UID: \"a4429813-b873-4168-b913-b2f3e9f92567\") " pod="openshift-marketplace/redhat-operators-4xwlm" Nov 25 10:19:11 crc kubenswrapper[4776]: I1125 10:19:11.309678 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4xwlm" Nov 25 10:19:11 crc kubenswrapper[4776]: I1125 10:19:11.662185 4776 scope.go:117] "RemoveContainer" containerID="81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255" Nov 25 10:19:11 crc kubenswrapper[4776]: E1125 10:19:11.662768 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:19:11 crc kubenswrapper[4776]: I1125 10:19:11.747817 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4xwlm"] Nov 25 10:19:12 crc kubenswrapper[4776]: I1125 10:19:12.009536 4776 generic.go:334] "Generic (PLEG): container finished" podID="a4429813-b873-4168-b913-b2f3e9f92567" containerID="a1097b1b96af60e5d83566954858bdaa6be9eb7047a44bdd9f48fee01dc52d6f" exitCode=0 Nov 25 10:19:12 crc kubenswrapper[4776]: I1125 10:19:12.009580 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xwlm" event={"ID":"a4429813-b873-4168-b913-b2f3e9f92567","Type":"ContainerDied","Data":"a1097b1b96af60e5d83566954858bdaa6be9eb7047a44bdd9f48fee01dc52d6f"} Nov 25 10:19:12 crc kubenswrapper[4776]: I1125 10:19:12.009829 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xwlm" event={"ID":"a4429813-b873-4168-b913-b2f3e9f92567","Type":"ContainerStarted","Data":"fc38324621e5c232da9f10c7be1f76bf13c6209cc3ade05087f35724345a94af"} Nov 25 10:19:12 crc kubenswrapper[4776]: I1125 10:19:12.011338 4776 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 10:19:13 crc kubenswrapper[4776]: I1125 10:19:13.018986 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xwlm" event={"ID":"a4429813-b873-4168-b913-b2f3e9f92567","Type":"ContainerStarted","Data":"0d3c6d88ce39ea2615c8970356fa697067a19fe4827ba3f658a3578adadd1621"} Nov 25 10:19:14 crc kubenswrapper[4776]: I1125 10:19:14.030030 4776 generic.go:334] "Generic (PLEG): container finished" podID="a4429813-b873-4168-b913-b2f3e9f92567" containerID="0d3c6d88ce39ea2615c8970356fa697067a19fe4827ba3f658a3578adadd1621" exitCode=0 Nov 25 10:19:14 crc kubenswrapper[4776]: I1125 10:19:14.030138 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xwlm" event={"ID":"a4429813-b873-4168-b913-b2f3e9f92567","Type":"ContainerDied","Data":"0d3c6d88ce39ea2615c8970356fa697067a19fe4827ba3f658a3578adadd1621"} Nov 25 10:19:15 crc kubenswrapper[4776]: I1125 10:19:15.038331 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xwlm" event={"ID":"a4429813-b873-4168-b913-b2f3e9f92567","Type":"ContainerStarted","Data":"3174a5749a7ab166010e11ea3fbda028e3c4cb12bdb9f5b155b1a5b5553a94d5"} Nov 25 10:19:15 crc kubenswrapper[4776]: I1125 10:19:15.058480 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4xwlm" podStartSLOduration=2.616971435 podStartE2EDuration="5.058456839s" podCreationTimestamp="2025-11-25 10:19:10 +0000 UTC" firstStartedPulling="2025-11-25 10:19:12.011009074 +0000 UTC m=+3297.052068627" lastFinishedPulling="2025-11-25 10:19:14.452494468 +0000 UTC m=+3299.493554031" observedRunningTime="2025-11-25 10:19:15.054487009 +0000 UTC m=+3300.095546582" watchObservedRunningTime="2025-11-25 10:19:15.058456839 +0000 UTC m=+3300.099516382" Nov 25 10:19:21 crc kubenswrapper[4776]: I1125 10:19:21.310964 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4xwlm" Nov 25 10:19:21 crc kubenswrapper[4776]: I1125 10:19:21.311517 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4xwlm" Nov 25 10:19:21 crc kubenswrapper[4776]: I1125 10:19:21.355962 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4xwlm" Nov 25 10:19:22 crc kubenswrapper[4776]: I1125 10:19:22.127277 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4xwlm" Nov 25 10:19:22 crc kubenswrapper[4776]: I1125 10:19:22.175325 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4xwlm"] Nov 25 10:19:24 crc kubenswrapper[4776]: I1125 10:19:24.101919 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4xwlm" podUID="a4429813-b873-4168-b913-b2f3e9f92567" containerName="registry-server" containerID="cri-o://3174a5749a7ab166010e11ea3fbda028e3c4cb12bdb9f5b155b1a5b5553a94d5" gracePeriod=2 Nov 25 10:19:24 crc kubenswrapper[4776]: I1125 10:19:24.481718 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4xwlm" Nov 25 10:19:24 crc kubenswrapper[4776]: I1125 10:19:24.545920 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4429813-b873-4168-b913-b2f3e9f92567-utilities\") pod \"a4429813-b873-4168-b913-b2f3e9f92567\" (UID: \"a4429813-b873-4168-b913-b2f3e9f92567\") " Nov 25 10:19:24 crc kubenswrapper[4776]: I1125 10:19:24.546276 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkbg2\" (UniqueName: \"kubernetes.io/projected/a4429813-b873-4168-b913-b2f3e9f92567-kube-api-access-bkbg2\") pod \"a4429813-b873-4168-b913-b2f3e9f92567\" (UID: \"a4429813-b873-4168-b913-b2f3e9f92567\") " Nov 25 10:19:24 crc kubenswrapper[4776]: I1125 10:19:24.546309 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4429813-b873-4168-b913-b2f3e9f92567-catalog-content\") pod \"a4429813-b873-4168-b913-b2f3e9f92567\" (UID: \"a4429813-b873-4168-b913-b2f3e9f92567\") " Nov 25 10:19:24 crc kubenswrapper[4776]: I1125 10:19:24.546777 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4429813-b873-4168-b913-b2f3e9f92567-utilities" (OuterVolumeSpecName: "utilities") pod "a4429813-b873-4168-b913-b2f3e9f92567" (UID: "a4429813-b873-4168-b913-b2f3e9f92567"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:19:24 crc kubenswrapper[4776]: I1125 10:19:24.551126 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4429813-b873-4168-b913-b2f3e9f92567-kube-api-access-bkbg2" (OuterVolumeSpecName: "kube-api-access-bkbg2") pod "a4429813-b873-4168-b913-b2f3e9f92567" (UID: "a4429813-b873-4168-b913-b2f3e9f92567"). InnerVolumeSpecName "kube-api-access-bkbg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:19:24 crc kubenswrapper[4776]: I1125 10:19:24.641447 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4429813-b873-4168-b913-b2f3e9f92567-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a4429813-b873-4168-b913-b2f3e9f92567" (UID: "a4429813-b873-4168-b913-b2f3e9f92567"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:19:24 crc kubenswrapper[4776]: I1125 10:19:24.647833 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4429813-b873-4168-b913-b2f3e9f92567-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:19:24 crc kubenswrapper[4776]: I1125 10:19:24.647884 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkbg2\" (UniqueName: \"kubernetes.io/projected/a4429813-b873-4168-b913-b2f3e9f92567-kube-api-access-bkbg2\") on node \"crc\" DevicePath \"\"" Nov 25 10:19:24 crc kubenswrapper[4776]: I1125 10:19:24.647904 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4429813-b873-4168-b913-b2f3e9f92567-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:19:24 crc kubenswrapper[4776]: I1125 10:19:24.662194 4776 scope.go:117] "RemoveContainer" containerID="81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255" Nov 25 10:19:24 crc kubenswrapper[4776]: E1125 10:19:24.662526 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:19:25 crc kubenswrapper[4776]: I1125 10:19:25.114118 4776 generic.go:334] "Generic (PLEG): container finished" podID="a4429813-b873-4168-b913-b2f3e9f92567" containerID="3174a5749a7ab166010e11ea3fbda028e3c4cb12bdb9f5b155b1a5b5553a94d5" exitCode=0 Nov 25 10:19:25 crc kubenswrapper[4776]: I1125 10:19:25.114184 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xwlm" event={"ID":"a4429813-b873-4168-b913-b2f3e9f92567","Type":"ContainerDied","Data":"3174a5749a7ab166010e11ea3fbda028e3c4cb12bdb9f5b155b1a5b5553a94d5"} Nov 25 10:19:25 crc kubenswrapper[4776]: I1125 10:19:25.114234 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4xwlm" event={"ID":"a4429813-b873-4168-b913-b2f3e9f92567","Type":"ContainerDied","Data":"fc38324621e5c232da9f10c7be1f76bf13c6209cc3ade05087f35724345a94af"} Nov 25 10:19:25 crc kubenswrapper[4776]: I1125 10:19:25.114256 4776 scope.go:117] "RemoveContainer" containerID="3174a5749a7ab166010e11ea3fbda028e3c4cb12bdb9f5b155b1a5b5553a94d5" Nov 25 10:19:25 crc kubenswrapper[4776]: I1125 10:19:25.115696 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4xwlm" Nov 25 10:19:25 crc kubenswrapper[4776]: I1125 10:19:25.150246 4776 scope.go:117] "RemoveContainer" containerID="0d3c6d88ce39ea2615c8970356fa697067a19fe4827ba3f658a3578adadd1621" Nov 25 10:19:25 crc kubenswrapper[4776]: I1125 10:19:25.181136 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4xwlm"] Nov 25 10:19:25 crc kubenswrapper[4776]: I1125 10:19:25.195930 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4xwlm"] Nov 25 10:19:25 crc kubenswrapper[4776]: I1125 10:19:25.199222 4776 scope.go:117] "RemoveContainer" containerID="a1097b1b96af60e5d83566954858bdaa6be9eb7047a44bdd9f48fee01dc52d6f" Nov 25 10:19:25 crc kubenswrapper[4776]: I1125 10:19:25.245675 4776 scope.go:117] "RemoveContainer" containerID="3174a5749a7ab166010e11ea3fbda028e3c4cb12bdb9f5b155b1a5b5553a94d5" Nov 25 10:19:25 crc kubenswrapper[4776]: E1125 10:19:25.246053 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3174a5749a7ab166010e11ea3fbda028e3c4cb12bdb9f5b155b1a5b5553a94d5\": container with ID starting with 3174a5749a7ab166010e11ea3fbda028e3c4cb12bdb9f5b155b1a5b5553a94d5 not found: ID does not exist" containerID="3174a5749a7ab166010e11ea3fbda028e3c4cb12bdb9f5b155b1a5b5553a94d5" Nov 25 10:19:25 crc kubenswrapper[4776]: I1125 10:19:25.246188 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3174a5749a7ab166010e11ea3fbda028e3c4cb12bdb9f5b155b1a5b5553a94d5"} err="failed to get container status \"3174a5749a7ab166010e11ea3fbda028e3c4cb12bdb9f5b155b1a5b5553a94d5\": rpc error: code = NotFound desc = could not find container \"3174a5749a7ab166010e11ea3fbda028e3c4cb12bdb9f5b155b1a5b5553a94d5\": container with ID starting with 3174a5749a7ab166010e11ea3fbda028e3c4cb12bdb9f5b155b1a5b5553a94d5 not found: ID does not exist" Nov 25 10:19:25 crc kubenswrapper[4776]: I1125 10:19:25.246268 4776 scope.go:117] "RemoveContainer" containerID="0d3c6d88ce39ea2615c8970356fa697067a19fe4827ba3f658a3578adadd1621" Nov 25 10:19:25 crc kubenswrapper[4776]: E1125 10:19:25.246596 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d3c6d88ce39ea2615c8970356fa697067a19fe4827ba3f658a3578adadd1621\": container with ID starting with 0d3c6d88ce39ea2615c8970356fa697067a19fe4827ba3f658a3578adadd1621 not found: ID does not exist" containerID="0d3c6d88ce39ea2615c8970356fa697067a19fe4827ba3f658a3578adadd1621" Nov 25 10:19:25 crc kubenswrapper[4776]: I1125 10:19:25.246615 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d3c6d88ce39ea2615c8970356fa697067a19fe4827ba3f658a3578adadd1621"} err="failed to get container status \"0d3c6d88ce39ea2615c8970356fa697067a19fe4827ba3f658a3578adadd1621\": rpc error: code = NotFound desc = could not find container \"0d3c6d88ce39ea2615c8970356fa697067a19fe4827ba3f658a3578adadd1621\": container with ID starting with 0d3c6d88ce39ea2615c8970356fa697067a19fe4827ba3f658a3578adadd1621 not found: ID does not exist" Nov 25 10:19:25 crc kubenswrapper[4776]: I1125 10:19:25.246629 4776 scope.go:117] "RemoveContainer" containerID="a1097b1b96af60e5d83566954858bdaa6be9eb7047a44bdd9f48fee01dc52d6f" Nov 25 10:19:25 crc kubenswrapper[4776]: E1125 10:19:25.247048 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1097b1b96af60e5d83566954858bdaa6be9eb7047a44bdd9f48fee01dc52d6f\": container with ID starting with a1097b1b96af60e5d83566954858bdaa6be9eb7047a44bdd9f48fee01dc52d6f not found: ID does not exist" containerID="a1097b1b96af60e5d83566954858bdaa6be9eb7047a44bdd9f48fee01dc52d6f" Nov 25 10:19:25 crc kubenswrapper[4776]: I1125 10:19:25.247107 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1097b1b96af60e5d83566954858bdaa6be9eb7047a44bdd9f48fee01dc52d6f"} err="failed to get container status \"a1097b1b96af60e5d83566954858bdaa6be9eb7047a44bdd9f48fee01dc52d6f\": rpc error: code = NotFound desc = could not find container \"a1097b1b96af60e5d83566954858bdaa6be9eb7047a44bdd9f48fee01dc52d6f\": container with ID starting with a1097b1b96af60e5d83566954858bdaa6be9eb7047a44bdd9f48fee01dc52d6f not found: ID does not exist" Nov 25 10:19:25 crc kubenswrapper[4776]: I1125 10:19:25.672847 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4429813-b873-4168-b913-b2f3e9f92567" path="/var/lib/kubelet/pods/a4429813-b873-4168-b913-b2f3e9f92567/volumes" Nov 25 10:19:37 crc kubenswrapper[4776]: I1125 10:19:37.764794 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dddh9"] Nov 25 10:19:37 crc kubenswrapper[4776]: E1125 10:19:37.765869 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4429813-b873-4168-b913-b2f3e9f92567" containerName="registry-server" Nov 25 10:19:37 crc kubenswrapper[4776]: I1125 10:19:37.765893 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4429813-b873-4168-b913-b2f3e9f92567" containerName="registry-server" Nov 25 10:19:37 crc kubenswrapper[4776]: E1125 10:19:37.765955 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4429813-b873-4168-b913-b2f3e9f92567" containerName="extract-content" Nov 25 10:19:37 crc kubenswrapper[4776]: I1125 10:19:37.765968 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4429813-b873-4168-b913-b2f3e9f92567" containerName="extract-content" Nov 25 10:19:37 crc kubenswrapper[4776]: E1125 10:19:37.765990 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4429813-b873-4168-b913-b2f3e9f92567" containerName="extract-utilities" Nov 25 10:19:37 crc kubenswrapper[4776]: I1125 10:19:37.766002 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4429813-b873-4168-b913-b2f3e9f92567" containerName="extract-utilities" Nov 25 10:19:37 crc kubenswrapper[4776]: I1125 10:19:37.766264 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4429813-b873-4168-b913-b2f3e9f92567" containerName="registry-server" Nov 25 10:19:37 crc kubenswrapper[4776]: I1125 10:19:37.767824 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dddh9" Nov 25 10:19:37 crc kubenswrapper[4776]: I1125 10:19:37.776230 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dddh9"] Nov 25 10:19:37 crc kubenswrapper[4776]: I1125 10:19:37.819850 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcc046b2-824d-4ab9-bc7f-3a37debd2fb1-utilities\") pod \"redhat-marketplace-dddh9\" (UID: \"fcc046b2-824d-4ab9-bc7f-3a37debd2fb1\") " pod="openshift-marketplace/redhat-marketplace-dddh9" Nov 25 10:19:37 crc kubenswrapper[4776]: I1125 10:19:37.819922 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcc046b2-824d-4ab9-bc7f-3a37debd2fb1-catalog-content\") pod \"redhat-marketplace-dddh9\" (UID: \"fcc046b2-824d-4ab9-bc7f-3a37debd2fb1\") " pod="openshift-marketplace/redhat-marketplace-dddh9" Nov 25 10:19:37 crc kubenswrapper[4776]: I1125 10:19:37.819945 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xdvx\" (UniqueName: \"kubernetes.io/projected/fcc046b2-824d-4ab9-bc7f-3a37debd2fb1-kube-api-access-8xdvx\") pod \"redhat-marketplace-dddh9\" (UID: \"fcc046b2-824d-4ab9-bc7f-3a37debd2fb1\") " pod="openshift-marketplace/redhat-marketplace-dddh9" Nov 25 10:19:37 crc kubenswrapper[4776]: I1125 10:19:37.921441 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcc046b2-824d-4ab9-bc7f-3a37debd2fb1-utilities\") pod \"redhat-marketplace-dddh9\" (UID: \"fcc046b2-824d-4ab9-bc7f-3a37debd2fb1\") " pod="openshift-marketplace/redhat-marketplace-dddh9" Nov 25 10:19:37 crc kubenswrapper[4776]: I1125 10:19:37.921519 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcc046b2-824d-4ab9-bc7f-3a37debd2fb1-catalog-content\") pod \"redhat-marketplace-dddh9\" (UID: \"fcc046b2-824d-4ab9-bc7f-3a37debd2fb1\") " pod="openshift-marketplace/redhat-marketplace-dddh9" Nov 25 10:19:37 crc kubenswrapper[4776]: I1125 10:19:37.921545 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xdvx\" (UniqueName: \"kubernetes.io/projected/fcc046b2-824d-4ab9-bc7f-3a37debd2fb1-kube-api-access-8xdvx\") pod \"redhat-marketplace-dddh9\" (UID: \"fcc046b2-824d-4ab9-bc7f-3a37debd2fb1\") " pod="openshift-marketplace/redhat-marketplace-dddh9" Nov 25 10:19:37 crc kubenswrapper[4776]: I1125 10:19:37.922011 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcc046b2-824d-4ab9-bc7f-3a37debd2fb1-utilities\") pod \"redhat-marketplace-dddh9\" (UID: \"fcc046b2-824d-4ab9-bc7f-3a37debd2fb1\") " pod="openshift-marketplace/redhat-marketplace-dddh9" Nov 25 10:19:37 crc kubenswrapper[4776]: I1125 10:19:37.922105 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcc046b2-824d-4ab9-bc7f-3a37debd2fb1-catalog-content\") pod \"redhat-marketplace-dddh9\" (UID: \"fcc046b2-824d-4ab9-bc7f-3a37debd2fb1\") " pod="openshift-marketplace/redhat-marketplace-dddh9" Nov 25 10:19:37 crc kubenswrapper[4776]: I1125 10:19:37.941211 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xdvx\" (UniqueName: \"kubernetes.io/projected/fcc046b2-824d-4ab9-bc7f-3a37debd2fb1-kube-api-access-8xdvx\") pod \"redhat-marketplace-dddh9\" (UID: \"fcc046b2-824d-4ab9-bc7f-3a37debd2fb1\") " pod="openshift-marketplace/redhat-marketplace-dddh9" Nov 25 10:19:38 crc kubenswrapper[4776]: I1125 10:19:38.084452 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dddh9" Nov 25 10:19:38 crc kubenswrapper[4776]: I1125 10:19:38.505723 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dddh9"] Nov 25 10:19:38 crc kubenswrapper[4776]: W1125 10:19:38.511905 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfcc046b2_824d_4ab9_bc7f_3a37debd2fb1.slice/crio-1a89630493dbc8317a04cb61e9351d5a69b4d5a4db2524efaa25857490aa415f WatchSource:0}: Error finding container 1a89630493dbc8317a04cb61e9351d5a69b4d5a4db2524efaa25857490aa415f: Status 404 returned error can't find the container with id 1a89630493dbc8317a04cb61e9351d5a69b4d5a4db2524efaa25857490aa415f Nov 25 10:19:39 crc kubenswrapper[4776]: I1125 10:19:39.227309 4776 generic.go:334] "Generic (PLEG): container finished" podID="fcc046b2-824d-4ab9-bc7f-3a37debd2fb1" containerID="da529372b4c63cd10d1e5915de96c6434ba6a6bc75b23f9674f37f1c44df89a2" exitCode=0 Nov 25 10:19:39 crc kubenswrapper[4776]: I1125 10:19:39.227363 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dddh9" event={"ID":"fcc046b2-824d-4ab9-bc7f-3a37debd2fb1","Type":"ContainerDied","Data":"da529372b4c63cd10d1e5915de96c6434ba6a6bc75b23f9674f37f1c44df89a2"} Nov 25 10:19:39 crc kubenswrapper[4776]: I1125 10:19:39.227389 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dddh9" event={"ID":"fcc046b2-824d-4ab9-bc7f-3a37debd2fb1","Type":"ContainerStarted","Data":"1a89630493dbc8317a04cb61e9351d5a69b4d5a4db2524efaa25857490aa415f"} Nov 25 10:19:39 crc kubenswrapper[4776]: I1125 10:19:39.663775 4776 scope.go:117] "RemoveContainer" containerID="81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255" Nov 25 10:19:39 crc kubenswrapper[4776]: E1125 10:19:39.664057 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:19:40 crc kubenswrapper[4776]: I1125 10:19:40.236461 4776 generic.go:334] "Generic (PLEG): container finished" podID="fcc046b2-824d-4ab9-bc7f-3a37debd2fb1" containerID="346f811fcadd9eb0c562275266fbd2969213923a420080bf9695c6c8e371020c" exitCode=0 Nov 25 10:19:40 crc kubenswrapper[4776]: I1125 10:19:40.236706 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dddh9" event={"ID":"fcc046b2-824d-4ab9-bc7f-3a37debd2fb1","Type":"ContainerDied","Data":"346f811fcadd9eb0c562275266fbd2969213923a420080bf9695c6c8e371020c"} Nov 25 10:19:41 crc kubenswrapper[4776]: I1125 10:19:41.245357 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dddh9" event={"ID":"fcc046b2-824d-4ab9-bc7f-3a37debd2fb1","Type":"ContainerStarted","Data":"c15475cb4cadb18e06a4b6f90fdce396477aaf12d96f6c074993a2f66819c7fd"} Nov 25 10:19:41 crc kubenswrapper[4776]: I1125 10:19:41.261596 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dddh9" podStartSLOduration=2.7824665790000003 podStartE2EDuration="4.261578407s" podCreationTimestamp="2025-11-25 10:19:37 +0000 UTC" firstStartedPulling="2025-11-25 10:19:39.228601847 +0000 UTC m=+3324.269661400" lastFinishedPulling="2025-11-25 10:19:40.707713675 +0000 UTC m=+3325.748773228" observedRunningTime="2025-11-25 10:19:41.259500964 +0000 UTC m=+3326.300560517" watchObservedRunningTime="2025-11-25 10:19:41.261578407 +0000 UTC m=+3326.302637960" Nov 25 10:19:48 crc kubenswrapper[4776]: I1125 10:19:48.085669 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dddh9" Nov 25 10:19:48 crc kubenswrapper[4776]: I1125 10:19:48.086702 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dddh9" Nov 25 10:19:48 crc kubenswrapper[4776]: I1125 10:19:48.138467 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dddh9" Nov 25 10:19:48 crc kubenswrapper[4776]: I1125 10:19:48.336785 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dddh9" Nov 25 10:19:48 crc kubenswrapper[4776]: I1125 10:19:48.379826 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dddh9"] Nov 25 10:19:50 crc kubenswrapper[4776]: I1125 10:19:50.315471 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dddh9" podUID="fcc046b2-824d-4ab9-bc7f-3a37debd2fb1" containerName="registry-server" containerID="cri-o://c15475cb4cadb18e06a4b6f90fdce396477aaf12d96f6c074993a2f66819c7fd" gracePeriod=2 Nov 25 10:19:50 crc kubenswrapper[4776]: I1125 10:19:50.661745 4776 scope.go:117] "RemoveContainer" containerID="81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255" Nov 25 10:19:50 crc kubenswrapper[4776]: E1125 10:19:50.662337 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:19:50 crc kubenswrapper[4776]: I1125 10:19:50.695945 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dddh9" Nov 25 10:19:50 crc kubenswrapper[4776]: I1125 10:19:50.857957 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xdvx\" (UniqueName: \"kubernetes.io/projected/fcc046b2-824d-4ab9-bc7f-3a37debd2fb1-kube-api-access-8xdvx\") pod \"fcc046b2-824d-4ab9-bc7f-3a37debd2fb1\" (UID: \"fcc046b2-824d-4ab9-bc7f-3a37debd2fb1\") " Nov 25 10:19:50 crc kubenswrapper[4776]: I1125 10:19:50.858012 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcc046b2-824d-4ab9-bc7f-3a37debd2fb1-utilities\") pod \"fcc046b2-824d-4ab9-bc7f-3a37debd2fb1\" (UID: \"fcc046b2-824d-4ab9-bc7f-3a37debd2fb1\") " Nov 25 10:19:50 crc kubenswrapper[4776]: I1125 10:19:50.858150 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcc046b2-824d-4ab9-bc7f-3a37debd2fb1-catalog-content\") pod \"fcc046b2-824d-4ab9-bc7f-3a37debd2fb1\" (UID: \"fcc046b2-824d-4ab9-bc7f-3a37debd2fb1\") " Nov 25 10:19:50 crc kubenswrapper[4776]: I1125 10:19:50.860934 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fcc046b2-824d-4ab9-bc7f-3a37debd2fb1-utilities" (OuterVolumeSpecName: "utilities") pod "fcc046b2-824d-4ab9-bc7f-3a37debd2fb1" (UID: "fcc046b2-824d-4ab9-bc7f-3a37debd2fb1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:19:50 crc kubenswrapper[4776]: I1125 10:19:50.863792 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcc046b2-824d-4ab9-bc7f-3a37debd2fb1-kube-api-access-8xdvx" (OuterVolumeSpecName: "kube-api-access-8xdvx") pod "fcc046b2-824d-4ab9-bc7f-3a37debd2fb1" (UID: "fcc046b2-824d-4ab9-bc7f-3a37debd2fb1"). InnerVolumeSpecName "kube-api-access-8xdvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:19:50 crc kubenswrapper[4776]: I1125 10:19:50.878144 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fcc046b2-824d-4ab9-bc7f-3a37debd2fb1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fcc046b2-824d-4ab9-bc7f-3a37debd2fb1" (UID: "fcc046b2-824d-4ab9-bc7f-3a37debd2fb1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:19:50 crc kubenswrapper[4776]: I1125 10:19:50.959777 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcc046b2-824d-4ab9-bc7f-3a37debd2fb1-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:19:50 crc kubenswrapper[4776]: I1125 10:19:50.959810 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xdvx\" (UniqueName: \"kubernetes.io/projected/fcc046b2-824d-4ab9-bc7f-3a37debd2fb1-kube-api-access-8xdvx\") on node \"crc\" DevicePath \"\"" Nov 25 10:19:50 crc kubenswrapper[4776]: I1125 10:19:50.959821 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcc046b2-824d-4ab9-bc7f-3a37debd2fb1-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:19:51 crc kubenswrapper[4776]: I1125 10:19:51.323418 4776 generic.go:334] "Generic (PLEG): container finished" podID="fcc046b2-824d-4ab9-bc7f-3a37debd2fb1" containerID="c15475cb4cadb18e06a4b6f90fdce396477aaf12d96f6c074993a2f66819c7fd" exitCode=0 Nov 25 10:19:51 crc kubenswrapper[4776]: I1125 10:19:51.323464 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dddh9" Nov 25 10:19:51 crc kubenswrapper[4776]: I1125 10:19:51.323463 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dddh9" event={"ID":"fcc046b2-824d-4ab9-bc7f-3a37debd2fb1","Type":"ContainerDied","Data":"c15475cb4cadb18e06a4b6f90fdce396477aaf12d96f6c074993a2f66819c7fd"} Nov 25 10:19:51 crc kubenswrapper[4776]: I1125 10:19:51.323628 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dddh9" event={"ID":"fcc046b2-824d-4ab9-bc7f-3a37debd2fb1","Type":"ContainerDied","Data":"1a89630493dbc8317a04cb61e9351d5a69b4d5a4db2524efaa25857490aa415f"} Nov 25 10:19:51 crc kubenswrapper[4776]: I1125 10:19:51.323668 4776 scope.go:117] "RemoveContainer" containerID="c15475cb4cadb18e06a4b6f90fdce396477aaf12d96f6c074993a2f66819c7fd" Nov 25 10:19:51 crc kubenswrapper[4776]: I1125 10:19:51.340091 4776 scope.go:117] "RemoveContainer" containerID="346f811fcadd9eb0c562275266fbd2969213923a420080bf9695c6c8e371020c" Nov 25 10:19:51 crc kubenswrapper[4776]: I1125 10:19:51.352835 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dddh9"] Nov 25 10:19:51 crc kubenswrapper[4776]: I1125 10:19:51.359496 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dddh9"] Nov 25 10:19:51 crc kubenswrapper[4776]: I1125 10:19:51.384904 4776 scope.go:117] "RemoveContainer" containerID="da529372b4c63cd10d1e5915de96c6434ba6a6bc75b23f9674f37f1c44df89a2" Nov 25 10:19:51 crc kubenswrapper[4776]: I1125 10:19:51.400664 4776 scope.go:117] "RemoveContainer" containerID="c15475cb4cadb18e06a4b6f90fdce396477aaf12d96f6c074993a2f66819c7fd" Nov 25 10:19:51 crc kubenswrapper[4776]: E1125 10:19:51.401238 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c15475cb4cadb18e06a4b6f90fdce396477aaf12d96f6c074993a2f66819c7fd\": container with ID starting with c15475cb4cadb18e06a4b6f90fdce396477aaf12d96f6c074993a2f66819c7fd not found: ID does not exist" containerID="c15475cb4cadb18e06a4b6f90fdce396477aaf12d96f6c074993a2f66819c7fd" Nov 25 10:19:51 crc kubenswrapper[4776]: I1125 10:19:51.401287 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c15475cb4cadb18e06a4b6f90fdce396477aaf12d96f6c074993a2f66819c7fd"} err="failed to get container status \"c15475cb4cadb18e06a4b6f90fdce396477aaf12d96f6c074993a2f66819c7fd\": rpc error: code = NotFound desc = could not find container \"c15475cb4cadb18e06a4b6f90fdce396477aaf12d96f6c074993a2f66819c7fd\": container with ID starting with c15475cb4cadb18e06a4b6f90fdce396477aaf12d96f6c074993a2f66819c7fd not found: ID does not exist" Nov 25 10:19:51 crc kubenswrapper[4776]: I1125 10:19:51.401312 4776 scope.go:117] "RemoveContainer" containerID="346f811fcadd9eb0c562275266fbd2969213923a420080bf9695c6c8e371020c" Nov 25 10:19:51 crc kubenswrapper[4776]: E1125 10:19:51.401746 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"346f811fcadd9eb0c562275266fbd2969213923a420080bf9695c6c8e371020c\": container with ID starting with 346f811fcadd9eb0c562275266fbd2969213923a420080bf9695c6c8e371020c not found: ID does not exist" containerID="346f811fcadd9eb0c562275266fbd2969213923a420080bf9695c6c8e371020c" Nov 25 10:19:51 crc kubenswrapper[4776]: I1125 10:19:51.401771 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"346f811fcadd9eb0c562275266fbd2969213923a420080bf9695c6c8e371020c"} err="failed to get container status \"346f811fcadd9eb0c562275266fbd2969213923a420080bf9695c6c8e371020c\": rpc error: code = NotFound desc = could not find container \"346f811fcadd9eb0c562275266fbd2969213923a420080bf9695c6c8e371020c\": container with ID starting with 346f811fcadd9eb0c562275266fbd2969213923a420080bf9695c6c8e371020c not found: ID does not exist" Nov 25 10:19:51 crc kubenswrapper[4776]: I1125 10:19:51.401785 4776 scope.go:117] "RemoveContainer" containerID="da529372b4c63cd10d1e5915de96c6434ba6a6bc75b23f9674f37f1c44df89a2" Nov 25 10:19:51 crc kubenswrapper[4776]: E1125 10:19:51.401991 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da529372b4c63cd10d1e5915de96c6434ba6a6bc75b23f9674f37f1c44df89a2\": container with ID starting with da529372b4c63cd10d1e5915de96c6434ba6a6bc75b23f9674f37f1c44df89a2 not found: ID does not exist" containerID="da529372b4c63cd10d1e5915de96c6434ba6a6bc75b23f9674f37f1c44df89a2" Nov 25 10:19:51 crc kubenswrapper[4776]: I1125 10:19:51.402012 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da529372b4c63cd10d1e5915de96c6434ba6a6bc75b23f9674f37f1c44df89a2"} err="failed to get container status \"da529372b4c63cd10d1e5915de96c6434ba6a6bc75b23f9674f37f1c44df89a2\": rpc error: code = NotFound desc = could not find container \"da529372b4c63cd10d1e5915de96c6434ba6a6bc75b23f9674f37f1c44df89a2\": container with ID starting with da529372b4c63cd10d1e5915de96c6434ba6a6bc75b23f9674f37f1c44df89a2 not found: ID does not exist" Nov 25 10:19:51 crc kubenswrapper[4776]: I1125 10:19:51.672233 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcc046b2-824d-4ab9-bc7f-3a37debd2fb1" path="/var/lib/kubelet/pods/fcc046b2-824d-4ab9-bc7f-3a37debd2fb1/volumes" Nov 25 10:20:04 crc kubenswrapper[4776]: I1125 10:20:04.662211 4776 scope.go:117] "RemoveContainer" containerID="81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255" Nov 25 10:20:04 crc kubenswrapper[4776]: E1125 10:20:04.663011 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:20:17 crc kubenswrapper[4776]: I1125 10:20:17.859493 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-prbh7"] Nov 25 10:20:17 crc kubenswrapper[4776]: E1125 10:20:17.861799 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcc046b2-824d-4ab9-bc7f-3a37debd2fb1" containerName="extract-utilities" Nov 25 10:20:17 crc kubenswrapper[4776]: I1125 10:20:17.861826 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcc046b2-824d-4ab9-bc7f-3a37debd2fb1" containerName="extract-utilities" Nov 25 10:20:17 crc kubenswrapper[4776]: E1125 10:20:17.861867 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcc046b2-824d-4ab9-bc7f-3a37debd2fb1" containerName="registry-server" Nov 25 10:20:17 crc kubenswrapper[4776]: I1125 10:20:17.861877 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcc046b2-824d-4ab9-bc7f-3a37debd2fb1" containerName="registry-server" Nov 25 10:20:17 crc kubenswrapper[4776]: E1125 10:20:17.861891 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcc046b2-824d-4ab9-bc7f-3a37debd2fb1" containerName="extract-content" Nov 25 10:20:17 crc kubenswrapper[4776]: I1125 10:20:17.861901 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcc046b2-824d-4ab9-bc7f-3a37debd2fb1" containerName="extract-content" Nov 25 10:20:17 crc kubenswrapper[4776]: I1125 10:20:17.862096 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcc046b2-824d-4ab9-bc7f-3a37debd2fb1" containerName="registry-server" Nov 25 10:20:17 crc kubenswrapper[4776]: I1125 10:20:17.863507 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-prbh7" Nov 25 10:20:17 crc kubenswrapper[4776]: I1125 10:20:17.869806 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-prbh7"] Nov 25 10:20:17 crc kubenswrapper[4776]: I1125 10:20:17.886177 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gz5p4\" (UniqueName: \"kubernetes.io/projected/5652ee1d-0df1-49d4-b377-9e7b97904751-kube-api-access-gz5p4\") pod \"community-operators-prbh7\" (UID: \"5652ee1d-0df1-49d4-b377-9e7b97904751\") " pod="openshift-marketplace/community-operators-prbh7" Nov 25 10:20:17 crc kubenswrapper[4776]: I1125 10:20:17.886237 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5652ee1d-0df1-49d4-b377-9e7b97904751-catalog-content\") pod \"community-operators-prbh7\" (UID: \"5652ee1d-0df1-49d4-b377-9e7b97904751\") " pod="openshift-marketplace/community-operators-prbh7" Nov 25 10:20:17 crc kubenswrapper[4776]: I1125 10:20:17.886264 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5652ee1d-0df1-49d4-b377-9e7b97904751-utilities\") pod \"community-operators-prbh7\" (UID: \"5652ee1d-0df1-49d4-b377-9e7b97904751\") " pod="openshift-marketplace/community-operators-prbh7" Nov 25 10:20:17 crc kubenswrapper[4776]: I1125 10:20:17.987402 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5652ee1d-0df1-49d4-b377-9e7b97904751-catalog-content\") pod \"community-operators-prbh7\" (UID: \"5652ee1d-0df1-49d4-b377-9e7b97904751\") " pod="openshift-marketplace/community-operators-prbh7" Nov 25 10:20:17 crc kubenswrapper[4776]: I1125 10:20:17.987455 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5652ee1d-0df1-49d4-b377-9e7b97904751-utilities\") pod \"community-operators-prbh7\" (UID: \"5652ee1d-0df1-49d4-b377-9e7b97904751\") " pod="openshift-marketplace/community-operators-prbh7" Nov 25 10:20:17 crc kubenswrapper[4776]: I1125 10:20:17.987567 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gz5p4\" (UniqueName: \"kubernetes.io/projected/5652ee1d-0df1-49d4-b377-9e7b97904751-kube-api-access-gz5p4\") pod \"community-operators-prbh7\" (UID: \"5652ee1d-0df1-49d4-b377-9e7b97904751\") " pod="openshift-marketplace/community-operators-prbh7" Nov 25 10:20:17 crc kubenswrapper[4776]: I1125 10:20:17.987911 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5652ee1d-0df1-49d4-b377-9e7b97904751-catalog-content\") pod \"community-operators-prbh7\" (UID: \"5652ee1d-0df1-49d4-b377-9e7b97904751\") " pod="openshift-marketplace/community-operators-prbh7" Nov 25 10:20:17 crc kubenswrapper[4776]: I1125 10:20:17.987918 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5652ee1d-0df1-49d4-b377-9e7b97904751-utilities\") pod \"community-operators-prbh7\" (UID: \"5652ee1d-0df1-49d4-b377-9e7b97904751\") " pod="openshift-marketplace/community-operators-prbh7" Nov 25 10:20:18 crc kubenswrapper[4776]: I1125 10:20:18.008402 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gz5p4\" (UniqueName: \"kubernetes.io/projected/5652ee1d-0df1-49d4-b377-9e7b97904751-kube-api-access-gz5p4\") pod \"community-operators-prbh7\" (UID: \"5652ee1d-0df1-49d4-b377-9e7b97904751\") " pod="openshift-marketplace/community-operators-prbh7" Nov 25 10:20:18 crc kubenswrapper[4776]: I1125 10:20:18.188810 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-prbh7" Nov 25 10:20:18 crc kubenswrapper[4776]: I1125 10:20:18.520736 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-prbh7"] Nov 25 10:20:18 crc kubenswrapper[4776]: I1125 10:20:18.549726 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-prbh7" event={"ID":"5652ee1d-0df1-49d4-b377-9e7b97904751","Type":"ContainerStarted","Data":"25e9dec16c611f7a091b9578a47124f0e1b287ce548d98d78a28c3c6a76f3146"} Nov 25 10:20:19 crc kubenswrapper[4776]: I1125 10:20:19.560221 4776 generic.go:334] "Generic (PLEG): container finished" podID="5652ee1d-0df1-49d4-b377-9e7b97904751" containerID="31b16e90c436ac9b63aac1ae5adb241d217f28c986360eaef7401ada5c083dd1" exitCode=0 Nov 25 10:20:19 crc kubenswrapper[4776]: I1125 10:20:19.560325 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-prbh7" event={"ID":"5652ee1d-0df1-49d4-b377-9e7b97904751","Type":"ContainerDied","Data":"31b16e90c436ac9b63aac1ae5adb241d217f28c986360eaef7401ada5c083dd1"} Nov 25 10:20:19 crc kubenswrapper[4776]: I1125 10:20:19.663181 4776 scope.go:117] "RemoveContainer" containerID="81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255" Nov 25 10:20:19 crc kubenswrapper[4776]: E1125 10:20:19.663442 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:20:21 crc kubenswrapper[4776]: I1125 10:20:21.574391 4776 generic.go:334] "Generic (PLEG): container finished" podID="5652ee1d-0df1-49d4-b377-9e7b97904751" containerID="571c686568353b7334c4154ba00391b09f1db293d1465f82068ad2ff97d81031" exitCode=0 Nov 25 10:20:21 crc kubenswrapper[4776]: I1125 10:20:21.574474 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-prbh7" event={"ID":"5652ee1d-0df1-49d4-b377-9e7b97904751","Type":"ContainerDied","Data":"571c686568353b7334c4154ba00391b09f1db293d1465f82068ad2ff97d81031"} Nov 25 10:20:22 crc kubenswrapper[4776]: I1125 10:20:22.582822 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-prbh7" event={"ID":"5652ee1d-0df1-49d4-b377-9e7b97904751","Type":"ContainerStarted","Data":"389eeb8d87d4d14eaa4eb781fc6193b5f7c03c43fa9fc85a46fc38f07079575a"} Nov 25 10:20:22 crc kubenswrapper[4776]: I1125 10:20:22.604400 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-prbh7" podStartSLOduration=3.213176316 podStartE2EDuration="5.604383118s" podCreationTimestamp="2025-11-25 10:20:17 +0000 UTC" firstStartedPulling="2025-11-25 10:20:19.563055807 +0000 UTC m=+3364.604115360" lastFinishedPulling="2025-11-25 10:20:21.954262609 +0000 UTC m=+3366.995322162" observedRunningTime="2025-11-25 10:20:22.600637544 +0000 UTC m=+3367.641697097" watchObservedRunningTime="2025-11-25 10:20:22.604383118 +0000 UTC m=+3367.645442671" Nov 25 10:20:28 crc kubenswrapper[4776]: I1125 10:20:28.189241 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-prbh7" Nov 25 10:20:28 crc kubenswrapper[4776]: I1125 10:20:28.189613 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-prbh7" Nov 25 10:20:28 crc kubenswrapper[4776]: I1125 10:20:28.239456 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-prbh7" Nov 25 10:20:28 crc kubenswrapper[4776]: I1125 10:20:28.673867 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-prbh7" Nov 25 10:20:28 crc kubenswrapper[4776]: I1125 10:20:28.719680 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-prbh7"] Nov 25 10:20:30 crc kubenswrapper[4776]: I1125 10:20:30.641969 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-prbh7" podUID="5652ee1d-0df1-49d4-b377-9e7b97904751" containerName="registry-server" containerID="cri-o://389eeb8d87d4d14eaa4eb781fc6193b5f7c03c43fa9fc85a46fc38f07079575a" gracePeriod=2 Nov 25 10:20:31 crc kubenswrapper[4776]: I1125 10:20:31.022800 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-prbh7" Nov 25 10:20:31 crc kubenswrapper[4776]: I1125 10:20:31.175867 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gz5p4\" (UniqueName: \"kubernetes.io/projected/5652ee1d-0df1-49d4-b377-9e7b97904751-kube-api-access-gz5p4\") pod \"5652ee1d-0df1-49d4-b377-9e7b97904751\" (UID: \"5652ee1d-0df1-49d4-b377-9e7b97904751\") " Nov 25 10:20:31 crc kubenswrapper[4776]: I1125 10:20:31.175929 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5652ee1d-0df1-49d4-b377-9e7b97904751-catalog-content\") pod \"5652ee1d-0df1-49d4-b377-9e7b97904751\" (UID: \"5652ee1d-0df1-49d4-b377-9e7b97904751\") " Nov 25 10:20:31 crc kubenswrapper[4776]: I1125 10:20:31.176005 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5652ee1d-0df1-49d4-b377-9e7b97904751-utilities\") pod \"5652ee1d-0df1-49d4-b377-9e7b97904751\" (UID: \"5652ee1d-0df1-49d4-b377-9e7b97904751\") " Nov 25 10:20:31 crc kubenswrapper[4776]: I1125 10:20:31.177088 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5652ee1d-0df1-49d4-b377-9e7b97904751-utilities" (OuterVolumeSpecName: "utilities") pod "5652ee1d-0df1-49d4-b377-9e7b97904751" (UID: "5652ee1d-0df1-49d4-b377-9e7b97904751"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:20:31 crc kubenswrapper[4776]: I1125 10:20:31.182508 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5652ee1d-0df1-49d4-b377-9e7b97904751-kube-api-access-gz5p4" (OuterVolumeSpecName: "kube-api-access-gz5p4") pod "5652ee1d-0df1-49d4-b377-9e7b97904751" (UID: "5652ee1d-0df1-49d4-b377-9e7b97904751"). InnerVolumeSpecName "kube-api-access-gz5p4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:20:31 crc kubenswrapper[4776]: I1125 10:20:31.233012 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5652ee1d-0df1-49d4-b377-9e7b97904751-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5652ee1d-0df1-49d4-b377-9e7b97904751" (UID: "5652ee1d-0df1-49d4-b377-9e7b97904751"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:20:31 crc kubenswrapper[4776]: I1125 10:20:31.277877 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gz5p4\" (UniqueName: \"kubernetes.io/projected/5652ee1d-0df1-49d4-b377-9e7b97904751-kube-api-access-gz5p4\") on node \"crc\" DevicePath \"\"" Nov 25 10:20:31 crc kubenswrapper[4776]: I1125 10:20:31.277952 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5652ee1d-0df1-49d4-b377-9e7b97904751-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:20:31 crc kubenswrapper[4776]: I1125 10:20:31.277962 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5652ee1d-0df1-49d4-b377-9e7b97904751-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:20:31 crc kubenswrapper[4776]: I1125 10:20:31.652334 4776 generic.go:334] "Generic (PLEG): container finished" podID="5652ee1d-0df1-49d4-b377-9e7b97904751" containerID="389eeb8d87d4d14eaa4eb781fc6193b5f7c03c43fa9fc85a46fc38f07079575a" exitCode=0 Nov 25 10:20:31 crc kubenswrapper[4776]: I1125 10:20:31.652383 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-prbh7" event={"ID":"5652ee1d-0df1-49d4-b377-9e7b97904751","Type":"ContainerDied","Data":"389eeb8d87d4d14eaa4eb781fc6193b5f7c03c43fa9fc85a46fc38f07079575a"} Nov 25 10:20:31 crc kubenswrapper[4776]: I1125 10:20:31.652412 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-prbh7" event={"ID":"5652ee1d-0df1-49d4-b377-9e7b97904751","Type":"ContainerDied","Data":"25e9dec16c611f7a091b9578a47124f0e1b287ce548d98d78a28c3c6a76f3146"} Nov 25 10:20:31 crc kubenswrapper[4776]: I1125 10:20:31.652431 4776 scope.go:117] "RemoveContainer" containerID="389eeb8d87d4d14eaa4eb781fc6193b5f7c03c43fa9fc85a46fc38f07079575a" Nov 25 10:20:31 crc kubenswrapper[4776]: I1125 10:20:31.652430 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-prbh7" Nov 25 10:20:31 crc kubenswrapper[4776]: I1125 10:20:31.675961 4776 scope.go:117] "RemoveContainer" containerID="571c686568353b7334c4154ba00391b09f1db293d1465f82068ad2ff97d81031" Nov 25 10:20:31 crc kubenswrapper[4776]: I1125 10:20:31.688018 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-prbh7"] Nov 25 10:20:31 crc kubenswrapper[4776]: I1125 10:20:31.694288 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-prbh7"] Nov 25 10:20:31 crc kubenswrapper[4776]: I1125 10:20:31.718772 4776 scope.go:117] "RemoveContainer" containerID="31b16e90c436ac9b63aac1ae5adb241d217f28c986360eaef7401ada5c083dd1" Nov 25 10:20:31 crc kubenswrapper[4776]: I1125 10:20:31.736680 4776 scope.go:117] "RemoveContainer" containerID="389eeb8d87d4d14eaa4eb781fc6193b5f7c03c43fa9fc85a46fc38f07079575a" Nov 25 10:20:31 crc kubenswrapper[4776]: E1125 10:20:31.737060 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"389eeb8d87d4d14eaa4eb781fc6193b5f7c03c43fa9fc85a46fc38f07079575a\": container with ID starting with 389eeb8d87d4d14eaa4eb781fc6193b5f7c03c43fa9fc85a46fc38f07079575a not found: ID does not exist" containerID="389eeb8d87d4d14eaa4eb781fc6193b5f7c03c43fa9fc85a46fc38f07079575a" Nov 25 10:20:31 crc kubenswrapper[4776]: I1125 10:20:31.737124 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"389eeb8d87d4d14eaa4eb781fc6193b5f7c03c43fa9fc85a46fc38f07079575a"} err="failed to get container status \"389eeb8d87d4d14eaa4eb781fc6193b5f7c03c43fa9fc85a46fc38f07079575a\": rpc error: code = NotFound desc = could not find container \"389eeb8d87d4d14eaa4eb781fc6193b5f7c03c43fa9fc85a46fc38f07079575a\": container with ID starting with 389eeb8d87d4d14eaa4eb781fc6193b5f7c03c43fa9fc85a46fc38f07079575a not found: ID does not exist" Nov 25 10:20:31 crc kubenswrapper[4776]: I1125 10:20:31.737152 4776 scope.go:117] "RemoveContainer" containerID="571c686568353b7334c4154ba00391b09f1db293d1465f82068ad2ff97d81031" Nov 25 10:20:31 crc kubenswrapper[4776]: E1125 10:20:31.737532 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"571c686568353b7334c4154ba00391b09f1db293d1465f82068ad2ff97d81031\": container with ID starting with 571c686568353b7334c4154ba00391b09f1db293d1465f82068ad2ff97d81031 not found: ID does not exist" containerID="571c686568353b7334c4154ba00391b09f1db293d1465f82068ad2ff97d81031" Nov 25 10:20:31 crc kubenswrapper[4776]: I1125 10:20:31.737560 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"571c686568353b7334c4154ba00391b09f1db293d1465f82068ad2ff97d81031"} err="failed to get container status \"571c686568353b7334c4154ba00391b09f1db293d1465f82068ad2ff97d81031\": rpc error: code = NotFound desc = could not find container \"571c686568353b7334c4154ba00391b09f1db293d1465f82068ad2ff97d81031\": container with ID starting with 571c686568353b7334c4154ba00391b09f1db293d1465f82068ad2ff97d81031 not found: ID does not exist" Nov 25 10:20:31 crc kubenswrapper[4776]: I1125 10:20:31.737574 4776 scope.go:117] "RemoveContainer" containerID="31b16e90c436ac9b63aac1ae5adb241d217f28c986360eaef7401ada5c083dd1" Nov 25 10:20:31 crc kubenswrapper[4776]: E1125 10:20:31.737766 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31b16e90c436ac9b63aac1ae5adb241d217f28c986360eaef7401ada5c083dd1\": container with ID starting with 31b16e90c436ac9b63aac1ae5adb241d217f28c986360eaef7401ada5c083dd1 not found: ID does not exist" containerID="31b16e90c436ac9b63aac1ae5adb241d217f28c986360eaef7401ada5c083dd1" Nov 25 10:20:31 crc kubenswrapper[4776]: I1125 10:20:31.737788 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31b16e90c436ac9b63aac1ae5adb241d217f28c986360eaef7401ada5c083dd1"} err="failed to get container status \"31b16e90c436ac9b63aac1ae5adb241d217f28c986360eaef7401ada5c083dd1\": rpc error: code = NotFound desc = could not find container \"31b16e90c436ac9b63aac1ae5adb241d217f28c986360eaef7401ada5c083dd1\": container with ID starting with 31b16e90c436ac9b63aac1ae5adb241d217f28c986360eaef7401ada5c083dd1 not found: ID does not exist" Nov 25 10:20:32 crc kubenswrapper[4776]: I1125 10:20:32.662032 4776 scope.go:117] "RemoveContainer" containerID="81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255" Nov 25 10:20:32 crc kubenswrapper[4776]: E1125 10:20:32.663121 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:20:33 crc kubenswrapper[4776]: I1125 10:20:33.678109 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5652ee1d-0df1-49d4-b377-9e7b97904751" path="/var/lib/kubelet/pods/5652ee1d-0df1-49d4-b377-9e7b97904751/volumes" Nov 25 10:20:40 crc kubenswrapper[4776]: I1125 10:20:40.064161 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5hq92"] Nov 25 10:20:40 crc kubenswrapper[4776]: E1125 10:20:40.065229 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5652ee1d-0df1-49d4-b377-9e7b97904751" containerName="extract-utilities" Nov 25 10:20:40 crc kubenswrapper[4776]: I1125 10:20:40.065248 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="5652ee1d-0df1-49d4-b377-9e7b97904751" containerName="extract-utilities" Nov 25 10:20:40 crc kubenswrapper[4776]: E1125 10:20:40.065273 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5652ee1d-0df1-49d4-b377-9e7b97904751" containerName="registry-server" Nov 25 10:20:40 crc kubenswrapper[4776]: I1125 10:20:40.065283 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="5652ee1d-0df1-49d4-b377-9e7b97904751" containerName="registry-server" Nov 25 10:20:40 crc kubenswrapper[4776]: E1125 10:20:40.065303 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5652ee1d-0df1-49d4-b377-9e7b97904751" containerName="extract-content" Nov 25 10:20:40 crc kubenswrapper[4776]: I1125 10:20:40.065314 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="5652ee1d-0df1-49d4-b377-9e7b97904751" containerName="extract-content" Nov 25 10:20:40 crc kubenswrapper[4776]: I1125 10:20:40.065544 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="5652ee1d-0df1-49d4-b377-9e7b97904751" containerName="registry-server" Nov 25 10:20:40 crc kubenswrapper[4776]: I1125 10:20:40.067096 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5hq92" Nov 25 10:20:40 crc kubenswrapper[4776]: I1125 10:20:40.077431 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5hq92"] Nov 25 10:20:40 crc kubenswrapper[4776]: I1125 10:20:40.223585 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e0c6366-cf02-4e5c-acc9-1fbae7134a87-utilities\") pod \"certified-operators-5hq92\" (UID: \"2e0c6366-cf02-4e5c-acc9-1fbae7134a87\") " pod="openshift-marketplace/certified-operators-5hq92" Nov 25 10:20:40 crc kubenswrapper[4776]: I1125 10:20:40.223632 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e0c6366-cf02-4e5c-acc9-1fbae7134a87-catalog-content\") pod \"certified-operators-5hq92\" (UID: \"2e0c6366-cf02-4e5c-acc9-1fbae7134a87\") " pod="openshift-marketplace/certified-operators-5hq92" Nov 25 10:20:40 crc kubenswrapper[4776]: I1125 10:20:40.223779 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgz8f\" (UniqueName: \"kubernetes.io/projected/2e0c6366-cf02-4e5c-acc9-1fbae7134a87-kube-api-access-qgz8f\") pod \"certified-operators-5hq92\" (UID: \"2e0c6366-cf02-4e5c-acc9-1fbae7134a87\") " pod="openshift-marketplace/certified-operators-5hq92" Nov 25 10:20:40 crc kubenswrapper[4776]: I1125 10:20:40.324898 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e0c6366-cf02-4e5c-acc9-1fbae7134a87-utilities\") pod \"certified-operators-5hq92\" (UID: \"2e0c6366-cf02-4e5c-acc9-1fbae7134a87\") " pod="openshift-marketplace/certified-operators-5hq92" Nov 25 10:20:40 crc kubenswrapper[4776]: I1125 10:20:40.324946 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e0c6366-cf02-4e5c-acc9-1fbae7134a87-catalog-content\") pod \"certified-operators-5hq92\" (UID: \"2e0c6366-cf02-4e5c-acc9-1fbae7134a87\") " pod="openshift-marketplace/certified-operators-5hq92" Nov 25 10:20:40 crc kubenswrapper[4776]: I1125 10:20:40.325002 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgz8f\" (UniqueName: \"kubernetes.io/projected/2e0c6366-cf02-4e5c-acc9-1fbae7134a87-kube-api-access-qgz8f\") pod \"certified-operators-5hq92\" (UID: \"2e0c6366-cf02-4e5c-acc9-1fbae7134a87\") " pod="openshift-marketplace/certified-operators-5hq92" Nov 25 10:20:40 crc kubenswrapper[4776]: I1125 10:20:40.325713 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e0c6366-cf02-4e5c-acc9-1fbae7134a87-utilities\") pod \"certified-operators-5hq92\" (UID: \"2e0c6366-cf02-4e5c-acc9-1fbae7134a87\") " pod="openshift-marketplace/certified-operators-5hq92" Nov 25 10:20:40 crc kubenswrapper[4776]: I1125 10:20:40.325925 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e0c6366-cf02-4e5c-acc9-1fbae7134a87-catalog-content\") pod \"certified-operators-5hq92\" (UID: \"2e0c6366-cf02-4e5c-acc9-1fbae7134a87\") " pod="openshift-marketplace/certified-operators-5hq92" Nov 25 10:20:40 crc kubenswrapper[4776]: I1125 10:20:40.351576 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgz8f\" (UniqueName: \"kubernetes.io/projected/2e0c6366-cf02-4e5c-acc9-1fbae7134a87-kube-api-access-qgz8f\") pod \"certified-operators-5hq92\" (UID: \"2e0c6366-cf02-4e5c-acc9-1fbae7134a87\") " pod="openshift-marketplace/certified-operators-5hq92" Nov 25 10:20:40 crc kubenswrapper[4776]: I1125 10:20:40.396739 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5hq92" Nov 25 10:20:40 crc kubenswrapper[4776]: I1125 10:20:40.843739 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5hq92"] Nov 25 10:20:41 crc kubenswrapper[4776]: I1125 10:20:41.743257 4776 generic.go:334] "Generic (PLEG): container finished" podID="2e0c6366-cf02-4e5c-acc9-1fbae7134a87" containerID="570579a0afa3cdfd0a12ba6c94dae074002cc00d859d935d0861d52c9cb5ab2d" exitCode=0 Nov 25 10:20:41 crc kubenswrapper[4776]: I1125 10:20:41.743373 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5hq92" event={"ID":"2e0c6366-cf02-4e5c-acc9-1fbae7134a87","Type":"ContainerDied","Data":"570579a0afa3cdfd0a12ba6c94dae074002cc00d859d935d0861d52c9cb5ab2d"} Nov 25 10:20:41 crc kubenswrapper[4776]: I1125 10:20:41.743598 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5hq92" event={"ID":"2e0c6366-cf02-4e5c-acc9-1fbae7134a87","Type":"ContainerStarted","Data":"e1d420eb9a4cb434d9e51845f1a64ca4b0fa143057071f33b0ca1e7d27e532e0"} Nov 25 10:20:43 crc kubenswrapper[4776]: I1125 10:20:43.757671 4776 generic.go:334] "Generic (PLEG): container finished" podID="2e0c6366-cf02-4e5c-acc9-1fbae7134a87" containerID="aa5ba0c0c73dbea795ad3dabe25524a7d7f7392e9cdfefb2e0c9fd54315489a9" exitCode=0 Nov 25 10:20:43 crc kubenswrapper[4776]: I1125 10:20:43.757789 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5hq92" event={"ID":"2e0c6366-cf02-4e5c-acc9-1fbae7134a87","Type":"ContainerDied","Data":"aa5ba0c0c73dbea795ad3dabe25524a7d7f7392e9cdfefb2e0c9fd54315489a9"} Nov 25 10:20:44 crc kubenswrapper[4776]: I1125 10:20:44.768239 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5hq92" event={"ID":"2e0c6366-cf02-4e5c-acc9-1fbae7134a87","Type":"ContainerStarted","Data":"cc73707266971ecd85ba423f0f6edcb293766679273f6982ee68e7ee39b670f2"} Nov 25 10:20:44 crc kubenswrapper[4776]: I1125 10:20:44.790222 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5hq92" podStartSLOduration=2.303217961 podStartE2EDuration="4.790200468s" podCreationTimestamp="2025-11-25 10:20:40 +0000 UTC" firstStartedPulling="2025-11-25 10:20:41.745219154 +0000 UTC m=+3386.786278707" lastFinishedPulling="2025-11-25 10:20:44.232201661 +0000 UTC m=+3389.273261214" observedRunningTime="2025-11-25 10:20:44.786602137 +0000 UTC m=+3389.827661750" watchObservedRunningTime="2025-11-25 10:20:44.790200468 +0000 UTC m=+3389.831260021" Nov 25 10:20:47 crc kubenswrapper[4776]: I1125 10:20:47.662334 4776 scope.go:117] "RemoveContainer" containerID="81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255" Nov 25 10:20:47 crc kubenswrapper[4776]: E1125 10:20:47.662901 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:20:50 crc kubenswrapper[4776]: I1125 10:20:50.397457 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5hq92" Nov 25 10:20:50 crc kubenswrapper[4776]: I1125 10:20:50.397526 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5hq92" Nov 25 10:20:50 crc kubenswrapper[4776]: I1125 10:20:50.441022 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5hq92" Nov 25 10:20:50 crc kubenswrapper[4776]: I1125 10:20:50.853740 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5hq92" Nov 25 10:20:50 crc kubenswrapper[4776]: I1125 10:20:50.896628 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5hq92"] Nov 25 10:20:52 crc kubenswrapper[4776]: I1125 10:20:52.823649 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5hq92" podUID="2e0c6366-cf02-4e5c-acc9-1fbae7134a87" containerName="registry-server" containerID="cri-o://cc73707266971ecd85ba423f0f6edcb293766679273f6982ee68e7ee39b670f2" gracePeriod=2 Nov 25 10:20:53 crc kubenswrapper[4776]: I1125 10:20:53.194571 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5hq92" Nov 25 10:20:53 crc kubenswrapper[4776]: I1125 10:20:53.324359 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e0c6366-cf02-4e5c-acc9-1fbae7134a87-utilities\") pod \"2e0c6366-cf02-4e5c-acc9-1fbae7134a87\" (UID: \"2e0c6366-cf02-4e5c-acc9-1fbae7134a87\") " Nov 25 10:20:53 crc kubenswrapper[4776]: I1125 10:20:53.324509 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgz8f\" (UniqueName: \"kubernetes.io/projected/2e0c6366-cf02-4e5c-acc9-1fbae7134a87-kube-api-access-qgz8f\") pod \"2e0c6366-cf02-4e5c-acc9-1fbae7134a87\" (UID: \"2e0c6366-cf02-4e5c-acc9-1fbae7134a87\") " Nov 25 10:20:53 crc kubenswrapper[4776]: I1125 10:20:53.324673 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e0c6366-cf02-4e5c-acc9-1fbae7134a87-catalog-content\") pod \"2e0c6366-cf02-4e5c-acc9-1fbae7134a87\" (UID: \"2e0c6366-cf02-4e5c-acc9-1fbae7134a87\") " Nov 25 10:20:53 crc kubenswrapper[4776]: I1125 10:20:53.325417 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e0c6366-cf02-4e5c-acc9-1fbae7134a87-utilities" (OuterVolumeSpecName: "utilities") pod "2e0c6366-cf02-4e5c-acc9-1fbae7134a87" (UID: "2e0c6366-cf02-4e5c-acc9-1fbae7134a87"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:20:53 crc kubenswrapper[4776]: I1125 10:20:53.340593 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e0c6366-cf02-4e5c-acc9-1fbae7134a87-kube-api-access-qgz8f" (OuterVolumeSpecName: "kube-api-access-qgz8f") pod "2e0c6366-cf02-4e5c-acc9-1fbae7134a87" (UID: "2e0c6366-cf02-4e5c-acc9-1fbae7134a87"). InnerVolumeSpecName "kube-api-access-qgz8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:20:53 crc kubenswrapper[4776]: I1125 10:20:53.390269 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e0c6366-cf02-4e5c-acc9-1fbae7134a87-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2e0c6366-cf02-4e5c-acc9-1fbae7134a87" (UID: "2e0c6366-cf02-4e5c-acc9-1fbae7134a87"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:20:53 crc kubenswrapper[4776]: I1125 10:20:53.427034 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e0c6366-cf02-4e5c-acc9-1fbae7134a87-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:20:53 crc kubenswrapper[4776]: I1125 10:20:53.427105 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e0c6366-cf02-4e5c-acc9-1fbae7134a87-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:20:53 crc kubenswrapper[4776]: I1125 10:20:53.427121 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgz8f\" (UniqueName: \"kubernetes.io/projected/2e0c6366-cf02-4e5c-acc9-1fbae7134a87-kube-api-access-qgz8f\") on node \"crc\" DevicePath \"\"" Nov 25 10:20:53 crc kubenswrapper[4776]: I1125 10:20:53.832782 4776 generic.go:334] "Generic (PLEG): container finished" podID="2e0c6366-cf02-4e5c-acc9-1fbae7134a87" containerID="cc73707266971ecd85ba423f0f6edcb293766679273f6982ee68e7ee39b670f2" exitCode=0 Nov 25 10:20:53 crc kubenswrapper[4776]: I1125 10:20:53.832836 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5hq92" event={"ID":"2e0c6366-cf02-4e5c-acc9-1fbae7134a87","Type":"ContainerDied","Data":"cc73707266971ecd85ba423f0f6edcb293766679273f6982ee68e7ee39b670f2"} Nov 25 10:20:53 crc kubenswrapper[4776]: I1125 10:20:53.833148 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5hq92" event={"ID":"2e0c6366-cf02-4e5c-acc9-1fbae7134a87","Type":"ContainerDied","Data":"e1d420eb9a4cb434d9e51845f1a64ca4b0fa143057071f33b0ca1e7d27e532e0"} Nov 25 10:20:53 crc kubenswrapper[4776]: I1125 10:20:53.833176 4776 scope.go:117] "RemoveContainer" containerID="cc73707266971ecd85ba423f0f6edcb293766679273f6982ee68e7ee39b670f2" Nov 25 10:20:53 crc kubenswrapper[4776]: I1125 10:20:53.832937 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5hq92" Nov 25 10:20:53 crc kubenswrapper[4776]: I1125 10:20:53.850694 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5hq92"] Nov 25 10:20:53 crc kubenswrapper[4776]: I1125 10:20:53.855139 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5hq92"] Nov 25 10:20:53 crc kubenswrapper[4776]: I1125 10:20:53.857019 4776 scope.go:117] "RemoveContainer" containerID="aa5ba0c0c73dbea795ad3dabe25524a7d7f7392e9cdfefb2e0c9fd54315489a9" Nov 25 10:20:53 crc kubenswrapper[4776]: I1125 10:20:53.875477 4776 scope.go:117] "RemoveContainer" containerID="570579a0afa3cdfd0a12ba6c94dae074002cc00d859d935d0861d52c9cb5ab2d" Nov 25 10:20:53 crc kubenswrapper[4776]: I1125 10:20:53.906001 4776 scope.go:117] "RemoveContainer" containerID="cc73707266971ecd85ba423f0f6edcb293766679273f6982ee68e7ee39b670f2" Nov 25 10:20:53 crc kubenswrapper[4776]: E1125 10:20:53.906633 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc73707266971ecd85ba423f0f6edcb293766679273f6982ee68e7ee39b670f2\": container with ID starting with cc73707266971ecd85ba423f0f6edcb293766679273f6982ee68e7ee39b670f2 not found: ID does not exist" containerID="cc73707266971ecd85ba423f0f6edcb293766679273f6982ee68e7ee39b670f2" Nov 25 10:20:53 crc kubenswrapper[4776]: I1125 10:20:53.906690 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc73707266971ecd85ba423f0f6edcb293766679273f6982ee68e7ee39b670f2"} err="failed to get container status \"cc73707266971ecd85ba423f0f6edcb293766679273f6982ee68e7ee39b670f2\": rpc error: code = NotFound desc = could not find container \"cc73707266971ecd85ba423f0f6edcb293766679273f6982ee68e7ee39b670f2\": container with ID starting with cc73707266971ecd85ba423f0f6edcb293766679273f6982ee68e7ee39b670f2 not found: ID does not exist" Nov 25 10:20:53 crc kubenswrapper[4776]: I1125 10:20:53.906716 4776 scope.go:117] "RemoveContainer" containerID="aa5ba0c0c73dbea795ad3dabe25524a7d7f7392e9cdfefb2e0c9fd54315489a9" Nov 25 10:20:53 crc kubenswrapper[4776]: E1125 10:20:53.907082 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa5ba0c0c73dbea795ad3dabe25524a7d7f7392e9cdfefb2e0c9fd54315489a9\": container with ID starting with aa5ba0c0c73dbea795ad3dabe25524a7d7f7392e9cdfefb2e0c9fd54315489a9 not found: ID does not exist" containerID="aa5ba0c0c73dbea795ad3dabe25524a7d7f7392e9cdfefb2e0c9fd54315489a9" Nov 25 10:20:53 crc kubenswrapper[4776]: I1125 10:20:53.907115 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa5ba0c0c73dbea795ad3dabe25524a7d7f7392e9cdfefb2e0c9fd54315489a9"} err="failed to get container status \"aa5ba0c0c73dbea795ad3dabe25524a7d7f7392e9cdfefb2e0c9fd54315489a9\": rpc error: code = NotFound desc = could not find container \"aa5ba0c0c73dbea795ad3dabe25524a7d7f7392e9cdfefb2e0c9fd54315489a9\": container with ID starting with aa5ba0c0c73dbea795ad3dabe25524a7d7f7392e9cdfefb2e0c9fd54315489a9 not found: ID does not exist" Nov 25 10:20:53 crc kubenswrapper[4776]: I1125 10:20:53.907138 4776 scope.go:117] "RemoveContainer" containerID="570579a0afa3cdfd0a12ba6c94dae074002cc00d859d935d0861d52c9cb5ab2d" Nov 25 10:20:53 crc kubenswrapper[4776]: E1125 10:20:53.907404 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"570579a0afa3cdfd0a12ba6c94dae074002cc00d859d935d0861d52c9cb5ab2d\": container with ID starting with 570579a0afa3cdfd0a12ba6c94dae074002cc00d859d935d0861d52c9cb5ab2d not found: ID does not exist" containerID="570579a0afa3cdfd0a12ba6c94dae074002cc00d859d935d0861d52c9cb5ab2d" Nov 25 10:20:53 crc kubenswrapper[4776]: I1125 10:20:53.907446 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"570579a0afa3cdfd0a12ba6c94dae074002cc00d859d935d0861d52c9cb5ab2d"} err="failed to get container status \"570579a0afa3cdfd0a12ba6c94dae074002cc00d859d935d0861d52c9cb5ab2d\": rpc error: code = NotFound desc = could not find container \"570579a0afa3cdfd0a12ba6c94dae074002cc00d859d935d0861d52c9cb5ab2d\": container with ID starting with 570579a0afa3cdfd0a12ba6c94dae074002cc00d859d935d0861d52c9cb5ab2d not found: ID does not exist" Nov 25 10:20:55 crc kubenswrapper[4776]: I1125 10:20:55.670565 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e0c6366-cf02-4e5c-acc9-1fbae7134a87" path="/var/lib/kubelet/pods/2e0c6366-cf02-4e5c-acc9-1fbae7134a87/volumes" Nov 25 10:21:02 crc kubenswrapper[4776]: I1125 10:21:02.663156 4776 scope.go:117] "RemoveContainer" containerID="81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255" Nov 25 10:21:02 crc kubenswrapper[4776]: E1125 10:21:02.663919 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:21:15 crc kubenswrapper[4776]: I1125 10:21:15.668041 4776 scope.go:117] "RemoveContainer" containerID="81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255" Nov 25 10:21:15 crc kubenswrapper[4776]: E1125 10:21:15.671338 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:21:28 crc kubenswrapper[4776]: I1125 10:21:28.663505 4776 scope.go:117] "RemoveContainer" containerID="81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255" Nov 25 10:21:29 crc kubenswrapper[4776]: I1125 10:21:29.082510 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"99a3b8d7554581a19173c727796026fc9d0ad051a1c359c17f1ea382bf424612"} Nov 25 10:23:47 crc kubenswrapper[4776]: I1125 10:23:47.818843 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:23:47 crc kubenswrapper[4776]: I1125 10:23:47.819450 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:24:17 crc kubenswrapper[4776]: I1125 10:24:17.818882 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:24:17 crc kubenswrapper[4776]: I1125 10:24:17.819624 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:24:47 crc kubenswrapper[4776]: I1125 10:24:47.819140 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:24:47 crc kubenswrapper[4776]: I1125 10:24:47.819772 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:24:47 crc kubenswrapper[4776]: I1125 10:24:47.819827 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 10:24:47 crc kubenswrapper[4776]: I1125 10:24:47.820369 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"99a3b8d7554581a19173c727796026fc9d0ad051a1c359c17f1ea382bf424612"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:24:47 crc kubenswrapper[4776]: I1125 10:24:47.820435 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://99a3b8d7554581a19173c727796026fc9d0ad051a1c359c17f1ea382bf424612" gracePeriod=600 Nov 25 10:24:48 crc kubenswrapper[4776]: I1125 10:24:48.495353 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="99a3b8d7554581a19173c727796026fc9d0ad051a1c359c17f1ea382bf424612" exitCode=0 Nov 25 10:24:48 crc kubenswrapper[4776]: I1125 10:24:48.495437 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"99a3b8d7554581a19173c727796026fc9d0ad051a1c359c17f1ea382bf424612"} Nov 25 10:24:48 crc kubenswrapper[4776]: I1125 10:24:48.495693 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf"} Nov 25 10:24:48 crc kubenswrapper[4776]: I1125 10:24:48.495715 4776 scope.go:117] "RemoveContainer" containerID="81b5de3a755c2613832445d2933e3da42d14c3b872857d493f8297429b1c1255" Nov 25 10:27:17 crc kubenswrapper[4776]: I1125 10:27:17.819145 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:27:17 crc kubenswrapper[4776]: I1125 10:27:17.819673 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:27:47 crc kubenswrapper[4776]: I1125 10:27:47.829628 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:27:47 crc kubenswrapper[4776]: I1125 10:27:47.833286 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:28:17 crc kubenswrapper[4776]: I1125 10:28:17.818856 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:28:17 crc kubenswrapper[4776]: I1125 10:28:17.819478 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:28:17 crc kubenswrapper[4776]: I1125 10:28:17.819529 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 10:28:17 crc kubenswrapper[4776]: I1125 10:28:17.820213 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:28:17 crc kubenswrapper[4776]: I1125 10:28:17.820269 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf" gracePeriod=600 Nov 25 10:28:17 crc kubenswrapper[4776]: E1125 10:28:17.951042 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:28:18 crc kubenswrapper[4776]: I1125 10:28:18.939050 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf" exitCode=0 Nov 25 10:28:18 crc kubenswrapper[4776]: I1125 10:28:18.939146 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf"} Nov 25 10:28:18 crc kubenswrapper[4776]: I1125 10:28:18.939267 4776 scope.go:117] "RemoveContainer" containerID="99a3b8d7554581a19173c727796026fc9d0ad051a1c359c17f1ea382bf424612" Nov 25 10:28:18 crc kubenswrapper[4776]: I1125 10:28:18.939837 4776 scope.go:117] "RemoveContainer" containerID="356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf" Nov 25 10:28:18 crc kubenswrapper[4776]: E1125 10:28:18.940976 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:28:33 crc kubenswrapper[4776]: I1125 10:28:33.662136 4776 scope.go:117] "RemoveContainer" containerID="356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf" Nov 25 10:28:33 crc kubenswrapper[4776]: E1125 10:28:33.664487 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:28:47 crc kubenswrapper[4776]: I1125 10:28:47.662711 4776 scope.go:117] "RemoveContainer" containerID="356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf" Nov 25 10:28:47 crc kubenswrapper[4776]: E1125 10:28:47.663739 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:29:00 crc kubenswrapper[4776]: I1125 10:29:00.661963 4776 scope.go:117] "RemoveContainer" containerID="356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf" Nov 25 10:29:00 crc kubenswrapper[4776]: E1125 10:29:00.662881 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:29:11 crc kubenswrapper[4776]: I1125 10:29:11.662149 4776 scope.go:117] "RemoveContainer" containerID="356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf" Nov 25 10:29:11 crc kubenswrapper[4776]: E1125 10:29:11.663001 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:29:23 crc kubenswrapper[4776]: I1125 10:29:23.662650 4776 scope.go:117] "RemoveContainer" containerID="356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf" Nov 25 10:29:23 crc kubenswrapper[4776]: E1125 10:29:23.663426 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:29:35 crc kubenswrapper[4776]: I1125 10:29:35.667953 4776 scope.go:117] "RemoveContainer" containerID="356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf" Nov 25 10:29:35 crc kubenswrapper[4776]: E1125 10:29:35.668829 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:29:50 crc kubenswrapper[4776]: I1125 10:29:50.662547 4776 scope.go:117] "RemoveContainer" containerID="356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf" Nov 25 10:29:50 crc kubenswrapper[4776]: E1125 10:29:50.663466 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:30:00 crc kubenswrapper[4776]: I1125 10:30:00.150365 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401110-zvmnb"] Nov 25 10:30:00 crc kubenswrapper[4776]: E1125 10:30:00.151211 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e0c6366-cf02-4e5c-acc9-1fbae7134a87" containerName="registry-server" Nov 25 10:30:00 crc kubenswrapper[4776]: I1125 10:30:00.151225 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e0c6366-cf02-4e5c-acc9-1fbae7134a87" containerName="registry-server" Nov 25 10:30:00 crc kubenswrapper[4776]: E1125 10:30:00.151240 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e0c6366-cf02-4e5c-acc9-1fbae7134a87" containerName="extract-utilities" Nov 25 10:30:00 crc kubenswrapper[4776]: I1125 10:30:00.151248 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e0c6366-cf02-4e5c-acc9-1fbae7134a87" containerName="extract-utilities" Nov 25 10:30:00 crc kubenswrapper[4776]: E1125 10:30:00.151259 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e0c6366-cf02-4e5c-acc9-1fbae7134a87" containerName="extract-content" Nov 25 10:30:00 crc kubenswrapper[4776]: I1125 10:30:00.151266 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e0c6366-cf02-4e5c-acc9-1fbae7134a87" containerName="extract-content" Nov 25 10:30:00 crc kubenswrapper[4776]: I1125 10:30:00.151423 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e0c6366-cf02-4e5c-acc9-1fbae7134a87" containerName="registry-server" Nov 25 10:30:00 crc kubenswrapper[4776]: I1125 10:30:00.151968 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-zvmnb" Nov 25 10:30:00 crc kubenswrapper[4776]: I1125 10:30:00.155354 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 10:30:00 crc kubenswrapper[4776]: I1125 10:30:00.155656 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401110-zvmnb"] Nov 25 10:30:00 crc kubenswrapper[4776]: I1125 10:30:00.160109 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 10:30:00 crc kubenswrapper[4776]: I1125 10:30:00.250603 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba-secret-volume\") pod \"collect-profiles-29401110-zvmnb\" (UID: \"e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-zvmnb" Nov 25 10:30:00 crc kubenswrapper[4776]: I1125 10:30:00.250689 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c84c\" (UniqueName: \"kubernetes.io/projected/e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba-kube-api-access-8c84c\") pod \"collect-profiles-29401110-zvmnb\" (UID: \"e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-zvmnb" Nov 25 10:30:00 crc kubenswrapper[4776]: I1125 10:30:00.250860 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba-config-volume\") pod \"collect-profiles-29401110-zvmnb\" (UID: \"e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-zvmnb" Nov 25 10:30:00 crc kubenswrapper[4776]: I1125 10:30:00.351683 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba-config-volume\") pod \"collect-profiles-29401110-zvmnb\" (UID: \"e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-zvmnb" Nov 25 10:30:00 crc kubenswrapper[4776]: I1125 10:30:00.351771 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba-secret-volume\") pod \"collect-profiles-29401110-zvmnb\" (UID: \"e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-zvmnb" Nov 25 10:30:00 crc kubenswrapper[4776]: I1125 10:30:00.351819 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c84c\" (UniqueName: \"kubernetes.io/projected/e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba-kube-api-access-8c84c\") pod \"collect-profiles-29401110-zvmnb\" (UID: \"e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-zvmnb" Nov 25 10:30:00 crc kubenswrapper[4776]: I1125 10:30:00.352916 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba-config-volume\") pod \"collect-profiles-29401110-zvmnb\" (UID: \"e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-zvmnb" Nov 25 10:30:00 crc kubenswrapper[4776]: I1125 10:30:00.362468 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba-secret-volume\") pod \"collect-profiles-29401110-zvmnb\" (UID: \"e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-zvmnb" Nov 25 10:30:00 crc kubenswrapper[4776]: I1125 10:30:00.369500 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c84c\" (UniqueName: \"kubernetes.io/projected/e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba-kube-api-access-8c84c\") pod \"collect-profiles-29401110-zvmnb\" (UID: \"e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-zvmnb" Nov 25 10:30:00 crc kubenswrapper[4776]: I1125 10:30:00.473596 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-zvmnb" Nov 25 10:30:01 crc kubenswrapper[4776]: I1125 10:30:01.204454 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401110-zvmnb"] Nov 25 10:30:01 crc kubenswrapper[4776]: I1125 10:30:01.889213 4776 generic.go:334] "Generic (PLEG): container finished" podID="e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba" containerID="139ec3a7186c7a3dbacfefd2816155567ee5fa7b87a7852026635b3aaaff1030" exitCode=0 Nov 25 10:30:01 crc kubenswrapper[4776]: I1125 10:30:01.889292 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-zvmnb" event={"ID":"e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba","Type":"ContainerDied","Data":"139ec3a7186c7a3dbacfefd2816155567ee5fa7b87a7852026635b3aaaff1030"} Nov 25 10:30:01 crc kubenswrapper[4776]: I1125 10:30:01.889530 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-zvmnb" event={"ID":"e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba","Type":"ContainerStarted","Data":"ae8d0d0f8510be10dc7644c538b5fa45bac118472527e8cdc01c5bda00be294b"} Nov 25 10:30:02 crc kubenswrapper[4776]: I1125 10:30:02.662591 4776 scope.go:117] "RemoveContainer" containerID="356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf" Nov 25 10:30:02 crc kubenswrapper[4776]: E1125 10:30:02.663106 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:30:03 crc kubenswrapper[4776]: I1125 10:30:03.151915 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-zvmnb" Nov 25 10:30:03 crc kubenswrapper[4776]: I1125 10:30:03.217005 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8c84c\" (UniqueName: \"kubernetes.io/projected/e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba-kube-api-access-8c84c\") pod \"e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba\" (UID: \"e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba\") " Nov 25 10:30:03 crc kubenswrapper[4776]: I1125 10:30:03.217089 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba-secret-volume\") pod \"e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba\" (UID: \"e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba\") " Nov 25 10:30:03 crc kubenswrapper[4776]: I1125 10:30:03.217122 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba-config-volume\") pod \"e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba\" (UID: \"e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba\") " Nov 25 10:30:03 crc kubenswrapper[4776]: I1125 10:30:03.218315 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba-config-volume" (OuterVolumeSpecName: "config-volume") pod "e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba" (UID: "e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:30:03 crc kubenswrapper[4776]: I1125 10:30:03.223256 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba" (UID: "e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:30:03 crc kubenswrapper[4776]: I1125 10:30:03.223884 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba-kube-api-access-8c84c" (OuterVolumeSpecName: "kube-api-access-8c84c") pod "e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba" (UID: "e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba"). InnerVolumeSpecName "kube-api-access-8c84c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:30:03 crc kubenswrapper[4776]: I1125 10:30:03.318428 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8c84c\" (UniqueName: \"kubernetes.io/projected/e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba-kube-api-access-8c84c\") on node \"crc\" DevicePath \"\"" Nov 25 10:30:03 crc kubenswrapper[4776]: I1125 10:30:03.318475 4776 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 10:30:03 crc kubenswrapper[4776]: I1125 10:30:03.318484 4776 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 10:30:03 crc kubenswrapper[4776]: I1125 10:30:03.904601 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-zvmnb" event={"ID":"e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba","Type":"ContainerDied","Data":"ae8d0d0f8510be10dc7644c538b5fa45bac118472527e8cdc01c5bda00be294b"} Nov 25 10:30:03 crc kubenswrapper[4776]: I1125 10:30:03.904640 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae8d0d0f8510be10dc7644c538b5fa45bac118472527e8cdc01c5bda00be294b" Nov 25 10:30:03 crc kubenswrapper[4776]: I1125 10:30:03.904673 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-zvmnb" Nov 25 10:30:04 crc kubenswrapper[4776]: I1125 10:30:04.223188 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401065-tf5xr"] Nov 25 10:30:04 crc kubenswrapper[4776]: I1125 10:30:04.227525 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401065-tf5xr"] Nov 25 10:30:05 crc kubenswrapper[4776]: I1125 10:30:05.673635 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffedd17d-0d6c-4f5c-88b5-a11e9931d1af" path="/var/lib/kubelet/pods/ffedd17d-0d6c-4f5c-88b5-a11e9931d1af/volumes" Nov 25 10:30:16 crc kubenswrapper[4776]: I1125 10:30:16.662583 4776 scope.go:117] "RemoveContainer" containerID="356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf" Nov 25 10:30:16 crc kubenswrapper[4776]: E1125 10:30:16.663231 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:30:30 crc kubenswrapper[4776]: I1125 10:30:30.393948 4776 scope.go:117] "RemoveContainer" containerID="190f0504e8bf2e09fc36c3fc20e9878eb5565206fdf350f5381773a2a111c697" Nov 25 10:30:31 crc kubenswrapper[4776]: I1125 10:30:31.662485 4776 scope.go:117] "RemoveContainer" containerID="356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf" Nov 25 10:30:31 crc kubenswrapper[4776]: E1125 10:30:31.663185 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:30:42 crc kubenswrapper[4776]: I1125 10:30:42.683041 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dnrsz"] Nov 25 10:30:42 crc kubenswrapper[4776]: E1125 10:30:42.684166 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba" containerName="collect-profiles" Nov 25 10:30:42 crc kubenswrapper[4776]: I1125 10:30:42.684184 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba" containerName="collect-profiles" Nov 25 10:30:42 crc kubenswrapper[4776]: I1125 10:30:42.684356 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba" containerName="collect-profiles" Nov 25 10:30:42 crc kubenswrapper[4776]: I1125 10:30:42.685588 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dnrsz" Nov 25 10:30:42 crc kubenswrapper[4776]: I1125 10:30:42.700787 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dnrsz"] Nov 25 10:30:42 crc kubenswrapper[4776]: I1125 10:30:42.703161 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcklt\" (UniqueName: \"kubernetes.io/projected/3a378c9a-dd8f-448b-b229-7e875605f913-kube-api-access-qcklt\") pod \"certified-operators-dnrsz\" (UID: \"3a378c9a-dd8f-448b-b229-7e875605f913\") " pod="openshift-marketplace/certified-operators-dnrsz" Nov 25 10:30:42 crc kubenswrapper[4776]: I1125 10:30:42.703340 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a378c9a-dd8f-448b-b229-7e875605f913-utilities\") pod \"certified-operators-dnrsz\" (UID: \"3a378c9a-dd8f-448b-b229-7e875605f913\") " pod="openshift-marketplace/certified-operators-dnrsz" Nov 25 10:30:42 crc kubenswrapper[4776]: I1125 10:30:42.703374 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a378c9a-dd8f-448b-b229-7e875605f913-catalog-content\") pod \"certified-operators-dnrsz\" (UID: \"3a378c9a-dd8f-448b-b229-7e875605f913\") " pod="openshift-marketplace/certified-operators-dnrsz" Nov 25 10:30:42 crc kubenswrapper[4776]: I1125 10:30:42.805221 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a378c9a-dd8f-448b-b229-7e875605f913-utilities\") pod \"certified-operators-dnrsz\" (UID: \"3a378c9a-dd8f-448b-b229-7e875605f913\") " pod="openshift-marketplace/certified-operators-dnrsz" Nov 25 10:30:42 crc kubenswrapper[4776]: I1125 10:30:42.805276 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a378c9a-dd8f-448b-b229-7e875605f913-catalog-content\") pod \"certified-operators-dnrsz\" (UID: \"3a378c9a-dd8f-448b-b229-7e875605f913\") " pod="openshift-marketplace/certified-operators-dnrsz" Nov 25 10:30:42 crc kubenswrapper[4776]: I1125 10:30:42.805350 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcklt\" (UniqueName: \"kubernetes.io/projected/3a378c9a-dd8f-448b-b229-7e875605f913-kube-api-access-qcklt\") pod \"certified-operators-dnrsz\" (UID: \"3a378c9a-dd8f-448b-b229-7e875605f913\") " pod="openshift-marketplace/certified-operators-dnrsz" Nov 25 10:30:42 crc kubenswrapper[4776]: I1125 10:30:42.805888 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a378c9a-dd8f-448b-b229-7e875605f913-catalog-content\") pod \"certified-operators-dnrsz\" (UID: \"3a378c9a-dd8f-448b-b229-7e875605f913\") " pod="openshift-marketplace/certified-operators-dnrsz" Nov 25 10:30:42 crc kubenswrapper[4776]: I1125 10:30:42.806270 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a378c9a-dd8f-448b-b229-7e875605f913-utilities\") pod \"certified-operators-dnrsz\" (UID: \"3a378c9a-dd8f-448b-b229-7e875605f913\") " pod="openshift-marketplace/certified-operators-dnrsz" Nov 25 10:30:42 crc kubenswrapper[4776]: I1125 10:30:42.834729 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcklt\" (UniqueName: \"kubernetes.io/projected/3a378c9a-dd8f-448b-b229-7e875605f913-kube-api-access-qcklt\") pod \"certified-operators-dnrsz\" (UID: \"3a378c9a-dd8f-448b-b229-7e875605f913\") " pod="openshift-marketplace/certified-operators-dnrsz" Nov 25 10:30:43 crc kubenswrapper[4776]: I1125 10:30:43.013387 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dnrsz" Nov 25 10:30:43 crc kubenswrapper[4776]: I1125 10:30:43.508998 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dnrsz"] Nov 25 10:30:43 crc kubenswrapper[4776]: I1125 10:30:43.662875 4776 scope.go:117] "RemoveContainer" containerID="356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf" Nov 25 10:30:43 crc kubenswrapper[4776]: E1125 10:30:43.663460 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:30:44 crc kubenswrapper[4776]: I1125 10:30:44.208311 4776 generic.go:334] "Generic (PLEG): container finished" podID="3a378c9a-dd8f-448b-b229-7e875605f913" containerID="b02e0fbc1495fb6366f882f12c7793445136165f11dd3c3d023650746afbbeb2" exitCode=0 Nov 25 10:30:44 crc kubenswrapper[4776]: I1125 10:30:44.208359 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnrsz" event={"ID":"3a378c9a-dd8f-448b-b229-7e875605f913","Type":"ContainerDied","Data":"b02e0fbc1495fb6366f882f12c7793445136165f11dd3c3d023650746afbbeb2"} Nov 25 10:30:44 crc kubenswrapper[4776]: I1125 10:30:44.208387 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnrsz" event={"ID":"3a378c9a-dd8f-448b-b229-7e875605f913","Type":"ContainerStarted","Data":"5726d909d1c95615e9411372e2bb2dbf9de513faba9230805e71e8dca1510d73"} Nov 25 10:30:44 crc kubenswrapper[4776]: I1125 10:30:44.210123 4776 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 10:30:45 crc kubenswrapper[4776]: I1125 10:30:45.216979 4776 generic.go:334] "Generic (PLEG): container finished" podID="3a378c9a-dd8f-448b-b229-7e875605f913" containerID="e1a11952018f61942dcd64f9288c012740a89b5dc631087a823380bac24aa360" exitCode=0 Nov 25 10:30:45 crc kubenswrapper[4776]: I1125 10:30:45.217048 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnrsz" event={"ID":"3a378c9a-dd8f-448b-b229-7e875605f913","Type":"ContainerDied","Data":"e1a11952018f61942dcd64f9288c012740a89b5dc631087a823380bac24aa360"} Nov 25 10:30:46 crc kubenswrapper[4776]: I1125 10:30:46.228180 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnrsz" event={"ID":"3a378c9a-dd8f-448b-b229-7e875605f913","Type":"ContainerStarted","Data":"35de714215a81f56597dd5c0cd0c86f2eaf12e6fe59e5b4634640309e11a9e99"} Nov 25 10:30:46 crc kubenswrapper[4776]: I1125 10:30:46.255938 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dnrsz" podStartSLOduration=2.824909983 podStartE2EDuration="4.255917497s" podCreationTimestamp="2025-11-25 10:30:42 +0000 UTC" firstStartedPulling="2025-11-25 10:30:44.209824927 +0000 UTC m=+3989.250884480" lastFinishedPulling="2025-11-25 10:30:45.640832441 +0000 UTC m=+3990.681891994" observedRunningTime="2025-11-25 10:30:46.250436199 +0000 UTC m=+3991.291495772" watchObservedRunningTime="2025-11-25 10:30:46.255917497 +0000 UTC m=+3991.296977050" Nov 25 10:30:53 crc kubenswrapper[4776]: I1125 10:30:53.014153 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dnrsz" Nov 25 10:30:53 crc kubenswrapper[4776]: I1125 10:30:53.014586 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dnrsz" Nov 25 10:30:53 crc kubenswrapper[4776]: I1125 10:30:53.056799 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dnrsz" Nov 25 10:30:53 crc kubenswrapper[4776]: I1125 10:30:53.345253 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dnrsz" Nov 25 10:30:53 crc kubenswrapper[4776]: I1125 10:30:53.384971 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dnrsz"] Nov 25 10:30:55 crc kubenswrapper[4776]: I1125 10:30:55.308389 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dnrsz" podUID="3a378c9a-dd8f-448b-b229-7e875605f913" containerName="registry-server" containerID="cri-o://35de714215a81f56597dd5c0cd0c86f2eaf12e6fe59e5b4634640309e11a9e99" gracePeriod=2 Nov 25 10:30:55 crc kubenswrapper[4776]: I1125 10:30:55.722121 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dnrsz" Nov 25 10:30:55 crc kubenswrapper[4776]: I1125 10:30:55.888054 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcklt\" (UniqueName: \"kubernetes.io/projected/3a378c9a-dd8f-448b-b229-7e875605f913-kube-api-access-qcklt\") pod \"3a378c9a-dd8f-448b-b229-7e875605f913\" (UID: \"3a378c9a-dd8f-448b-b229-7e875605f913\") " Nov 25 10:30:55 crc kubenswrapper[4776]: I1125 10:30:55.888170 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a378c9a-dd8f-448b-b229-7e875605f913-utilities\") pod \"3a378c9a-dd8f-448b-b229-7e875605f913\" (UID: \"3a378c9a-dd8f-448b-b229-7e875605f913\") " Nov 25 10:30:55 crc kubenswrapper[4776]: I1125 10:30:55.888317 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a378c9a-dd8f-448b-b229-7e875605f913-catalog-content\") pod \"3a378c9a-dd8f-448b-b229-7e875605f913\" (UID: \"3a378c9a-dd8f-448b-b229-7e875605f913\") " Nov 25 10:30:55 crc kubenswrapper[4776]: I1125 10:30:55.889322 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a378c9a-dd8f-448b-b229-7e875605f913-utilities" (OuterVolumeSpecName: "utilities") pod "3a378c9a-dd8f-448b-b229-7e875605f913" (UID: "3a378c9a-dd8f-448b-b229-7e875605f913"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:30:55 crc kubenswrapper[4776]: I1125 10:30:55.893532 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a378c9a-dd8f-448b-b229-7e875605f913-kube-api-access-qcklt" (OuterVolumeSpecName: "kube-api-access-qcklt") pod "3a378c9a-dd8f-448b-b229-7e875605f913" (UID: "3a378c9a-dd8f-448b-b229-7e875605f913"). InnerVolumeSpecName "kube-api-access-qcklt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:30:55 crc kubenswrapper[4776]: I1125 10:30:55.943846 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a378c9a-dd8f-448b-b229-7e875605f913-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3a378c9a-dd8f-448b-b229-7e875605f913" (UID: "3a378c9a-dd8f-448b-b229-7e875605f913"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:30:55 crc kubenswrapper[4776]: I1125 10:30:55.989687 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a378c9a-dd8f-448b-b229-7e875605f913-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:30:55 crc kubenswrapper[4776]: I1125 10:30:55.989736 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcklt\" (UniqueName: \"kubernetes.io/projected/3a378c9a-dd8f-448b-b229-7e875605f913-kube-api-access-qcklt\") on node \"crc\" DevicePath \"\"" Nov 25 10:30:55 crc kubenswrapper[4776]: I1125 10:30:55.989749 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a378c9a-dd8f-448b-b229-7e875605f913-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:30:56 crc kubenswrapper[4776]: I1125 10:30:56.319729 4776 generic.go:334] "Generic (PLEG): container finished" podID="3a378c9a-dd8f-448b-b229-7e875605f913" containerID="35de714215a81f56597dd5c0cd0c86f2eaf12e6fe59e5b4634640309e11a9e99" exitCode=0 Nov 25 10:30:56 crc kubenswrapper[4776]: I1125 10:30:56.319799 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dnrsz" Nov 25 10:30:56 crc kubenswrapper[4776]: I1125 10:30:56.319795 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnrsz" event={"ID":"3a378c9a-dd8f-448b-b229-7e875605f913","Type":"ContainerDied","Data":"35de714215a81f56597dd5c0cd0c86f2eaf12e6fe59e5b4634640309e11a9e99"} Nov 25 10:30:56 crc kubenswrapper[4776]: I1125 10:30:56.319964 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dnrsz" event={"ID":"3a378c9a-dd8f-448b-b229-7e875605f913","Type":"ContainerDied","Data":"5726d909d1c95615e9411372e2bb2dbf9de513faba9230805e71e8dca1510d73"} Nov 25 10:30:56 crc kubenswrapper[4776]: I1125 10:30:56.319999 4776 scope.go:117] "RemoveContainer" containerID="35de714215a81f56597dd5c0cd0c86f2eaf12e6fe59e5b4634640309e11a9e99" Nov 25 10:30:56 crc kubenswrapper[4776]: I1125 10:30:56.354737 4776 scope.go:117] "RemoveContainer" containerID="e1a11952018f61942dcd64f9288c012740a89b5dc631087a823380bac24aa360" Nov 25 10:30:56 crc kubenswrapper[4776]: I1125 10:30:56.360123 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dnrsz"] Nov 25 10:30:56 crc kubenswrapper[4776]: I1125 10:30:56.377170 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dnrsz"] Nov 25 10:30:56 crc kubenswrapper[4776]: I1125 10:30:56.385786 4776 scope.go:117] "RemoveContainer" containerID="b02e0fbc1495fb6366f882f12c7793445136165f11dd3c3d023650746afbbeb2" Nov 25 10:30:56 crc kubenswrapper[4776]: I1125 10:30:56.407844 4776 scope.go:117] "RemoveContainer" containerID="35de714215a81f56597dd5c0cd0c86f2eaf12e6fe59e5b4634640309e11a9e99" Nov 25 10:30:56 crc kubenswrapper[4776]: E1125 10:30:56.408337 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35de714215a81f56597dd5c0cd0c86f2eaf12e6fe59e5b4634640309e11a9e99\": container with ID starting with 35de714215a81f56597dd5c0cd0c86f2eaf12e6fe59e5b4634640309e11a9e99 not found: ID does not exist" containerID="35de714215a81f56597dd5c0cd0c86f2eaf12e6fe59e5b4634640309e11a9e99" Nov 25 10:30:56 crc kubenswrapper[4776]: I1125 10:30:56.408399 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35de714215a81f56597dd5c0cd0c86f2eaf12e6fe59e5b4634640309e11a9e99"} err="failed to get container status \"35de714215a81f56597dd5c0cd0c86f2eaf12e6fe59e5b4634640309e11a9e99\": rpc error: code = NotFound desc = could not find container \"35de714215a81f56597dd5c0cd0c86f2eaf12e6fe59e5b4634640309e11a9e99\": container with ID starting with 35de714215a81f56597dd5c0cd0c86f2eaf12e6fe59e5b4634640309e11a9e99 not found: ID does not exist" Nov 25 10:30:56 crc kubenswrapper[4776]: I1125 10:30:56.408441 4776 scope.go:117] "RemoveContainer" containerID="e1a11952018f61942dcd64f9288c012740a89b5dc631087a823380bac24aa360" Nov 25 10:30:56 crc kubenswrapper[4776]: E1125 10:30:56.408787 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1a11952018f61942dcd64f9288c012740a89b5dc631087a823380bac24aa360\": container with ID starting with e1a11952018f61942dcd64f9288c012740a89b5dc631087a823380bac24aa360 not found: ID does not exist" containerID="e1a11952018f61942dcd64f9288c012740a89b5dc631087a823380bac24aa360" Nov 25 10:30:56 crc kubenswrapper[4776]: I1125 10:30:56.408870 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1a11952018f61942dcd64f9288c012740a89b5dc631087a823380bac24aa360"} err="failed to get container status \"e1a11952018f61942dcd64f9288c012740a89b5dc631087a823380bac24aa360\": rpc error: code = NotFound desc = could not find container \"e1a11952018f61942dcd64f9288c012740a89b5dc631087a823380bac24aa360\": container with ID starting with e1a11952018f61942dcd64f9288c012740a89b5dc631087a823380bac24aa360 not found: ID does not exist" Nov 25 10:30:56 crc kubenswrapper[4776]: I1125 10:30:56.408930 4776 scope.go:117] "RemoveContainer" containerID="b02e0fbc1495fb6366f882f12c7793445136165f11dd3c3d023650746afbbeb2" Nov 25 10:30:56 crc kubenswrapper[4776]: E1125 10:30:56.409403 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b02e0fbc1495fb6366f882f12c7793445136165f11dd3c3d023650746afbbeb2\": container with ID starting with b02e0fbc1495fb6366f882f12c7793445136165f11dd3c3d023650746afbbeb2 not found: ID does not exist" containerID="b02e0fbc1495fb6366f882f12c7793445136165f11dd3c3d023650746afbbeb2" Nov 25 10:30:56 crc kubenswrapper[4776]: I1125 10:30:56.409442 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b02e0fbc1495fb6366f882f12c7793445136165f11dd3c3d023650746afbbeb2"} err="failed to get container status \"b02e0fbc1495fb6366f882f12c7793445136165f11dd3c3d023650746afbbeb2\": rpc error: code = NotFound desc = could not find container \"b02e0fbc1495fb6366f882f12c7793445136165f11dd3c3d023650746afbbeb2\": container with ID starting with b02e0fbc1495fb6366f882f12c7793445136165f11dd3c3d023650746afbbeb2 not found: ID does not exist" Nov 25 10:30:57 crc kubenswrapper[4776]: I1125 10:30:57.672478 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a378c9a-dd8f-448b-b229-7e875605f913" path="/var/lib/kubelet/pods/3a378c9a-dd8f-448b-b229-7e875605f913/volumes" Nov 25 10:30:58 crc kubenswrapper[4776]: I1125 10:30:58.663587 4776 scope.go:117] "RemoveContainer" containerID="356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf" Nov 25 10:30:58 crc kubenswrapper[4776]: E1125 10:30:58.664371 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:31:09 crc kubenswrapper[4776]: I1125 10:31:09.662430 4776 scope.go:117] "RemoveContainer" containerID="356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf" Nov 25 10:31:09 crc kubenswrapper[4776]: E1125 10:31:09.664709 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:31:21 crc kubenswrapper[4776]: I1125 10:31:21.662994 4776 scope.go:117] "RemoveContainer" containerID="356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf" Nov 25 10:31:21 crc kubenswrapper[4776]: E1125 10:31:21.663944 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:31:32 crc kubenswrapper[4776]: I1125 10:31:32.661881 4776 scope.go:117] "RemoveContainer" containerID="356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf" Nov 25 10:31:32 crc kubenswrapper[4776]: E1125 10:31:32.662609 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:31:45 crc kubenswrapper[4776]: I1125 10:31:45.666933 4776 scope.go:117] "RemoveContainer" containerID="356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf" Nov 25 10:31:45 crc kubenswrapper[4776]: E1125 10:31:45.667890 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:31:59 crc kubenswrapper[4776]: I1125 10:31:59.662718 4776 scope.go:117] "RemoveContainer" containerID="356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf" Nov 25 10:31:59 crc kubenswrapper[4776]: E1125 10:31:59.663449 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:32:10 crc kubenswrapper[4776]: I1125 10:32:10.662351 4776 scope.go:117] "RemoveContainer" containerID="356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf" Nov 25 10:32:10 crc kubenswrapper[4776]: E1125 10:32:10.663900 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:32:23 crc kubenswrapper[4776]: I1125 10:32:23.662482 4776 scope.go:117] "RemoveContainer" containerID="356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf" Nov 25 10:32:23 crc kubenswrapper[4776]: E1125 10:32:23.663290 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:32:35 crc kubenswrapper[4776]: I1125 10:32:35.666167 4776 scope.go:117] "RemoveContainer" containerID="356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf" Nov 25 10:32:35 crc kubenswrapper[4776]: E1125 10:32:35.666711 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.289465 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8lkx9"] Nov 25 10:32:45 crc kubenswrapper[4776]: E1125 10:32:45.290577 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a378c9a-dd8f-448b-b229-7e875605f913" containerName="registry-server" Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.290598 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a378c9a-dd8f-448b-b229-7e875605f913" containerName="registry-server" Nov 25 10:32:45 crc kubenswrapper[4776]: E1125 10:32:45.290624 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a378c9a-dd8f-448b-b229-7e875605f913" containerName="extract-content" Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.290633 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a378c9a-dd8f-448b-b229-7e875605f913" containerName="extract-content" Nov 25 10:32:45 crc kubenswrapper[4776]: E1125 10:32:45.290661 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a378c9a-dd8f-448b-b229-7e875605f913" containerName="extract-utilities" Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.290669 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a378c9a-dd8f-448b-b229-7e875605f913" containerName="extract-utilities" Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.290923 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a378c9a-dd8f-448b-b229-7e875605f913" containerName="registry-server" Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.294822 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8lkx9" Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.312105 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8lkx9"] Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.322677 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06ff01b2-9e6d-459e-82a9-18d63a507559-catalog-content\") pod \"community-operators-8lkx9\" (UID: \"06ff01b2-9e6d-459e-82a9-18d63a507559\") " pod="openshift-marketplace/community-operators-8lkx9" Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.322818 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06ff01b2-9e6d-459e-82a9-18d63a507559-utilities\") pod \"community-operators-8lkx9\" (UID: \"06ff01b2-9e6d-459e-82a9-18d63a507559\") " pod="openshift-marketplace/community-operators-8lkx9" Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.322905 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl68m\" (UniqueName: \"kubernetes.io/projected/06ff01b2-9e6d-459e-82a9-18d63a507559-kube-api-access-kl68m\") pod \"community-operators-8lkx9\" (UID: \"06ff01b2-9e6d-459e-82a9-18d63a507559\") " pod="openshift-marketplace/community-operators-8lkx9" Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.423506 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl68m\" (UniqueName: \"kubernetes.io/projected/06ff01b2-9e6d-459e-82a9-18d63a507559-kube-api-access-kl68m\") pod \"community-operators-8lkx9\" (UID: \"06ff01b2-9e6d-459e-82a9-18d63a507559\") " pod="openshift-marketplace/community-operators-8lkx9" Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.423622 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06ff01b2-9e6d-459e-82a9-18d63a507559-catalog-content\") pod \"community-operators-8lkx9\" (UID: \"06ff01b2-9e6d-459e-82a9-18d63a507559\") " pod="openshift-marketplace/community-operators-8lkx9" Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.423671 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06ff01b2-9e6d-459e-82a9-18d63a507559-utilities\") pod \"community-operators-8lkx9\" (UID: \"06ff01b2-9e6d-459e-82a9-18d63a507559\") " pod="openshift-marketplace/community-operators-8lkx9" Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.424292 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06ff01b2-9e6d-459e-82a9-18d63a507559-catalog-content\") pod \"community-operators-8lkx9\" (UID: \"06ff01b2-9e6d-459e-82a9-18d63a507559\") " pod="openshift-marketplace/community-operators-8lkx9" Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.424361 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06ff01b2-9e6d-459e-82a9-18d63a507559-utilities\") pod \"community-operators-8lkx9\" (UID: \"06ff01b2-9e6d-459e-82a9-18d63a507559\") " pod="openshift-marketplace/community-operators-8lkx9" Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.448780 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl68m\" (UniqueName: \"kubernetes.io/projected/06ff01b2-9e6d-459e-82a9-18d63a507559-kube-api-access-kl68m\") pod \"community-operators-8lkx9\" (UID: \"06ff01b2-9e6d-459e-82a9-18d63a507559\") " pod="openshift-marketplace/community-operators-8lkx9" Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.489568 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tqb6t"] Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.491868 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tqb6t" Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.502204 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tqb6t"] Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.526585 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbcd062b-3c33-49ea-9f28-30d316e43273-utilities\") pod \"redhat-operators-tqb6t\" (UID: \"dbcd062b-3c33-49ea-9f28-30d316e43273\") " pod="openshift-marketplace/redhat-operators-tqb6t" Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.526680 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbcd062b-3c33-49ea-9f28-30d316e43273-catalog-content\") pod \"redhat-operators-tqb6t\" (UID: \"dbcd062b-3c33-49ea-9f28-30d316e43273\") " pod="openshift-marketplace/redhat-operators-tqb6t" Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.526706 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmvbc\" (UniqueName: \"kubernetes.io/projected/dbcd062b-3c33-49ea-9f28-30d316e43273-kube-api-access-lmvbc\") pod \"redhat-operators-tqb6t\" (UID: \"dbcd062b-3c33-49ea-9f28-30d316e43273\") " pod="openshift-marketplace/redhat-operators-tqb6t" Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.623419 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8lkx9" Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.627652 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbcd062b-3c33-49ea-9f28-30d316e43273-utilities\") pod \"redhat-operators-tqb6t\" (UID: \"dbcd062b-3c33-49ea-9f28-30d316e43273\") " pod="openshift-marketplace/redhat-operators-tqb6t" Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.627703 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbcd062b-3c33-49ea-9f28-30d316e43273-catalog-content\") pod \"redhat-operators-tqb6t\" (UID: \"dbcd062b-3c33-49ea-9f28-30d316e43273\") " pod="openshift-marketplace/redhat-operators-tqb6t" Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.627724 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmvbc\" (UniqueName: \"kubernetes.io/projected/dbcd062b-3c33-49ea-9f28-30d316e43273-kube-api-access-lmvbc\") pod \"redhat-operators-tqb6t\" (UID: \"dbcd062b-3c33-49ea-9f28-30d316e43273\") " pod="openshift-marketplace/redhat-operators-tqb6t" Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.628407 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbcd062b-3c33-49ea-9f28-30d316e43273-utilities\") pod \"redhat-operators-tqb6t\" (UID: \"dbcd062b-3c33-49ea-9f28-30d316e43273\") " pod="openshift-marketplace/redhat-operators-tqb6t" Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.628460 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbcd062b-3c33-49ea-9f28-30d316e43273-catalog-content\") pod \"redhat-operators-tqb6t\" (UID: \"dbcd062b-3c33-49ea-9f28-30d316e43273\") " pod="openshift-marketplace/redhat-operators-tqb6t" Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.656241 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmvbc\" (UniqueName: \"kubernetes.io/projected/dbcd062b-3c33-49ea-9f28-30d316e43273-kube-api-access-lmvbc\") pod \"redhat-operators-tqb6t\" (UID: \"dbcd062b-3c33-49ea-9f28-30d316e43273\") " pod="openshift-marketplace/redhat-operators-tqb6t" Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.825684 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tqb6t" Nov 25 10:32:45 crc kubenswrapper[4776]: I1125 10:32:45.925236 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8lkx9"] Nov 25 10:32:45 crc kubenswrapper[4776]: W1125 10:32:45.943236 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06ff01b2_9e6d_459e_82a9_18d63a507559.slice/crio-c2672a03e6c6d306b6eb1900c56aa43040e01529fe0bc7ac6036580ec1cfd627 WatchSource:0}: Error finding container c2672a03e6c6d306b6eb1900c56aa43040e01529fe0bc7ac6036580ec1cfd627: Status 404 returned error can't find the container with id c2672a03e6c6d306b6eb1900c56aa43040e01529fe0bc7ac6036580ec1cfd627 Nov 25 10:32:46 crc kubenswrapper[4776]: I1125 10:32:46.109612 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tqb6t"] Nov 25 10:32:46 crc kubenswrapper[4776]: W1125 10:32:46.114921 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddbcd062b_3c33_49ea_9f28_30d316e43273.slice/crio-1a5552c4b6ec0ee5470a0cfdd0382881f74f2e210a78442eeefaa39482ecd8d7 WatchSource:0}: Error finding container 1a5552c4b6ec0ee5470a0cfdd0382881f74f2e210a78442eeefaa39482ecd8d7: Status 404 returned error can't find the container with id 1a5552c4b6ec0ee5470a0cfdd0382881f74f2e210a78442eeefaa39482ecd8d7 Nov 25 10:32:46 crc kubenswrapper[4776]: I1125 10:32:46.497886 4776 generic.go:334] "Generic (PLEG): container finished" podID="06ff01b2-9e6d-459e-82a9-18d63a507559" containerID="06bdd1d0d1aaf05a491b633334c4d623288663f19b20ed5a30bb560e22d63661" exitCode=0 Nov 25 10:32:46 crc kubenswrapper[4776]: I1125 10:32:46.498007 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8lkx9" event={"ID":"06ff01b2-9e6d-459e-82a9-18d63a507559","Type":"ContainerDied","Data":"06bdd1d0d1aaf05a491b633334c4d623288663f19b20ed5a30bb560e22d63661"} Nov 25 10:32:46 crc kubenswrapper[4776]: I1125 10:32:46.498305 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8lkx9" event={"ID":"06ff01b2-9e6d-459e-82a9-18d63a507559","Type":"ContainerStarted","Data":"c2672a03e6c6d306b6eb1900c56aa43040e01529fe0bc7ac6036580ec1cfd627"} Nov 25 10:32:46 crc kubenswrapper[4776]: I1125 10:32:46.500742 4776 generic.go:334] "Generic (PLEG): container finished" podID="dbcd062b-3c33-49ea-9f28-30d316e43273" containerID="0315c666685e64377a951b2145043bb0de8fec43d33e20e4b2577e423a97f240" exitCode=0 Nov 25 10:32:46 crc kubenswrapper[4776]: I1125 10:32:46.500769 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqb6t" event={"ID":"dbcd062b-3c33-49ea-9f28-30d316e43273","Type":"ContainerDied","Data":"0315c666685e64377a951b2145043bb0de8fec43d33e20e4b2577e423a97f240"} Nov 25 10:32:46 crc kubenswrapper[4776]: I1125 10:32:46.500796 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqb6t" event={"ID":"dbcd062b-3c33-49ea-9f28-30d316e43273","Type":"ContainerStarted","Data":"1a5552c4b6ec0ee5470a0cfdd0382881f74f2e210a78442eeefaa39482ecd8d7"} Nov 25 10:32:47 crc kubenswrapper[4776]: I1125 10:32:47.509010 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8lkx9" event={"ID":"06ff01b2-9e6d-459e-82a9-18d63a507559","Type":"ContainerStarted","Data":"445832d11b64048c358a7eec24816feeac4fe8e2335374fc4c67d6e349783b8e"} Nov 25 10:32:48 crc kubenswrapper[4776]: I1125 10:32:48.492772 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-l2dzw"] Nov 25 10:32:48 crc kubenswrapper[4776]: I1125 10:32:48.494963 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l2dzw" Nov 25 10:32:48 crc kubenswrapper[4776]: I1125 10:32:48.516125 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l2dzw"] Nov 25 10:32:48 crc kubenswrapper[4776]: I1125 10:32:48.530874 4776 generic.go:334] "Generic (PLEG): container finished" podID="06ff01b2-9e6d-459e-82a9-18d63a507559" containerID="445832d11b64048c358a7eec24816feeac4fe8e2335374fc4c67d6e349783b8e" exitCode=0 Nov 25 10:32:48 crc kubenswrapper[4776]: I1125 10:32:48.530998 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8lkx9" event={"ID":"06ff01b2-9e6d-459e-82a9-18d63a507559","Type":"ContainerDied","Data":"445832d11b64048c358a7eec24816feeac4fe8e2335374fc4c67d6e349783b8e"} Nov 25 10:32:48 crc kubenswrapper[4776]: I1125 10:32:48.594971 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/881b453e-0a07-45e9-9ee2-178c9a0489e3-catalog-content\") pod \"redhat-marketplace-l2dzw\" (UID: \"881b453e-0a07-45e9-9ee2-178c9a0489e3\") " pod="openshift-marketplace/redhat-marketplace-l2dzw" Nov 25 10:32:48 crc kubenswrapper[4776]: I1125 10:32:48.595081 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/881b453e-0a07-45e9-9ee2-178c9a0489e3-utilities\") pod \"redhat-marketplace-l2dzw\" (UID: \"881b453e-0a07-45e9-9ee2-178c9a0489e3\") " pod="openshift-marketplace/redhat-marketplace-l2dzw" Nov 25 10:32:48 crc kubenswrapper[4776]: I1125 10:32:48.595134 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghmjb\" (UniqueName: \"kubernetes.io/projected/881b453e-0a07-45e9-9ee2-178c9a0489e3-kube-api-access-ghmjb\") pod \"redhat-marketplace-l2dzw\" (UID: \"881b453e-0a07-45e9-9ee2-178c9a0489e3\") " pod="openshift-marketplace/redhat-marketplace-l2dzw" Nov 25 10:32:48 crc kubenswrapper[4776]: I1125 10:32:48.696142 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/881b453e-0a07-45e9-9ee2-178c9a0489e3-catalog-content\") pod \"redhat-marketplace-l2dzw\" (UID: \"881b453e-0a07-45e9-9ee2-178c9a0489e3\") " pod="openshift-marketplace/redhat-marketplace-l2dzw" Nov 25 10:32:48 crc kubenswrapper[4776]: I1125 10:32:48.696218 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/881b453e-0a07-45e9-9ee2-178c9a0489e3-utilities\") pod \"redhat-marketplace-l2dzw\" (UID: \"881b453e-0a07-45e9-9ee2-178c9a0489e3\") " pod="openshift-marketplace/redhat-marketplace-l2dzw" Nov 25 10:32:48 crc kubenswrapper[4776]: I1125 10:32:48.696268 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghmjb\" (UniqueName: \"kubernetes.io/projected/881b453e-0a07-45e9-9ee2-178c9a0489e3-kube-api-access-ghmjb\") pod \"redhat-marketplace-l2dzw\" (UID: \"881b453e-0a07-45e9-9ee2-178c9a0489e3\") " pod="openshift-marketplace/redhat-marketplace-l2dzw" Nov 25 10:32:48 crc kubenswrapper[4776]: I1125 10:32:48.696756 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/881b453e-0a07-45e9-9ee2-178c9a0489e3-utilities\") pod \"redhat-marketplace-l2dzw\" (UID: \"881b453e-0a07-45e9-9ee2-178c9a0489e3\") " pod="openshift-marketplace/redhat-marketplace-l2dzw" Nov 25 10:32:48 crc kubenswrapper[4776]: I1125 10:32:48.696923 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/881b453e-0a07-45e9-9ee2-178c9a0489e3-catalog-content\") pod \"redhat-marketplace-l2dzw\" (UID: \"881b453e-0a07-45e9-9ee2-178c9a0489e3\") " pod="openshift-marketplace/redhat-marketplace-l2dzw" Nov 25 10:32:48 crc kubenswrapper[4776]: I1125 10:32:48.872841 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghmjb\" (UniqueName: \"kubernetes.io/projected/881b453e-0a07-45e9-9ee2-178c9a0489e3-kube-api-access-ghmjb\") pod \"redhat-marketplace-l2dzw\" (UID: \"881b453e-0a07-45e9-9ee2-178c9a0489e3\") " pod="openshift-marketplace/redhat-marketplace-l2dzw" Nov 25 10:32:49 crc kubenswrapper[4776]: I1125 10:32:49.120825 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l2dzw" Nov 25 10:32:49 crc kubenswrapper[4776]: I1125 10:32:49.543871 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8lkx9" event={"ID":"06ff01b2-9e6d-459e-82a9-18d63a507559","Type":"ContainerStarted","Data":"5e2c5f8d1125f2c38b289292aba7e37ea7d15be7a821820b6885c9eb590f0e4d"} Nov 25 10:32:49 crc kubenswrapper[4776]: I1125 10:32:49.567635 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l2dzw"] Nov 25 10:32:49 crc kubenswrapper[4776]: I1125 10:32:49.568871 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8lkx9" podStartSLOduration=2.114036701 podStartE2EDuration="4.568852183s" podCreationTimestamp="2025-11-25 10:32:45 +0000 UTC" firstStartedPulling="2025-11-25 10:32:46.499419317 +0000 UTC m=+4111.540478870" lastFinishedPulling="2025-11-25 10:32:48.954234809 +0000 UTC m=+4113.995294352" observedRunningTime="2025-11-25 10:32:49.562938794 +0000 UTC m=+4114.603998347" watchObservedRunningTime="2025-11-25 10:32:49.568852183 +0000 UTC m=+4114.609911746" Nov 25 10:32:49 crc kubenswrapper[4776]: I1125 10:32:49.663926 4776 scope.go:117] "RemoveContainer" containerID="356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf" Nov 25 10:32:49 crc kubenswrapper[4776]: E1125 10:32:49.664229 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:32:50 crc kubenswrapper[4776]: I1125 10:32:50.555286 4776 generic.go:334] "Generic (PLEG): container finished" podID="881b453e-0a07-45e9-9ee2-178c9a0489e3" containerID="e9a8ad8fe15a245e8ac36ceec79497b2262f26ca317919e2d4b2c9b09a893591" exitCode=0 Nov 25 10:32:50 crc kubenswrapper[4776]: I1125 10:32:50.555488 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l2dzw" event={"ID":"881b453e-0a07-45e9-9ee2-178c9a0489e3","Type":"ContainerDied","Data":"e9a8ad8fe15a245e8ac36ceec79497b2262f26ca317919e2d4b2c9b09a893591"} Nov 25 10:32:50 crc kubenswrapper[4776]: I1125 10:32:50.556406 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l2dzw" event={"ID":"881b453e-0a07-45e9-9ee2-178c9a0489e3","Type":"ContainerStarted","Data":"31f31013e1e634b3501424f05e6137e3b8d4e77aa301640ea408ffcbb148d47b"} Nov 25 10:32:54 crc kubenswrapper[4776]: I1125 10:32:54.586128 4776 generic.go:334] "Generic (PLEG): container finished" podID="881b453e-0a07-45e9-9ee2-178c9a0489e3" containerID="fb3761dfc0f044e1cdd6d6b6260552bb3fa4df09a36df687a4869c9ada57e32f" exitCode=0 Nov 25 10:32:54 crc kubenswrapper[4776]: I1125 10:32:54.586219 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l2dzw" event={"ID":"881b453e-0a07-45e9-9ee2-178c9a0489e3","Type":"ContainerDied","Data":"fb3761dfc0f044e1cdd6d6b6260552bb3fa4df09a36df687a4869c9ada57e32f"} Nov 25 10:32:54 crc kubenswrapper[4776]: I1125 10:32:54.589026 4776 generic.go:334] "Generic (PLEG): container finished" podID="dbcd062b-3c33-49ea-9f28-30d316e43273" containerID="40365c4b18b60e9ee19374b72e0fd9a8838770606898037ee7485253ad172e8c" exitCode=0 Nov 25 10:32:54 crc kubenswrapper[4776]: I1125 10:32:54.589094 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqb6t" event={"ID":"dbcd062b-3c33-49ea-9f28-30d316e43273","Type":"ContainerDied","Data":"40365c4b18b60e9ee19374b72e0fd9a8838770606898037ee7485253ad172e8c"} Nov 25 10:32:55 crc kubenswrapper[4776]: I1125 10:32:55.598683 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l2dzw" event={"ID":"881b453e-0a07-45e9-9ee2-178c9a0489e3","Type":"ContainerStarted","Data":"4e1d3f63e2625a9e8bc112c3216d2999bd9514e57058f86cc8dcc9d54568e2fd"} Nov 25 10:32:55 crc kubenswrapper[4776]: I1125 10:32:55.601515 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqb6t" event={"ID":"dbcd062b-3c33-49ea-9f28-30d316e43273","Type":"ContainerStarted","Data":"e77b469ae67457d49b6659b5386c8d363e8819ee62a34d206e15cfff06a5230e"} Nov 25 10:32:55 crc kubenswrapper[4776]: I1125 10:32:55.617585 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-l2dzw" podStartSLOduration=5.998078164 podStartE2EDuration="7.617567022s" podCreationTimestamp="2025-11-25 10:32:48 +0000 UTC" firstStartedPulling="2025-11-25 10:32:53.357968121 +0000 UTC m=+4118.399027684" lastFinishedPulling="2025-11-25 10:32:54.977456989 +0000 UTC m=+4120.018516542" observedRunningTime="2025-11-25 10:32:55.613174191 +0000 UTC m=+4120.654233744" watchObservedRunningTime="2025-11-25 10:32:55.617567022 +0000 UTC m=+4120.658626575" Nov 25 10:32:55 crc kubenswrapper[4776]: I1125 10:32:55.624452 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8lkx9" Nov 25 10:32:55 crc kubenswrapper[4776]: I1125 10:32:55.624557 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8lkx9" Nov 25 10:32:55 crc kubenswrapper[4776]: I1125 10:32:55.638585 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tqb6t" podStartSLOduration=2.092480988 podStartE2EDuration="10.638566239s" podCreationTimestamp="2025-11-25 10:32:45 +0000 UTC" firstStartedPulling="2025-11-25 10:32:46.502941775 +0000 UTC m=+4111.544001328" lastFinishedPulling="2025-11-25 10:32:55.049027026 +0000 UTC m=+4120.090086579" observedRunningTime="2025-11-25 10:32:55.633118952 +0000 UTC m=+4120.674178505" watchObservedRunningTime="2025-11-25 10:32:55.638566239 +0000 UTC m=+4120.679625792" Nov 25 10:32:55 crc kubenswrapper[4776]: I1125 10:32:55.688425 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8lkx9" Nov 25 10:32:55 crc kubenswrapper[4776]: I1125 10:32:55.826883 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tqb6t" Nov 25 10:32:55 crc kubenswrapper[4776]: I1125 10:32:55.826933 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tqb6t" Nov 25 10:32:56 crc kubenswrapper[4776]: I1125 10:32:56.667573 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8lkx9" Nov 25 10:32:56 crc kubenswrapper[4776]: I1125 10:32:56.866151 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tqb6t" podUID="dbcd062b-3c33-49ea-9f28-30d316e43273" containerName="registry-server" probeResult="failure" output=< Nov 25 10:32:56 crc kubenswrapper[4776]: timeout: failed to connect service ":50051" within 1s Nov 25 10:32:56 crc kubenswrapper[4776]: > Nov 25 10:32:58 crc kubenswrapper[4776]: I1125 10:32:58.274592 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8lkx9"] Nov 25 10:32:59 crc kubenswrapper[4776]: I1125 10:32:59.460545 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-l2dzw" Nov 25 10:32:59 crc kubenswrapper[4776]: I1125 10:32:59.462332 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-l2dzw" Nov 25 10:32:59 crc kubenswrapper[4776]: I1125 10:32:59.606991 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-l2dzw" Nov 25 10:32:59 crc kubenswrapper[4776]: I1125 10:32:59.628960 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8lkx9" podUID="06ff01b2-9e6d-459e-82a9-18d63a507559" containerName="registry-server" containerID="cri-o://5e2c5f8d1125f2c38b289292aba7e37ea7d15be7a821820b6885c9eb590f0e4d" gracePeriod=2 Nov 25 10:33:00 crc kubenswrapper[4776]: I1125 10:33:00.690564 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-l2dzw" Nov 25 10:33:01 crc kubenswrapper[4776]: I1125 10:33:01.874519 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l2dzw"] Nov 25 10:33:02 crc kubenswrapper[4776]: I1125 10:33:02.651290 4776 generic.go:334] "Generic (PLEG): container finished" podID="06ff01b2-9e6d-459e-82a9-18d63a507559" containerID="5e2c5f8d1125f2c38b289292aba7e37ea7d15be7a821820b6885c9eb590f0e4d" exitCode=0 Nov 25 10:33:02 crc kubenswrapper[4776]: I1125 10:33:02.651754 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-l2dzw" podUID="881b453e-0a07-45e9-9ee2-178c9a0489e3" containerName="registry-server" containerID="cri-o://4e1d3f63e2625a9e8bc112c3216d2999bd9514e57058f86cc8dcc9d54568e2fd" gracePeriod=2 Nov 25 10:33:02 crc kubenswrapper[4776]: I1125 10:33:02.651989 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8lkx9" event={"ID":"06ff01b2-9e6d-459e-82a9-18d63a507559","Type":"ContainerDied","Data":"5e2c5f8d1125f2c38b289292aba7e37ea7d15be7a821820b6885c9eb590f0e4d"} Nov 25 10:33:03 crc kubenswrapper[4776]: I1125 10:33:03.662928 4776 scope.go:117] "RemoveContainer" containerID="356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf" Nov 25 10:33:03 crc kubenswrapper[4776]: E1125 10:33:03.663187 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.432726 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l2dzw" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.441602 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8lkx9" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.541852 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/881b453e-0a07-45e9-9ee2-178c9a0489e3-utilities\") pod \"881b453e-0a07-45e9-9ee2-178c9a0489e3\" (UID: \"881b453e-0a07-45e9-9ee2-178c9a0489e3\") " Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.541919 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghmjb\" (UniqueName: \"kubernetes.io/projected/881b453e-0a07-45e9-9ee2-178c9a0489e3-kube-api-access-ghmjb\") pod \"881b453e-0a07-45e9-9ee2-178c9a0489e3\" (UID: \"881b453e-0a07-45e9-9ee2-178c9a0489e3\") " Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.541948 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/881b453e-0a07-45e9-9ee2-178c9a0489e3-catalog-content\") pod \"881b453e-0a07-45e9-9ee2-178c9a0489e3\" (UID: \"881b453e-0a07-45e9-9ee2-178c9a0489e3\") " Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.543717 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/881b453e-0a07-45e9-9ee2-178c9a0489e3-utilities" (OuterVolumeSpecName: "utilities") pod "881b453e-0a07-45e9-9ee2-178c9a0489e3" (UID: "881b453e-0a07-45e9-9ee2-178c9a0489e3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.553348 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/881b453e-0a07-45e9-9ee2-178c9a0489e3-kube-api-access-ghmjb" (OuterVolumeSpecName: "kube-api-access-ghmjb") pod "881b453e-0a07-45e9-9ee2-178c9a0489e3" (UID: "881b453e-0a07-45e9-9ee2-178c9a0489e3"). InnerVolumeSpecName "kube-api-access-ghmjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.561305 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/881b453e-0a07-45e9-9ee2-178c9a0489e3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "881b453e-0a07-45e9-9ee2-178c9a0489e3" (UID: "881b453e-0a07-45e9-9ee2-178c9a0489e3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.643380 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kl68m\" (UniqueName: \"kubernetes.io/projected/06ff01b2-9e6d-459e-82a9-18d63a507559-kube-api-access-kl68m\") pod \"06ff01b2-9e6d-459e-82a9-18d63a507559\" (UID: \"06ff01b2-9e6d-459e-82a9-18d63a507559\") " Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.643524 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06ff01b2-9e6d-459e-82a9-18d63a507559-catalog-content\") pod \"06ff01b2-9e6d-459e-82a9-18d63a507559\" (UID: \"06ff01b2-9e6d-459e-82a9-18d63a507559\") " Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.643555 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06ff01b2-9e6d-459e-82a9-18d63a507559-utilities\") pod \"06ff01b2-9e6d-459e-82a9-18d63a507559\" (UID: \"06ff01b2-9e6d-459e-82a9-18d63a507559\") " Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.643955 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/881b453e-0a07-45e9-9ee2-178c9a0489e3-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.644140 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghmjb\" (UniqueName: \"kubernetes.io/projected/881b453e-0a07-45e9-9ee2-178c9a0489e3-kube-api-access-ghmjb\") on node \"crc\" DevicePath \"\"" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.644156 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/881b453e-0a07-45e9-9ee2-178c9a0489e3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.644497 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06ff01b2-9e6d-459e-82a9-18d63a507559-utilities" (OuterVolumeSpecName: "utilities") pod "06ff01b2-9e6d-459e-82a9-18d63a507559" (UID: "06ff01b2-9e6d-459e-82a9-18d63a507559"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.651173 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06ff01b2-9e6d-459e-82a9-18d63a507559-kube-api-access-kl68m" (OuterVolumeSpecName: "kube-api-access-kl68m") pod "06ff01b2-9e6d-459e-82a9-18d63a507559" (UID: "06ff01b2-9e6d-459e-82a9-18d63a507559"). InnerVolumeSpecName "kube-api-access-kl68m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.676617 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8lkx9" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.677460 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8lkx9" event={"ID":"06ff01b2-9e6d-459e-82a9-18d63a507559","Type":"ContainerDied","Data":"c2672a03e6c6d306b6eb1900c56aa43040e01529fe0bc7ac6036580ec1cfd627"} Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.677504 4776 scope.go:117] "RemoveContainer" containerID="5e2c5f8d1125f2c38b289292aba7e37ea7d15be7a821820b6885c9eb590f0e4d" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.681528 4776 generic.go:334] "Generic (PLEG): container finished" podID="881b453e-0a07-45e9-9ee2-178c9a0489e3" containerID="4e1d3f63e2625a9e8bc112c3216d2999bd9514e57058f86cc8dcc9d54568e2fd" exitCode=0 Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.681572 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l2dzw" event={"ID":"881b453e-0a07-45e9-9ee2-178c9a0489e3","Type":"ContainerDied","Data":"4e1d3f63e2625a9e8bc112c3216d2999bd9514e57058f86cc8dcc9d54568e2fd"} Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.681599 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l2dzw" event={"ID":"881b453e-0a07-45e9-9ee2-178c9a0489e3","Type":"ContainerDied","Data":"31f31013e1e634b3501424f05e6137e3b8d4e77aa301640ea408ffcbb148d47b"} Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.681617 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l2dzw" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.698231 4776 scope.go:117] "RemoveContainer" containerID="445832d11b64048c358a7eec24816feeac4fe8e2335374fc4c67d6e349783b8e" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.701416 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06ff01b2-9e6d-459e-82a9-18d63a507559-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "06ff01b2-9e6d-459e-82a9-18d63a507559" (UID: "06ff01b2-9e6d-459e-82a9-18d63a507559"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.712569 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l2dzw"] Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.723735 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-l2dzw"] Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.745157 4776 scope.go:117] "RemoveContainer" containerID="06bdd1d0d1aaf05a491b633334c4d623288663f19b20ed5a30bb560e22d63661" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.745439 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kl68m\" (UniqueName: \"kubernetes.io/projected/06ff01b2-9e6d-459e-82a9-18d63a507559-kube-api-access-kl68m\") on node \"crc\" DevicePath \"\"" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.745476 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06ff01b2-9e6d-459e-82a9-18d63a507559-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.745490 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06ff01b2-9e6d-459e-82a9-18d63a507559-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.759232 4776 scope.go:117] "RemoveContainer" containerID="4e1d3f63e2625a9e8bc112c3216d2999bd9514e57058f86cc8dcc9d54568e2fd" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.776306 4776 scope.go:117] "RemoveContainer" containerID="fb3761dfc0f044e1cdd6d6b6260552bb3fa4df09a36df687a4869c9ada57e32f" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.791462 4776 scope.go:117] "RemoveContainer" containerID="e9a8ad8fe15a245e8ac36ceec79497b2262f26ca317919e2d4b2c9b09a893591" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.806846 4776 scope.go:117] "RemoveContainer" containerID="4e1d3f63e2625a9e8bc112c3216d2999bd9514e57058f86cc8dcc9d54568e2fd" Nov 25 10:33:05 crc kubenswrapper[4776]: E1125 10:33:05.807331 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e1d3f63e2625a9e8bc112c3216d2999bd9514e57058f86cc8dcc9d54568e2fd\": container with ID starting with 4e1d3f63e2625a9e8bc112c3216d2999bd9514e57058f86cc8dcc9d54568e2fd not found: ID does not exist" containerID="4e1d3f63e2625a9e8bc112c3216d2999bd9514e57058f86cc8dcc9d54568e2fd" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.807361 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e1d3f63e2625a9e8bc112c3216d2999bd9514e57058f86cc8dcc9d54568e2fd"} err="failed to get container status \"4e1d3f63e2625a9e8bc112c3216d2999bd9514e57058f86cc8dcc9d54568e2fd\": rpc error: code = NotFound desc = could not find container \"4e1d3f63e2625a9e8bc112c3216d2999bd9514e57058f86cc8dcc9d54568e2fd\": container with ID starting with 4e1d3f63e2625a9e8bc112c3216d2999bd9514e57058f86cc8dcc9d54568e2fd not found: ID does not exist" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.807383 4776 scope.go:117] "RemoveContainer" containerID="fb3761dfc0f044e1cdd6d6b6260552bb3fa4df09a36df687a4869c9ada57e32f" Nov 25 10:33:05 crc kubenswrapper[4776]: E1125 10:33:05.807879 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb3761dfc0f044e1cdd6d6b6260552bb3fa4df09a36df687a4869c9ada57e32f\": container with ID starting with fb3761dfc0f044e1cdd6d6b6260552bb3fa4df09a36df687a4869c9ada57e32f not found: ID does not exist" containerID="fb3761dfc0f044e1cdd6d6b6260552bb3fa4df09a36df687a4869c9ada57e32f" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.807920 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb3761dfc0f044e1cdd6d6b6260552bb3fa4df09a36df687a4869c9ada57e32f"} err="failed to get container status \"fb3761dfc0f044e1cdd6d6b6260552bb3fa4df09a36df687a4869c9ada57e32f\": rpc error: code = NotFound desc = could not find container \"fb3761dfc0f044e1cdd6d6b6260552bb3fa4df09a36df687a4869c9ada57e32f\": container with ID starting with fb3761dfc0f044e1cdd6d6b6260552bb3fa4df09a36df687a4869c9ada57e32f not found: ID does not exist" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.807947 4776 scope.go:117] "RemoveContainer" containerID="e9a8ad8fe15a245e8ac36ceec79497b2262f26ca317919e2d4b2c9b09a893591" Nov 25 10:33:05 crc kubenswrapper[4776]: E1125 10:33:05.808237 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9a8ad8fe15a245e8ac36ceec79497b2262f26ca317919e2d4b2c9b09a893591\": container with ID starting with e9a8ad8fe15a245e8ac36ceec79497b2262f26ca317919e2d4b2c9b09a893591 not found: ID does not exist" containerID="e9a8ad8fe15a245e8ac36ceec79497b2262f26ca317919e2d4b2c9b09a893591" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.808259 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9a8ad8fe15a245e8ac36ceec79497b2262f26ca317919e2d4b2c9b09a893591"} err="failed to get container status \"e9a8ad8fe15a245e8ac36ceec79497b2262f26ca317919e2d4b2c9b09a893591\": rpc error: code = NotFound desc = could not find container \"e9a8ad8fe15a245e8ac36ceec79497b2262f26ca317919e2d4b2c9b09a893591\": container with ID starting with e9a8ad8fe15a245e8ac36ceec79497b2262f26ca317919e2d4b2c9b09a893591 not found: ID does not exist" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.891901 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tqb6t" Nov 25 10:33:05 crc kubenswrapper[4776]: I1125 10:33:05.943453 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tqb6t" Nov 25 10:33:06 crc kubenswrapper[4776]: I1125 10:33:06.002993 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8lkx9"] Nov 25 10:33:06 crc kubenswrapper[4776]: I1125 10:33:06.013430 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8lkx9"] Nov 25 10:33:07 crc kubenswrapper[4776]: I1125 10:33:07.671572 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06ff01b2-9e6d-459e-82a9-18d63a507559" path="/var/lib/kubelet/pods/06ff01b2-9e6d-459e-82a9-18d63a507559/volumes" Nov 25 10:33:07 crc kubenswrapper[4776]: I1125 10:33:07.672400 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="881b453e-0a07-45e9-9ee2-178c9a0489e3" path="/var/lib/kubelet/pods/881b453e-0a07-45e9-9ee2-178c9a0489e3/volumes" Nov 25 10:33:07 crc kubenswrapper[4776]: I1125 10:33:07.891680 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tqb6t"] Nov 25 10:33:08 crc kubenswrapper[4776]: I1125 10:33:08.266062 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5pmz9"] Nov 25 10:33:08 crc kubenswrapper[4776]: I1125 10:33:08.266434 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5pmz9" podUID="6e5ec892-fe87-4b96-8cf1-cd81510223c2" containerName="registry-server" containerID="cri-o://bafa88355c4379e175b5e32901bb1406224a3638a16138f21dfda8d0184d235f" gracePeriod=2 Nov 25 10:33:08 crc kubenswrapper[4776]: I1125 10:33:08.715333 4776 generic.go:334] "Generic (PLEG): container finished" podID="6e5ec892-fe87-4b96-8cf1-cd81510223c2" containerID="bafa88355c4379e175b5e32901bb1406224a3638a16138f21dfda8d0184d235f" exitCode=0 Nov 25 10:33:08 crc kubenswrapper[4776]: I1125 10:33:08.715378 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5pmz9" event={"ID":"6e5ec892-fe87-4b96-8cf1-cd81510223c2","Type":"ContainerDied","Data":"bafa88355c4379e175b5e32901bb1406224a3638a16138f21dfda8d0184d235f"} Nov 25 10:33:08 crc kubenswrapper[4776]: I1125 10:33:08.715405 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5pmz9" event={"ID":"6e5ec892-fe87-4b96-8cf1-cd81510223c2","Type":"ContainerDied","Data":"5fd6015dfcf80b26481da08305ce53e411ae6669f8a5a9fb6ed8498ec0eeb7c9"} Nov 25 10:33:08 crc kubenswrapper[4776]: I1125 10:33:08.715416 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5fd6015dfcf80b26481da08305ce53e411ae6669f8a5a9fb6ed8498ec0eeb7c9" Nov 25 10:33:08 crc kubenswrapper[4776]: I1125 10:33:08.735577 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5pmz9" Nov 25 10:33:08 crc kubenswrapper[4776]: I1125 10:33:08.886564 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e5ec892-fe87-4b96-8cf1-cd81510223c2-catalog-content\") pod \"6e5ec892-fe87-4b96-8cf1-cd81510223c2\" (UID: \"6e5ec892-fe87-4b96-8cf1-cd81510223c2\") " Nov 25 10:33:08 crc kubenswrapper[4776]: I1125 10:33:08.887020 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cfxn\" (UniqueName: \"kubernetes.io/projected/6e5ec892-fe87-4b96-8cf1-cd81510223c2-kube-api-access-7cfxn\") pod \"6e5ec892-fe87-4b96-8cf1-cd81510223c2\" (UID: \"6e5ec892-fe87-4b96-8cf1-cd81510223c2\") " Nov 25 10:33:08 crc kubenswrapper[4776]: I1125 10:33:08.887091 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e5ec892-fe87-4b96-8cf1-cd81510223c2-utilities\") pod \"6e5ec892-fe87-4b96-8cf1-cd81510223c2\" (UID: \"6e5ec892-fe87-4b96-8cf1-cd81510223c2\") " Nov 25 10:33:08 crc kubenswrapper[4776]: I1125 10:33:08.887647 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e5ec892-fe87-4b96-8cf1-cd81510223c2-utilities" (OuterVolumeSpecName: "utilities") pod "6e5ec892-fe87-4b96-8cf1-cd81510223c2" (UID: "6e5ec892-fe87-4b96-8cf1-cd81510223c2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:33:08 crc kubenswrapper[4776]: I1125 10:33:08.892486 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e5ec892-fe87-4b96-8cf1-cd81510223c2-kube-api-access-7cfxn" (OuterVolumeSpecName: "kube-api-access-7cfxn") pod "6e5ec892-fe87-4b96-8cf1-cd81510223c2" (UID: "6e5ec892-fe87-4b96-8cf1-cd81510223c2"). InnerVolumeSpecName "kube-api-access-7cfxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:33:08 crc kubenswrapper[4776]: I1125 10:33:08.968282 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e5ec892-fe87-4b96-8cf1-cd81510223c2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6e5ec892-fe87-4b96-8cf1-cd81510223c2" (UID: "6e5ec892-fe87-4b96-8cf1-cd81510223c2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:33:08 crc kubenswrapper[4776]: I1125 10:33:08.988798 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e5ec892-fe87-4b96-8cf1-cd81510223c2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:33:08 crc kubenswrapper[4776]: I1125 10:33:08.988833 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cfxn\" (UniqueName: \"kubernetes.io/projected/6e5ec892-fe87-4b96-8cf1-cd81510223c2-kube-api-access-7cfxn\") on node \"crc\" DevicePath \"\"" Nov 25 10:33:08 crc kubenswrapper[4776]: I1125 10:33:08.988848 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e5ec892-fe87-4b96-8cf1-cd81510223c2-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:33:09 crc kubenswrapper[4776]: I1125 10:33:09.720168 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5pmz9" Nov 25 10:33:09 crc kubenswrapper[4776]: I1125 10:33:09.740695 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5pmz9"] Nov 25 10:33:09 crc kubenswrapper[4776]: I1125 10:33:09.752011 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5pmz9"] Nov 25 10:33:09 crc kubenswrapper[4776]: E1125 10:33:09.825425 4776 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e5ec892_fe87_4b96_8cf1_cd81510223c2.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e5ec892_fe87_4b96_8cf1_cd81510223c2.slice/crio-5fd6015dfcf80b26481da08305ce53e411ae6669f8a5a9fb6ed8498ec0eeb7c9\": RecentStats: unable to find data in memory cache]" Nov 25 10:33:11 crc kubenswrapper[4776]: I1125 10:33:11.671728 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e5ec892-fe87-4b96-8cf1-cd81510223c2" path="/var/lib/kubelet/pods/6e5ec892-fe87-4b96-8cf1-cd81510223c2/volumes" Nov 25 10:33:15 crc kubenswrapper[4776]: I1125 10:33:15.666255 4776 scope.go:117] "RemoveContainer" containerID="356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf" Nov 25 10:33:15 crc kubenswrapper[4776]: E1125 10:33:15.668565 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:33:27 crc kubenswrapper[4776]: I1125 10:33:27.662548 4776 scope.go:117] "RemoveContainer" containerID="356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf" Nov 25 10:33:27 crc kubenswrapper[4776]: I1125 10:33:27.873603 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"8434b5e583fe39c8904344b2019db9ec4e80e39c2b18f85675be819ed4c15c9b"} Nov 25 10:33:30 crc kubenswrapper[4776]: I1125 10:33:30.937236 4776 scope.go:117] "RemoveContainer" containerID="c0b73e794a052f0a8f8878e647049945c4acef32db1870f61712db52b3d31824" Nov 25 10:33:30 crc kubenswrapper[4776]: I1125 10:33:30.959679 4776 scope.go:117] "RemoveContainer" containerID="bafa88355c4379e175b5e32901bb1406224a3638a16138f21dfda8d0184d235f" Nov 25 10:33:31 crc kubenswrapper[4776]: I1125 10:33:31.002385 4776 scope.go:117] "RemoveContainer" containerID="f8e5993f36165eff18d03e7547b4e2f10e329241eebfa3ac4c016b22aa9e558d" Nov 25 10:35:47 crc kubenswrapper[4776]: I1125 10:35:47.819236 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:35:47 crc kubenswrapper[4776]: I1125 10:35:47.821663 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:36:17 crc kubenswrapper[4776]: I1125 10:36:17.819020 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:36:17 crc kubenswrapper[4776]: I1125 10:36:17.821739 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:36:47 crc kubenswrapper[4776]: I1125 10:36:47.818003 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:36:47 crc kubenswrapper[4776]: I1125 10:36:47.819184 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:36:47 crc kubenswrapper[4776]: I1125 10:36:47.819241 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 10:36:47 crc kubenswrapper[4776]: I1125 10:36:47.820130 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8434b5e583fe39c8904344b2019db9ec4e80e39c2b18f85675be819ed4c15c9b"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:36:47 crc kubenswrapper[4776]: I1125 10:36:47.820213 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://8434b5e583fe39c8904344b2019db9ec4e80e39c2b18f85675be819ed4c15c9b" gracePeriod=600 Nov 25 10:36:48 crc kubenswrapper[4776]: I1125 10:36:48.637461 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="8434b5e583fe39c8904344b2019db9ec4e80e39c2b18f85675be819ed4c15c9b" exitCode=0 Nov 25 10:36:48 crc kubenswrapper[4776]: I1125 10:36:48.637511 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"8434b5e583fe39c8904344b2019db9ec4e80e39c2b18f85675be819ed4c15c9b"} Nov 25 10:36:48 crc kubenswrapper[4776]: I1125 10:36:48.637975 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a"} Nov 25 10:36:48 crc kubenswrapper[4776]: I1125 10:36:48.637999 4776 scope.go:117] "RemoveContainer" containerID="356ad6fcadf2be3615b760e616b30cd67ba266270bc561d3cdbb706ae1ce02cf" Nov 25 10:39:17 crc kubenswrapper[4776]: I1125 10:39:17.818323 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:39:17 crc kubenswrapper[4776]: I1125 10:39:17.818927 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:39:47 crc kubenswrapper[4776]: I1125 10:39:47.818822 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:39:47 crc kubenswrapper[4776]: I1125 10:39:47.819592 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:40:17 crc kubenswrapper[4776]: I1125 10:40:17.818025 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:40:17 crc kubenswrapper[4776]: I1125 10:40:17.818635 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:40:17 crc kubenswrapper[4776]: I1125 10:40:17.818688 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 10:40:17 crc kubenswrapper[4776]: I1125 10:40:17.819758 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:40:17 crc kubenswrapper[4776]: I1125 10:40:17.819829 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a" gracePeriod=600 Nov 25 10:40:17 crc kubenswrapper[4776]: E1125 10:40:17.892805 4776 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa47ebcc_a95e_4693_876d_7284c28c3ade.slice/crio-f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a.scope\": RecentStats: unable to find data in memory cache]" Nov 25 10:40:17 crc kubenswrapper[4776]: E1125 10:40:17.963347 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:40:18 crc kubenswrapper[4776]: I1125 10:40:18.222136 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a" exitCode=0 Nov 25 10:40:18 crc kubenswrapper[4776]: I1125 10:40:18.222204 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a"} Nov 25 10:40:18 crc kubenswrapper[4776]: I1125 10:40:18.222530 4776 scope.go:117] "RemoveContainer" containerID="8434b5e583fe39c8904344b2019db9ec4e80e39c2b18f85675be819ed4c15c9b" Nov 25 10:40:18 crc kubenswrapper[4776]: I1125 10:40:18.222986 4776 scope.go:117] "RemoveContainer" containerID="f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a" Nov 25 10:40:18 crc kubenswrapper[4776]: E1125 10:40:18.223240 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:40:32 crc kubenswrapper[4776]: I1125 10:40:32.662014 4776 scope.go:117] "RemoveContainer" containerID="f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a" Nov 25 10:40:32 crc kubenswrapper[4776]: E1125 10:40:32.663177 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:40:45 crc kubenswrapper[4776]: I1125 10:40:45.667400 4776 scope.go:117] "RemoveContainer" containerID="f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a" Nov 25 10:40:45 crc kubenswrapper[4776]: E1125 10:40:45.668015 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:40:59 crc kubenswrapper[4776]: I1125 10:40:59.662250 4776 scope.go:117] "RemoveContainer" containerID="f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a" Nov 25 10:40:59 crc kubenswrapper[4776]: E1125 10:40:59.663220 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:41:10 crc kubenswrapper[4776]: I1125 10:41:10.582333 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rmpxx"] Nov 25 10:41:10 crc kubenswrapper[4776]: E1125 10:41:10.583183 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="881b453e-0a07-45e9-9ee2-178c9a0489e3" containerName="registry-server" Nov 25 10:41:10 crc kubenswrapper[4776]: I1125 10:41:10.583198 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="881b453e-0a07-45e9-9ee2-178c9a0489e3" containerName="registry-server" Nov 25 10:41:10 crc kubenswrapper[4776]: E1125 10:41:10.583219 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06ff01b2-9e6d-459e-82a9-18d63a507559" containerName="registry-server" Nov 25 10:41:10 crc kubenswrapper[4776]: I1125 10:41:10.583226 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="06ff01b2-9e6d-459e-82a9-18d63a507559" containerName="registry-server" Nov 25 10:41:10 crc kubenswrapper[4776]: E1125 10:41:10.583248 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="881b453e-0a07-45e9-9ee2-178c9a0489e3" containerName="extract-utilities" Nov 25 10:41:10 crc kubenswrapper[4776]: I1125 10:41:10.583256 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="881b453e-0a07-45e9-9ee2-178c9a0489e3" containerName="extract-utilities" Nov 25 10:41:10 crc kubenswrapper[4776]: E1125 10:41:10.583270 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e5ec892-fe87-4b96-8cf1-cd81510223c2" containerName="extract-utilities" Nov 25 10:41:10 crc kubenswrapper[4776]: I1125 10:41:10.583277 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e5ec892-fe87-4b96-8cf1-cd81510223c2" containerName="extract-utilities" Nov 25 10:41:10 crc kubenswrapper[4776]: E1125 10:41:10.583286 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e5ec892-fe87-4b96-8cf1-cd81510223c2" containerName="registry-server" Nov 25 10:41:10 crc kubenswrapper[4776]: I1125 10:41:10.583293 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e5ec892-fe87-4b96-8cf1-cd81510223c2" containerName="registry-server" Nov 25 10:41:10 crc kubenswrapper[4776]: E1125 10:41:10.583307 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06ff01b2-9e6d-459e-82a9-18d63a507559" containerName="extract-utilities" Nov 25 10:41:10 crc kubenswrapper[4776]: I1125 10:41:10.583314 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="06ff01b2-9e6d-459e-82a9-18d63a507559" containerName="extract-utilities" Nov 25 10:41:10 crc kubenswrapper[4776]: E1125 10:41:10.583329 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e5ec892-fe87-4b96-8cf1-cd81510223c2" containerName="extract-content" Nov 25 10:41:10 crc kubenswrapper[4776]: I1125 10:41:10.583336 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e5ec892-fe87-4b96-8cf1-cd81510223c2" containerName="extract-content" Nov 25 10:41:10 crc kubenswrapper[4776]: E1125 10:41:10.583348 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="881b453e-0a07-45e9-9ee2-178c9a0489e3" containerName="extract-content" Nov 25 10:41:10 crc kubenswrapper[4776]: I1125 10:41:10.583356 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="881b453e-0a07-45e9-9ee2-178c9a0489e3" containerName="extract-content" Nov 25 10:41:10 crc kubenswrapper[4776]: E1125 10:41:10.583374 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06ff01b2-9e6d-459e-82a9-18d63a507559" containerName="extract-content" Nov 25 10:41:10 crc kubenswrapper[4776]: I1125 10:41:10.583381 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="06ff01b2-9e6d-459e-82a9-18d63a507559" containerName="extract-content" Nov 25 10:41:10 crc kubenswrapper[4776]: I1125 10:41:10.583527 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="06ff01b2-9e6d-459e-82a9-18d63a507559" containerName="registry-server" Nov 25 10:41:10 crc kubenswrapper[4776]: I1125 10:41:10.583550 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="881b453e-0a07-45e9-9ee2-178c9a0489e3" containerName="registry-server" Nov 25 10:41:10 crc kubenswrapper[4776]: I1125 10:41:10.583566 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e5ec892-fe87-4b96-8cf1-cd81510223c2" containerName="registry-server" Nov 25 10:41:10 crc kubenswrapper[4776]: I1125 10:41:10.584758 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rmpxx" Nov 25 10:41:10 crc kubenswrapper[4776]: I1125 10:41:10.593142 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rmpxx"] Nov 25 10:41:10 crc kubenswrapper[4776]: I1125 10:41:10.662559 4776 scope.go:117] "RemoveContainer" containerID="f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a" Nov 25 10:41:10 crc kubenswrapper[4776]: E1125 10:41:10.662786 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:41:10 crc kubenswrapper[4776]: I1125 10:41:10.709630 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc31a27-feda-42e2-ad49-e136fbe4730e-utilities\") pod \"certified-operators-rmpxx\" (UID: \"afc31a27-feda-42e2-ad49-e136fbe4730e\") " pod="openshift-marketplace/certified-operators-rmpxx" Nov 25 10:41:10 crc kubenswrapper[4776]: I1125 10:41:10.709683 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc31a27-feda-42e2-ad49-e136fbe4730e-catalog-content\") pod \"certified-operators-rmpxx\" (UID: \"afc31a27-feda-42e2-ad49-e136fbe4730e\") " pod="openshift-marketplace/certified-operators-rmpxx" Nov 25 10:41:10 crc kubenswrapper[4776]: I1125 10:41:10.710081 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6sj2\" (UniqueName: \"kubernetes.io/projected/afc31a27-feda-42e2-ad49-e136fbe4730e-kube-api-access-x6sj2\") pod \"certified-operators-rmpxx\" (UID: \"afc31a27-feda-42e2-ad49-e136fbe4730e\") " pod="openshift-marketplace/certified-operators-rmpxx" Nov 25 10:41:10 crc kubenswrapper[4776]: I1125 10:41:10.811190 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6sj2\" (UniqueName: \"kubernetes.io/projected/afc31a27-feda-42e2-ad49-e136fbe4730e-kube-api-access-x6sj2\") pod \"certified-operators-rmpxx\" (UID: \"afc31a27-feda-42e2-ad49-e136fbe4730e\") " pod="openshift-marketplace/certified-operators-rmpxx" Nov 25 10:41:10 crc kubenswrapper[4776]: I1125 10:41:10.811295 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc31a27-feda-42e2-ad49-e136fbe4730e-utilities\") pod \"certified-operators-rmpxx\" (UID: \"afc31a27-feda-42e2-ad49-e136fbe4730e\") " pod="openshift-marketplace/certified-operators-rmpxx" Nov 25 10:41:10 crc kubenswrapper[4776]: I1125 10:41:10.811334 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc31a27-feda-42e2-ad49-e136fbe4730e-catalog-content\") pod \"certified-operators-rmpxx\" (UID: \"afc31a27-feda-42e2-ad49-e136fbe4730e\") " pod="openshift-marketplace/certified-operators-rmpxx" Nov 25 10:41:10 crc kubenswrapper[4776]: I1125 10:41:10.811842 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc31a27-feda-42e2-ad49-e136fbe4730e-catalog-content\") pod \"certified-operators-rmpxx\" (UID: \"afc31a27-feda-42e2-ad49-e136fbe4730e\") " pod="openshift-marketplace/certified-operators-rmpxx" Nov 25 10:41:10 crc kubenswrapper[4776]: I1125 10:41:10.812415 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc31a27-feda-42e2-ad49-e136fbe4730e-utilities\") pod \"certified-operators-rmpxx\" (UID: \"afc31a27-feda-42e2-ad49-e136fbe4730e\") " pod="openshift-marketplace/certified-operators-rmpxx" Nov 25 10:41:10 crc kubenswrapper[4776]: I1125 10:41:10.834376 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6sj2\" (UniqueName: \"kubernetes.io/projected/afc31a27-feda-42e2-ad49-e136fbe4730e-kube-api-access-x6sj2\") pod \"certified-operators-rmpxx\" (UID: \"afc31a27-feda-42e2-ad49-e136fbe4730e\") " pod="openshift-marketplace/certified-operators-rmpxx" Nov 25 10:41:10 crc kubenswrapper[4776]: I1125 10:41:10.907604 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rmpxx" Nov 25 10:41:11 crc kubenswrapper[4776]: I1125 10:41:11.144719 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rmpxx"] Nov 25 10:41:11 crc kubenswrapper[4776]: I1125 10:41:11.736274 4776 generic.go:334] "Generic (PLEG): container finished" podID="afc31a27-feda-42e2-ad49-e136fbe4730e" containerID="f323834dd9db12ac7e2792e0ed828f09b7ffa14aa5b01d6d0e31319a8002dab3" exitCode=0 Nov 25 10:41:11 crc kubenswrapper[4776]: I1125 10:41:11.736393 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rmpxx" event={"ID":"afc31a27-feda-42e2-ad49-e136fbe4730e","Type":"ContainerDied","Data":"f323834dd9db12ac7e2792e0ed828f09b7ffa14aa5b01d6d0e31319a8002dab3"} Nov 25 10:41:11 crc kubenswrapper[4776]: I1125 10:41:11.736669 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rmpxx" event={"ID":"afc31a27-feda-42e2-ad49-e136fbe4730e","Type":"ContainerStarted","Data":"3d09bd26e1a1f26213832a056977b9c512d7c2a16aaf521836b25525da3aa5e0"} Nov 25 10:41:11 crc kubenswrapper[4776]: I1125 10:41:11.739312 4776 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 10:41:12 crc kubenswrapper[4776]: I1125 10:41:12.747262 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rmpxx" event={"ID":"afc31a27-feda-42e2-ad49-e136fbe4730e","Type":"ContainerStarted","Data":"02cdc0c02e3f7f9a2aafa5c7de8f8ca16074a608517d2787bd86ba8263b100cf"} Nov 25 10:41:13 crc kubenswrapper[4776]: I1125 10:41:13.755795 4776 generic.go:334] "Generic (PLEG): container finished" podID="afc31a27-feda-42e2-ad49-e136fbe4730e" containerID="02cdc0c02e3f7f9a2aafa5c7de8f8ca16074a608517d2787bd86ba8263b100cf" exitCode=0 Nov 25 10:41:13 crc kubenswrapper[4776]: I1125 10:41:13.755903 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rmpxx" event={"ID":"afc31a27-feda-42e2-ad49-e136fbe4730e","Type":"ContainerDied","Data":"02cdc0c02e3f7f9a2aafa5c7de8f8ca16074a608517d2787bd86ba8263b100cf"} Nov 25 10:41:14 crc kubenswrapper[4776]: I1125 10:41:14.765690 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rmpxx" event={"ID":"afc31a27-feda-42e2-ad49-e136fbe4730e","Type":"ContainerStarted","Data":"ad9e979ab121fa2fe88b5864e0a4d7122f614c0d84a82ea36732236f79fc7380"} Nov 25 10:41:14 crc kubenswrapper[4776]: I1125 10:41:14.797636 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rmpxx" podStartSLOduration=2.354316993 podStartE2EDuration="4.797616232s" podCreationTimestamp="2025-11-25 10:41:10 +0000 UTC" firstStartedPulling="2025-11-25 10:41:11.738809077 +0000 UTC m=+4616.779868670" lastFinishedPulling="2025-11-25 10:41:14.182108346 +0000 UTC m=+4619.223167909" observedRunningTime="2025-11-25 10:41:14.79157446 +0000 UTC m=+4619.832634013" watchObservedRunningTime="2025-11-25 10:41:14.797616232 +0000 UTC m=+4619.838675775" Nov 25 10:41:20 crc kubenswrapper[4776]: I1125 10:41:20.908327 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rmpxx" Nov 25 10:41:20 crc kubenswrapper[4776]: I1125 10:41:20.909108 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rmpxx" Nov 25 10:41:20 crc kubenswrapper[4776]: I1125 10:41:20.988268 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rmpxx" Nov 25 10:41:21 crc kubenswrapper[4776]: I1125 10:41:21.894999 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rmpxx" Nov 25 10:41:21 crc kubenswrapper[4776]: I1125 10:41:21.943096 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rmpxx"] Nov 25 10:41:23 crc kubenswrapper[4776]: I1125 10:41:23.858477 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rmpxx" podUID="afc31a27-feda-42e2-ad49-e136fbe4730e" containerName="registry-server" containerID="cri-o://ad9e979ab121fa2fe88b5864e0a4d7122f614c0d84a82ea36732236f79fc7380" gracePeriod=2 Nov 25 10:41:24 crc kubenswrapper[4776]: I1125 10:41:24.663896 4776 scope.go:117] "RemoveContainer" containerID="f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a" Nov 25 10:41:24 crc kubenswrapper[4776]: E1125 10:41:24.664562 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:41:24 crc kubenswrapper[4776]: I1125 10:41:24.868555 4776 generic.go:334] "Generic (PLEG): container finished" podID="afc31a27-feda-42e2-ad49-e136fbe4730e" containerID="ad9e979ab121fa2fe88b5864e0a4d7122f614c0d84a82ea36732236f79fc7380" exitCode=0 Nov 25 10:41:24 crc kubenswrapper[4776]: I1125 10:41:24.868673 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rmpxx" event={"ID":"afc31a27-feda-42e2-ad49-e136fbe4730e","Type":"ContainerDied","Data":"ad9e979ab121fa2fe88b5864e0a4d7122f614c0d84a82ea36732236f79fc7380"} Nov 25 10:41:25 crc kubenswrapper[4776]: I1125 10:41:25.049943 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rmpxx" Nov 25 10:41:25 crc kubenswrapper[4776]: I1125 10:41:25.240796 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc31a27-feda-42e2-ad49-e136fbe4730e-catalog-content\") pod \"afc31a27-feda-42e2-ad49-e136fbe4730e\" (UID: \"afc31a27-feda-42e2-ad49-e136fbe4730e\") " Nov 25 10:41:25 crc kubenswrapper[4776]: I1125 10:41:25.240872 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc31a27-feda-42e2-ad49-e136fbe4730e-utilities\") pod \"afc31a27-feda-42e2-ad49-e136fbe4730e\" (UID: \"afc31a27-feda-42e2-ad49-e136fbe4730e\") " Nov 25 10:41:25 crc kubenswrapper[4776]: I1125 10:41:25.240900 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6sj2\" (UniqueName: \"kubernetes.io/projected/afc31a27-feda-42e2-ad49-e136fbe4730e-kube-api-access-x6sj2\") pod \"afc31a27-feda-42e2-ad49-e136fbe4730e\" (UID: \"afc31a27-feda-42e2-ad49-e136fbe4730e\") " Nov 25 10:41:25 crc kubenswrapper[4776]: I1125 10:41:25.242375 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afc31a27-feda-42e2-ad49-e136fbe4730e-utilities" (OuterVolumeSpecName: "utilities") pod "afc31a27-feda-42e2-ad49-e136fbe4730e" (UID: "afc31a27-feda-42e2-ad49-e136fbe4730e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:41:25 crc kubenswrapper[4776]: I1125 10:41:25.254618 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afc31a27-feda-42e2-ad49-e136fbe4730e-kube-api-access-x6sj2" (OuterVolumeSpecName: "kube-api-access-x6sj2") pod "afc31a27-feda-42e2-ad49-e136fbe4730e" (UID: "afc31a27-feda-42e2-ad49-e136fbe4730e"). InnerVolumeSpecName "kube-api-access-x6sj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:41:25 crc kubenswrapper[4776]: I1125 10:41:25.305759 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afc31a27-feda-42e2-ad49-e136fbe4730e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "afc31a27-feda-42e2-ad49-e136fbe4730e" (UID: "afc31a27-feda-42e2-ad49-e136fbe4730e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:41:25 crc kubenswrapper[4776]: I1125 10:41:25.343501 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afc31a27-feda-42e2-ad49-e136fbe4730e-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:41:25 crc kubenswrapper[4776]: I1125 10:41:25.343568 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afc31a27-feda-42e2-ad49-e136fbe4730e-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:41:25 crc kubenswrapper[4776]: I1125 10:41:25.343585 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6sj2\" (UniqueName: \"kubernetes.io/projected/afc31a27-feda-42e2-ad49-e136fbe4730e-kube-api-access-x6sj2\") on node \"crc\" DevicePath \"\"" Nov 25 10:41:25 crc kubenswrapper[4776]: I1125 10:41:25.880304 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rmpxx" event={"ID":"afc31a27-feda-42e2-ad49-e136fbe4730e","Type":"ContainerDied","Data":"3d09bd26e1a1f26213832a056977b9c512d7c2a16aaf521836b25525da3aa5e0"} Nov 25 10:41:25 crc kubenswrapper[4776]: I1125 10:41:25.880361 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rmpxx" Nov 25 10:41:25 crc kubenswrapper[4776]: I1125 10:41:25.880416 4776 scope.go:117] "RemoveContainer" containerID="ad9e979ab121fa2fe88b5864e0a4d7122f614c0d84a82ea36732236f79fc7380" Nov 25 10:41:25 crc kubenswrapper[4776]: I1125 10:41:25.908609 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rmpxx"] Nov 25 10:41:25 crc kubenswrapper[4776]: I1125 10:41:25.912485 4776 scope.go:117] "RemoveContainer" containerID="02cdc0c02e3f7f9a2aafa5c7de8f8ca16074a608517d2787bd86ba8263b100cf" Nov 25 10:41:25 crc kubenswrapper[4776]: I1125 10:41:25.915290 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rmpxx"] Nov 25 10:41:25 crc kubenswrapper[4776]: I1125 10:41:25.937300 4776 scope.go:117] "RemoveContainer" containerID="f323834dd9db12ac7e2792e0ed828f09b7ffa14aa5b01d6d0e31319a8002dab3" Nov 25 10:41:27 crc kubenswrapper[4776]: I1125 10:41:27.676357 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afc31a27-feda-42e2-ad49-e136fbe4730e" path="/var/lib/kubelet/pods/afc31a27-feda-42e2-ad49-e136fbe4730e/volumes" Nov 25 10:41:39 crc kubenswrapper[4776]: I1125 10:41:39.662877 4776 scope.go:117] "RemoveContainer" containerID="f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a" Nov 25 10:41:39 crc kubenswrapper[4776]: E1125 10:41:39.663689 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:41:50 crc kubenswrapper[4776]: I1125 10:41:50.662418 4776 scope.go:117] "RemoveContainer" containerID="f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a" Nov 25 10:41:50 crc kubenswrapper[4776]: E1125 10:41:50.664100 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:42:04 crc kubenswrapper[4776]: I1125 10:42:04.662375 4776 scope.go:117] "RemoveContainer" containerID="f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a" Nov 25 10:42:04 crc kubenswrapper[4776]: E1125 10:42:04.663130 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:42:15 crc kubenswrapper[4776]: I1125 10:42:15.666313 4776 scope.go:117] "RemoveContainer" containerID="f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a" Nov 25 10:42:15 crc kubenswrapper[4776]: E1125 10:42:15.667063 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:42:29 crc kubenswrapper[4776]: I1125 10:42:29.662866 4776 scope.go:117] "RemoveContainer" containerID="f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a" Nov 25 10:42:29 crc kubenswrapper[4776]: E1125 10:42:29.663645 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:42:40 crc kubenswrapper[4776]: I1125 10:42:40.662392 4776 scope.go:117] "RemoveContainer" containerID="f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a" Nov 25 10:42:40 crc kubenswrapper[4776]: E1125 10:42:40.663090 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:42:52 crc kubenswrapper[4776]: I1125 10:42:52.662786 4776 scope.go:117] "RemoveContainer" containerID="f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a" Nov 25 10:42:52 crc kubenswrapper[4776]: E1125 10:42:52.665480 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:43:03 crc kubenswrapper[4776]: I1125 10:43:03.661842 4776 scope.go:117] "RemoveContainer" containerID="f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a" Nov 25 10:43:03 crc kubenswrapper[4776]: E1125 10:43:03.662715 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:43:18 crc kubenswrapper[4776]: I1125 10:43:18.662307 4776 scope.go:117] "RemoveContainer" containerID="f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a" Nov 25 10:43:18 crc kubenswrapper[4776]: E1125 10:43:18.663238 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:43:26 crc kubenswrapper[4776]: I1125 10:43:26.694218 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b45qs"] Nov 25 10:43:26 crc kubenswrapper[4776]: E1125 10:43:26.695009 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc31a27-feda-42e2-ad49-e136fbe4730e" containerName="extract-content" Nov 25 10:43:26 crc kubenswrapper[4776]: I1125 10:43:26.695021 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc31a27-feda-42e2-ad49-e136fbe4730e" containerName="extract-content" Nov 25 10:43:26 crc kubenswrapper[4776]: E1125 10:43:26.695044 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc31a27-feda-42e2-ad49-e136fbe4730e" containerName="registry-server" Nov 25 10:43:26 crc kubenswrapper[4776]: I1125 10:43:26.695050 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc31a27-feda-42e2-ad49-e136fbe4730e" containerName="registry-server" Nov 25 10:43:26 crc kubenswrapper[4776]: E1125 10:43:26.695062 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc31a27-feda-42e2-ad49-e136fbe4730e" containerName="extract-utilities" Nov 25 10:43:26 crc kubenswrapper[4776]: I1125 10:43:26.695088 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc31a27-feda-42e2-ad49-e136fbe4730e" containerName="extract-utilities" Nov 25 10:43:26 crc kubenswrapper[4776]: I1125 10:43:26.695219 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="afc31a27-feda-42e2-ad49-e136fbe4730e" containerName="registry-server" Nov 25 10:43:26 crc kubenswrapper[4776]: I1125 10:43:26.696328 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b45qs" Nov 25 10:43:26 crc kubenswrapper[4776]: I1125 10:43:26.705967 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b45qs"] Nov 25 10:43:26 crc kubenswrapper[4776]: I1125 10:43:26.808748 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/185c5b2e-90e4-4717-b079-59de084f079a-utilities\") pod \"redhat-marketplace-b45qs\" (UID: \"185c5b2e-90e4-4717-b079-59de084f079a\") " pod="openshift-marketplace/redhat-marketplace-b45qs" Nov 25 10:43:26 crc kubenswrapper[4776]: I1125 10:43:26.808824 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w94x\" (UniqueName: \"kubernetes.io/projected/185c5b2e-90e4-4717-b079-59de084f079a-kube-api-access-8w94x\") pod \"redhat-marketplace-b45qs\" (UID: \"185c5b2e-90e4-4717-b079-59de084f079a\") " pod="openshift-marketplace/redhat-marketplace-b45qs" Nov 25 10:43:26 crc kubenswrapper[4776]: I1125 10:43:26.808855 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/185c5b2e-90e4-4717-b079-59de084f079a-catalog-content\") pod \"redhat-marketplace-b45qs\" (UID: \"185c5b2e-90e4-4717-b079-59de084f079a\") " pod="openshift-marketplace/redhat-marketplace-b45qs" Nov 25 10:43:26 crc kubenswrapper[4776]: I1125 10:43:26.910274 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/185c5b2e-90e4-4717-b079-59de084f079a-utilities\") pod \"redhat-marketplace-b45qs\" (UID: \"185c5b2e-90e4-4717-b079-59de084f079a\") " pod="openshift-marketplace/redhat-marketplace-b45qs" Nov 25 10:43:26 crc kubenswrapper[4776]: I1125 10:43:26.910352 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w94x\" (UniqueName: \"kubernetes.io/projected/185c5b2e-90e4-4717-b079-59de084f079a-kube-api-access-8w94x\") pod \"redhat-marketplace-b45qs\" (UID: \"185c5b2e-90e4-4717-b079-59de084f079a\") " pod="openshift-marketplace/redhat-marketplace-b45qs" Nov 25 10:43:26 crc kubenswrapper[4776]: I1125 10:43:26.910375 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/185c5b2e-90e4-4717-b079-59de084f079a-catalog-content\") pod \"redhat-marketplace-b45qs\" (UID: \"185c5b2e-90e4-4717-b079-59de084f079a\") " pod="openshift-marketplace/redhat-marketplace-b45qs" Nov 25 10:43:26 crc kubenswrapper[4776]: I1125 10:43:26.911045 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/185c5b2e-90e4-4717-b079-59de084f079a-catalog-content\") pod \"redhat-marketplace-b45qs\" (UID: \"185c5b2e-90e4-4717-b079-59de084f079a\") " pod="openshift-marketplace/redhat-marketplace-b45qs" Nov 25 10:43:26 crc kubenswrapper[4776]: I1125 10:43:26.911055 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/185c5b2e-90e4-4717-b079-59de084f079a-utilities\") pod \"redhat-marketplace-b45qs\" (UID: \"185c5b2e-90e4-4717-b079-59de084f079a\") " pod="openshift-marketplace/redhat-marketplace-b45qs" Nov 25 10:43:26 crc kubenswrapper[4776]: I1125 10:43:26.935986 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w94x\" (UniqueName: \"kubernetes.io/projected/185c5b2e-90e4-4717-b079-59de084f079a-kube-api-access-8w94x\") pod \"redhat-marketplace-b45qs\" (UID: \"185c5b2e-90e4-4717-b079-59de084f079a\") " pod="openshift-marketplace/redhat-marketplace-b45qs" Nov 25 10:43:27 crc kubenswrapper[4776]: I1125 10:43:27.021861 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b45qs" Nov 25 10:43:27 crc kubenswrapper[4776]: I1125 10:43:27.480094 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b45qs"] Nov 25 10:43:27 crc kubenswrapper[4776]: W1125 10:43:27.494013 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod185c5b2e_90e4_4717_b079_59de084f079a.slice/crio-4296a425b39f8c59fd35056104636360e3231feec821b745e75fa9f30a7692f1 WatchSource:0}: Error finding container 4296a425b39f8c59fd35056104636360e3231feec821b745e75fa9f30a7692f1: Status 404 returned error can't find the container with id 4296a425b39f8c59fd35056104636360e3231feec821b745e75fa9f30a7692f1 Nov 25 10:43:27 crc kubenswrapper[4776]: I1125 10:43:27.812354 4776 generic.go:334] "Generic (PLEG): container finished" podID="185c5b2e-90e4-4717-b079-59de084f079a" containerID="f847e7992a3f301648b35b0b96a9db50db4ba00d6f9fd90966ecf0aa67c4534c" exitCode=0 Nov 25 10:43:27 crc kubenswrapper[4776]: I1125 10:43:27.812420 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b45qs" event={"ID":"185c5b2e-90e4-4717-b079-59de084f079a","Type":"ContainerDied","Data":"f847e7992a3f301648b35b0b96a9db50db4ba00d6f9fd90966ecf0aa67c4534c"} Nov 25 10:43:27 crc kubenswrapper[4776]: I1125 10:43:27.812725 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b45qs" event={"ID":"185c5b2e-90e4-4717-b079-59de084f079a","Type":"ContainerStarted","Data":"4296a425b39f8c59fd35056104636360e3231feec821b745e75fa9f30a7692f1"} Nov 25 10:43:28 crc kubenswrapper[4776]: I1125 10:43:28.822535 4776 generic.go:334] "Generic (PLEG): container finished" podID="185c5b2e-90e4-4717-b079-59de084f079a" containerID="8540a355ba40eddb4976630bb3786298b778501d20cd97c495b919120d61544d" exitCode=0 Nov 25 10:43:28 crc kubenswrapper[4776]: I1125 10:43:28.822579 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b45qs" event={"ID":"185c5b2e-90e4-4717-b079-59de084f079a","Type":"ContainerDied","Data":"8540a355ba40eddb4976630bb3786298b778501d20cd97c495b919120d61544d"} Nov 25 10:43:29 crc kubenswrapper[4776]: I1125 10:43:29.831470 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b45qs" event={"ID":"185c5b2e-90e4-4717-b079-59de084f079a","Type":"ContainerStarted","Data":"e5de3304db34056558328ce6b4abae36952190992544d677c8c39076aa3a0d66"} Nov 25 10:43:29 crc kubenswrapper[4776]: I1125 10:43:29.852061 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b45qs" podStartSLOduration=2.40928516 podStartE2EDuration="3.852042594s" podCreationTimestamp="2025-11-25 10:43:26 +0000 UTC" firstStartedPulling="2025-11-25 10:43:27.815318991 +0000 UTC m=+4752.856378544" lastFinishedPulling="2025-11-25 10:43:29.258076425 +0000 UTC m=+4754.299135978" observedRunningTime="2025-11-25 10:43:29.845995012 +0000 UTC m=+4754.887054595" watchObservedRunningTime="2025-11-25 10:43:29.852042594 +0000 UTC m=+4754.893102147" Nov 25 10:43:33 crc kubenswrapper[4776]: I1125 10:43:33.663717 4776 scope.go:117] "RemoveContainer" containerID="f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a" Nov 25 10:43:33 crc kubenswrapper[4776]: E1125 10:43:33.664632 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:43:35 crc kubenswrapper[4776]: I1125 10:43:35.842792 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mknpb"] Nov 25 10:43:35 crc kubenswrapper[4776]: I1125 10:43:35.845677 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mknpb" Nov 25 10:43:35 crc kubenswrapper[4776]: I1125 10:43:35.847062 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e010e173-e0bc-423e-9bba-2ef28d32d129-utilities\") pod \"community-operators-mknpb\" (UID: \"e010e173-e0bc-423e-9bba-2ef28d32d129\") " pod="openshift-marketplace/community-operators-mknpb" Nov 25 10:43:35 crc kubenswrapper[4776]: I1125 10:43:35.847181 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e010e173-e0bc-423e-9bba-2ef28d32d129-catalog-content\") pod \"community-operators-mknpb\" (UID: \"e010e173-e0bc-423e-9bba-2ef28d32d129\") " pod="openshift-marketplace/community-operators-mknpb" Nov 25 10:43:35 crc kubenswrapper[4776]: I1125 10:43:35.847265 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvx8p\" (UniqueName: \"kubernetes.io/projected/e010e173-e0bc-423e-9bba-2ef28d32d129-kube-api-access-qvx8p\") pod \"community-operators-mknpb\" (UID: \"e010e173-e0bc-423e-9bba-2ef28d32d129\") " pod="openshift-marketplace/community-operators-mknpb" Nov 25 10:43:35 crc kubenswrapper[4776]: I1125 10:43:35.856772 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mknpb"] Nov 25 10:43:35 crc kubenswrapper[4776]: I1125 10:43:35.948377 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e010e173-e0bc-423e-9bba-2ef28d32d129-utilities\") pod \"community-operators-mknpb\" (UID: \"e010e173-e0bc-423e-9bba-2ef28d32d129\") " pod="openshift-marketplace/community-operators-mknpb" Nov 25 10:43:35 crc kubenswrapper[4776]: I1125 10:43:35.948419 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e010e173-e0bc-423e-9bba-2ef28d32d129-catalog-content\") pod \"community-operators-mknpb\" (UID: \"e010e173-e0bc-423e-9bba-2ef28d32d129\") " pod="openshift-marketplace/community-operators-mknpb" Nov 25 10:43:35 crc kubenswrapper[4776]: I1125 10:43:35.948466 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvx8p\" (UniqueName: \"kubernetes.io/projected/e010e173-e0bc-423e-9bba-2ef28d32d129-kube-api-access-qvx8p\") pod \"community-operators-mknpb\" (UID: \"e010e173-e0bc-423e-9bba-2ef28d32d129\") " pod="openshift-marketplace/community-operators-mknpb" Nov 25 10:43:35 crc kubenswrapper[4776]: I1125 10:43:35.949275 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e010e173-e0bc-423e-9bba-2ef28d32d129-utilities\") pod \"community-operators-mknpb\" (UID: \"e010e173-e0bc-423e-9bba-2ef28d32d129\") " pod="openshift-marketplace/community-operators-mknpb" Nov 25 10:43:35 crc kubenswrapper[4776]: I1125 10:43:35.949330 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e010e173-e0bc-423e-9bba-2ef28d32d129-catalog-content\") pod \"community-operators-mknpb\" (UID: \"e010e173-e0bc-423e-9bba-2ef28d32d129\") " pod="openshift-marketplace/community-operators-mknpb" Nov 25 10:43:35 crc kubenswrapper[4776]: I1125 10:43:35.966870 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvx8p\" (UniqueName: \"kubernetes.io/projected/e010e173-e0bc-423e-9bba-2ef28d32d129-kube-api-access-qvx8p\") pod \"community-operators-mknpb\" (UID: \"e010e173-e0bc-423e-9bba-2ef28d32d129\") " pod="openshift-marketplace/community-operators-mknpb" Nov 25 10:43:36 crc kubenswrapper[4776]: I1125 10:43:36.183314 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mknpb" Nov 25 10:43:36 crc kubenswrapper[4776]: I1125 10:43:36.649018 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mknpb"] Nov 25 10:43:36 crc kubenswrapper[4776]: I1125 10:43:36.891876 4776 generic.go:334] "Generic (PLEG): container finished" podID="e010e173-e0bc-423e-9bba-2ef28d32d129" containerID="7ccea64a6dc16d85ea9ded610580c9c9276d74826f7e0ab83e7389beadc0ca88" exitCode=0 Nov 25 10:43:36 crc kubenswrapper[4776]: I1125 10:43:36.891924 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mknpb" event={"ID":"e010e173-e0bc-423e-9bba-2ef28d32d129","Type":"ContainerDied","Data":"7ccea64a6dc16d85ea9ded610580c9c9276d74826f7e0ab83e7389beadc0ca88"} Nov 25 10:43:36 crc kubenswrapper[4776]: I1125 10:43:36.891951 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mknpb" event={"ID":"e010e173-e0bc-423e-9bba-2ef28d32d129","Type":"ContainerStarted","Data":"42f8d1e78b2698c019c02020a09a19f5a5e5bf44c7c52b0de0ec1f953db726b3"} Nov 25 10:43:37 crc kubenswrapper[4776]: I1125 10:43:37.022622 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b45qs" Nov 25 10:43:37 crc kubenswrapper[4776]: I1125 10:43:37.023159 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b45qs" Nov 25 10:43:37 crc kubenswrapper[4776]: I1125 10:43:37.061051 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b45qs" Nov 25 10:43:37 crc kubenswrapper[4776]: I1125 10:43:37.906236 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mknpb" event={"ID":"e010e173-e0bc-423e-9bba-2ef28d32d129","Type":"ContainerStarted","Data":"58090c73a39c980d9cfb1d7debc1a3b8e065c1b8698b803052cd7f124bb08423"} Nov 25 10:43:37 crc kubenswrapper[4776]: I1125 10:43:37.953632 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b45qs" Nov 25 10:43:38 crc kubenswrapper[4776]: I1125 10:43:38.913807 4776 generic.go:334] "Generic (PLEG): container finished" podID="e010e173-e0bc-423e-9bba-2ef28d32d129" containerID="58090c73a39c980d9cfb1d7debc1a3b8e065c1b8698b803052cd7f124bb08423" exitCode=0 Nov 25 10:43:38 crc kubenswrapper[4776]: I1125 10:43:38.913855 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mknpb" event={"ID":"e010e173-e0bc-423e-9bba-2ef28d32d129","Type":"ContainerDied","Data":"58090c73a39c980d9cfb1d7debc1a3b8e065c1b8698b803052cd7f124bb08423"} Nov 25 10:43:39 crc kubenswrapper[4776]: I1125 10:43:39.414302 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b45qs"] Nov 25 10:43:39 crc kubenswrapper[4776]: I1125 10:43:39.920915 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mknpb" event={"ID":"e010e173-e0bc-423e-9bba-2ef28d32d129","Type":"ContainerStarted","Data":"646ab527eb55bc70bb44a3b15624dedd63e259867372573c4d6f7846f56ad443"} Nov 25 10:43:39 crc kubenswrapper[4776]: I1125 10:43:39.921111 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b45qs" podUID="185c5b2e-90e4-4717-b079-59de084f079a" containerName="registry-server" containerID="cri-o://e5de3304db34056558328ce6b4abae36952190992544d677c8c39076aa3a0d66" gracePeriod=2 Nov 25 10:43:39 crc kubenswrapper[4776]: I1125 10:43:39.946097 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mknpb" podStartSLOduration=2.487906369 podStartE2EDuration="4.946058801s" podCreationTimestamp="2025-11-25 10:43:35 +0000 UTC" firstStartedPulling="2025-11-25 10:43:36.893819108 +0000 UTC m=+4761.934878661" lastFinishedPulling="2025-11-25 10:43:39.35197154 +0000 UTC m=+4764.393031093" observedRunningTime="2025-11-25 10:43:39.945182149 +0000 UTC m=+4764.986241702" watchObservedRunningTime="2025-11-25 10:43:39.946058801 +0000 UTC m=+4764.987118354" Nov 25 10:43:40 crc kubenswrapper[4776]: I1125 10:43:40.292458 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b45qs" Nov 25 10:43:40 crc kubenswrapper[4776]: I1125 10:43:40.435850 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8w94x\" (UniqueName: \"kubernetes.io/projected/185c5b2e-90e4-4717-b079-59de084f079a-kube-api-access-8w94x\") pod \"185c5b2e-90e4-4717-b079-59de084f079a\" (UID: \"185c5b2e-90e4-4717-b079-59de084f079a\") " Nov 25 10:43:40 crc kubenswrapper[4776]: I1125 10:43:40.435967 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/185c5b2e-90e4-4717-b079-59de084f079a-catalog-content\") pod \"185c5b2e-90e4-4717-b079-59de084f079a\" (UID: \"185c5b2e-90e4-4717-b079-59de084f079a\") " Nov 25 10:43:40 crc kubenswrapper[4776]: I1125 10:43:40.436148 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/185c5b2e-90e4-4717-b079-59de084f079a-utilities\") pod \"185c5b2e-90e4-4717-b079-59de084f079a\" (UID: \"185c5b2e-90e4-4717-b079-59de084f079a\") " Nov 25 10:43:40 crc kubenswrapper[4776]: I1125 10:43:40.437423 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/185c5b2e-90e4-4717-b079-59de084f079a-utilities" (OuterVolumeSpecName: "utilities") pod "185c5b2e-90e4-4717-b079-59de084f079a" (UID: "185c5b2e-90e4-4717-b079-59de084f079a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:43:40 crc kubenswrapper[4776]: I1125 10:43:40.441226 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/185c5b2e-90e4-4717-b079-59de084f079a-kube-api-access-8w94x" (OuterVolumeSpecName: "kube-api-access-8w94x") pod "185c5b2e-90e4-4717-b079-59de084f079a" (UID: "185c5b2e-90e4-4717-b079-59de084f079a"). InnerVolumeSpecName "kube-api-access-8w94x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:43:40 crc kubenswrapper[4776]: I1125 10:43:40.462261 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/185c5b2e-90e4-4717-b079-59de084f079a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "185c5b2e-90e4-4717-b079-59de084f079a" (UID: "185c5b2e-90e4-4717-b079-59de084f079a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:43:40 crc kubenswrapper[4776]: I1125 10:43:40.538266 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/185c5b2e-90e4-4717-b079-59de084f079a-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:43:40 crc kubenswrapper[4776]: I1125 10:43:40.538313 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8w94x\" (UniqueName: \"kubernetes.io/projected/185c5b2e-90e4-4717-b079-59de084f079a-kube-api-access-8w94x\") on node \"crc\" DevicePath \"\"" Nov 25 10:43:40 crc kubenswrapper[4776]: I1125 10:43:40.538323 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/185c5b2e-90e4-4717-b079-59de084f079a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:43:40 crc kubenswrapper[4776]: I1125 10:43:40.930120 4776 generic.go:334] "Generic (PLEG): container finished" podID="185c5b2e-90e4-4717-b079-59de084f079a" containerID="e5de3304db34056558328ce6b4abae36952190992544d677c8c39076aa3a0d66" exitCode=0 Nov 25 10:43:40 crc kubenswrapper[4776]: I1125 10:43:40.930188 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b45qs" Nov 25 10:43:40 crc kubenswrapper[4776]: I1125 10:43:40.930206 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b45qs" event={"ID":"185c5b2e-90e4-4717-b079-59de084f079a","Type":"ContainerDied","Data":"e5de3304db34056558328ce6b4abae36952190992544d677c8c39076aa3a0d66"} Nov 25 10:43:40 crc kubenswrapper[4776]: I1125 10:43:40.930267 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b45qs" event={"ID":"185c5b2e-90e4-4717-b079-59de084f079a","Type":"ContainerDied","Data":"4296a425b39f8c59fd35056104636360e3231feec821b745e75fa9f30a7692f1"} Nov 25 10:43:40 crc kubenswrapper[4776]: I1125 10:43:40.930296 4776 scope.go:117] "RemoveContainer" containerID="e5de3304db34056558328ce6b4abae36952190992544d677c8c39076aa3a0d66" Nov 25 10:43:40 crc kubenswrapper[4776]: I1125 10:43:40.948767 4776 scope.go:117] "RemoveContainer" containerID="8540a355ba40eddb4976630bb3786298b778501d20cd97c495b919120d61544d" Nov 25 10:43:40 crc kubenswrapper[4776]: I1125 10:43:40.964651 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b45qs"] Nov 25 10:43:40 crc kubenswrapper[4776]: I1125 10:43:40.969172 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b45qs"] Nov 25 10:43:40 crc kubenswrapper[4776]: I1125 10:43:40.980466 4776 scope.go:117] "RemoveContainer" containerID="f847e7992a3f301648b35b0b96a9db50db4ba00d6f9fd90966ecf0aa67c4534c" Nov 25 10:43:41 crc kubenswrapper[4776]: I1125 10:43:41.014583 4776 scope.go:117] "RemoveContainer" containerID="e5de3304db34056558328ce6b4abae36952190992544d677c8c39076aa3a0d66" Nov 25 10:43:41 crc kubenswrapper[4776]: E1125 10:43:41.016629 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5de3304db34056558328ce6b4abae36952190992544d677c8c39076aa3a0d66\": container with ID starting with e5de3304db34056558328ce6b4abae36952190992544d677c8c39076aa3a0d66 not found: ID does not exist" containerID="e5de3304db34056558328ce6b4abae36952190992544d677c8c39076aa3a0d66" Nov 25 10:43:41 crc kubenswrapper[4776]: I1125 10:43:41.016662 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5de3304db34056558328ce6b4abae36952190992544d677c8c39076aa3a0d66"} err="failed to get container status \"e5de3304db34056558328ce6b4abae36952190992544d677c8c39076aa3a0d66\": rpc error: code = NotFound desc = could not find container \"e5de3304db34056558328ce6b4abae36952190992544d677c8c39076aa3a0d66\": container with ID starting with e5de3304db34056558328ce6b4abae36952190992544d677c8c39076aa3a0d66 not found: ID does not exist" Nov 25 10:43:41 crc kubenswrapper[4776]: I1125 10:43:41.016704 4776 scope.go:117] "RemoveContainer" containerID="8540a355ba40eddb4976630bb3786298b778501d20cd97c495b919120d61544d" Nov 25 10:43:41 crc kubenswrapper[4776]: E1125 10:43:41.016963 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8540a355ba40eddb4976630bb3786298b778501d20cd97c495b919120d61544d\": container with ID starting with 8540a355ba40eddb4976630bb3786298b778501d20cd97c495b919120d61544d not found: ID does not exist" containerID="8540a355ba40eddb4976630bb3786298b778501d20cd97c495b919120d61544d" Nov 25 10:43:41 crc kubenswrapper[4776]: I1125 10:43:41.016981 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8540a355ba40eddb4976630bb3786298b778501d20cd97c495b919120d61544d"} err="failed to get container status \"8540a355ba40eddb4976630bb3786298b778501d20cd97c495b919120d61544d\": rpc error: code = NotFound desc = could not find container \"8540a355ba40eddb4976630bb3786298b778501d20cd97c495b919120d61544d\": container with ID starting with 8540a355ba40eddb4976630bb3786298b778501d20cd97c495b919120d61544d not found: ID does not exist" Nov 25 10:43:41 crc kubenswrapper[4776]: I1125 10:43:41.017000 4776 scope.go:117] "RemoveContainer" containerID="f847e7992a3f301648b35b0b96a9db50db4ba00d6f9fd90966ecf0aa67c4534c" Nov 25 10:43:41 crc kubenswrapper[4776]: E1125 10:43:41.017215 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f847e7992a3f301648b35b0b96a9db50db4ba00d6f9fd90966ecf0aa67c4534c\": container with ID starting with f847e7992a3f301648b35b0b96a9db50db4ba00d6f9fd90966ecf0aa67c4534c not found: ID does not exist" containerID="f847e7992a3f301648b35b0b96a9db50db4ba00d6f9fd90966ecf0aa67c4534c" Nov 25 10:43:41 crc kubenswrapper[4776]: I1125 10:43:41.017236 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f847e7992a3f301648b35b0b96a9db50db4ba00d6f9fd90966ecf0aa67c4534c"} err="failed to get container status \"f847e7992a3f301648b35b0b96a9db50db4ba00d6f9fd90966ecf0aa67c4534c\": rpc error: code = NotFound desc = could not find container \"f847e7992a3f301648b35b0b96a9db50db4ba00d6f9fd90966ecf0aa67c4534c\": container with ID starting with f847e7992a3f301648b35b0b96a9db50db4ba00d6f9fd90966ecf0aa67c4534c not found: ID does not exist" Nov 25 10:43:41 crc kubenswrapper[4776]: I1125 10:43:41.683135 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="185c5b2e-90e4-4717-b079-59de084f079a" path="/var/lib/kubelet/pods/185c5b2e-90e4-4717-b079-59de084f079a/volumes" Nov 25 10:43:46 crc kubenswrapper[4776]: I1125 10:43:46.183978 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mknpb" Nov 25 10:43:46 crc kubenswrapper[4776]: I1125 10:43:46.184360 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mknpb" Nov 25 10:43:46 crc kubenswrapper[4776]: I1125 10:43:46.232015 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mknpb" Nov 25 10:43:46 crc kubenswrapper[4776]: I1125 10:43:46.663381 4776 scope.go:117] "RemoveContainer" containerID="f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a" Nov 25 10:43:46 crc kubenswrapper[4776]: E1125 10:43:46.663808 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:43:47 crc kubenswrapper[4776]: I1125 10:43:47.034722 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mknpb" Nov 25 10:43:47 crc kubenswrapper[4776]: I1125 10:43:47.095624 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mknpb"] Nov 25 10:43:48 crc kubenswrapper[4776]: I1125 10:43:48.991505 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mknpb" podUID="e010e173-e0bc-423e-9bba-2ef28d32d129" containerName="registry-server" containerID="cri-o://646ab527eb55bc70bb44a3b15624dedd63e259867372573c4d6f7846f56ad443" gracePeriod=2 Nov 25 10:43:50 crc kubenswrapper[4776]: I1125 10:43:50.002642 4776 generic.go:334] "Generic (PLEG): container finished" podID="e010e173-e0bc-423e-9bba-2ef28d32d129" containerID="646ab527eb55bc70bb44a3b15624dedd63e259867372573c4d6f7846f56ad443" exitCode=0 Nov 25 10:43:50 crc kubenswrapper[4776]: I1125 10:43:50.002687 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mknpb" event={"ID":"e010e173-e0bc-423e-9bba-2ef28d32d129","Type":"ContainerDied","Data":"646ab527eb55bc70bb44a3b15624dedd63e259867372573c4d6f7846f56ad443"} Nov 25 10:43:50 crc kubenswrapper[4776]: I1125 10:43:50.743426 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mknpb" Nov 25 10:43:50 crc kubenswrapper[4776]: I1125 10:43:50.891603 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e010e173-e0bc-423e-9bba-2ef28d32d129-utilities\") pod \"e010e173-e0bc-423e-9bba-2ef28d32d129\" (UID: \"e010e173-e0bc-423e-9bba-2ef28d32d129\") " Nov 25 10:43:50 crc kubenswrapper[4776]: I1125 10:43:50.891677 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvx8p\" (UniqueName: \"kubernetes.io/projected/e010e173-e0bc-423e-9bba-2ef28d32d129-kube-api-access-qvx8p\") pod \"e010e173-e0bc-423e-9bba-2ef28d32d129\" (UID: \"e010e173-e0bc-423e-9bba-2ef28d32d129\") " Nov 25 10:43:50 crc kubenswrapper[4776]: I1125 10:43:50.891791 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e010e173-e0bc-423e-9bba-2ef28d32d129-catalog-content\") pod \"e010e173-e0bc-423e-9bba-2ef28d32d129\" (UID: \"e010e173-e0bc-423e-9bba-2ef28d32d129\") " Nov 25 10:43:50 crc kubenswrapper[4776]: I1125 10:43:50.893261 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e010e173-e0bc-423e-9bba-2ef28d32d129-utilities" (OuterVolumeSpecName: "utilities") pod "e010e173-e0bc-423e-9bba-2ef28d32d129" (UID: "e010e173-e0bc-423e-9bba-2ef28d32d129"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:43:50 crc kubenswrapper[4776]: I1125 10:43:50.904212 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e010e173-e0bc-423e-9bba-2ef28d32d129-kube-api-access-qvx8p" (OuterVolumeSpecName: "kube-api-access-qvx8p") pod "e010e173-e0bc-423e-9bba-2ef28d32d129" (UID: "e010e173-e0bc-423e-9bba-2ef28d32d129"). InnerVolumeSpecName "kube-api-access-qvx8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:43:50 crc kubenswrapper[4776]: I1125 10:43:50.967084 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e010e173-e0bc-423e-9bba-2ef28d32d129-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e010e173-e0bc-423e-9bba-2ef28d32d129" (UID: "e010e173-e0bc-423e-9bba-2ef28d32d129"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:43:50 crc kubenswrapper[4776]: I1125 10:43:50.993567 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e010e173-e0bc-423e-9bba-2ef28d32d129-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:43:50 crc kubenswrapper[4776]: I1125 10:43:50.993608 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvx8p\" (UniqueName: \"kubernetes.io/projected/e010e173-e0bc-423e-9bba-2ef28d32d129-kube-api-access-qvx8p\") on node \"crc\" DevicePath \"\"" Nov 25 10:43:50 crc kubenswrapper[4776]: I1125 10:43:50.993622 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e010e173-e0bc-423e-9bba-2ef28d32d129-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:43:51 crc kubenswrapper[4776]: I1125 10:43:51.017823 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mknpb" event={"ID":"e010e173-e0bc-423e-9bba-2ef28d32d129","Type":"ContainerDied","Data":"42f8d1e78b2698c019c02020a09a19f5a5e5bf44c7c52b0de0ec1f953db726b3"} Nov 25 10:43:51 crc kubenswrapper[4776]: I1125 10:43:51.017887 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mknpb" Nov 25 10:43:51 crc kubenswrapper[4776]: I1125 10:43:51.017894 4776 scope.go:117] "RemoveContainer" containerID="646ab527eb55bc70bb44a3b15624dedd63e259867372573c4d6f7846f56ad443" Nov 25 10:43:51 crc kubenswrapper[4776]: I1125 10:43:51.053980 4776 scope.go:117] "RemoveContainer" containerID="58090c73a39c980d9cfb1d7debc1a3b8e065c1b8698b803052cd7f124bb08423" Nov 25 10:43:51 crc kubenswrapper[4776]: I1125 10:43:51.054903 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mknpb"] Nov 25 10:43:51 crc kubenswrapper[4776]: I1125 10:43:51.065148 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mknpb"] Nov 25 10:43:51 crc kubenswrapper[4776]: I1125 10:43:51.091650 4776 scope.go:117] "RemoveContainer" containerID="7ccea64a6dc16d85ea9ded610580c9c9276d74826f7e0ab83e7389beadc0ca88" Nov 25 10:43:51 crc kubenswrapper[4776]: I1125 10:43:51.677109 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e010e173-e0bc-423e-9bba-2ef28d32d129" path="/var/lib/kubelet/pods/e010e173-e0bc-423e-9bba-2ef28d32d129/volumes" Nov 25 10:43:57 crc kubenswrapper[4776]: I1125 10:43:57.662892 4776 scope.go:117] "RemoveContainer" containerID="f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a" Nov 25 10:43:57 crc kubenswrapper[4776]: E1125 10:43:57.663510 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:44:09 crc kubenswrapper[4776]: I1125 10:44:09.663979 4776 scope.go:117] "RemoveContainer" containerID="f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a" Nov 25 10:44:09 crc kubenswrapper[4776]: E1125 10:44:09.664739 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:44:22 crc kubenswrapper[4776]: I1125 10:44:22.662685 4776 scope.go:117] "RemoveContainer" containerID="f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a" Nov 25 10:44:22 crc kubenswrapper[4776]: E1125 10:44:22.663581 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:44:36 crc kubenswrapper[4776]: I1125 10:44:36.662388 4776 scope.go:117] "RemoveContainer" containerID="f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a" Nov 25 10:44:36 crc kubenswrapper[4776]: E1125 10:44:36.663200 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.014195 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-xhbgd"] Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.018376 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-xhbgd"] Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.118655 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-slnjh"] Nov 25 10:44:39 crc kubenswrapper[4776]: E1125 10:44:39.118949 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e010e173-e0bc-423e-9bba-2ef28d32d129" containerName="extract-content" Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.118965 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="e010e173-e0bc-423e-9bba-2ef28d32d129" containerName="extract-content" Nov 25 10:44:39 crc kubenswrapper[4776]: E1125 10:44:39.118973 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e010e173-e0bc-423e-9bba-2ef28d32d129" containerName="extract-utilities" Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.118980 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="e010e173-e0bc-423e-9bba-2ef28d32d129" containerName="extract-utilities" Nov 25 10:44:39 crc kubenswrapper[4776]: E1125 10:44:39.118999 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="185c5b2e-90e4-4717-b079-59de084f079a" containerName="registry-server" Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.119005 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="185c5b2e-90e4-4717-b079-59de084f079a" containerName="registry-server" Nov 25 10:44:39 crc kubenswrapper[4776]: E1125 10:44:39.119021 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="185c5b2e-90e4-4717-b079-59de084f079a" containerName="extract-utilities" Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.119028 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="185c5b2e-90e4-4717-b079-59de084f079a" containerName="extract-utilities" Nov 25 10:44:39 crc kubenswrapper[4776]: E1125 10:44:39.119039 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e010e173-e0bc-423e-9bba-2ef28d32d129" containerName="registry-server" Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.119044 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="e010e173-e0bc-423e-9bba-2ef28d32d129" containerName="registry-server" Nov 25 10:44:39 crc kubenswrapper[4776]: E1125 10:44:39.119059 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="185c5b2e-90e4-4717-b079-59de084f079a" containerName="extract-content" Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.119077 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="185c5b2e-90e4-4717-b079-59de084f079a" containerName="extract-content" Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.119244 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="e010e173-e0bc-423e-9bba-2ef28d32d129" containerName="registry-server" Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.119257 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="185c5b2e-90e4-4717-b079-59de084f079a" containerName="registry-server" Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.119749 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-slnjh" Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.122500 4776 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-cnl7m" Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.123152 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.123231 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.123234 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.126753 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-slnjh"] Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.159665 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vq9mg\" (UniqueName: \"kubernetes.io/projected/92c92eb6-45b3-40fe-a16a-fae6ccea4bd9-kube-api-access-vq9mg\") pod \"crc-storage-crc-slnjh\" (UID: \"92c92eb6-45b3-40fe-a16a-fae6ccea4bd9\") " pod="crc-storage/crc-storage-crc-slnjh" Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.159707 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/92c92eb6-45b3-40fe-a16a-fae6ccea4bd9-node-mnt\") pod \"crc-storage-crc-slnjh\" (UID: \"92c92eb6-45b3-40fe-a16a-fae6ccea4bd9\") " pod="crc-storage/crc-storage-crc-slnjh" Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.159745 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/92c92eb6-45b3-40fe-a16a-fae6ccea4bd9-crc-storage\") pod \"crc-storage-crc-slnjh\" (UID: \"92c92eb6-45b3-40fe-a16a-fae6ccea4bd9\") " pod="crc-storage/crc-storage-crc-slnjh" Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.260746 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/92c92eb6-45b3-40fe-a16a-fae6ccea4bd9-crc-storage\") pod \"crc-storage-crc-slnjh\" (UID: \"92c92eb6-45b3-40fe-a16a-fae6ccea4bd9\") " pod="crc-storage/crc-storage-crc-slnjh" Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.260894 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vq9mg\" (UniqueName: \"kubernetes.io/projected/92c92eb6-45b3-40fe-a16a-fae6ccea4bd9-kube-api-access-vq9mg\") pod \"crc-storage-crc-slnjh\" (UID: \"92c92eb6-45b3-40fe-a16a-fae6ccea4bd9\") " pod="crc-storage/crc-storage-crc-slnjh" Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.260926 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/92c92eb6-45b3-40fe-a16a-fae6ccea4bd9-node-mnt\") pod \"crc-storage-crc-slnjh\" (UID: \"92c92eb6-45b3-40fe-a16a-fae6ccea4bd9\") " pod="crc-storage/crc-storage-crc-slnjh" Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.261286 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/92c92eb6-45b3-40fe-a16a-fae6ccea4bd9-node-mnt\") pod \"crc-storage-crc-slnjh\" (UID: \"92c92eb6-45b3-40fe-a16a-fae6ccea4bd9\") " pod="crc-storage/crc-storage-crc-slnjh" Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.262478 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/92c92eb6-45b3-40fe-a16a-fae6ccea4bd9-crc-storage\") pod \"crc-storage-crc-slnjh\" (UID: \"92c92eb6-45b3-40fe-a16a-fae6ccea4bd9\") " pod="crc-storage/crc-storage-crc-slnjh" Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.285458 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vq9mg\" (UniqueName: \"kubernetes.io/projected/92c92eb6-45b3-40fe-a16a-fae6ccea4bd9-kube-api-access-vq9mg\") pod \"crc-storage-crc-slnjh\" (UID: \"92c92eb6-45b3-40fe-a16a-fae6ccea4bd9\") " pod="crc-storage/crc-storage-crc-slnjh" Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.436636 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-slnjh" Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.672951 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed" path="/var/lib/kubelet/pods/f04ab01c-290a-4e8f-ba7c-d4bf438ca9ed/volumes" Nov 25 10:44:39 crc kubenswrapper[4776]: I1125 10:44:39.890158 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-slnjh"] Nov 25 10:44:40 crc kubenswrapper[4776]: I1125 10:44:40.421328 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-slnjh" event={"ID":"92c92eb6-45b3-40fe-a16a-fae6ccea4bd9","Type":"ContainerStarted","Data":"ce68541cafb7bb6289aa3b59502c1fa50c272aad7b8a1d1b1df0b65e3cae30fe"} Nov 25 10:44:41 crc kubenswrapper[4776]: I1125 10:44:41.443846 4776 generic.go:334] "Generic (PLEG): container finished" podID="92c92eb6-45b3-40fe-a16a-fae6ccea4bd9" containerID="525ac307b87c2ac45fc15b62f0e9864a3c0ff4e0cb50536916647e40747636b2" exitCode=0 Nov 25 10:44:41 crc kubenswrapper[4776]: I1125 10:44:41.444214 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-slnjh" event={"ID":"92c92eb6-45b3-40fe-a16a-fae6ccea4bd9","Type":"ContainerDied","Data":"525ac307b87c2ac45fc15b62f0e9864a3c0ff4e0cb50536916647e40747636b2"} Nov 25 10:44:42 crc kubenswrapper[4776]: I1125 10:44:42.750978 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-slnjh" Nov 25 10:44:42 crc kubenswrapper[4776]: I1125 10:44:42.817466 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vq9mg\" (UniqueName: \"kubernetes.io/projected/92c92eb6-45b3-40fe-a16a-fae6ccea4bd9-kube-api-access-vq9mg\") pod \"92c92eb6-45b3-40fe-a16a-fae6ccea4bd9\" (UID: \"92c92eb6-45b3-40fe-a16a-fae6ccea4bd9\") " Nov 25 10:44:42 crc kubenswrapper[4776]: I1125 10:44:42.817624 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/92c92eb6-45b3-40fe-a16a-fae6ccea4bd9-crc-storage\") pod \"92c92eb6-45b3-40fe-a16a-fae6ccea4bd9\" (UID: \"92c92eb6-45b3-40fe-a16a-fae6ccea4bd9\") " Nov 25 10:44:42 crc kubenswrapper[4776]: I1125 10:44:42.817659 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/92c92eb6-45b3-40fe-a16a-fae6ccea4bd9-node-mnt\") pod \"92c92eb6-45b3-40fe-a16a-fae6ccea4bd9\" (UID: \"92c92eb6-45b3-40fe-a16a-fae6ccea4bd9\") " Nov 25 10:44:42 crc kubenswrapper[4776]: I1125 10:44:42.818082 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/92c92eb6-45b3-40fe-a16a-fae6ccea4bd9-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "92c92eb6-45b3-40fe-a16a-fae6ccea4bd9" (UID: "92c92eb6-45b3-40fe-a16a-fae6ccea4bd9"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:44:42 crc kubenswrapper[4776]: I1125 10:44:42.823748 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92c92eb6-45b3-40fe-a16a-fae6ccea4bd9-kube-api-access-vq9mg" (OuterVolumeSpecName: "kube-api-access-vq9mg") pod "92c92eb6-45b3-40fe-a16a-fae6ccea4bd9" (UID: "92c92eb6-45b3-40fe-a16a-fae6ccea4bd9"). InnerVolumeSpecName "kube-api-access-vq9mg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:44:42 crc kubenswrapper[4776]: I1125 10:44:42.836296 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92c92eb6-45b3-40fe-a16a-fae6ccea4bd9-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "92c92eb6-45b3-40fe-a16a-fae6ccea4bd9" (UID: "92c92eb6-45b3-40fe-a16a-fae6ccea4bd9"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:44:42 crc kubenswrapper[4776]: I1125 10:44:42.919204 4776 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/92c92eb6-45b3-40fe-a16a-fae6ccea4bd9-node-mnt\") on node \"crc\" DevicePath \"\"" Nov 25 10:44:42 crc kubenswrapper[4776]: I1125 10:44:42.919265 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vq9mg\" (UniqueName: \"kubernetes.io/projected/92c92eb6-45b3-40fe-a16a-fae6ccea4bd9-kube-api-access-vq9mg\") on node \"crc\" DevicePath \"\"" Nov 25 10:44:42 crc kubenswrapper[4776]: I1125 10:44:42.919279 4776 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/92c92eb6-45b3-40fe-a16a-fae6ccea4bd9-crc-storage\") on node \"crc\" DevicePath \"\"" Nov 25 10:44:43 crc kubenswrapper[4776]: I1125 10:44:43.465360 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-slnjh" event={"ID":"92c92eb6-45b3-40fe-a16a-fae6ccea4bd9","Type":"ContainerDied","Data":"ce68541cafb7bb6289aa3b59502c1fa50c272aad7b8a1d1b1df0b65e3cae30fe"} Nov 25 10:44:43 crc kubenswrapper[4776]: I1125 10:44:43.465397 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce68541cafb7bb6289aa3b59502c1fa50c272aad7b8a1d1b1df0b65e3cae30fe" Nov 25 10:44:43 crc kubenswrapper[4776]: I1125 10:44:43.465648 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-slnjh" Nov 25 10:44:45 crc kubenswrapper[4776]: I1125 10:44:45.033108 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-slnjh"] Nov 25 10:44:45 crc kubenswrapper[4776]: I1125 10:44:45.038110 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-slnjh"] Nov 25 10:44:45 crc kubenswrapper[4776]: I1125 10:44:45.152708 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-9rplj"] Nov 25 10:44:45 crc kubenswrapper[4776]: E1125 10:44:45.153309 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92c92eb6-45b3-40fe-a16a-fae6ccea4bd9" containerName="storage" Nov 25 10:44:45 crc kubenswrapper[4776]: I1125 10:44:45.153349 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="92c92eb6-45b3-40fe-a16a-fae6ccea4bd9" containerName="storage" Nov 25 10:44:45 crc kubenswrapper[4776]: I1125 10:44:45.153766 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="92c92eb6-45b3-40fe-a16a-fae6ccea4bd9" containerName="storage" Nov 25 10:44:45 crc kubenswrapper[4776]: I1125 10:44:45.154702 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-9rplj" Nov 25 10:44:45 crc kubenswrapper[4776]: I1125 10:44:45.159082 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Nov 25 10:44:45 crc kubenswrapper[4776]: I1125 10:44:45.159143 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Nov 25 10:44:45 crc kubenswrapper[4776]: I1125 10:44:45.159059 4776 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-cnl7m" Nov 25 10:44:45 crc kubenswrapper[4776]: I1125 10:44:45.159148 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Nov 25 10:44:45 crc kubenswrapper[4776]: I1125 10:44:45.164037 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-9rplj"] Nov 25 10:44:45 crc kubenswrapper[4776]: I1125 10:44:45.259701 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/dc2a6e2b-acfe-4977-ba59-cfaac4babfcd-node-mnt\") pod \"crc-storage-crc-9rplj\" (UID: \"dc2a6e2b-acfe-4977-ba59-cfaac4babfcd\") " pod="crc-storage/crc-storage-crc-9rplj" Nov 25 10:44:45 crc kubenswrapper[4776]: I1125 10:44:45.259760 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dxg6\" (UniqueName: \"kubernetes.io/projected/dc2a6e2b-acfe-4977-ba59-cfaac4babfcd-kube-api-access-9dxg6\") pod \"crc-storage-crc-9rplj\" (UID: \"dc2a6e2b-acfe-4977-ba59-cfaac4babfcd\") " pod="crc-storage/crc-storage-crc-9rplj" Nov 25 10:44:45 crc kubenswrapper[4776]: I1125 10:44:45.260118 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/dc2a6e2b-acfe-4977-ba59-cfaac4babfcd-crc-storage\") pod \"crc-storage-crc-9rplj\" (UID: \"dc2a6e2b-acfe-4977-ba59-cfaac4babfcd\") " pod="crc-storage/crc-storage-crc-9rplj" Nov 25 10:44:45 crc kubenswrapper[4776]: I1125 10:44:45.362112 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/dc2a6e2b-acfe-4977-ba59-cfaac4babfcd-node-mnt\") pod \"crc-storage-crc-9rplj\" (UID: \"dc2a6e2b-acfe-4977-ba59-cfaac4babfcd\") " pod="crc-storage/crc-storage-crc-9rplj" Nov 25 10:44:45 crc kubenswrapper[4776]: I1125 10:44:45.362207 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dxg6\" (UniqueName: \"kubernetes.io/projected/dc2a6e2b-acfe-4977-ba59-cfaac4babfcd-kube-api-access-9dxg6\") pod \"crc-storage-crc-9rplj\" (UID: \"dc2a6e2b-acfe-4977-ba59-cfaac4babfcd\") " pod="crc-storage/crc-storage-crc-9rplj" Nov 25 10:44:45 crc kubenswrapper[4776]: I1125 10:44:45.362281 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/dc2a6e2b-acfe-4977-ba59-cfaac4babfcd-crc-storage\") pod \"crc-storage-crc-9rplj\" (UID: \"dc2a6e2b-acfe-4977-ba59-cfaac4babfcd\") " pod="crc-storage/crc-storage-crc-9rplj" Nov 25 10:44:45 crc kubenswrapper[4776]: I1125 10:44:45.362525 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/dc2a6e2b-acfe-4977-ba59-cfaac4babfcd-node-mnt\") pod \"crc-storage-crc-9rplj\" (UID: \"dc2a6e2b-acfe-4977-ba59-cfaac4babfcd\") " pod="crc-storage/crc-storage-crc-9rplj" Nov 25 10:44:45 crc kubenswrapper[4776]: I1125 10:44:45.363264 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/dc2a6e2b-acfe-4977-ba59-cfaac4babfcd-crc-storage\") pod \"crc-storage-crc-9rplj\" (UID: \"dc2a6e2b-acfe-4977-ba59-cfaac4babfcd\") " pod="crc-storage/crc-storage-crc-9rplj" Nov 25 10:44:45 crc kubenswrapper[4776]: I1125 10:44:45.387207 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dxg6\" (UniqueName: \"kubernetes.io/projected/dc2a6e2b-acfe-4977-ba59-cfaac4babfcd-kube-api-access-9dxg6\") pod \"crc-storage-crc-9rplj\" (UID: \"dc2a6e2b-acfe-4977-ba59-cfaac4babfcd\") " pod="crc-storage/crc-storage-crc-9rplj" Nov 25 10:44:45 crc kubenswrapper[4776]: I1125 10:44:45.487595 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-9rplj" Nov 25 10:44:45 crc kubenswrapper[4776]: I1125 10:44:45.695315 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92c92eb6-45b3-40fe-a16a-fae6ccea4bd9" path="/var/lib/kubelet/pods/92c92eb6-45b3-40fe-a16a-fae6ccea4bd9/volumes" Nov 25 10:44:45 crc kubenswrapper[4776]: I1125 10:44:45.706761 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-9rplj"] Nov 25 10:44:46 crc kubenswrapper[4776]: I1125 10:44:46.491260 4776 generic.go:334] "Generic (PLEG): container finished" podID="dc2a6e2b-acfe-4977-ba59-cfaac4babfcd" containerID="d1f50f09a82196d5f44e56e093b26cc3f8f6f2904e7e2976cb235012628a4e19" exitCode=0 Nov 25 10:44:46 crc kubenswrapper[4776]: I1125 10:44:46.491344 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-9rplj" event={"ID":"dc2a6e2b-acfe-4977-ba59-cfaac4babfcd","Type":"ContainerDied","Data":"d1f50f09a82196d5f44e56e093b26cc3f8f6f2904e7e2976cb235012628a4e19"} Nov 25 10:44:46 crc kubenswrapper[4776]: I1125 10:44:46.491563 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-9rplj" event={"ID":"dc2a6e2b-acfe-4977-ba59-cfaac4babfcd","Type":"ContainerStarted","Data":"52004d696020907e4f2ccfb5fdf8c1623724f1180ebc28b4618d0de72148f0c3"} Nov 25 10:44:47 crc kubenswrapper[4776]: I1125 10:44:47.790532 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-9rplj" Nov 25 10:44:47 crc kubenswrapper[4776]: I1125 10:44:47.898920 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dxg6\" (UniqueName: \"kubernetes.io/projected/dc2a6e2b-acfe-4977-ba59-cfaac4babfcd-kube-api-access-9dxg6\") pod \"dc2a6e2b-acfe-4977-ba59-cfaac4babfcd\" (UID: \"dc2a6e2b-acfe-4977-ba59-cfaac4babfcd\") " Nov 25 10:44:47 crc kubenswrapper[4776]: I1125 10:44:47.899113 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/dc2a6e2b-acfe-4977-ba59-cfaac4babfcd-node-mnt\") pod \"dc2a6e2b-acfe-4977-ba59-cfaac4babfcd\" (UID: \"dc2a6e2b-acfe-4977-ba59-cfaac4babfcd\") " Nov 25 10:44:47 crc kubenswrapper[4776]: I1125 10:44:47.899147 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/dc2a6e2b-acfe-4977-ba59-cfaac4babfcd-crc-storage\") pod \"dc2a6e2b-acfe-4977-ba59-cfaac4babfcd\" (UID: \"dc2a6e2b-acfe-4977-ba59-cfaac4babfcd\") " Nov 25 10:44:47 crc kubenswrapper[4776]: I1125 10:44:47.899283 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dc2a6e2b-acfe-4977-ba59-cfaac4babfcd-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "dc2a6e2b-acfe-4977-ba59-cfaac4babfcd" (UID: "dc2a6e2b-acfe-4977-ba59-cfaac4babfcd"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:44:47 crc kubenswrapper[4776]: I1125 10:44:47.899860 4776 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/dc2a6e2b-acfe-4977-ba59-cfaac4babfcd-node-mnt\") on node \"crc\" DevicePath \"\"" Nov 25 10:44:47 crc kubenswrapper[4776]: I1125 10:44:47.904709 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc2a6e2b-acfe-4977-ba59-cfaac4babfcd-kube-api-access-9dxg6" (OuterVolumeSpecName: "kube-api-access-9dxg6") pod "dc2a6e2b-acfe-4977-ba59-cfaac4babfcd" (UID: "dc2a6e2b-acfe-4977-ba59-cfaac4babfcd"). InnerVolumeSpecName "kube-api-access-9dxg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:44:47 crc kubenswrapper[4776]: I1125 10:44:47.917765 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc2a6e2b-acfe-4977-ba59-cfaac4babfcd-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "dc2a6e2b-acfe-4977-ba59-cfaac4babfcd" (UID: "dc2a6e2b-acfe-4977-ba59-cfaac4babfcd"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:44:48 crc kubenswrapper[4776]: I1125 10:44:48.001192 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dxg6\" (UniqueName: \"kubernetes.io/projected/dc2a6e2b-acfe-4977-ba59-cfaac4babfcd-kube-api-access-9dxg6\") on node \"crc\" DevicePath \"\"" Nov 25 10:44:48 crc kubenswrapper[4776]: I1125 10:44:48.001310 4776 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/dc2a6e2b-acfe-4977-ba59-cfaac4babfcd-crc-storage\") on node \"crc\" DevicePath \"\"" Nov 25 10:44:48 crc kubenswrapper[4776]: I1125 10:44:48.508306 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-9rplj" event={"ID":"dc2a6e2b-acfe-4977-ba59-cfaac4babfcd","Type":"ContainerDied","Data":"52004d696020907e4f2ccfb5fdf8c1623724f1180ebc28b4618d0de72148f0c3"} Nov 25 10:44:48 crc kubenswrapper[4776]: I1125 10:44:48.508369 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52004d696020907e4f2ccfb5fdf8c1623724f1180ebc28b4618d0de72148f0c3" Nov 25 10:44:48 crc kubenswrapper[4776]: I1125 10:44:48.508377 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-9rplj" Nov 25 10:44:50 crc kubenswrapper[4776]: I1125 10:44:50.662689 4776 scope.go:117] "RemoveContainer" containerID="f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a" Nov 25 10:44:50 crc kubenswrapper[4776]: E1125 10:44:50.663311 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:45:00 crc kubenswrapper[4776]: I1125 10:45:00.155560 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401125-hhgzx"] Nov 25 10:45:00 crc kubenswrapper[4776]: E1125 10:45:00.156546 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc2a6e2b-acfe-4977-ba59-cfaac4babfcd" containerName="storage" Nov 25 10:45:00 crc kubenswrapper[4776]: I1125 10:45:00.156562 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc2a6e2b-acfe-4977-ba59-cfaac4babfcd" containerName="storage" Nov 25 10:45:00 crc kubenswrapper[4776]: I1125 10:45:00.156761 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc2a6e2b-acfe-4977-ba59-cfaac4babfcd" containerName="storage" Nov 25 10:45:00 crc kubenswrapper[4776]: I1125 10:45:00.157397 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hhgzx" Nov 25 10:45:00 crc kubenswrapper[4776]: I1125 10:45:00.161680 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 10:45:00 crc kubenswrapper[4776]: I1125 10:45:00.162080 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 10:45:00 crc kubenswrapper[4776]: I1125 10:45:00.165243 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401125-hhgzx"] Nov 25 10:45:00 crc kubenswrapper[4776]: I1125 10:45:00.285906 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4d6f9eeb-c801-495d-97e5-7c83d46e754a-secret-volume\") pod \"collect-profiles-29401125-hhgzx\" (UID: \"4d6f9eeb-c801-495d-97e5-7c83d46e754a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hhgzx" Nov 25 10:45:00 crc kubenswrapper[4776]: I1125 10:45:00.285966 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-499c2\" (UniqueName: \"kubernetes.io/projected/4d6f9eeb-c801-495d-97e5-7c83d46e754a-kube-api-access-499c2\") pod \"collect-profiles-29401125-hhgzx\" (UID: \"4d6f9eeb-c801-495d-97e5-7c83d46e754a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hhgzx" Nov 25 10:45:00 crc kubenswrapper[4776]: I1125 10:45:00.286139 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4d6f9eeb-c801-495d-97e5-7c83d46e754a-config-volume\") pod \"collect-profiles-29401125-hhgzx\" (UID: \"4d6f9eeb-c801-495d-97e5-7c83d46e754a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hhgzx" Nov 25 10:45:00 crc kubenswrapper[4776]: I1125 10:45:00.387671 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4d6f9eeb-c801-495d-97e5-7c83d46e754a-config-volume\") pod \"collect-profiles-29401125-hhgzx\" (UID: \"4d6f9eeb-c801-495d-97e5-7c83d46e754a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hhgzx" Nov 25 10:45:00 crc kubenswrapper[4776]: I1125 10:45:00.387740 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4d6f9eeb-c801-495d-97e5-7c83d46e754a-secret-volume\") pod \"collect-profiles-29401125-hhgzx\" (UID: \"4d6f9eeb-c801-495d-97e5-7c83d46e754a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hhgzx" Nov 25 10:45:00 crc kubenswrapper[4776]: I1125 10:45:00.387775 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-499c2\" (UniqueName: \"kubernetes.io/projected/4d6f9eeb-c801-495d-97e5-7c83d46e754a-kube-api-access-499c2\") pod \"collect-profiles-29401125-hhgzx\" (UID: \"4d6f9eeb-c801-495d-97e5-7c83d46e754a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hhgzx" Nov 25 10:45:00 crc kubenswrapper[4776]: I1125 10:45:00.388479 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4d6f9eeb-c801-495d-97e5-7c83d46e754a-config-volume\") pod \"collect-profiles-29401125-hhgzx\" (UID: \"4d6f9eeb-c801-495d-97e5-7c83d46e754a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hhgzx" Nov 25 10:45:00 crc kubenswrapper[4776]: I1125 10:45:00.397475 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4d6f9eeb-c801-495d-97e5-7c83d46e754a-secret-volume\") pod \"collect-profiles-29401125-hhgzx\" (UID: \"4d6f9eeb-c801-495d-97e5-7c83d46e754a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hhgzx" Nov 25 10:45:00 crc kubenswrapper[4776]: I1125 10:45:00.411740 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-499c2\" (UniqueName: \"kubernetes.io/projected/4d6f9eeb-c801-495d-97e5-7c83d46e754a-kube-api-access-499c2\") pod \"collect-profiles-29401125-hhgzx\" (UID: \"4d6f9eeb-c801-495d-97e5-7c83d46e754a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hhgzx" Nov 25 10:45:00 crc kubenswrapper[4776]: I1125 10:45:00.484446 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hhgzx" Nov 25 10:45:00 crc kubenswrapper[4776]: I1125 10:45:00.891093 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401125-hhgzx"] Nov 25 10:45:01 crc kubenswrapper[4776]: I1125 10:45:01.650400 4776 generic.go:334] "Generic (PLEG): container finished" podID="4d6f9eeb-c801-495d-97e5-7c83d46e754a" containerID="ef36b1748f5b0da835396b431acf96950ff611d61b980b26afd69b0e39ccff7a" exitCode=0 Nov 25 10:45:01 crc kubenswrapper[4776]: I1125 10:45:01.650464 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hhgzx" event={"ID":"4d6f9eeb-c801-495d-97e5-7c83d46e754a","Type":"ContainerDied","Data":"ef36b1748f5b0da835396b431acf96950ff611d61b980b26afd69b0e39ccff7a"} Nov 25 10:45:01 crc kubenswrapper[4776]: I1125 10:45:01.650675 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hhgzx" event={"ID":"4d6f9eeb-c801-495d-97e5-7c83d46e754a","Type":"ContainerStarted","Data":"e5eeec41ad123fad7ecde10edbbfd1328be6bd3989ce3778598e65e3e5adacd3"} Nov 25 10:45:01 crc kubenswrapper[4776]: I1125 10:45:01.663721 4776 scope.go:117] "RemoveContainer" containerID="f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a" Nov 25 10:45:01 crc kubenswrapper[4776]: E1125 10:45:01.663974 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:45:02 crc kubenswrapper[4776]: I1125 10:45:02.929815 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hhgzx" Nov 25 10:45:03 crc kubenswrapper[4776]: I1125 10:45:03.029428 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4d6f9eeb-c801-495d-97e5-7c83d46e754a-secret-volume\") pod \"4d6f9eeb-c801-495d-97e5-7c83d46e754a\" (UID: \"4d6f9eeb-c801-495d-97e5-7c83d46e754a\") " Nov 25 10:45:03 crc kubenswrapper[4776]: I1125 10:45:03.029499 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-499c2\" (UniqueName: \"kubernetes.io/projected/4d6f9eeb-c801-495d-97e5-7c83d46e754a-kube-api-access-499c2\") pod \"4d6f9eeb-c801-495d-97e5-7c83d46e754a\" (UID: \"4d6f9eeb-c801-495d-97e5-7c83d46e754a\") " Nov 25 10:45:03 crc kubenswrapper[4776]: I1125 10:45:03.029651 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4d6f9eeb-c801-495d-97e5-7c83d46e754a-config-volume\") pod \"4d6f9eeb-c801-495d-97e5-7c83d46e754a\" (UID: \"4d6f9eeb-c801-495d-97e5-7c83d46e754a\") " Nov 25 10:45:03 crc kubenswrapper[4776]: I1125 10:45:03.030461 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d6f9eeb-c801-495d-97e5-7c83d46e754a-config-volume" (OuterVolumeSpecName: "config-volume") pod "4d6f9eeb-c801-495d-97e5-7c83d46e754a" (UID: "4d6f9eeb-c801-495d-97e5-7c83d46e754a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:45:03 crc kubenswrapper[4776]: I1125 10:45:03.034319 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d6f9eeb-c801-495d-97e5-7c83d46e754a-kube-api-access-499c2" (OuterVolumeSpecName: "kube-api-access-499c2") pod "4d6f9eeb-c801-495d-97e5-7c83d46e754a" (UID: "4d6f9eeb-c801-495d-97e5-7c83d46e754a"). InnerVolumeSpecName "kube-api-access-499c2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:45:03 crc kubenswrapper[4776]: I1125 10:45:03.034318 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d6f9eeb-c801-495d-97e5-7c83d46e754a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4d6f9eeb-c801-495d-97e5-7c83d46e754a" (UID: "4d6f9eeb-c801-495d-97e5-7c83d46e754a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:45:03 crc kubenswrapper[4776]: I1125 10:45:03.130784 4776 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4d6f9eeb-c801-495d-97e5-7c83d46e754a-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 10:45:03 crc kubenswrapper[4776]: I1125 10:45:03.130816 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-499c2\" (UniqueName: \"kubernetes.io/projected/4d6f9eeb-c801-495d-97e5-7c83d46e754a-kube-api-access-499c2\") on node \"crc\" DevicePath \"\"" Nov 25 10:45:03 crc kubenswrapper[4776]: I1125 10:45:03.130826 4776 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4d6f9eeb-c801-495d-97e5-7c83d46e754a-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 10:45:03 crc kubenswrapper[4776]: I1125 10:45:03.667533 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hhgzx" Nov 25 10:45:03 crc kubenswrapper[4776]: I1125 10:45:03.678912 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401125-hhgzx" event={"ID":"4d6f9eeb-c801-495d-97e5-7c83d46e754a","Type":"ContainerDied","Data":"e5eeec41ad123fad7ecde10edbbfd1328be6bd3989ce3778598e65e3e5adacd3"} Nov 25 10:45:03 crc kubenswrapper[4776]: I1125 10:45:03.678973 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5eeec41ad123fad7ecde10edbbfd1328be6bd3989ce3778598e65e3e5adacd3" Nov 25 10:45:04 crc kubenswrapper[4776]: I1125 10:45:04.003346 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401080-m5xwz"] Nov 25 10:45:04 crc kubenswrapper[4776]: I1125 10:45:04.008547 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401080-m5xwz"] Nov 25 10:45:05 crc kubenswrapper[4776]: I1125 10:45:05.672408 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad146178-73be-4f2b-a2d1-87905a70e2f0" path="/var/lib/kubelet/pods/ad146178-73be-4f2b-a2d1-87905a70e2f0/volumes" Nov 25 10:45:12 crc kubenswrapper[4776]: I1125 10:45:12.662886 4776 scope.go:117] "RemoveContainer" containerID="f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a" Nov 25 10:45:12 crc kubenswrapper[4776]: E1125 10:45:12.663543 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:45:27 crc kubenswrapper[4776]: I1125 10:45:27.663109 4776 scope.go:117] "RemoveContainer" containerID="f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a" Nov 25 10:45:27 crc kubenswrapper[4776]: I1125 10:45:27.873489 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"62f6bd2b248ceb776fb1c929efd14083d136607e3bcf021f3f4b235fbcacf0e1"} Nov 25 10:45:31 crc kubenswrapper[4776]: I1125 10:45:31.298307 4776 scope.go:117] "RemoveContainer" containerID="0bca6dcd4ef9e2845f88814b62bcb81d949aab1925dddb1c40352bbd094805bc" Nov 25 10:45:31 crc kubenswrapper[4776]: I1125 10:45:31.356941 4776 scope.go:117] "RemoveContainer" containerID="5976e99e2d279f9b29baf268695259602ba8e471cdba2c32bd0654801018f07d" Nov 25 10:45:57 crc kubenswrapper[4776]: I1125 10:45:57.815951 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wdmpv"] Nov 25 10:45:57 crc kubenswrapper[4776]: E1125 10:45:57.819858 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d6f9eeb-c801-495d-97e5-7c83d46e754a" containerName="collect-profiles" Nov 25 10:45:57 crc kubenswrapper[4776]: I1125 10:45:57.820147 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d6f9eeb-c801-495d-97e5-7c83d46e754a" containerName="collect-profiles" Nov 25 10:45:57 crc kubenswrapper[4776]: I1125 10:45:57.820774 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d6f9eeb-c801-495d-97e5-7c83d46e754a" containerName="collect-profiles" Nov 25 10:45:57 crc kubenswrapper[4776]: I1125 10:45:57.823372 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wdmpv" Nov 25 10:45:57 crc kubenswrapper[4776]: I1125 10:45:57.841502 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5ecae99-1e2c-475f-9857-ba3caac5b8f9-catalog-content\") pod \"redhat-operators-wdmpv\" (UID: \"f5ecae99-1e2c-475f-9857-ba3caac5b8f9\") " pod="openshift-marketplace/redhat-operators-wdmpv" Nov 25 10:45:57 crc kubenswrapper[4776]: I1125 10:45:57.841573 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t89z9\" (UniqueName: \"kubernetes.io/projected/f5ecae99-1e2c-475f-9857-ba3caac5b8f9-kube-api-access-t89z9\") pod \"redhat-operators-wdmpv\" (UID: \"f5ecae99-1e2c-475f-9857-ba3caac5b8f9\") " pod="openshift-marketplace/redhat-operators-wdmpv" Nov 25 10:45:57 crc kubenswrapper[4776]: I1125 10:45:57.841743 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5ecae99-1e2c-475f-9857-ba3caac5b8f9-utilities\") pod \"redhat-operators-wdmpv\" (UID: \"f5ecae99-1e2c-475f-9857-ba3caac5b8f9\") " pod="openshift-marketplace/redhat-operators-wdmpv" Nov 25 10:45:57 crc kubenswrapper[4776]: I1125 10:45:57.854263 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wdmpv"] Nov 25 10:45:57 crc kubenswrapper[4776]: I1125 10:45:57.943205 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5ecae99-1e2c-475f-9857-ba3caac5b8f9-catalog-content\") pod \"redhat-operators-wdmpv\" (UID: \"f5ecae99-1e2c-475f-9857-ba3caac5b8f9\") " pod="openshift-marketplace/redhat-operators-wdmpv" Nov 25 10:45:57 crc kubenswrapper[4776]: I1125 10:45:57.943301 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t89z9\" (UniqueName: \"kubernetes.io/projected/f5ecae99-1e2c-475f-9857-ba3caac5b8f9-kube-api-access-t89z9\") pod \"redhat-operators-wdmpv\" (UID: \"f5ecae99-1e2c-475f-9857-ba3caac5b8f9\") " pod="openshift-marketplace/redhat-operators-wdmpv" Nov 25 10:45:57 crc kubenswrapper[4776]: I1125 10:45:57.943368 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5ecae99-1e2c-475f-9857-ba3caac5b8f9-utilities\") pod \"redhat-operators-wdmpv\" (UID: \"f5ecae99-1e2c-475f-9857-ba3caac5b8f9\") " pod="openshift-marketplace/redhat-operators-wdmpv" Nov 25 10:45:57 crc kubenswrapper[4776]: I1125 10:45:57.944335 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5ecae99-1e2c-475f-9857-ba3caac5b8f9-utilities\") pod \"redhat-operators-wdmpv\" (UID: \"f5ecae99-1e2c-475f-9857-ba3caac5b8f9\") " pod="openshift-marketplace/redhat-operators-wdmpv" Nov 25 10:45:57 crc kubenswrapper[4776]: I1125 10:45:57.944852 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5ecae99-1e2c-475f-9857-ba3caac5b8f9-catalog-content\") pod \"redhat-operators-wdmpv\" (UID: \"f5ecae99-1e2c-475f-9857-ba3caac5b8f9\") " pod="openshift-marketplace/redhat-operators-wdmpv" Nov 25 10:45:57 crc kubenswrapper[4776]: I1125 10:45:57.964246 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t89z9\" (UniqueName: \"kubernetes.io/projected/f5ecae99-1e2c-475f-9857-ba3caac5b8f9-kube-api-access-t89z9\") pod \"redhat-operators-wdmpv\" (UID: \"f5ecae99-1e2c-475f-9857-ba3caac5b8f9\") " pod="openshift-marketplace/redhat-operators-wdmpv" Nov 25 10:45:58 crc kubenswrapper[4776]: I1125 10:45:58.151444 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wdmpv" Nov 25 10:45:58 crc kubenswrapper[4776]: I1125 10:45:58.620707 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wdmpv"] Nov 25 10:45:59 crc kubenswrapper[4776]: I1125 10:45:59.123388 4776 generic.go:334] "Generic (PLEG): container finished" podID="f5ecae99-1e2c-475f-9857-ba3caac5b8f9" containerID="2f8ac631d4107dce658afc9b77852924f03cde9dfc12b1961f9e9ecebf766baf" exitCode=0 Nov 25 10:45:59 crc kubenswrapper[4776]: I1125 10:45:59.123445 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wdmpv" event={"ID":"f5ecae99-1e2c-475f-9857-ba3caac5b8f9","Type":"ContainerDied","Data":"2f8ac631d4107dce658afc9b77852924f03cde9dfc12b1961f9e9ecebf766baf"} Nov 25 10:45:59 crc kubenswrapper[4776]: I1125 10:45:59.123761 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wdmpv" event={"ID":"f5ecae99-1e2c-475f-9857-ba3caac5b8f9","Type":"ContainerStarted","Data":"7bb0c3496fd353e7bfd3a917c07bb2de11c152979d4b722162141d9587e7f330"} Nov 25 10:46:00 crc kubenswrapper[4776]: I1125 10:46:00.131611 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wdmpv" event={"ID":"f5ecae99-1e2c-475f-9857-ba3caac5b8f9","Type":"ContainerStarted","Data":"58a06d27d5ba20f1bf0bcae3ad21d9810c1edfe975b7b6c253464658e19d392d"} Nov 25 10:46:01 crc kubenswrapper[4776]: I1125 10:46:01.142829 4776 generic.go:334] "Generic (PLEG): container finished" podID="f5ecae99-1e2c-475f-9857-ba3caac5b8f9" containerID="58a06d27d5ba20f1bf0bcae3ad21d9810c1edfe975b7b6c253464658e19d392d" exitCode=0 Nov 25 10:46:01 crc kubenswrapper[4776]: I1125 10:46:01.142931 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wdmpv" event={"ID":"f5ecae99-1e2c-475f-9857-ba3caac5b8f9","Type":"ContainerDied","Data":"58a06d27d5ba20f1bf0bcae3ad21d9810c1edfe975b7b6c253464658e19d392d"} Nov 25 10:46:02 crc kubenswrapper[4776]: I1125 10:46:02.153733 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wdmpv" event={"ID":"f5ecae99-1e2c-475f-9857-ba3caac5b8f9","Type":"ContainerStarted","Data":"fb6004912bbc17234b8102cc6664345028d776b6982b2a1732fd6d37dda22438"} Nov 25 10:46:02 crc kubenswrapper[4776]: I1125 10:46:02.177480 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wdmpv" podStartSLOduration=2.736737949 podStartE2EDuration="5.177461602s" podCreationTimestamp="2025-11-25 10:45:57 +0000 UTC" firstStartedPulling="2025-11-25 10:45:59.125255811 +0000 UTC m=+4904.166315364" lastFinishedPulling="2025-11-25 10:46:01.565979464 +0000 UTC m=+4906.607039017" observedRunningTime="2025-11-25 10:46:02.173226706 +0000 UTC m=+4907.214286259" watchObservedRunningTime="2025-11-25 10:46:02.177461602 +0000 UTC m=+4907.218521165" Nov 25 10:46:08 crc kubenswrapper[4776]: I1125 10:46:08.152235 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wdmpv" Nov 25 10:46:08 crc kubenswrapper[4776]: I1125 10:46:08.152614 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wdmpv" Nov 25 10:46:08 crc kubenswrapper[4776]: I1125 10:46:08.201209 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wdmpv" Nov 25 10:46:08 crc kubenswrapper[4776]: I1125 10:46:08.246355 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wdmpv" Nov 25 10:46:08 crc kubenswrapper[4776]: I1125 10:46:08.435096 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wdmpv"] Nov 25 10:46:10 crc kubenswrapper[4776]: I1125 10:46:10.216603 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wdmpv" podUID="f5ecae99-1e2c-475f-9857-ba3caac5b8f9" containerName="registry-server" containerID="cri-o://fb6004912bbc17234b8102cc6664345028d776b6982b2a1732fd6d37dda22438" gracePeriod=2 Nov 25 10:46:11 crc kubenswrapper[4776]: I1125 10:46:11.718353 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wdmpv" Nov 25 10:46:11 crc kubenswrapper[4776]: I1125 10:46:11.848734 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5ecae99-1e2c-475f-9857-ba3caac5b8f9-catalog-content\") pod \"f5ecae99-1e2c-475f-9857-ba3caac5b8f9\" (UID: \"f5ecae99-1e2c-475f-9857-ba3caac5b8f9\") " Nov 25 10:46:11 crc kubenswrapper[4776]: I1125 10:46:11.848795 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t89z9\" (UniqueName: \"kubernetes.io/projected/f5ecae99-1e2c-475f-9857-ba3caac5b8f9-kube-api-access-t89z9\") pod \"f5ecae99-1e2c-475f-9857-ba3caac5b8f9\" (UID: \"f5ecae99-1e2c-475f-9857-ba3caac5b8f9\") " Nov 25 10:46:11 crc kubenswrapper[4776]: I1125 10:46:11.848936 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5ecae99-1e2c-475f-9857-ba3caac5b8f9-utilities\") pod \"f5ecae99-1e2c-475f-9857-ba3caac5b8f9\" (UID: \"f5ecae99-1e2c-475f-9857-ba3caac5b8f9\") " Nov 25 10:46:11 crc kubenswrapper[4776]: I1125 10:46:11.851176 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5ecae99-1e2c-475f-9857-ba3caac5b8f9-utilities" (OuterVolumeSpecName: "utilities") pod "f5ecae99-1e2c-475f-9857-ba3caac5b8f9" (UID: "f5ecae99-1e2c-475f-9857-ba3caac5b8f9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:46:11 crc kubenswrapper[4776]: I1125 10:46:11.855414 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5ecae99-1e2c-475f-9857-ba3caac5b8f9-kube-api-access-t89z9" (OuterVolumeSpecName: "kube-api-access-t89z9") pod "f5ecae99-1e2c-475f-9857-ba3caac5b8f9" (UID: "f5ecae99-1e2c-475f-9857-ba3caac5b8f9"). InnerVolumeSpecName "kube-api-access-t89z9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:46:11 crc kubenswrapper[4776]: I1125 10:46:11.943818 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5ecae99-1e2c-475f-9857-ba3caac5b8f9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f5ecae99-1e2c-475f-9857-ba3caac5b8f9" (UID: "f5ecae99-1e2c-475f-9857-ba3caac5b8f9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:46:11 crc kubenswrapper[4776]: I1125 10:46:11.952861 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5ecae99-1e2c-475f-9857-ba3caac5b8f9-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:11 crc kubenswrapper[4776]: I1125 10:46:11.952916 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5ecae99-1e2c-475f-9857-ba3caac5b8f9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:11 crc kubenswrapper[4776]: I1125 10:46:11.952939 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t89z9\" (UniqueName: \"kubernetes.io/projected/f5ecae99-1e2c-475f-9857-ba3caac5b8f9-kube-api-access-t89z9\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:12 crc kubenswrapper[4776]: I1125 10:46:12.243486 4776 generic.go:334] "Generic (PLEG): container finished" podID="f5ecae99-1e2c-475f-9857-ba3caac5b8f9" containerID="fb6004912bbc17234b8102cc6664345028d776b6982b2a1732fd6d37dda22438" exitCode=0 Nov 25 10:46:12 crc kubenswrapper[4776]: I1125 10:46:12.243553 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wdmpv" event={"ID":"f5ecae99-1e2c-475f-9857-ba3caac5b8f9","Type":"ContainerDied","Data":"fb6004912bbc17234b8102cc6664345028d776b6982b2a1732fd6d37dda22438"} Nov 25 10:46:12 crc kubenswrapper[4776]: I1125 10:46:12.243595 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wdmpv" event={"ID":"f5ecae99-1e2c-475f-9857-ba3caac5b8f9","Type":"ContainerDied","Data":"7bb0c3496fd353e7bfd3a917c07bb2de11c152979d4b722162141d9587e7f330"} Nov 25 10:46:12 crc kubenswrapper[4776]: I1125 10:46:12.243628 4776 scope.go:117] "RemoveContainer" containerID="fb6004912bbc17234b8102cc6664345028d776b6982b2a1732fd6d37dda22438" Nov 25 10:46:12 crc kubenswrapper[4776]: I1125 10:46:12.243850 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wdmpv" Nov 25 10:46:12 crc kubenswrapper[4776]: I1125 10:46:12.271733 4776 scope.go:117] "RemoveContainer" containerID="58a06d27d5ba20f1bf0bcae3ad21d9810c1edfe975b7b6c253464658e19d392d" Nov 25 10:46:12 crc kubenswrapper[4776]: I1125 10:46:12.302084 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wdmpv"] Nov 25 10:46:12 crc kubenswrapper[4776]: I1125 10:46:12.303968 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wdmpv"] Nov 25 10:46:12 crc kubenswrapper[4776]: I1125 10:46:12.316324 4776 scope.go:117] "RemoveContainer" containerID="2f8ac631d4107dce658afc9b77852924f03cde9dfc12b1961f9e9ecebf766baf" Nov 25 10:46:12 crc kubenswrapper[4776]: I1125 10:46:12.335511 4776 scope.go:117] "RemoveContainer" containerID="fb6004912bbc17234b8102cc6664345028d776b6982b2a1732fd6d37dda22438" Nov 25 10:46:12 crc kubenswrapper[4776]: E1125 10:46:12.335998 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb6004912bbc17234b8102cc6664345028d776b6982b2a1732fd6d37dda22438\": container with ID starting with fb6004912bbc17234b8102cc6664345028d776b6982b2a1732fd6d37dda22438 not found: ID does not exist" containerID="fb6004912bbc17234b8102cc6664345028d776b6982b2a1732fd6d37dda22438" Nov 25 10:46:12 crc kubenswrapper[4776]: I1125 10:46:12.336056 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb6004912bbc17234b8102cc6664345028d776b6982b2a1732fd6d37dda22438"} err="failed to get container status \"fb6004912bbc17234b8102cc6664345028d776b6982b2a1732fd6d37dda22438\": rpc error: code = NotFound desc = could not find container \"fb6004912bbc17234b8102cc6664345028d776b6982b2a1732fd6d37dda22438\": container with ID starting with fb6004912bbc17234b8102cc6664345028d776b6982b2a1732fd6d37dda22438 not found: ID does not exist" Nov 25 10:46:12 crc kubenswrapper[4776]: I1125 10:46:12.336105 4776 scope.go:117] "RemoveContainer" containerID="58a06d27d5ba20f1bf0bcae3ad21d9810c1edfe975b7b6c253464658e19d392d" Nov 25 10:46:12 crc kubenswrapper[4776]: E1125 10:46:12.336390 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58a06d27d5ba20f1bf0bcae3ad21d9810c1edfe975b7b6c253464658e19d392d\": container with ID starting with 58a06d27d5ba20f1bf0bcae3ad21d9810c1edfe975b7b6c253464658e19d392d not found: ID does not exist" containerID="58a06d27d5ba20f1bf0bcae3ad21d9810c1edfe975b7b6c253464658e19d392d" Nov 25 10:46:12 crc kubenswrapper[4776]: I1125 10:46:12.336419 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58a06d27d5ba20f1bf0bcae3ad21d9810c1edfe975b7b6c253464658e19d392d"} err="failed to get container status \"58a06d27d5ba20f1bf0bcae3ad21d9810c1edfe975b7b6c253464658e19d392d\": rpc error: code = NotFound desc = could not find container \"58a06d27d5ba20f1bf0bcae3ad21d9810c1edfe975b7b6c253464658e19d392d\": container with ID starting with 58a06d27d5ba20f1bf0bcae3ad21d9810c1edfe975b7b6c253464658e19d392d not found: ID does not exist" Nov 25 10:46:12 crc kubenswrapper[4776]: I1125 10:46:12.336438 4776 scope.go:117] "RemoveContainer" containerID="2f8ac631d4107dce658afc9b77852924f03cde9dfc12b1961f9e9ecebf766baf" Nov 25 10:46:12 crc kubenswrapper[4776]: E1125 10:46:12.336648 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f8ac631d4107dce658afc9b77852924f03cde9dfc12b1961f9e9ecebf766baf\": container with ID starting with 2f8ac631d4107dce658afc9b77852924f03cde9dfc12b1961f9e9ecebf766baf not found: ID does not exist" containerID="2f8ac631d4107dce658afc9b77852924f03cde9dfc12b1961f9e9ecebf766baf" Nov 25 10:46:12 crc kubenswrapper[4776]: I1125 10:46:12.336672 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f8ac631d4107dce658afc9b77852924f03cde9dfc12b1961f9e9ecebf766baf"} err="failed to get container status \"2f8ac631d4107dce658afc9b77852924f03cde9dfc12b1961f9e9ecebf766baf\": rpc error: code = NotFound desc = could not find container \"2f8ac631d4107dce658afc9b77852924f03cde9dfc12b1961f9e9ecebf766baf\": container with ID starting with 2f8ac631d4107dce658afc9b77852924f03cde9dfc12b1961f9e9ecebf766baf not found: ID does not exist" Nov 25 10:46:13 crc kubenswrapper[4776]: I1125 10:46:13.676446 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5ecae99-1e2c-475f-9857-ba3caac5b8f9" path="/var/lib/kubelet/pods/f5ecae99-1e2c-475f-9857-ba3caac5b8f9/volumes" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.580417 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-866449bdb9-psppk"] Nov 25 10:46:49 crc kubenswrapper[4776]: E1125 10:46:49.582559 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5ecae99-1e2c-475f-9857-ba3caac5b8f9" containerName="registry-server" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.585843 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5ecae99-1e2c-475f-9857-ba3caac5b8f9" containerName="registry-server" Nov 25 10:46:49 crc kubenswrapper[4776]: E1125 10:46:49.586010 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5ecae99-1e2c-475f-9857-ba3caac5b8f9" containerName="extract-content" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.586112 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5ecae99-1e2c-475f-9857-ba3caac5b8f9" containerName="extract-content" Nov 25 10:46:49 crc kubenswrapper[4776]: E1125 10:46:49.586231 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5ecae99-1e2c-475f-9857-ba3caac5b8f9" containerName="extract-utilities" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.586321 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5ecae99-1e2c-475f-9857-ba3caac5b8f9" containerName="extract-utilities" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.586782 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5ecae99-1e2c-475f-9857-ba3caac5b8f9" containerName="registry-server" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.587802 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-866449bdb9-psppk" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.591234 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.591248 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.591281 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.591392 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-rhlr7" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.594221 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-866449bdb9-psppk"] Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.626941 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55c86457d7-z64jx"] Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.628347 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55c86457d7-z64jx" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.631016 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.643916 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55c86457d7-z64jx"] Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.790273 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ca04eb1-c148-4a93-bf34-fcf325fcd09c-config\") pod \"dnsmasq-dns-866449bdb9-psppk\" (UID: \"8ca04eb1-c148-4a93-bf34-fcf325fcd09c\") " pod="openstack/dnsmasq-dns-866449bdb9-psppk" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.790370 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01e82c9a-afc8-477c-a2c0-faf51a20fc6c-dns-svc\") pod \"dnsmasq-dns-55c86457d7-z64jx\" (UID: \"01e82c9a-afc8-477c-a2c0-faf51a20fc6c\") " pod="openstack/dnsmasq-dns-55c86457d7-z64jx" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.790431 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01e82c9a-afc8-477c-a2c0-faf51a20fc6c-config\") pod \"dnsmasq-dns-55c86457d7-z64jx\" (UID: \"01e82c9a-afc8-477c-a2c0-faf51a20fc6c\") " pod="openstack/dnsmasq-dns-55c86457d7-z64jx" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.790501 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fqfq\" (UniqueName: \"kubernetes.io/projected/8ca04eb1-c148-4a93-bf34-fcf325fcd09c-kube-api-access-9fqfq\") pod \"dnsmasq-dns-866449bdb9-psppk\" (UID: \"8ca04eb1-c148-4a93-bf34-fcf325fcd09c\") " pod="openstack/dnsmasq-dns-866449bdb9-psppk" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.790526 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgpv5\" (UniqueName: \"kubernetes.io/projected/01e82c9a-afc8-477c-a2c0-faf51a20fc6c-kube-api-access-pgpv5\") pod \"dnsmasq-dns-55c86457d7-z64jx\" (UID: \"01e82c9a-afc8-477c-a2c0-faf51a20fc6c\") " pod="openstack/dnsmasq-dns-55c86457d7-z64jx" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.868082 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55c86457d7-z64jx"] Nov 25 10:46:49 crc kubenswrapper[4776]: E1125 10:46:49.868640 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-pgpv5], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-55c86457d7-z64jx" podUID="01e82c9a-afc8-477c-a2c0-faf51a20fc6c" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.892367 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01e82c9a-afc8-477c-a2c0-faf51a20fc6c-config\") pod \"dnsmasq-dns-55c86457d7-z64jx\" (UID: \"01e82c9a-afc8-477c-a2c0-faf51a20fc6c\") " pod="openstack/dnsmasq-dns-55c86457d7-z64jx" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.892436 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fqfq\" (UniqueName: \"kubernetes.io/projected/8ca04eb1-c148-4a93-bf34-fcf325fcd09c-kube-api-access-9fqfq\") pod \"dnsmasq-dns-866449bdb9-psppk\" (UID: \"8ca04eb1-c148-4a93-bf34-fcf325fcd09c\") " pod="openstack/dnsmasq-dns-866449bdb9-psppk" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.892459 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgpv5\" (UniqueName: \"kubernetes.io/projected/01e82c9a-afc8-477c-a2c0-faf51a20fc6c-kube-api-access-pgpv5\") pod \"dnsmasq-dns-55c86457d7-z64jx\" (UID: \"01e82c9a-afc8-477c-a2c0-faf51a20fc6c\") " pod="openstack/dnsmasq-dns-55c86457d7-z64jx" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.892504 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ca04eb1-c148-4a93-bf34-fcf325fcd09c-config\") pod \"dnsmasq-dns-866449bdb9-psppk\" (UID: \"8ca04eb1-c148-4a93-bf34-fcf325fcd09c\") " pod="openstack/dnsmasq-dns-866449bdb9-psppk" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.892538 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01e82c9a-afc8-477c-a2c0-faf51a20fc6c-dns-svc\") pod \"dnsmasq-dns-55c86457d7-z64jx\" (UID: \"01e82c9a-afc8-477c-a2c0-faf51a20fc6c\") " pod="openstack/dnsmasq-dns-55c86457d7-z64jx" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.893322 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01e82c9a-afc8-477c-a2c0-faf51a20fc6c-config\") pod \"dnsmasq-dns-55c86457d7-z64jx\" (UID: \"01e82c9a-afc8-477c-a2c0-faf51a20fc6c\") " pod="openstack/dnsmasq-dns-55c86457d7-z64jx" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.893421 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01e82c9a-afc8-477c-a2c0-faf51a20fc6c-dns-svc\") pod \"dnsmasq-dns-55c86457d7-z64jx\" (UID: \"01e82c9a-afc8-477c-a2c0-faf51a20fc6c\") " pod="openstack/dnsmasq-dns-55c86457d7-z64jx" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.893950 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ca04eb1-c148-4a93-bf34-fcf325fcd09c-config\") pod \"dnsmasq-dns-866449bdb9-psppk\" (UID: \"8ca04eb1-c148-4a93-bf34-fcf325fcd09c\") " pod="openstack/dnsmasq-dns-866449bdb9-psppk" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.899267 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f4c6c447c-7w4tr"] Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.909182 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f4c6c447c-7w4tr" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.914513 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f4c6c447c-7w4tr"] Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.936097 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fqfq\" (UniqueName: \"kubernetes.io/projected/8ca04eb1-c148-4a93-bf34-fcf325fcd09c-kube-api-access-9fqfq\") pod \"dnsmasq-dns-866449bdb9-psppk\" (UID: \"8ca04eb1-c148-4a93-bf34-fcf325fcd09c\") " pod="openstack/dnsmasq-dns-866449bdb9-psppk" Nov 25 10:46:49 crc kubenswrapper[4776]: I1125 10:46:49.940784 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgpv5\" (UniqueName: \"kubernetes.io/projected/01e82c9a-afc8-477c-a2c0-faf51a20fc6c-kube-api-access-pgpv5\") pod \"dnsmasq-dns-55c86457d7-z64jx\" (UID: \"01e82c9a-afc8-477c-a2c0-faf51a20fc6c\") " pod="openstack/dnsmasq-dns-55c86457d7-z64jx" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.094918 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6933bcb-92b8-42b6-8151-dd628024188e-dns-svc\") pod \"dnsmasq-dns-f4c6c447c-7w4tr\" (UID: \"a6933bcb-92b8-42b6-8151-dd628024188e\") " pod="openstack/dnsmasq-dns-f4c6c447c-7w4tr" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.095124 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6933bcb-92b8-42b6-8151-dd628024188e-config\") pod \"dnsmasq-dns-f4c6c447c-7w4tr\" (UID: \"a6933bcb-92b8-42b6-8151-dd628024188e\") " pod="openstack/dnsmasq-dns-f4c6c447c-7w4tr" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.095224 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j2vx\" (UniqueName: \"kubernetes.io/projected/a6933bcb-92b8-42b6-8151-dd628024188e-kube-api-access-2j2vx\") pod \"dnsmasq-dns-f4c6c447c-7w4tr\" (UID: \"a6933bcb-92b8-42b6-8151-dd628024188e\") " pod="openstack/dnsmasq-dns-f4c6c447c-7w4tr" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.196563 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j2vx\" (UniqueName: \"kubernetes.io/projected/a6933bcb-92b8-42b6-8151-dd628024188e-kube-api-access-2j2vx\") pod \"dnsmasq-dns-f4c6c447c-7w4tr\" (UID: \"a6933bcb-92b8-42b6-8151-dd628024188e\") " pod="openstack/dnsmasq-dns-f4c6c447c-7w4tr" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.196643 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6933bcb-92b8-42b6-8151-dd628024188e-dns-svc\") pod \"dnsmasq-dns-f4c6c447c-7w4tr\" (UID: \"a6933bcb-92b8-42b6-8151-dd628024188e\") " pod="openstack/dnsmasq-dns-f4c6c447c-7w4tr" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.196724 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6933bcb-92b8-42b6-8151-dd628024188e-config\") pod \"dnsmasq-dns-f4c6c447c-7w4tr\" (UID: \"a6933bcb-92b8-42b6-8151-dd628024188e\") " pod="openstack/dnsmasq-dns-f4c6c447c-7w4tr" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.197547 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6933bcb-92b8-42b6-8151-dd628024188e-config\") pod \"dnsmasq-dns-f4c6c447c-7w4tr\" (UID: \"a6933bcb-92b8-42b6-8151-dd628024188e\") " pod="openstack/dnsmasq-dns-f4c6c447c-7w4tr" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.197548 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6933bcb-92b8-42b6-8151-dd628024188e-dns-svc\") pod \"dnsmasq-dns-f4c6c447c-7w4tr\" (UID: \"a6933bcb-92b8-42b6-8151-dd628024188e\") " pod="openstack/dnsmasq-dns-f4c6c447c-7w4tr" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.208781 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-866449bdb9-psppk"] Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.209356 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-866449bdb9-psppk" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.238118 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j2vx\" (UniqueName: \"kubernetes.io/projected/a6933bcb-92b8-42b6-8151-dd628024188e-kube-api-access-2j2vx\") pod \"dnsmasq-dns-f4c6c447c-7w4tr\" (UID: \"a6933bcb-92b8-42b6-8151-dd628024188e\") " pod="openstack/dnsmasq-dns-f4c6c447c-7w4tr" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.263640 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59c6c64b5c-dd86r"] Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.265491 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59c6c64b5c-dd86r" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.273377 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59c6c64b5c-dd86r"] Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.275319 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f4c6c447c-7w4tr" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.299057 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bdc8091-6682-49da-a5dd-743dc013934d-config\") pod \"dnsmasq-dns-59c6c64b5c-dd86r\" (UID: \"0bdc8091-6682-49da-a5dd-743dc013934d\") " pod="openstack/dnsmasq-dns-59c6c64b5c-dd86r" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.299225 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsngw\" (UniqueName: \"kubernetes.io/projected/0bdc8091-6682-49da-a5dd-743dc013934d-kube-api-access-xsngw\") pod \"dnsmasq-dns-59c6c64b5c-dd86r\" (UID: \"0bdc8091-6682-49da-a5dd-743dc013934d\") " pod="openstack/dnsmasq-dns-59c6c64b5c-dd86r" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.299283 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bdc8091-6682-49da-a5dd-743dc013934d-dns-svc\") pod \"dnsmasq-dns-59c6c64b5c-dd86r\" (UID: \"0bdc8091-6682-49da-a5dd-743dc013934d\") " pod="openstack/dnsmasq-dns-59c6c64b5c-dd86r" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.400965 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bdc8091-6682-49da-a5dd-743dc013934d-config\") pod \"dnsmasq-dns-59c6c64b5c-dd86r\" (UID: \"0bdc8091-6682-49da-a5dd-743dc013934d\") " pod="openstack/dnsmasq-dns-59c6c64b5c-dd86r" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.401137 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsngw\" (UniqueName: \"kubernetes.io/projected/0bdc8091-6682-49da-a5dd-743dc013934d-kube-api-access-xsngw\") pod \"dnsmasq-dns-59c6c64b5c-dd86r\" (UID: \"0bdc8091-6682-49da-a5dd-743dc013934d\") " pod="openstack/dnsmasq-dns-59c6c64b5c-dd86r" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.401177 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bdc8091-6682-49da-a5dd-743dc013934d-dns-svc\") pod \"dnsmasq-dns-59c6c64b5c-dd86r\" (UID: \"0bdc8091-6682-49da-a5dd-743dc013934d\") " pod="openstack/dnsmasq-dns-59c6c64b5c-dd86r" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.402160 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bdc8091-6682-49da-a5dd-743dc013934d-config\") pod \"dnsmasq-dns-59c6c64b5c-dd86r\" (UID: \"0bdc8091-6682-49da-a5dd-743dc013934d\") " pod="openstack/dnsmasq-dns-59c6c64b5c-dd86r" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.402251 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bdc8091-6682-49da-a5dd-743dc013934d-dns-svc\") pod \"dnsmasq-dns-59c6c64b5c-dd86r\" (UID: \"0bdc8091-6682-49da-a5dd-743dc013934d\") " pod="openstack/dnsmasq-dns-59c6c64b5c-dd86r" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.424329 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsngw\" (UniqueName: \"kubernetes.io/projected/0bdc8091-6682-49da-a5dd-743dc013934d-kube-api-access-xsngw\") pod \"dnsmasq-dns-59c6c64b5c-dd86r\" (UID: \"0bdc8091-6682-49da-a5dd-743dc013934d\") " pod="openstack/dnsmasq-dns-59c6c64b5c-dd86r" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.540576 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55c86457d7-z64jx" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.555180 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55c86457d7-z64jx" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.603195 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgpv5\" (UniqueName: \"kubernetes.io/projected/01e82c9a-afc8-477c-a2c0-faf51a20fc6c-kube-api-access-pgpv5\") pod \"01e82c9a-afc8-477c-a2c0-faf51a20fc6c\" (UID: \"01e82c9a-afc8-477c-a2c0-faf51a20fc6c\") " Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.603580 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01e82c9a-afc8-477c-a2c0-faf51a20fc6c-dns-svc\") pod \"01e82c9a-afc8-477c-a2c0-faf51a20fc6c\" (UID: \"01e82c9a-afc8-477c-a2c0-faf51a20fc6c\") " Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.603685 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01e82c9a-afc8-477c-a2c0-faf51a20fc6c-config\") pod \"01e82c9a-afc8-477c-a2c0-faf51a20fc6c\" (UID: \"01e82c9a-afc8-477c-a2c0-faf51a20fc6c\") " Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.604178 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01e82c9a-afc8-477c-a2c0-faf51a20fc6c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "01e82c9a-afc8-477c-a2c0-faf51a20fc6c" (UID: "01e82c9a-afc8-477c-a2c0-faf51a20fc6c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.605247 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01e82c9a-afc8-477c-a2c0-faf51a20fc6c-config" (OuterVolumeSpecName: "config") pod "01e82c9a-afc8-477c-a2c0-faf51a20fc6c" (UID: "01e82c9a-afc8-477c-a2c0-faf51a20fc6c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.605574 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01e82c9a-afc8-477c-a2c0-faf51a20fc6c-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.605592 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01e82c9a-afc8-477c-a2c0-faf51a20fc6c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.610057 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01e82c9a-afc8-477c-a2c0-faf51a20fc6c-kube-api-access-pgpv5" (OuterVolumeSpecName: "kube-api-access-pgpv5") pod "01e82c9a-afc8-477c-a2c0-faf51a20fc6c" (UID: "01e82c9a-afc8-477c-a2c0-faf51a20fc6c"). InnerVolumeSpecName "kube-api-access-pgpv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.651364 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59c6c64b5c-dd86r" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.706648 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgpv5\" (UniqueName: \"kubernetes.io/projected/01e82c9a-afc8-477c-a2c0-faf51a20fc6c-kube-api-access-pgpv5\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.753029 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-866449bdb9-psppk"] Nov 25 10:46:50 crc kubenswrapper[4776]: I1125 10:46:50.868006 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f4c6c447c-7w4tr"] Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.086148 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.088297 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.091393 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.091423 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.091685 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.092169 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.093006 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.095941 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.096575 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-g7vlf" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.097832 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.143214 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59c6c64b5c-dd86r"] Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.218437 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ea47bcae-f9de-4486-93eb-8dd959573bd2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.218488 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ea47bcae-f9de-4486-93eb-8dd959573bd2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.218543 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ea47bcae-f9de-4486-93eb-8dd959573bd2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.218561 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ea47bcae-f9de-4486-93eb-8dd959573bd2-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.218588 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjxjw\" (UniqueName: \"kubernetes.io/projected/ea47bcae-f9de-4486-93eb-8dd959573bd2-kube-api-access-sjxjw\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.218641 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ea47bcae-f9de-4486-93eb-8dd959573bd2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.218696 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-947e4d82-272e-4d92-87b2-c028bbea2317\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-947e4d82-272e-4d92-87b2-c028bbea2317\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.218735 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ea47bcae-f9de-4486-93eb-8dd959573bd2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.218757 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ea47bcae-f9de-4486-93eb-8dd959573bd2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.218780 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ea47bcae-f9de-4486-93eb-8dd959573bd2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.218802 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea47bcae-f9de-4486-93eb-8dd959573bd2-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.319600 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-947e4d82-272e-4d92-87b2-c028bbea2317\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-947e4d82-272e-4d92-87b2-c028bbea2317\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.319645 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ea47bcae-f9de-4486-93eb-8dd959573bd2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.319672 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ea47bcae-f9de-4486-93eb-8dd959573bd2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.319706 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ea47bcae-f9de-4486-93eb-8dd959573bd2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.319736 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea47bcae-f9de-4486-93eb-8dd959573bd2-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.319792 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ea47bcae-f9de-4486-93eb-8dd959573bd2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.319809 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ea47bcae-f9de-4486-93eb-8dd959573bd2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.319846 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ea47bcae-f9de-4486-93eb-8dd959573bd2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.319863 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ea47bcae-f9de-4486-93eb-8dd959573bd2-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.319884 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjxjw\" (UniqueName: \"kubernetes.io/projected/ea47bcae-f9de-4486-93eb-8dd959573bd2-kube-api-access-sjxjw\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.319900 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ea47bcae-f9de-4486-93eb-8dd959573bd2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.320765 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ea47bcae-f9de-4486-93eb-8dd959573bd2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.321552 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ea47bcae-f9de-4486-93eb-8dd959573bd2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.321647 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea47bcae-f9de-4486-93eb-8dd959573bd2-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.321691 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ea47bcae-f9de-4486-93eb-8dd959573bd2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.321839 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ea47bcae-f9de-4486-93eb-8dd959573bd2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.322137 4776 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.322158 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-947e4d82-272e-4d92-87b2-c028bbea2317\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-947e4d82-272e-4d92-87b2-c028bbea2317\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/fe5a02d6afa727d9853c90ed2daf9d081979f7c5457f86edf7a6da3605d7b0a0/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.331857 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ea47bcae-f9de-4486-93eb-8dd959573bd2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.331885 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ea47bcae-f9de-4486-93eb-8dd959573bd2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.332048 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ea47bcae-f9de-4486-93eb-8dd959573bd2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.332182 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ea47bcae-f9de-4486-93eb-8dd959573bd2-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.340190 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjxjw\" (UniqueName: \"kubernetes.io/projected/ea47bcae-f9de-4486-93eb-8dd959573bd2-kube-api-access-sjxjw\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.365981 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.367160 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.372201 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.372201 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.372245 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.372280 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.372325 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.372654 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.374126 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-gwftv" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.377247 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-947e4d82-272e-4d92-87b2-c028bbea2317\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-947e4d82-272e-4d92-87b2-c028bbea2317\") pod \"rabbitmq-cell1-server-0\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.387790 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.412505 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.521846 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2bcr\" (UniqueName: \"kubernetes.io/projected/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-kube-api-access-c2bcr\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.521916 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.521954 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.521976 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-181a479f-9029-4e9e-a4e1-faa600369e96\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-181a479f-9029-4e9e-a4e1-faa600369e96\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.521999 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.522039 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.522100 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.522141 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.522188 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-config-data\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.522209 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.522244 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.554362 4776 generic.go:334] "Generic (PLEG): container finished" podID="0bdc8091-6682-49da-a5dd-743dc013934d" containerID="75889533bb71117d5bf6b3d7ef073f97a4f09f60d30a706091fc8145e0f0bb1f" exitCode=0 Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.554433 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59c6c64b5c-dd86r" event={"ID":"0bdc8091-6682-49da-a5dd-743dc013934d","Type":"ContainerDied","Data":"75889533bb71117d5bf6b3d7ef073f97a4f09f60d30a706091fc8145e0f0bb1f"} Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.554462 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59c6c64b5c-dd86r" event={"ID":"0bdc8091-6682-49da-a5dd-743dc013934d","Type":"ContainerStarted","Data":"2d8f19e4a6b66dffd7117e5b238d820bc05e8cf75787c665f957f8b0e830aae8"} Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.559373 4776 generic.go:334] "Generic (PLEG): container finished" podID="a6933bcb-92b8-42b6-8151-dd628024188e" containerID="bd6abeb8dc2fd5f7d60bca5177b879c7d65ee7c3946acb6509b7fbafe7dd88e2" exitCode=0 Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.559438 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f4c6c447c-7w4tr" event={"ID":"a6933bcb-92b8-42b6-8151-dd628024188e","Type":"ContainerDied","Data":"bd6abeb8dc2fd5f7d60bca5177b879c7d65ee7c3946acb6509b7fbafe7dd88e2"} Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.559469 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f4c6c447c-7w4tr" event={"ID":"a6933bcb-92b8-42b6-8151-dd628024188e","Type":"ContainerStarted","Data":"7496d46bf760315ff75958036bc7abeed5ac25b06a203b53483034f64bcdb37a"} Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.565557 4776 generic.go:334] "Generic (PLEG): container finished" podID="8ca04eb1-c148-4a93-bf34-fcf325fcd09c" containerID="d9832cbbbe4c51a9baada8c2ce5ce5f63ceaa392e7c381c2f0766cfbf716f584" exitCode=0 Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.565634 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55c86457d7-z64jx" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.566368 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-866449bdb9-psppk" event={"ID":"8ca04eb1-c148-4a93-bf34-fcf325fcd09c","Type":"ContainerDied","Data":"d9832cbbbe4c51a9baada8c2ce5ce5f63ceaa392e7c381c2f0766cfbf716f584"} Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.566423 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-866449bdb9-psppk" event={"ID":"8ca04eb1-c148-4a93-bf34-fcf325fcd09c","Type":"ContainerStarted","Data":"94ac5b5894cf55ca6e1aab0efba89a14964a9deefc7bd150fedc5932d67748e7"} Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.625927 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2bcr\" (UniqueName: \"kubernetes.io/projected/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-kube-api-access-c2bcr\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.625986 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.626009 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.626029 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-181a479f-9029-4e9e-a4e1-faa600369e96\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-181a479f-9029-4e9e-a4e1-faa600369e96\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.626053 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.626102 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.626140 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.626187 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.626205 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-config-data\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.626224 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.626242 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.627288 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.636608 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.636746 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.636970 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.638945 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-config-data\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.639243 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.640038 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.640692 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.644890 4776 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.645138 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-181a479f-9029-4e9e-a4e1-faa600369e96\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-181a479f-9029-4e9e-a4e1-faa600369e96\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a722d3c24f2b44f3b5ae75c49f6505dc76718dc72741f13e90c9eddd35738e5d/globalmount\"" pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.650654 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.652744 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2bcr\" (UniqueName: \"kubernetes.io/projected/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-kube-api-access-c2bcr\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.690836 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55c86457d7-z64jx"] Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.700402 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55c86457d7-z64jx"] Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.702019 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-181a479f-9029-4e9e-a4e1-faa600369e96\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-181a479f-9029-4e9e-a4e1-faa600369e96\") pod \"rabbitmq-server-0\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " pod="openstack/rabbitmq-server-0" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.896213 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.920907 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-866449bdb9-psppk" Nov 25 10:46:51 crc kubenswrapper[4776]: I1125 10:46:51.984218 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.032458 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fqfq\" (UniqueName: \"kubernetes.io/projected/8ca04eb1-c148-4a93-bf34-fcf325fcd09c-kube-api-access-9fqfq\") pod \"8ca04eb1-c148-4a93-bf34-fcf325fcd09c\" (UID: \"8ca04eb1-c148-4a93-bf34-fcf325fcd09c\") " Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.032528 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ca04eb1-c148-4a93-bf34-fcf325fcd09c-config\") pod \"8ca04eb1-c148-4a93-bf34-fcf325fcd09c\" (UID: \"8ca04eb1-c148-4a93-bf34-fcf325fcd09c\") " Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.037432 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ca04eb1-c148-4a93-bf34-fcf325fcd09c-kube-api-access-9fqfq" (OuterVolumeSpecName: "kube-api-access-9fqfq") pod "8ca04eb1-c148-4a93-bf34-fcf325fcd09c" (UID: "8ca04eb1-c148-4a93-bf34-fcf325fcd09c"). InnerVolumeSpecName "kube-api-access-9fqfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.053694 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ca04eb1-c148-4a93-bf34-fcf325fcd09c-config" (OuterVolumeSpecName: "config") pod "8ca04eb1-c148-4a93-bf34-fcf325fcd09c" (UID: "8ca04eb1-c148-4a93-bf34-fcf325fcd09c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.135146 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fqfq\" (UniqueName: \"kubernetes.io/projected/8ca04eb1-c148-4a93-bf34-fcf325fcd09c-kube-api-access-9fqfq\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.135176 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ca04eb1-c148-4a93-bf34-fcf325fcd09c-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.256758 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 25 10:46:52 crc kubenswrapper[4776]: E1125 10:46:52.257181 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ca04eb1-c148-4a93-bf34-fcf325fcd09c" containerName="init" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.257207 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ca04eb1-c148-4a93-bf34-fcf325fcd09c" containerName="init" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.257384 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ca04eb1-c148-4a93-bf34-fcf325fcd09c" containerName="init" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.258143 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.259703 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.260181 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-nw5nm" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.260466 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.262017 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.266711 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.267284 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.337970 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ed1eddec-2d91-434e-803e-69cf7501f99d-config-data-default\") pod \"openstack-galera-0\" (UID: \"ed1eddec-2d91-434e-803e-69cf7501f99d\") " pod="openstack/openstack-galera-0" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.338019 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed1eddec-2d91-434e-803e-69cf7501f99d-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ed1eddec-2d91-434e-803e-69cf7501f99d\") " pod="openstack/openstack-galera-0" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.338102 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ed1eddec-2d91-434e-803e-69cf7501f99d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ed1eddec-2d91-434e-803e-69cf7501f99d\") " pod="openstack/openstack-galera-0" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.338125 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ed1eddec-2d91-434e-803e-69cf7501f99d-kolla-config\") pod \"openstack-galera-0\" (UID: \"ed1eddec-2d91-434e-803e-69cf7501f99d\") " pod="openstack/openstack-galera-0" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.338143 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed1eddec-2d91-434e-803e-69cf7501f99d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ed1eddec-2d91-434e-803e-69cf7501f99d\") " pod="openstack/openstack-galera-0" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.338161 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4xxj\" (UniqueName: \"kubernetes.io/projected/ed1eddec-2d91-434e-803e-69cf7501f99d-kube-api-access-x4xxj\") pod \"openstack-galera-0\" (UID: \"ed1eddec-2d91-434e-803e-69cf7501f99d\") " pod="openstack/openstack-galera-0" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.338207 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-55e5cb7b-187d-47cc-8838-10cab2c4943e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-55e5cb7b-187d-47cc-8838-10cab2c4943e\") pod \"openstack-galera-0\" (UID: \"ed1eddec-2d91-434e-803e-69cf7501f99d\") " pod="openstack/openstack-galera-0" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.338227 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed1eddec-2d91-434e-803e-69cf7501f99d-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ed1eddec-2d91-434e-803e-69cf7501f99d\") " pod="openstack/openstack-galera-0" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.417791 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:46:52 crc kubenswrapper[4776]: W1125 10:46:52.424582 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89c88cd7_7cd2_43f8_bb00_0908d26c4a9e.slice/crio-36a01b98e90356b2e66b08dda7acf48f19d9fb68b86725f1a836fbc4ae1b8585 WatchSource:0}: Error finding container 36a01b98e90356b2e66b08dda7acf48f19d9fb68b86725f1a836fbc4ae1b8585: Status 404 returned error can't find the container with id 36a01b98e90356b2e66b08dda7acf48f19d9fb68b86725f1a836fbc4ae1b8585 Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.438830 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed1eddec-2d91-434e-803e-69cf7501f99d-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ed1eddec-2d91-434e-803e-69cf7501f99d\") " pod="openstack/openstack-galera-0" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.438898 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ed1eddec-2d91-434e-803e-69cf7501f99d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ed1eddec-2d91-434e-803e-69cf7501f99d\") " pod="openstack/openstack-galera-0" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.438917 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ed1eddec-2d91-434e-803e-69cf7501f99d-kolla-config\") pod \"openstack-galera-0\" (UID: \"ed1eddec-2d91-434e-803e-69cf7501f99d\") " pod="openstack/openstack-galera-0" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.438931 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed1eddec-2d91-434e-803e-69cf7501f99d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ed1eddec-2d91-434e-803e-69cf7501f99d\") " pod="openstack/openstack-galera-0" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.438945 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4xxj\" (UniqueName: \"kubernetes.io/projected/ed1eddec-2d91-434e-803e-69cf7501f99d-kube-api-access-x4xxj\") pod \"openstack-galera-0\" (UID: \"ed1eddec-2d91-434e-803e-69cf7501f99d\") " pod="openstack/openstack-galera-0" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.438984 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-55e5cb7b-187d-47cc-8838-10cab2c4943e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-55e5cb7b-187d-47cc-8838-10cab2c4943e\") pod \"openstack-galera-0\" (UID: \"ed1eddec-2d91-434e-803e-69cf7501f99d\") " pod="openstack/openstack-galera-0" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.439000 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed1eddec-2d91-434e-803e-69cf7501f99d-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ed1eddec-2d91-434e-803e-69cf7501f99d\") " pod="openstack/openstack-galera-0" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.439053 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ed1eddec-2d91-434e-803e-69cf7501f99d-config-data-default\") pod \"openstack-galera-0\" (UID: \"ed1eddec-2d91-434e-803e-69cf7501f99d\") " pod="openstack/openstack-galera-0" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.439342 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ed1eddec-2d91-434e-803e-69cf7501f99d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ed1eddec-2d91-434e-803e-69cf7501f99d\") " pod="openstack/openstack-galera-0" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.439952 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ed1eddec-2d91-434e-803e-69cf7501f99d-kolla-config\") pod \"openstack-galera-0\" (UID: \"ed1eddec-2d91-434e-803e-69cf7501f99d\") " pod="openstack/openstack-galera-0" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.440005 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ed1eddec-2d91-434e-803e-69cf7501f99d-config-data-default\") pod \"openstack-galera-0\" (UID: \"ed1eddec-2d91-434e-803e-69cf7501f99d\") " pod="openstack/openstack-galera-0" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.440510 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed1eddec-2d91-434e-803e-69cf7501f99d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ed1eddec-2d91-434e-803e-69cf7501f99d\") " pod="openstack/openstack-galera-0" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.449307 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed1eddec-2d91-434e-803e-69cf7501f99d-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ed1eddec-2d91-434e-803e-69cf7501f99d\") " pod="openstack/openstack-galera-0" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.449571 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed1eddec-2d91-434e-803e-69cf7501f99d-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ed1eddec-2d91-434e-803e-69cf7501f99d\") " pod="openstack/openstack-galera-0" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.454084 4776 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.454120 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-55e5cb7b-187d-47cc-8838-10cab2c4943e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-55e5cb7b-187d-47cc-8838-10cab2c4943e\") pod \"openstack-galera-0\" (UID: \"ed1eddec-2d91-434e-803e-69cf7501f99d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8675beb3b87c466a761573165a2765e7ed98858f9741e4cab62bc570685f9f1e/globalmount\"" pod="openstack/openstack-galera-0" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.554416 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4xxj\" (UniqueName: \"kubernetes.io/projected/ed1eddec-2d91-434e-803e-69cf7501f99d-kube-api-access-x4xxj\") pod \"openstack-galera-0\" (UID: \"ed1eddec-2d91-434e-803e-69cf7501f99d\") " pod="openstack/openstack-galera-0" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.577562 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-866449bdb9-psppk" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.577597 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-866449bdb9-psppk" event={"ID":"8ca04eb1-c148-4a93-bf34-fcf325fcd09c","Type":"ContainerDied","Data":"94ac5b5894cf55ca6e1aab0efba89a14964a9deefc7bd150fedc5932d67748e7"} Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.577653 4776 scope.go:117] "RemoveContainer" containerID="d9832cbbbe4c51a9baada8c2ce5ce5f63ceaa392e7c381c2f0766cfbf716f584" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.578931 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ea47bcae-f9de-4486-93eb-8dd959573bd2","Type":"ContainerStarted","Data":"9f3b9a0fa9d5d030967d7a1af74f0da7d30b115c2c0f1c6bb6a398d12deb5b4f"} Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.581391 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59c6c64b5c-dd86r" event={"ID":"0bdc8091-6682-49da-a5dd-743dc013934d","Type":"ContainerStarted","Data":"186b4dd940c3d859642772cffd1be431d22aa57aa04f371b7bce8b6627a54a91"} Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.581715 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59c6c64b5c-dd86r" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.582267 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-55e5cb7b-187d-47cc-8838-10cab2c4943e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-55e5cb7b-187d-47cc-8838-10cab2c4943e\") pod \"openstack-galera-0\" (UID: \"ed1eddec-2d91-434e-803e-69cf7501f99d\") " pod="openstack/openstack-galera-0" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.590098 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e","Type":"ContainerStarted","Data":"36a01b98e90356b2e66b08dda7acf48f19d9fb68b86725f1a836fbc4ae1b8585"} Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.612772 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59c6c64b5c-dd86r" podStartSLOduration=2.612754324 podStartE2EDuration="2.612754324s" podCreationTimestamp="2025-11-25 10:46:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:46:52.610715672 +0000 UTC m=+4957.651775225" watchObservedRunningTime="2025-11-25 10:46:52.612754324 +0000 UTC m=+4957.653813877" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.614843 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f4c6c447c-7w4tr" event={"ID":"a6933bcb-92b8-42b6-8151-dd628024188e","Type":"ContainerStarted","Data":"c4702f24d04ffd9cb225f965d1b67d88b5e2685015af701afa63f259c67b8b04"} Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.615571 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f4c6c447c-7w4tr" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.657192 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-866449bdb9-psppk"] Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.667362 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-866449bdb9-psppk"] Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.681105 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f4c6c447c-7w4tr" podStartSLOduration=3.6810844190000003 podStartE2EDuration="3.681084419s" podCreationTimestamp="2025-11-25 10:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:46:52.674662468 +0000 UTC m=+4957.715722021" watchObservedRunningTime="2025-11-25 10:46:52.681084419 +0000 UTC m=+4957.722143972" Nov 25 10:46:52 crc kubenswrapper[4776]: I1125 10:46:52.871949 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.310689 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.622668 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ea47bcae-f9de-4486-93eb-8dd959573bd2","Type":"ContainerStarted","Data":"5bd67c8d057c8f9c8a1eee8383f029cacd61b21bdacb1a5c0b25926cbc4ae826"} Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.623923 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e","Type":"ContainerStarted","Data":"bc062b2a63abbd08250341b4b871dd718116d42e8b4841c259c448e93b33e490"} Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.627033 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ed1eddec-2d91-434e-803e-69cf7501f99d","Type":"ContainerStarted","Data":"2d05e40aeae9a5d826042435b85194f074432d4f341911ba75f41a89ed852e02"} Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.627153 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ed1eddec-2d91-434e-803e-69cf7501f99d","Type":"ContainerStarted","Data":"fe9fef434cde32a675de2fe122db497a8dd3bc538cdd981ca04e19e3147750ff"} Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.672392 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01e82c9a-afc8-477c-a2c0-faf51a20fc6c" path="/var/lib/kubelet/pods/01e82c9a-afc8-477c-a2c0-faf51a20fc6c/volumes" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.672869 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ca04eb1-c148-4a93-bf34-fcf325fcd09c" path="/var/lib/kubelet/pods/8ca04eb1-c148-4a93-bf34-fcf325fcd09c/volumes" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.702226 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.703817 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.705757 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-vs4kj" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.706179 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.707660 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.709407 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.741477 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.861117 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3aa288-f360-4925-8229-5a3593b31be7-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"8c3aa288-f360-4925-8229-5a3593b31be7\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.861193 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c3aa288-f360-4925-8229-5a3593b31be7-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"8c3aa288-f360-4925-8229-5a3593b31be7\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.861217 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8c3aa288-f360-4925-8229-5a3593b31be7-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"8c3aa288-f360-4925-8229-5a3593b31be7\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.861244 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c3aa288-f360-4925-8229-5a3593b31be7-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"8c3aa288-f360-4925-8229-5a3593b31be7\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.861433 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8c3aa288-f360-4925-8229-5a3593b31be7-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"8c3aa288-f360-4925-8229-5a3593b31be7\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.861561 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8c3aa288-f360-4925-8229-5a3593b31be7-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"8c3aa288-f360-4925-8229-5a3593b31be7\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.861592 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-361ed85b-e323-4015-bd22-5c48d87b50f1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-361ed85b-e323-4015-bd22-5c48d87b50f1\") pod \"openstack-cell1-galera-0\" (UID: \"8c3aa288-f360-4925-8229-5a3593b31be7\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.861734 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z47gc\" (UniqueName: \"kubernetes.io/projected/8c3aa288-f360-4925-8229-5a3593b31be7-kube-api-access-z47gc\") pod \"openstack-cell1-galera-0\" (UID: \"8c3aa288-f360-4925-8229-5a3593b31be7\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.963291 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c3aa288-f360-4925-8229-5a3593b31be7-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"8c3aa288-f360-4925-8229-5a3593b31be7\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.963660 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8c3aa288-f360-4925-8229-5a3593b31be7-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"8c3aa288-f360-4925-8229-5a3593b31be7\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.963804 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8c3aa288-f360-4925-8229-5a3593b31be7-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"8c3aa288-f360-4925-8229-5a3593b31be7\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.963916 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-361ed85b-e323-4015-bd22-5c48d87b50f1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-361ed85b-e323-4015-bd22-5c48d87b50f1\") pod \"openstack-cell1-galera-0\" (UID: \"8c3aa288-f360-4925-8229-5a3593b31be7\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.964016 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z47gc\" (UniqueName: \"kubernetes.io/projected/8c3aa288-f360-4925-8229-5a3593b31be7-kube-api-access-z47gc\") pod \"openstack-cell1-galera-0\" (UID: \"8c3aa288-f360-4925-8229-5a3593b31be7\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.964097 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8c3aa288-f360-4925-8229-5a3593b31be7-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"8c3aa288-f360-4925-8229-5a3593b31be7\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.964241 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3aa288-f360-4925-8229-5a3593b31be7-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"8c3aa288-f360-4925-8229-5a3593b31be7\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.964389 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c3aa288-f360-4925-8229-5a3593b31be7-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"8c3aa288-f360-4925-8229-5a3593b31be7\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.964491 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8c3aa288-f360-4925-8229-5a3593b31be7-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"8c3aa288-f360-4925-8229-5a3593b31be7\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.964504 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8c3aa288-f360-4925-8229-5a3593b31be7-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"8c3aa288-f360-4925-8229-5a3593b31be7\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.965447 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8c3aa288-f360-4925-8229-5a3593b31be7-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"8c3aa288-f360-4925-8229-5a3593b31be7\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.966514 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c3aa288-f360-4925-8229-5a3593b31be7-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"8c3aa288-f360-4925-8229-5a3593b31be7\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.968614 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c3aa288-f360-4925-8229-5a3593b31be7-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"8c3aa288-f360-4925-8229-5a3593b31be7\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.968805 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c3aa288-f360-4925-8229-5a3593b31be7-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"8c3aa288-f360-4925-8229-5a3593b31be7\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.969013 4776 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 10:46:53 crc kubenswrapper[4776]: I1125 10:46:53.969076 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-361ed85b-e323-4015-bd22-5c48d87b50f1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-361ed85b-e323-4015-bd22-5c48d87b50f1\") pod \"openstack-cell1-galera-0\" (UID: \"8c3aa288-f360-4925-8229-5a3593b31be7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/dff6e92245e20c6b0577253209086a3cb95a20af263baf82ac2ac8cd589d98c8/globalmount\"" pod="openstack/openstack-cell1-galera-0" Nov 25 10:46:54 crc kubenswrapper[4776]: I1125 10:46:54.007960 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 25 10:46:54 crc kubenswrapper[4776]: I1125 10:46:54.010404 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 25 10:46:54 crc kubenswrapper[4776]: I1125 10:46:54.013703 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 25 10:46:54 crc kubenswrapper[4776]: I1125 10:46:54.014077 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 25 10:46:54 crc kubenswrapper[4776]: I1125 10:46:54.014236 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-9wzm8" Nov 25 10:46:54 crc kubenswrapper[4776]: I1125 10:46:54.017452 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z47gc\" (UniqueName: \"kubernetes.io/projected/8c3aa288-f360-4925-8229-5a3593b31be7-kube-api-access-z47gc\") pod \"openstack-cell1-galera-0\" (UID: \"8c3aa288-f360-4925-8229-5a3593b31be7\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:46:54 crc kubenswrapper[4776]: I1125 10:46:54.020817 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-361ed85b-e323-4015-bd22-5c48d87b50f1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-361ed85b-e323-4015-bd22-5c48d87b50f1\") pod \"openstack-cell1-galera-0\" (UID: \"8c3aa288-f360-4925-8229-5a3593b31be7\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:46:54 crc kubenswrapper[4776]: I1125 10:46:54.039653 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 25 10:46:54 crc kubenswrapper[4776]: I1125 10:46:54.067849 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxv74\" (UniqueName: \"kubernetes.io/projected/4f2267d1-b0d6-4a66-8dde-beb9d280b4e1-kube-api-access-qxv74\") pod \"memcached-0\" (UID: \"4f2267d1-b0d6-4a66-8dde-beb9d280b4e1\") " pod="openstack/memcached-0" Nov 25 10:46:54 crc kubenswrapper[4776]: I1125 10:46:54.067930 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4f2267d1-b0d6-4a66-8dde-beb9d280b4e1-kolla-config\") pod \"memcached-0\" (UID: \"4f2267d1-b0d6-4a66-8dde-beb9d280b4e1\") " pod="openstack/memcached-0" Nov 25 10:46:54 crc kubenswrapper[4776]: I1125 10:46:54.068059 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f2267d1-b0d6-4a66-8dde-beb9d280b4e1-memcached-tls-certs\") pod \"memcached-0\" (UID: \"4f2267d1-b0d6-4a66-8dde-beb9d280b4e1\") " pod="openstack/memcached-0" Nov 25 10:46:54 crc kubenswrapper[4776]: I1125 10:46:54.068423 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4f2267d1-b0d6-4a66-8dde-beb9d280b4e1-config-data\") pod \"memcached-0\" (UID: \"4f2267d1-b0d6-4a66-8dde-beb9d280b4e1\") " pod="openstack/memcached-0" Nov 25 10:46:54 crc kubenswrapper[4776]: I1125 10:46:54.068561 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f2267d1-b0d6-4a66-8dde-beb9d280b4e1-combined-ca-bundle\") pod \"memcached-0\" (UID: \"4f2267d1-b0d6-4a66-8dde-beb9d280b4e1\") " pod="openstack/memcached-0" Nov 25 10:46:54 crc kubenswrapper[4776]: I1125 10:46:54.170268 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4f2267d1-b0d6-4a66-8dde-beb9d280b4e1-kolla-config\") pod \"memcached-0\" (UID: \"4f2267d1-b0d6-4a66-8dde-beb9d280b4e1\") " pod="openstack/memcached-0" Nov 25 10:46:54 crc kubenswrapper[4776]: I1125 10:46:54.170323 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f2267d1-b0d6-4a66-8dde-beb9d280b4e1-memcached-tls-certs\") pod \"memcached-0\" (UID: \"4f2267d1-b0d6-4a66-8dde-beb9d280b4e1\") " pod="openstack/memcached-0" Nov 25 10:46:54 crc kubenswrapper[4776]: I1125 10:46:54.170352 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4f2267d1-b0d6-4a66-8dde-beb9d280b4e1-config-data\") pod \"memcached-0\" (UID: \"4f2267d1-b0d6-4a66-8dde-beb9d280b4e1\") " pod="openstack/memcached-0" Nov 25 10:46:54 crc kubenswrapper[4776]: I1125 10:46:54.170402 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f2267d1-b0d6-4a66-8dde-beb9d280b4e1-combined-ca-bundle\") pod \"memcached-0\" (UID: \"4f2267d1-b0d6-4a66-8dde-beb9d280b4e1\") " pod="openstack/memcached-0" Nov 25 10:46:54 crc kubenswrapper[4776]: I1125 10:46:54.170448 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxv74\" (UniqueName: \"kubernetes.io/projected/4f2267d1-b0d6-4a66-8dde-beb9d280b4e1-kube-api-access-qxv74\") pod \"memcached-0\" (UID: \"4f2267d1-b0d6-4a66-8dde-beb9d280b4e1\") " pod="openstack/memcached-0" Nov 25 10:46:54 crc kubenswrapper[4776]: I1125 10:46:54.171400 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4f2267d1-b0d6-4a66-8dde-beb9d280b4e1-kolla-config\") pod \"memcached-0\" (UID: \"4f2267d1-b0d6-4a66-8dde-beb9d280b4e1\") " pod="openstack/memcached-0" Nov 25 10:46:54 crc kubenswrapper[4776]: I1125 10:46:54.171512 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4f2267d1-b0d6-4a66-8dde-beb9d280b4e1-config-data\") pod \"memcached-0\" (UID: \"4f2267d1-b0d6-4a66-8dde-beb9d280b4e1\") " pod="openstack/memcached-0" Nov 25 10:46:54 crc kubenswrapper[4776]: I1125 10:46:54.175419 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f2267d1-b0d6-4a66-8dde-beb9d280b4e1-combined-ca-bundle\") pod \"memcached-0\" (UID: \"4f2267d1-b0d6-4a66-8dde-beb9d280b4e1\") " pod="openstack/memcached-0" Nov 25 10:46:54 crc kubenswrapper[4776]: I1125 10:46:54.189531 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f2267d1-b0d6-4a66-8dde-beb9d280b4e1-memcached-tls-certs\") pod \"memcached-0\" (UID: \"4f2267d1-b0d6-4a66-8dde-beb9d280b4e1\") " pod="openstack/memcached-0" Nov 25 10:46:54 crc kubenswrapper[4776]: I1125 10:46:54.192426 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxv74\" (UniqueName: \"kubernetes.io/projected/4f2267d1-b0d6-4a66-8dde-beb9d280b4e1-kube-api-access-qxv74\") pod \"memcached-0\" (UID: \"4f2267d1-b0d6-4a66-8dde-beb9d280b4e1\") " pod="openstack/memcached-0" Nov 25 10:46:54 crc kubenswrapper[4776]: I1125 10:46:54.319942 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 25 10:46:54 crc kubenswrapper[4776]: I1125 10:46:54.370797 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 25 10:46:54 crc kubenswrapper[4776]: I1125 10:46:54.725032 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 25 10:46:54 crc kubenswrapper[4776]: I1125 10:46:54.802974 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 25 10:46:54 crc kubenswrapper[4776]: W1125 10:46:54.812920 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f2267d1_b0d6_4a66_8dde_beb9d280b4e1.slice/crio-da3aedb780864919af4942b3df90461e7db4f1bd07da4a2e4c303f2b415d36c3 WatchSource:0}: Error finding container da3aedb780864919af4942b3df90461e7db4f1bd07da4a2e4c303f2b415d36c3: Status 404 returned error can't find the container with id da3aedb780864919af4942b3df90461e7db4f1bd07da4a2e4c303f2b415d36c3 Nov 25 10:46:55 crc kubenswrapper[4776]: I1125 10:46:55.643273 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"4f2267d1-b0d6-4a66-8dde-beb9d280b4e1","Type":"ContainerStarted","Data":"18c0d1b8ac80ddfd5ecc4e418abf67ed7ae4ca0836490f15bff41ff33b1ae056"} Nov 25 10:46:55 crc kubenswrapper[4776]: I1125 10:46:55.643324 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"4f2267d1-b0d6-4a66-8dde-beb9d280b4e1","Type":"ContainerStarted","Data":"da3aedb780864919af4942b3df90461e7db4f1bd07da4a2e4c303f2b415d36c3"} Nov 25 10:46:55 crc kubenswrapper[4776]: I1125 10:46:55.643418 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 25 10:46:55 crc kubenswrapper[4776]: I1125 10:46:55.646801 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"8c3aa288-f360-4925-8229-5a3593b31be7","Type":"ContainerStarted","Data":"bda87215fadd3800912d76daf42f3b584d9da28174db36320de048684f9090a0"} Nov 25 10:46:55 crc kubenswrapper[4776]: I1125 10:46:55.646839 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"8c3aa288-f360-4925-8229-5a3593b31be7","Type":"ContainerStarted","Data":"43c8d3fb92f0b214c86165a52221f02e8db945de22fa7334214c808830f7b941"} Nov 25 10:46:55 crc kubenswrapper[4776]: I1125 10:46:55.664588 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.664572056 podStartE2EDuration="2.664572056s" podCreationTimestamp="2025-11-25 10:46:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:46:55.663853958 +0000 UTC m=+4960.704913541" watchObservedRunningTime="2025-11-25 10:46:55.664572056 +0000 UTC m=+4960.705631609" Nov 25 10:46:57 crc kubenswrapper[4776]: I1125 10:46:57.663128 4776 generic.go:334] "Generic (PLEG): container finished" podID="ed1eddec-2d91-434e-803e-69cf7501f99d" containerID="2d05e40aeae9a5d826042435b85194f074432d4f341911ba75f41a89ed852e02" exitCode=0 Nov 25 10:46:57 crc kubenswrapper[4776]: I1125 10:46:57.689238 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ed1eddec-2d91-434e-803e-69cf7501f99d","Type":"ContainerDied","Data":"2d05e40aeae9a5d826042435b85194f074432d4f341911ba75f41a89ed852e02"} Nov 25 10:46:58 crc kubenswrapper[4776]: I1125 10:46:58.675385 4776 generic.go:334] "Generic (PLEG): container finished" podID="8c3aa288-f360-4925-8229-5a3593b31be7" containerID="bda87215fadd3800912d76daf42f3b584d9da28174db36320de048684f9090a0" exitCode=0 Nov 25 10:46:58 crc kubenswrapper[4776]: I1125 10:46:58.675449 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"8c3aa288-f360-4925-8229-5a3593b31be7","Type":"ContainerDied","Data":"bda87215fadd3800912d76daf42f3b584d9da28174db36320de048684f9090a0"} Nov 25 10:46:58 crc kubenswrapper[4776]: I1125 10:46:58.678260 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ed1eddec-2d91-434e-803e-69cf7501f99d","Type":"ContainerStarted","Data":"1e8b11616eb37e16580db8a622cb42f0511584ec4388d56c86f1f948217e7d66"} Nov 25 10:46:58 crc kubenswrapper[4776]: I1125 10:46:58.718355 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=7.718337897 podStartE2EDuration="7.718337897s" podCreationTimestamp="2025-11-25 10:46:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:46:58.716890511 +0000 UTC m=+4963.757950064" watchObservedRunningTime="2025-11-25 10:46:58.718337897 +0000 UTC m=+4963.759397450" Nov 25 10:46:59 crc kubenswrapper[4776]: I1125 10:46:59.372782 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 25 10:46:59 crc kubenswrapper[4776]: I1125 10:46:59.686027 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"8c3aa288-f360-4925-8229-5a3593b31be7","Type":"ContainerStarted","Data":"964e88807bbaef68a1c78b0a1b848cdf839c48ed4d931c66d10e5c6d9c67ef2f"} Nov 25 10:46:59 crc kubenswrapper[4776]: I1125 10:46:59.713341 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=7.713325772 podStartE2EDuration="7.713325772s" podCreationTimestamp="2025-11-25 10:46:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:46:59.706254745 +0000 UTC m=+4964.747314288" watchObservedRunningTime="2025-11-25 10:46:59.713325772 +0000 UTC m=+4964.754385315" Nov 25 10:47:00 crc kubenswrapper[4776]: I1125 10:47:00.277270 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f4c6c447c-7w4tr" Nov 25 10:47:00 crc kubenswrapper[4776]: I1125 10:47:00.653286 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59c6c64b5c-dd86r" Nov 25 10:47:00 crc kubenswrapper[4776]: I1125 10:47:00.698244 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f4c6c447c-7w4tr"] Nov 25 10:47:00 crc kubenswrapper[4776]: I1125 10:47:00.698517 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f4c6c447c-7w4tr" podUID="a6933bcb-92b8-42b6-8151-dd628024188e" containerName="dnsmasq-dns" containerID="cri-o://c4702f24d04ffd9cb225f965d1b67d88b5e2685015af701afa63f259c67b8b04" gracePeriod=10 Nov 25 10:47:01 crc kubenswrapper[4776]: I1125 10:47:01.121057 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f4c6c447c-7w4tr" Nov 25 10:47:01 crc kubenswrapper[4776]: I1125 10:47:01.271825 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6933bcb-92b8-42b6-8151-dd628024188e-dns-svc\") pod \"a6933bcb-92b8-42b6-8151-dd628024188e\" (UID: \"a6933bcb-92b8-42b6-8151-dd628024188e\") " Nov 25 10:47:01 crc kubenswrapper[4776]: I1125 10:47:01.271876 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6933bcb-92b8-42b6-8151-dd628024188e-config\") pod \"a6933bcb-92b8-42b6-8151-dd628024188e\" (UID: \"a6933bcb-92b8-42b6-8151-dd628024188e\") " Nov 25 10:47:01 crc kubenswrapper[4776]: I1125 10:47:01.272041 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2j2vx\" (UniqueName: \"kubernetes.io/projected/a6933bcb-92b8-42b6-8151-dd628024188e-kube-api-access-2j2vx\") pod \"a6933bcb-92b8-42b6-8151-dd628024188e\" (UID: \"a6933bcb-92b8-42b6-8151-dd628024188e\") " Nov 25 10:47:01 crc kubenswrapper[4776]: I1125 10:47:01.278461 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6933bcb-92b8-42b6-8151-dd628024188e-kube-api-access-2j2vx" (OuterVolumeSpecName: "kube-api-access-2j2vx") pod "a6933bcb-92b8-42b6-8151-dd628024188e" (UID: "a6933bcb-92b8-42b6-8151-dd628024188e"). InnerVolumeSpecName "kube-api-access-2j2vx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:47:01 crc kubenswrapper[4776]: I1125 10:47:01.307327 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6933bcb-92b8-42b6-8151-dd628024188e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a6933bcb-92b8-42b6-8151-dd628024188e" (UID: "a6933bcb-92b8-42b6-8151-dd628024188e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:01 crc kubenswrapper[4776]: I1125 10:47:01.310590 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6933bcb-92b8-42b6-8151-dd628024188e-config" (OuterVolumeSpecName: "config") pod "a6933bcb-92b8-42b6-8151-dd628024188e" (UID: "a6933bcb-92b8-42b6-8151-dd628024188e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:01 crc kubenswrapper[4776]: I1125 10:47:01.374219 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6933bcb-92b8-42b6-8151-dd628024188e-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:01 crc kubenswrapper[4776]: I1125 10:47:01.374270 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6933bcb-92b8-42b6-8151-dd628024188e-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:01 crc kubenswrapper[4776]: I1125 10:47:01.374286 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2j2vx\" (UniqueName: \"kubernetes.io/projected/a6933bcb-92b8-42b6-8151-dd628024188e-kube-api-access-2j2vx\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:01 crc kubenswrapper[4776]: I1125 10:47:01.701666 4776 generic.go:334] "Generic (PLEG): container finished" podID="a6933bcb-92b8-42b6-8151-dd628024188e" containerID="c4702f24d04ffd9cb225f965d1b67d88b5e2685015af701afa63f259c67b8b04" exitCode=0 Nov 25 10:47:01 crc kubenswrapper[4776]: I1125 10:47:01.701709 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f4c6c447c-7w4tr" event={"ID":"a6933bcb-92b8-42b6-8151-dd628024188e","Type":"ContainerDied","Data":"c4702f24d04ffd9cb225f965d1b67d88b5e2685015af701afa63f259c67b8b04"} Nov 25 10:47:01 crc kubenswrapper[4776]: I1125 10:47:01.701735 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f4c6c447c-7w4tr" event={"ID":"a6933bcb-92b8-42b6-8151-dd628024188e","Type":"ContainerDied","Data":"7496d46bf760315ff75958036bc7abeed5ac25b06a203b53483034f64bcdb37a"} Nov 25 10:47:01 crc kubenswrapper[4776]: I1125 10:47:01.701753 4776 scope.go:117] "RemoveContainer" containerID="c4702f24d04ffd9cb225f965d1b67d88b5e2685015af701afa63f259c67b8b04" Nov 25 10:47:01 crc kubenswrapper[4776]: I1125 10:47:01.701929 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f4c6c447c-7w4tr" Nov 25 10:47:01 crc kubenswrapper[4776]: I1125 10:47:01.724563 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f4c6c447c-7w4tr"] Nov 25 10:47:01 crc kubenswrapper[4776]: I1125 10:47:01.728514 4776 scope.go:117] "RemoveContainer" containerID="bd6abeb8dc2fd5f7d60bca5177b879c7d65ee7c3946acb6509b7fbafe7dd88e2" Nov 25 10:47:01 crc kubenswrapper[4776]: I1125 10:47:01.730383 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f4c6c447c-7w4tr"] Nov 25 10:47:01 crc kubenswrapper[4776]: I1125 10:47:01.747090 4776 scope.go:117] "RemoveContainer" containerID="c4702f24d04ffd9cb225f965d1b67d88b5e2685015af701afa63f259c67b8b04" Nov 25 10:47:01 crc kubenswrapper[4776]: E1125 10:47:01.747544 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4702f24d04ffd9cb225f965d1b67d88b5e2685015af701afa63f259c67b8b04\": container with ID starting with c4702f24d04ffd9cb225f965d1b67d88b5e2685015af701afa63f259c67b8b04 not found: ID does not exist" containerID="c4702f24d04ffd9cb225f965d1b67d88b5e2685015af701afa63f259c67b8b04" Nov 25 10:47:01 crc kubenswrapper[4776]: I1125 10:47:01.747580 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4702f24d04ffd9cb225f965d1b67d88b5e2685015af701afa63f259c67b8b04"} err="failed to get container status \"c4702f24d04ffd9cb225f965d1b67d88b5e2685015af701afa63f259c67b8b04\": rpc error: code = NotFound desc = could not find container \"c4702f24d04ffd9cb225f965d1b67d88b5e2685015af701afa63f259c67b8b04\": container with ID starting with c4702f24d04ffd9cb225f965d1b67d88b5e2685015af701afa63f259c67b8b04 not found: ID does not exist" Nov 25 10:47:01 crc kubenswrapper[4776]: I1125 10:47:01.747601 4776 scope.go:117] "RemoveContainer" containerID="bd6abeb8dc2fd5f7d60bca5177b879c7d65ee7c3946acb6509b7fbafe7dd88e2" Nov 25 10:47:01 crc kubenswrapper[4776]: E1125 10:47:01.747931 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd6abeb8dc2fd5f7d60bca5177b879c7d65ee7c3946acb6509b7fbafe7dd88e2\": container with ID starting with bd6abeb8dc2fd5f7d60bca5177b879c7d65ee7c3946acb6509b7fbafe7dd88e2 not found: ID does not exist" containerID="bd6abeb8dc2fd5f7d60bca5177b879c7d65ee7c3946acb6509b7fbafe7dd88e2" Nov 25 10:47:01 crc kubenswrapper[4776]: I1125 10:47:01.748003 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd6abeb8dc2fd5f7d60bca5177b879c7d65ee7c3946acb6509b7fbafe7dd88e2"} err="failed to get container status \"bd6abeb8dc2fd5f7d60bca5177b879c7d65ee7c3946acb6509b7fbafe7dd88e2\": rpc error: code = NotFound desc = could not find container \"bd6abeb8dc2fd5f7d60bca5177b879c7d65ee7c3946acb6509b7fbafe7dd88e2\": container with ID starting with bd6abeb8dc2fd5f7d60bca5177b879c7d65ee7c3946acb6509b7fbafe7dd88e2 not found: ID does not exist" Nov 25 10:47:02 crc kubenswrapper[4776]: I1125 10:47:02.873637 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 25 10:47:02 crc kubenswrapper[4776]: I1125 10:47:02.874637 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 25 10:47:02 crc kubenswrapper[4776]: I1125 10:47:02.947981 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 25 10:47:03 crc kubenswrapper[4776]: I1125 10:47:03.677156 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6933bcb-92b8-42b6-8151-dd628024188e" path="/var/lib/kubelet/pods/a6933bcb-92b8-42b6-8151-dd628024188e/volumes" Nov 25 10:47:03 crc kubenswrapper[4776]: I1125 10:47:03.786005 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 25 10:47:04 crc kubenswrapper[4776]: I1125 10:47:04.320844 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:04 crc kubenswrapper[4776]: I1125 10:47:04.321213 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:06 crc kubenswrapper[4776]: I1125 10:47:06.518829 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:06 crc kubenswrapper[4776]: I1125 10:47:06.588126 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 25 10:47:25 crc kubenswrapper[4776]: I1125 10:47:25.867137 4776 generic.go:334] "Generic (PLEG): container finished" podID="ea47bcae-f9de-4486-93eb-8dd959573bd2" containerID="5bd67c8d057c8f9c8a1eee8383f029cacd61b21bdacb1a5c0b25926cbc4ae826" exitCode=0 Nov 25 10:47:25 crc kubenswrapper[4776]: I1125 10:47:25.867225 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ea47bcae-f9de-4486-93eb-8dd959573bd2","Type":"ContainerDied","Data":"5bd67c8d057c8f9c8a1eee8383f029cacd61b21bdacb1a5c0b25926cbc4ae826"} Nov 25 10:47:25 crc kubenswrapper[4776]: I1125 10:47:25.870731 4776 generic.go:334] "Generic (PLEG): container finished" podID="89c88cd7-7cd2-43f8-bb00-0908d26c4a9e" containerID="bc062b2a63abbd08250341b4b871dd718116d42e8b4841c259c448e93b33e490" exitCode=0 Nov 25 10:47:25 crc kubenswrapper[4776]: I1125 10:47:25.870780 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e","Type":"ContainerDied","Data":"bc062b2a63abbd08250341b4b871dd718116d42e8b4841c259c448e93b33e490"} Nov 25 10:47:26 crc kubenswrapper[4776]: I1125 10:47:26.878839 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ea47bcae-f9de-4486-93eb-8dd959573bd2","Type":"ContainerStarted","Data":"df6d6cf305233fa8e00e44108ac511a2f03ea528426db3e56a43d666fd5c53e4"} Nov 25 10:47:26 crc kubenswrapper[4776]: I1125 10:47:26.879362 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:26 crc kubenswrapper[4776]: I1125 10:47:26.880928 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e","Type":"ContainerStarted","Data":"51e2e393deba56f5c516ba944e0a685a84a2cb389a5ab6cf916dbed780313ffe"} Nov 25 10:47:26 crc kubenswrapper[4776]: I1125 10:47:26.881122 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 25 10:47:26 crc kubenswrapper[4776]: I1125 10:47:26.902527 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.902508684 podStartE2EDuration="36.902508684s" podCreationTimestamp="2025-11-25 10:46:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:47:26.899388466 +0000 UTC m=+4991.940448029" watchObservedRunningTime="2025-11-25 10:47:26.902508684 +0000 UTC m=+4991.943568227" Nov 25 10:47:26 crc kubenswrapper[4776]: I1125 10:47:26.927371 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.927355398 podStartE2EDuration="36.927355398s" podCreationTimestamp="2025-11-25 10:46:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:47:26.923421089 +0000 UTC m=+4991.964480642" watchObservedRunningTime="2025-11-25 10:47:26.927355398 +0000 UTC m=+4991.968414951" Nov 25 10:47:41 crc kubenswrapper[4776]: I1125 10:47:41.417452 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:41 crc kubenswrapper[4776]: I1125 10:47:41.986313 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 25 10:47:45 crc kubenswrapper[4776]: I1125 10:47:45.350537 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54564445dc-5wx2z"] Nov 25 10:47:45 crc kubenswrapper[4776]: E1125 10:47:45.353010 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6933bcb-92b8-42b6-8151-dd628024188e" containerName="dnsmasq-dns" Nov 25 10:47:45 crc kubenswrapper[4776]: I1125 10:47:45.355988 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6933bcb-92b8-42b6-8151-dd628024188e" containerName="dnsmasq-dns" Nov 25 10:47:45 crc kubenswrapper[4776]: E1125 10:47:45.356198 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6933bcb-92b8-42b6-8151-dd628024188e" containerName="init" Nov 25 10:47:45 crc kubenswrapper[4776]: I1125 10:47:45.356318 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6933bcb-92b8-42b6-8151-dd628024188e" containerName="init" Nov 25 10:47:45 crc kubenswrapper[4776]: I1125 10:47:45.356827 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6933bcb-92b8-42b6-8151-dd628024188e" containerName="dnsmasq-dns" Nov 25 10:47:45 crc kubenswrapper[4776]: I1125 10:47:45.357909 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54564445dc-5wx2z" Nov 25 10:47:45 crc kubenswrapper[4776]: I1125 10:47:45.360514 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54564445dc-5wx2z"] Nov 25 10:47:45 crc kubenswrapper[4776]: I1125 10:47:45.449205 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac07dcc7-2c32-40b9-802b-8380d6c2ef3e-config\") pod \"dnsmasq-dns-54564445dc-5wx2z\" (UID: \"ac07dcc7-2c32-40b9-802b-8380d6c2ef3e\") " pod="openstack/dnsmasq-dns-54564445dc-5wx2z" Nov 25 10:47:45 crc kubenswrapper[4776]: I1125 10:47:45.449671 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bx5pl\" (UniqueName: \"kubernetes.io/projected/ac07dcc7-2c32-40b9-802b-8380d6c2ef3e-kube-api-access-bx5pl\") pod \"dnsmasq-dns-54564445dc-5wx2z\" (UID: \"ac07dcc7-2c32-40b9-802b-8380d6c2ef3e\") " pod="openstack/dnsmasq-dns-54564445dc-5wx2z" Nov 25 10:47:45 crc kubenswrapper[4776]: I1125 10:47:45.449974 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac07dcc7-2c32-40b9-802b-8380d6c2ef3e-dns-svc\") pod \"dnsmasq-dns-54564445dc-5wx2z\" (UID: \"ac07dcc7-2c32-40b9-802b-8380d6c2ef3e\") " pod="openstack/dnsmasq-dns-54564445dc-5wx2z" Nov 25 10:47:45 crc kubenswrapper[4776]: I1125 10:47:45.551692 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac07dcc7-2c32-40b9-802b-8380d6c2ef3e-config\") pod \"dnsmasq-dns-54564445dc-5wx2z\" (UID: \"ac07dcc7-2c32-40b9-802b-8380d6c2ef3e\") " pod="openstack/dnsmasq-dns-54564445dc-5wx2z" Nov 25 10:47:45 crc kubenswrapper[4776]: I1125 10:47:45.551747 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bx5pl\" (UniqueName: \"kubernetes.io/projected/ac07dcc7-2c32-40b9-802b-8380d6c2ef3e-kube-api-access-bx5pl\") pod \"dnsmasq-dns-54564445dc-5wx2z\" (UID: \"ac07dcc7-2c32-40b9-802b-8380d6c2ef3e\") " pod="openstack/dnsmasq-dns-54564445dc-5wx2z" Nov 25 10:47:45 crc kubenswrapper[4776]: I1125 10:47:45.551823 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac07dcc7-2c32-40b9-802b-8380d6c2ef3e-dns-svc\") pod \"dnsmasq-dns-54564445dc-5wx2z\" (UID: \"ac07dcc7-2c32-40b9-802b-8380d6c2ef3e\") " pod="openstack/dnsmasq-dns-54564445dc-5wx2z" Nov 25 10:47:45 crc kubenswrapper[4776]: I1125 10:47:45.552621 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac07dcc7-2c32-40b9-802b-8380d6c2ef3e-dns-svc\") pod \"dnsmasq-dns-54564445dc-5wx2z\" (UID: \"ac07dcc7-2c32-40b9-802b-8380d6c2ef3e\") " pod="openstack/dnsmasq-dns-54564445dc-5wx2z" Nov 25 10:47:45 crc kubenswrapper[4776]: I1125 10:47:45.552658 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac07dcc7-2c32-40b9-802b-8380d6c2ef3e-config\") pod \"dnsmasq-dns-54564445dc-5wx2z\" (UID: \"ac07dcc7-2c32-40b9-802b-8380d6c2ef3e\") " pod="openstack/dnsmasq-dns-54564445dc-5wx2z" Nov 25 10:47:45 crc kubenswrapper[4776]: I1125 10:47:45.570417 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bx5pl\" (UniqueName: \"kubernetes.io/projected/ac07dcc7-2c32-40b9-802b-8380d6c2ef3e-kube-api-access-bx5pl\") pod \"dnsmasq-dns-54564445dc-5wx2z\" (UID: \"ac07dcc7-2c32-40b9-802b-8380d6c2ef3e\") " pod="openstack/dnsmasq-dns-54564445dc-5wx2z" Nov 25 10:47:45 crc kubenswrapper[4776]: I1125 10:47:45.680929 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54564445dc-5wx2z" Nov 25 10:47:45 crc kubenswrapper[4776]: I1125 10:47:45.898808 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54564445dc-5wx2z"] Nov 25 10:47:45 crc kubenswrapper[4776]: W1125 10:47:45.901186 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac07dcc7_2c32_40b9_802b_8380d6c2ef3e.slice/crio-15247c76da0d56702119c5f9ea2db2fda9961944ae06d1f7f6e72b7c349ede45 WatchSource:0}: Error finding container 15247c76da0d56702119c5f9ea2db2fda9961944ae06d1f7f6e72b7c349ede45: Status 404 returned error can't find the container with id 15247c76da0d56702119c5f9ea2db2fda9961944ae06d1f7f6e72b7c349ede45 Nov 25 10:47:45 crc kubenswrapper[4776]: I1125 10:47:45.978250 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:47:46 crc kubenswrapper[4776]: I1125 10:47:46.028882 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54564445dc-5wx2z" event={"ID":"ac07dcc7-2c32-40b9-802b-8380d6c2ef3e","Type":"ContainerStarted","Data":"15247c76da0d56702119c5f9ea2db2fda9961944ae06d1f7f6e72b7c349ede45"} Nov 25 10:47:46 crc kubenswrapper[4776]: I1125 10:47:46.659078 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:47:47 crc kubenswrapper[4776]: I1125 10:47:47.042911 4776 generic.go:334] "Generic (PLEG): container finished" podID="ac07dcc7-2c32-40b9-802b-8380d6c2ef3e" containerID="a9828407e3e3001e79a68285213e3c6a1a99a8ffe58d0b51a724488e05c7e7e6" exitCode=0 Nov 25 10:47:47 crc kubenswrapper[4776]: I1125 10:47:47.043035 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54564445dc-5wx2z" event={"ID":"ac07dcc7-2c32-40b9-802b-8380d6c2ef3e","Type":"ContainerDied","Data":"a9828407e3e3001e79a68285213e3c6a1a99a8ffe58d0b51a724488e05c7e7e6"} Nov 25 10:47:47 crc kubenswrapper[4776]: I1125 10:47:47.818592 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:47:47 crc kubenswrapper[4776]: I1125 10:47:47.818966 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:47:48 crc kubenswrapper[4776]: I1125 10:47:48.052106 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54564445dc-5wx2z" event={"ID":"ac07dcc7-2c32-40b9-802b-8380d6c2ef3e","Type":"ContainerStarted","Data":"e428da7d3af50e7504fbf05eed836a8b5431da41599a55c00486121248dd4b70"} Nov 25 10:47:48 crc kubenswrapper[4776]: I1125 10:47:48.052296 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-54564445dc-5wx2z" Nov 25 10:47:48 crc kubenswrapper[4776]: I1125 10:47:48.069540 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-54564445dc-5wx2z" podStartSLOduration=3.0695178 podStartE2EDuration="3.0695178s" podCreationTimestamp="2025-11-25 10:47:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:47:48.067032668 +0000 UTC m=+5013.108092221" watchObservedRunningTime="2025-11-25 10:47:48.0695178 +0000 UTC m=+5013.110577353" Nov 25 10:47:49 crc kubenswrapper[4776]: I1125 10:47:49.872781 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="89c88cd7-7cd2-43f8-bb00-0908d26c4a9e" containerName="rabbitmq" containerID="cri-o://51e2e393deba56f5c516ba944e0a685a84a2cb389a5ab6cf916dbed780313ffe" gracePeriod=604797 Nov 25 10:47:50 crc kubenswrapper[4776]: I1125 10:47:50.308427 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="ea47bcae-f9de-4486-93eb-8dd959573bd2" containerName="rabbitmq" containerID="cri-o://df6d6cf305233fa8e00e44108ac511a2f03ea528426db3e56a43d666fd5c53e4" gracePeriod=604797 Nov 25 10:47:51 crc kubenswrapper[4776]: I1125 10:47:51.413702 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="ea47bcae-f9de-4486-93eb-8dd959573bd2" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.247:5671: connect: connection refused" Nov 25 10:47:51 crc kubenswrapper[4776]: I1125 10:47:51.984813 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="89c88cd7-7cd2-43f8-bb00-0908d26c4a9e" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.248:5671: connect: connection refused" Nov 25 10:47:55 crc kubenswrapper[4776]: I1125 10:47:55.683017 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-54564445dc-5wx2z" Nov 25 10:47:55 crc kubenswrapper[4776]: I1125 10:47:55.740521 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59c6c64b5c-dd86r"] Nov 25 10:47:55 crc kubenswrapper[4776]: I1125 10:47:55.740797 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59c6c64b5c-dd86r" podUID="0bdc8091-6682-49da-a5dd-743dc013934d" containerName="dnsmasq-dns" containerID="cri-o://186b4dd940c3d859642772cffd1be431d22aa57aa04f371b7bce8b6627a54a91" gracePeriod=10 Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.116042 4776 generic.go:334] "Generic (PLEG): container finished" podID="0bdc8091-6682-49da-a5dd-743dc013934d" containerID="186b4dd940c3d859642772cffd1be431d22aa57aa04f371b7bce8b6627a54a91" exitCode=0 Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.116357 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59c6c64b5c-dd86r" event={"ID":"0bdc8091-6682-49da-a5dd-743dc013934d","Type":"ContainerDied","Data":"186b4dd940c3d859642772cffd1be431d22aa57aa04f371b7bce8b6627a54a91"} Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.116386 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59c6c64b5c-dd86r" event={"ID":"0bdc8091-6682-49da-a5dd-743dc013934d","Type":"ContainerDied","Data":"2d8f19e4a6b66dffd7117e5b238d820bc05e8cf75787c665f957f8b0e830aae8"} Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.116400 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d8f19e4a6b66dffd7117e5b238d820bc05e8cf75787c665f957f8b0e830aae8" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.157737 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59c6c64b5c-dd86r" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.342687 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bdc8091-6682-49da-a5dd-743dc013934d-config\") pod \"0bdc8091-6682-49da-a5dd-743dc013934d\" (UID: \"0bdc8091-6682-49da-a5dd-743dc013934d\") " Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.343132 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bdc8091-6682-49da-a5dd-743dc013934d-dns-svc\") pod \"0bdc8091-6682-49da-a5dd-743dc013934d\" (UID: \"0bdc8091-6682-49da-a5dd-743dc013934d\") " Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.343182 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsngw\" (UniqueName: \"kubernetes.io/projected/0bdc8091-6682-49da-a5dd-743dc013934d-kube-api-access-xsngw\") pod \"0bdc8091-6682-49da-a5dd-743dc013934d\" (UID: \"0bdc8091-6682-49da-a5dd-743dc013934d\") " Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.350545 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bdc8091-6682-49da-a5dd-743dc013934d-kube-api-access-xsngw" (OuterVolumeSpecName: "kube-api-access-xsngw") pod "0bdc8091-6682-49da-a5dd-743dc013934d" (UID: "0bdc8091-6682-49da-a5dd-743dc013934d"). InnerVolumeSpecName "kube-api-access-xsngw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.387663 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bdc8091-6682-49da-a5dd-743dc013934d-config" (OuterVolumeSpecName: "config") pod "0bdc8091-6682-49da-a5dd-743dc013934d" (UID: "0bdc8091-6682-49da-a5dd-743dc013934d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.394928 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0bdc8091-6682-49da-a5dd-743dc013934d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0bdc8091-6682-49da-a5dd-743dc013934d" (UID: "0bdc8091-6682-49da-a5dd-743dc013934d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.444337 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bdc8091-6682-49da-a5dd-743dc013934d-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.444454 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bdc8091-6682-49da-a5dd-743dc013934d-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.444519 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsngw\" (UniqueName: \"kubernetes.io/projected/0bdc8091-6682-49da-a5dd-743dc013934d-kube-api-access-xsngw\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.686860 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.852847 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2bcr\" (UniqueName: \"kubernetes.io/projected/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-kube-api-access-c2bcr\") pod \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.852891 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-rabbitmq-plugins\") pod \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.852927 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-erlang-cookie-secret\") pod \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.852959 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-rabbitmq-erlang-cookie\") pod \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.852983 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-config-data\") pod \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.853000 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-rabbitmq-tls\") pod \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.853035 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-pod-info\") pod \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.854194 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-181a479f-9029-4e9e-a4e1-faa600369e96\") pod \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.854204 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "89c88cd7-7cd2-43f8-bb00-0908d26c4a9e" (UID: "89c88cd7-7cd2-43f8-bb00-0908d26c4a9e"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.854326 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-server-conf\") pod \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.854376 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-rabbitmq-confd\") pod \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.854423 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-plugins-conf\") pod \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\" (UID: \"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e\") " Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.854801 4776 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.854825 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "89c88cd7-7cd2-43f8-bb00-0908d26c4a9e" (UID: "89c88cd7-7cd2-43f8-bb00-0908d26c4a9e"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.855345 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "89c88cd7-7cd2-43f8-bb00-0908d26c4a9e" (UID: "89c88cd7-7cd2-43f8-bb00-0908d26c4a9e"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.859116 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-pod-info" (OuterVolumeSpecName: "pod-info") pod "89c88cd7-7cd2-43f8-bb00-0908d26c4a9e" (UID: "89c88cd7-7cd2-43f8-bb00-0908d26c4a9e"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.859184 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "89c88cd7-7cd2-43f8-bb00-0908d26c4a9e" (UID: "89c88cd7-7cd2-43f8-bb00-0908d26c4a9e"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.862743 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "89c88cd7-7cd2-43f8-bb00-0908d26c4a9e" (UID: "89c88cd7-7cd2-43f8-bb00-0908d26c4a9e"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.863520 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-kube-api-access-c2bcr" (OuterVolumeSpecName: "kube-api-access-c2bcr") pod "89c88cd7-7cd2-43f8-bb00-0908d26c4a9e" (UID: "89c88cd7-7cd2-43f8-bb00-0908d26c4a9e"). InnerVolumeSpecName "kube-api-access-c2bcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.895909 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-181a479f-9029-4e9e-a4e1-faa600369e96" (OuterVolumeSpecName: "persistence") pod "89c88cd7-7cd2-43f8-bb00-0908d26c4a9e" (UID: "89c88cd7-7cd2-43f8-bb00-0908d26c4a9e"). InnerVolumeSpecName "pvc-181a479f-9029-4e9e-a4e1-faa600369e96". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.902392 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-config-data" (OuterVolumeSpecName: "config-data") pod "89c88cd7-7cd2-43f8-bb00-0908d26c4a9e" (UID: "89c88cd7-7cd2-43f8-bb00-0908d26c4a9e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.908481 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-server-conf" (OuterVolumeSpecName: "server-conf") pod "89c88cd7-7cd2-43f8-bb00-0908d26c4a9e" (UID: "89c88cd7-7cd2-43f8-bb00-0908d26c4a9e"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.952005 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "89c88cd7-7cd2-43f8-bb00-0908d26c4a9e" (UID: "89c88cd7-7cd2-43f8-bb00-0908d26c4a9e"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.956805 4776 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-server-conf\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.956834 4776 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.956844 4776 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.956854 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2bcr\" (UniqueName: \"kubernetes.io/projected/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-kube-api-access-c2bcr\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.956865 4776 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.956874 4776 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.956882 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.956890 4776 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.956898 4776 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e-pod-info\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.956928 4776 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-181a479f-9029-4e9e-a4e1-faa600369e96\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-181a479f-9029-4e9e-a4e1-faa600369e96\") on node \"crc\" " Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.973170 4776 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 25 10:47:56 crc kubenswrapper[4776]: I1125 10:47:56.973311 4776 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-181a479f-9029-4e9e-a4e1-faa600369e96" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-181a479f-9029-4e9e-a4e1-faa600369e96") on node "crc" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.002126 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.058665 4776 reconciler_common.go:293] "Volume detached for volume \"pvc-181a479f-9029-4e9e-a4e1-faa600369e96\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-181a479f-9029-4e9e-a4e1-faa600369e96\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.126995 4776 generic.go:334] "Generic (PLEG): container finished" podID="ea47bcae-f9de-4486-93eb-8dd959573bd2" containerID="df6d6cf305233fa8e00e44108ac511a2f03ea528426db3e56a43d666fd5c53e4" exitCode=0 Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.127092 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ea47bcae-f9de-4486-93eb-8dd959573bd2","Type":"ContainerDied","Data":"df6d6cf305233fa8e00e44108ac511a2f03ea528426db3e56a43d666fd5c53e4"} Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.127118 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ea47bcae-f9de-4486-93eb-8dd959573bd2","Type":"ContainerDied","Data":"9f3b9a0fa9d5d030967d7a1af74f0da7d30b115c2c0f1c6bb6a398d12deb5b4f"} Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.127134 4776 scope.go:117] "RemoveContainer" containerID="df6d6cf305233fa8e00e44108ac511a2f03ea528426db3e56a43d666fd5c53e4" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.127229 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.129690 4776 generic.go:334] "Generic (PLEG): container finished" podID="89c88cd7-7cd2-43f8-bb00-0908d26c4a9e" containerID="51e2e393deba56f5c516ba944e0a685a84a2cb389a5ab6cf916dbed780313ffe" exitCode=0 Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.129774 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59c6c64b5c-dd86r" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.130262 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e","Type":"ContainerDied","Data":"51e2e393deba56f5c516ba944e0a685a84a2cb389a5ab6cf916dbed780313ffe"} Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.130322 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.130331 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"89c88cd7-7cd2-43f8-bb00-0908d26c4a9e","Type":"ContainerDied","Data":"36a01b98e90356b2e66b08dda7acf48f19d9fb68b86725f1a836fbc4ae1b8585"} Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.150025 4776 scope.go:117] "RemoveContainer" containerID="5bd67c8d057c8f9c8a1eee8383f029cacd61b21bdacb1a5c0b25926cbc4ae826" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.161146 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ea47bcae-f9de-4486-93eb-8dd959573bd2-rabbitmq-confd\") pod \"ea47bcae-f9de-4486-93eb-8dd959573bd2\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.161211 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ea47bcae-f9de-4486-93eb-8dd959573bd2-pod-info\") pod \"ea47bcae-f9de-4486-93eb-8dd959573bd2\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.161274 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea47bcae-f9de-4486-93eb-8dd959573bd2-config-data\") pod \"ea47bcae-f9de-4486-93eb-8dd959573bd2\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.161320 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjxjw\" (UniqueName: \"kubernetes.io/projected/ea47bcae-f9de-4486-93eb-8dd959573bd2-kube-api-access-sjxjw\") pod \"ea47bcae-f9de-4486-93eb-8dd959573bd2\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.162233 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-947e4d82-272e-4d92-87b2-c028bbea2317\") pod \"ea47bcae-f9de-4486-93eb-8dd959573bd2\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.162288 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ea47bcae-f9de-4486-93eb-8dd959573bd2-erlang-cookie-secret\") pod \"ea47bcae-f9de-4486-93eb-8dd959573bd2\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.162336 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ea47bcae-f9de-4486-93eb-8dd959573bd2-plugins-conf\") pod \"ea47bcae-f9de-4486-93eb-8dd959573bd2\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.162366 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ea47bcae-f9de-4486-93eb-8dd959573bd2-rabbitmq-plugins\") pod \"ea47bcae-f9de-4486-93eb-8dd959573bd2\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.162419 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ea47bcae-f9de-4486-93eb-8dd959573bd2-server-conf\") pod \"ea47bcae-f9de-4486-93eb-8dd959573bd2\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.162460 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ea47bcae-f9de-4486-93eb-8dd959573bd2-rabbitmq-erlang-cookie\") pod \"ea47bcae-f9de-4486-93eb-8dd959573bd2\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.162510 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ea47bcae-f9de-4486-93eb-8dd959573bd2-rabbitmq-tls\") pod \"ea47bcae-f9de-4486-93eb-8dd959573bd2\" (UID: \"ea47bcae-f9de-4486-93eb-8dd959573bd2\") " Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.163853 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea47bcae-f9de-4486-93eb-8dd959573bd2-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "ea47bcae-f9de-4486-93eb-8dd959573bd2" (UID: "ea47bcae-f9de-4486-93eb-8dd959573bd2"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.164205 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea47bcae-f9de-4486-93eb-8dd959573bd2-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "ea47bcae-f9de-4486-93eb-8dd959573bd2" (UID: "ea47bcae-f9de-4486-93eb-8dd959573bd2"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.164320 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea47bcae-f9de-4486-93eb-8dd959573bd2-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "ea47bcae-f9de-4486-93eb-8dd959573bd2" (UID: "ea47bcae-f9de-4486-93eb-8dd959573bd2"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.167754 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea47bcae-f9de-4486-93eb-8dd959573bd2-kube-api-access-sjxjw" (OuterVolumeSpecName: "kube-api-access-sjxjw") pod "ea47bcae-f9de-4486-93eb-8dd959573bd2" (UID: "ea47bcae-f9de-4486-93eb-8dd959573bd2"). InnerVolumeSpecName "kube-api-access-sjxjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.168369 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea47bcae-f9de-4486-93eb-8dd959573bd2-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "ea47bcae-f9de-4486-93eb-8dd959573bd2" (UID: "ea47bcae-f9de-4486-93eb-8dd959573bd2"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.168971 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/ea47bcae-f9de-4486-93eb-8dd959573bd2-pod-info" (OuterVolumeSpecName: "pod-info") pod "ea47bcae-f9de-4486-93eb-8dd959573bd2" (UID: "ea47bcae-f9de-4486-93eb-8dd959573bd2"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.179470 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea47bcae-f9de-4486-93eb-8dd959573bd2-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "ea47bcae-f9de-4486-93eb-8dd959573bd2" (UID: "ea47bcae-f9de-4486-93eb-8dd959573bd2"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.181469 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-947e4d82-272e-4d92-87b2-c028bbea2317" (OuterVolumeSpecName: "persistence") pod "ea47bcae-f9de-4486-93eb-8dd959573bd2" (UID: "ea47bcae-f9de-4486-93eb-8dd959573bd2"). InnerVolumeSpecName "pvc-947e4d82-272e-4d92-87b2-c028bbea2317". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.189519 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59c6c64b5c-dd86r"] Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.200208 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59c6c64b5c-dd86r"] Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.202368 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea47bcae-f9de-4486-93eb-8dd959573bd2-config-data" (OuterVolumeSpecName: "config-data") pod "ea47bcae-f9de-4486-93eb-8dd959573bd2" (UID: "ea47bcae-f9de-4486-93eb-8dd959573bd2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.211012 4776 scope.go:117] "RemoveContainer" containerID="df6d6cf305233fa8e00e44108ac511a2f03ea528426db3e56a43d666fd5c53e4" Nov 25 10:47:57 crc kubenswrapper[4776]: E1125 10:47:57.212015 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df6d6cf305233fa8e00e44108ac511a2f03ea528426db3e56a43d666fd5c53e4\": container with ID starting with df6d6cf305233fa8e00e44108ac511a2f03ea528426db3e56a43d666fd5c53e4 not found: ID does not exist" containerID="df6d6cf305233fa8e00e44108ac511a2f03ea528426db3e56a43d666fd5c53e4" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.212057 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df6d6cf305233fa8e00e44108ac511a2f03ea528426db3e56a43d666fd5c53e4"} err="failed to get container status \"df6d6cf305233fa8e00e44108ac511a2f03ea528426db3e56a43d666fd5c53e4\": rpc error: code = NotFound desc = could not find container \"df6d6cf305233fa8e00e44108ac511a2f03ea528426db3e56a43d666fd5c53e4\": container with ID starting with df6d6cf305233fa8e00e44108ac511a2f03ea528426db3e56a43d666fd5c53e4 not found: ID does not exist" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.212094 4776 scope.go:117] "RemoveContainer" containerID="5bd67c8d057c8f9c8a1eee8383f029cacd61b21bdacb1a5c0b25926cbc4ae826" Nov 25 10:47:57 crc kubenswrapper[4776]: E1125 10:47:57.212363 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bd67c8d057c8f9c8a1eee8383f029cacd61b21bdacb1a5c0b25926cbc4ae826\": container with ID starting with 5bd67c8d057c8f9c8a1eee8383f029cacd61b21bdacb1a5c0b25926cbc4ae826 not found: ID does not exist" containerID="5bd67c8d057c8f9c8a1eee8383f029cacd61b21bdacb1a5c0b25926cbc4ae826" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.212388 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bd67c8d057c8f9c8a1eee8383f029cacd61b21bdacb1a5c0b25926cbc4ae826"} err="failed to get container status \"5bd67c8d057c8f9c8a1eee8383f029cacd61b21bdacb1a5c0b25926cbc4ae826\": rpc error: code = NotFound desc = could not find container \"5bd67c8d057c8f9c8a1eee8383f029cacd61b21bdacb1a5c0b25926cbc4ae826\": container with ID starting with 5bd67c8d057c8f9c8a1eee8383f029cacd61b21bdacb1a5c0b25926cbc4ae826 not found: ID does not exist" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.212410 4776 scope.go:117] "RemoveContainer" containerID="51e2e393deba56f5c516ba944e0a685a84a2cb389a5ab6cf916dbed780313ffe" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.224374 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.231990 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.239100 4776 scope.go:117] "RemoveContainer" containerID="bc062b2a63abbd08250341b4b871dd718116d42e8b4841c259c448e93b33e490" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.242590 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea47bcae-f9de-4486-93eb-8dd959573bd2-server-conf" (OuterVolumeSpecName: "server-conf") pod "ea47bcae-f9de-4486-93eb-8dd959573bd2" (UID: "ea47bcae-f9de-4486-93eb-8dd959573bd2"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.250244 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:47:57 crc kubenswrapper[4776]: E1125 10:47:57.250641 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89c88cd7-7cd2-43f8-bb00-0908d26c4a9e" containerName="rabbitmq" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.250658 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="89c88cd7-7cd2-43f8-bb00-0908d26c4a9e" containerName="rabbitmq" Nov 25 10:47:57 crc kubenswrapper[4776]: E1125 10:47:57.250674 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea47bcae-f9de-4486-93eb-8dd959573bd2" containerName="rabbitmq" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.250681 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea47bcae-f9de-4486-93eb-8dd959573bd2" containerName="rabbitmq" Nov 25 10:47:57 crc kubenswrapper[4776]: E1125 10:47:57.250697 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bdc8091-6682-49da-a5dd-743dc013934d" containerName="dnsmasq-dns" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.250705 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bdc8091-6682-49da-a5dd-743dc013934d" containerName="dnsmasq-dns" Nov 25 10:47:57 crc kubenswrapper[4776]: E1125 10:47:57.250715 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea47bcae-f9de-4486-93eb-8dd959573bd2" containerName="setup-container" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.250723 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea47bcae-f9de-4486-93eb-8dd959573bd2" containerName="setup-container" Nov 25 10:47:57 crc kubenswrapper[4776]: E1125 10:47:57.250735 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89c88cd7-7cd2-43f8-bb00-0908d26c4a9e" containerName="setup-container" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.250741 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="89c88cd7-7cd2-43f8-bb00-0908d26c4a9e" containerName="setup-container" Nov 25 10:47:57 crc kubenswrapper[4776]: E1125 10:47:57.250762 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bdc8091-6682-49da-a5dd-743dc013934d" containerName="init" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.250768 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bdc8091-6682-49da-a5dd-743dc013934d" containerName="init" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.250924 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="89c88cd7-7cd2-43f8-bb00-0908d26c4a9e" containerName="rabbitmq" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.250938 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bdc8091-6682-49da-a5dd-743dc013934d" containerName="dnsmasq-dns" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.250950 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea47bcae-f9de-4486-93eb-8dd959573bd2" containerName="rabbitmq" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.251768 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.253584 4776 scope.go:117] "RemoveContainer" containerID="51e2e393deba56f5c516ba944e0a685a84a2cb389a5ab6cf916dbed780313ffe" Nov 25 10:47:57 crc kubenswrapper[4776]: E1125 10:47:57.254012 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51e2e393deba56f5c516ba944e0a685a84a2cb389a5ab6cf916dbed780313ffe\": container with ID starting with 51e2e393deba56f5c516ba944e0a685a84a2cb389a5ab6cf916dbed780313ffe not found: ID does not exist" containerID="51e2e393deba56f5c516ba944e0a685a84a2cb389a5ab6cf916dbed780313ffe" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.254048 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51e2e393deba56f5c516ba944e0a685a84a2cb389a5ab6cf916dbed780313ffe"} err="failed to get container status \"51e2e393deba56f5c516ba944e0a685a84a2cb389a5ab6cf916dbed780313ffe\": rpc error: code = NotFound desc = could not find container \"51e2e393deba56f5c516ba944e0a685a84a2cb389a5ab6cf916dbed780313ffe\": container with ID starting with 51e2e393deba56f5c516ba944e0a685a84a2cb389a5ab6cf916dbed780313ffe not found: ID does not exist" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.254565 4776 scope.go:117] "RemoveContainer" containerID="bc062b2a63abbd08250341b4b871dd718116d42e8b4841c259c448e93b33e490" Nov 25 10:47:57 crc kubenswrapper[4776]: E1125 10:47:57.254887 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc062b2a63abbd08250341b4b871dd718116d42e8b4841c259c448e93b33e490\": container with ID starting with bc062b2a63abbd08250341b4b871dd718116d42e8b4841c259c448e93b33e490 not found: ID does not exist" containerID="bc062b2a63abbd08250341b4b871dd718116d42e8b4841c259c448e93b33e490" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.254953 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc062b2a63abbd08250341b4b871dd718116d42e8b4841c259c448e93b33e490"} err="failed to get container status \"bc062b2a63abbd08250341b4b871dd718116d42e8b4841c259c448e93b33e490\": rpc error: code = NotFound desc = could not find container \"bc062b2a63abbd08250341b4b871dd718116d42e8b4841c259c448e93b33e490\": container with ID starting with bc062b2a63abbd08250341b4b871dd718116d42e8b4841c259c448e93b33e490 not found: ID does not exist" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.255303 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.255338 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-gwftv" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.255428 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.256989 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.258107 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.258300 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.258330 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.260675 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.264329 4776 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ea47bcae-f9de-4486-93eb-8dd959573bd2-pod-info\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.264354 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea47bcae-f9de-4486-93eb-8dd959573bd2-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.264369 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjxjw\" (UniqueName: \"kubernetes.io/projected/ea47bcae-f9de-4486-93eb-8dd959573bd2-kube-api-access-sjxjw\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.264401 4776 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-947e4d82-272e-4d92-87b2-c028bbea2317\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-947e4d82-272e-4d92-87b2-c028bbea2317\") on node \"crc\" " Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.264418 4776 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ea47bcae-f9de-4486-93eb-8dd959573bd2-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.264432 4776 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ea47bcae-f9de-4486-93eb-8dd959573bd2-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.264444 4776 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ea47bcae-f9de-4486-93eb-8dd959573bd2-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.264457 4776 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ea47bcae-f9de-4486-93eb-8dd959573bd2-server-conf\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.264468 4776 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ea47bcae-f9de-4486-93eb-8dd959573bd2-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.264480 4776 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ea47bcae-f9de-4486-93eb-8dd959573bd2-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.286029 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea47bcae-f9de-4486-93eb-8dd959573bd2-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "ea47bcae-f9de-4486-93eb-8dd959573bd2" (UID: "ea47bcae-f9de-4486-93eb-8dd959573bd2"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.289148 4776 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.289345 4776 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-947e4d82-272e-4d92-87b2-c028bbea2317" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-947e4d82-272e-4d92-87b2-c028bbea2317") on node "crc" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.366959 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.367082 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-pod-info\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.367150 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.367181 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.367241 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-181a479f-9029-4e9e-a4e1-faa600369e96\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-181a479f-9029-4e9e-a4e1-faa600369e96\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.367288 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-server-conf\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.367366 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.367398 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-config-data\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.367443 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wt4xl\" (UniqueName: \"kubernetes.io/projected/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-kube-api-access-wt4xl\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.367530 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.367609 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.367790 4776 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ea47bcae-f9de-4486-93eb-8dd959573bd2-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.367810 4776 reconciler_common.go:293] "Volume detached for volume \"pvc-947e4d82-272e-4d92-87b2-c028bbea2317\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-947e4d82-272e-4d92-87b2-c028bbea2317\") on node \"crc\" DevicePath \"\"" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.470294 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.470375 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.470411 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.470444 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-pod-info\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.470503 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.470805 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.470842 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-181a479f-9029-4e9e-a4e1-faa600369e96\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-181a479f-9029-4e9e-a4e1-faa600369e96\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.471357 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-server-conf\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.471430 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.471464 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-config-data\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.471506 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wt4xl\" (UniqueName: \"kubernetes.io/projected/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-kube-api-access-wt4xl\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.471921 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.472795 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.472818 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-config-data\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.473313 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.474774 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.474801 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.474833 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.475341 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-server-conf\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.476475 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.478191 4776 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.478277 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-181a479f-9029-4e9e-a4e1-faa600369e96\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-181a479f-9029-4e9e-a4e1-faa600369e96\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a722d3c24f2b44f3b5ae75c49f6505dc76718dc72741f13e90c9eddd35738e5d/globalmount\"" pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.479553 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-pod-info\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.483908 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.493550 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wt4xl\" (UniqueName: \"kubernetes.io/projected/188d5ae0-1ff7-44fd-b0db-5500b52f2b63-kube-api-access-wt4xl\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.514786 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.516525 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.522926 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.523225 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.523327 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-g7vlf" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.523782 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.524026 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.524310 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.527728 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.532236 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.543103 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-181a479f-9029-4e9e-a4e1-faa600369e96\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-181a479f-9029-4e9e-a4e1-faa600369e96\") pod \"rabbitmq-server-0\" (UID: \"188d5ae0-1ff7-44fd-b0db-5500b52f2b63\") " pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.570231 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.699518 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-947e4d82-272e-4d92-87b2-c028bbea2317\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-947e4d82-272e-4d92-87b2-c028bbea2317\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.699623 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/147a21c0-308e-4792-ae3e-bfc852327d0b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.699710 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/147a21c0-308e-4792-ae3e-bfc852327d0b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.699767 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/147a21c0-308e-4792-ae3e-bfc852327d0b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.700253 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/147a21c0-308e-4792-ae3e-bfc852327d0b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.700292 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2ghm\" (UniqueName: \"kubernetes.io/projected/147a21c0-308e-4792-ae3e-bfc852327d0b-kube-api-access-m2ghm\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.700349 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/147a21c0-308e-4792-ae3e-bfc852327d0b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.700375 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/147a21c0-308e-4792-ae3e-bfc852327d0b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.700412 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/147a21c0-308e-4792-ae3e-bfc852327d0b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.700461 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/147a21c0-308e-4792-ae3e-bfc852327d0b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.700494 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/147a21c0-308e-4792-ae3e-bfc852327d0b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.731350 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bdc8091-6682-49da-a5dd-743dc013934d" path="/var/lib/kubelet/pods/0bdc8091-6682-49da-a5dd-743dc013934d/volumes" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.732403 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89c88cd7-7cd2-43f8-bb00-0908d26c4a9e" path="/var/lib/kubelet/pods/89c88cd7-7cd2-43f8-bb00-0908d26c4a9e/volumes" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.733745 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea47bcae-f9de-4486-93eb-8dd959573bd2" path="/var/lib/kubelet/pods/ea47bcae-f9de-4486-93eb-8dd959573bd2/volumes" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.802696 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/147a21c0-308e-4792-ae3e-bfc852327d0b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.802764 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2ghm\" (UniqueName: \"kubernetes.io/projected/147a21c0-308e-4792-ae3e-bfc852327d0b-kube-api-access-m2ghm\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.802845 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/147a21c0-308e-4792-ae3e-bfc852327d0b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.802867 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/147a21c0-308e-4792-ae3e-bfc852327d0b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.802899 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/147a21c0-308e-4792-ae3e-bfc852327d0b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.802924 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/147a21c0-308e-4792-ae3e-bfc852327d0b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.802943 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/147a21c0-308e-4792-ae3e-bfc852327d0b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.802994 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-947e4d82-272e-4d92-87b2-c028bbea2317\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-947e4d82-272e-4d92-87b2-c028bbea2317\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.803025 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/147a21c0-308e-4792-ae3e-bfc852327d0b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.803093 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/147a21c0-308e-4792-ae3e-bfc852327d0b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.803122 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/147a21c0-308e-4792-ae3e-bfc852327d0b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.804332 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/147a21c0-308e-4792-ae3e-bfc852327d0b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.804508 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/147a21c0-308e-4792-ae3e-bfc852327d0b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.804831 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/147a21c0-308e-4792-ae3e-bfc852327d0b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.805050 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/147a21c0-308e-4792-ae3e-bfc852327d0b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.807408 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/147a21c0-308e-4792-ae3e-bfc852327d0b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.808254 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/147a21c0-308e-4792-ae3e-bfc852327d0b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.808356 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/147a21c0-308e-4792-ae3e-bfc852327d0b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.808753 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/147a21c0-308e-4792-ae3e-bfc852327d0b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.809684 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/147a21c0-308e-4792-ae3e-bfc852327d0b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.810320 4776 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.810345 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-947e4d82-272e-4d92-87b2-c028bbea2317\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-947e4d82-272e-4d92-87b2-c028bbea2317\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/fe5a02d6afa727d9853c90ed2daf9d081979f7c5457f86edf7a6da3605d7b0a0/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.822844 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2ghm\" (UniqueName: \"kubernetes.io/projected/147a21c0-308e-4792-ae3e-bfc852327d0b-kube-api-access-m2ghm\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:57 crc kubenswrapper[4776]: I1125 10:47:57.838607 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-947e4d82-272e-4d92-87b2-c028bbea2317\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-947e4d82-272e-4d92-87b2-c028bbea2317\") pod \"rabbitmq-cell1-server-0\" (UID: \"147a21c0-308e-4792-ae3e-bfc852327d0b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:58 crc kubenswrapper[4776]: I1125 10:47:58.029115 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:47:58 crc kubenswrapper[4776]: I1125 10:47:58.137395 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:47:58 crc kubenswrapper[4776]: I1125 10:47:58.138197 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"188d5ae0-1ff7-44fd-b0db-5500b52f2b63","Type":"ContainerStarted","Data":"5d2b7f8e49988436b342e6bbd8d1d1d80245369e6c58e415dce5ebdd9f4eea22"} Nov 25 10:47:58 crc kubenswrapper[4776]: I1125 10:47:58.344901 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:47:58 crc kubenswrapper[4776]: W1125 10:47:58.352142 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod147a21c0_308e_4792_ae3e_bfc852327d0b.slice/crio-8369917ba0249b5a384160e786b734bb5516d87a014a58c56afc21b38c9fedf8 WatchSource:0}: Error finding container 8369917ba0249b5a384160e786b734bb5516d87a014a58c56afc21b38c9fedf8: Status 404 returned error can't find the container with id 8369917ba0249b5a384160e786b734bb5516d87a014a58c56afc21b38c9fedf8 Nov 25 10:47:59 crc kubenswrapper[4776]: I1125 10:47:59.152252 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"147a21c0-308e-4792-ae3e-bfc852327d0b","Type":"ContainerStarted","Data":"8369917ba0249b5a384160e786b734bb5516d87a014a58c56afc21b38c9fedf8"} Nov 25 10:48:00 crc kubenswrapper[4776]: I1125 10:48:00.163932 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"188d5ae0-1ff7-44fd-b0db-5500b52f2b63","Type":"ContainerStarted","Data":"3cf4e18161cd09c00f08106ebb2e78d76a201dd09b4fdeaa67e5e56515eef281"} Nov 25 10:48:02 crc kubenswrapper[4776]: I1125 10:48:02.183802 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"147a21c0-308e-4792-ae3e-bfc852327d0b","Type":"ContainerStarted","Data":"97369e4d6eecbd9d180eacfeeb704b8350ed5c1b660b88bec06f249112b26093"} Nov 25 10:48:17 crc kubenswrapper[4776]: I1125 10:48:17.819046 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:48:17 crc kubenswrapper[4776]: I1125 10:48:17.819610 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:48:32 crc kubenswrapper[4776]: I1125 10:48:32.435732 4776 generic.go:334] "Generic (PLEG): container finished" podID="188d5ae0-1ff7-44fd-b0db-5500b52f2b63" containerID="3cf4e18161cd09c00f08106ebb2e78d76a201dd09b4fdeaa67e5e56515eef281" exitCode=0 Nov 25 10:48:32 crc kubenswrapper[4776]: I1125 10:48:32.435839 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"188d5ae0-1ff7-44fd-b0db-5500b52f2b63","Type":"ContainerDied","Data":"3cf4e18161cd09c00f08106ebb2e78d76a201dd09b4fdeaa67e5e56515eef281"} Nov 25 10:48:33 crc kubenswrapper[4776]: I1125 10:48:33.446421 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"188d5ae0-1ff7-44fd-b0db-5500b52f2b63","Type":"ContainerStarted","Data":"8025cd92b821ff9f8aa741317d6a135f1e0a4a3b859dd67017824d395a7d2c53"} Nov 25 10:48:33 crc kubenswrapper[4776]: I1125 10:48:33.446994 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 25 10:48:33 crc kubenswrapper[4776]: I1125 10:48:33.473617 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.47359965 podStartE2EDuration="36.47359965s" podCreationTimestamp="2025-11-25 10:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:48:33.47040516 +0000 UTC m=+5058.511464713" watchObservedRunningTime="2025-11-25 10:48:33.47359965 +0000 UTC m=+5058.514659203" Nov 25 10:48:34 crc kubenswrapper[4776]: I1125 10:48:34.453875 4776 generic.go:334] "Generic (PLEG): container finished" podID="147a21c0-308e-4792-ae3e-bfc852327d0b" containerID="97369e4d6eecbd9d180eacfeeb704b8350ed5c1b660b88bec06f249112b26093" exitCode=0 Nov 25 10:48:34 crc kubenswrapper[4776]: I1125 10:48:34.453974 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"147a21c0-308e-4792-ae3e-bfc852327d0b","Type":"ContainerDied","Data":"97369e4d6eecbd9d180eacfeeb704b8350ed5c1b660b88bec06f249112b26093"} Nov 25 10:48:35 crc kubenswrapper[4776]: I1125 10:48:35.466903 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"147a21c0-308e-4792-ae3e-bfc852327d0b","Type":"ContainerStarted","Data":"6813ef9ea172834308e2fee188dfa19beb958b5acfbdcb3cee1d07805f93e0a5"} Nov 25 10:48:35 crc kubenswrapper[4776]: I1125 10:48:35.468896 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:48:35 crc kubenswrapper[4776]: I1125 10:48:35.514788 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.514767275 podStartE2EDuration="38.514767275s" podCreationTimestamp="2025-11-25 10:47:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:48:35.510633121 +0000 UTC m=+5060.551692674" watchObservedRunningTime="2025-11-25 10:48:35.514767275 +0000 UTC m=+5060.555826828" Nov 25 10:48:47 crc kubenswrapper[4776]: I1125 10:48:47.574437 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 25 10:48:47 crc kubenswrapper[4776]: I1125 10:48:47.820369 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:48:47 crc kubenswrapper[4776]: I1125 10:48:47.820437 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:48:47 crc kubenswrapper[4776]: I1125 10:48:47.820487 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 10:48:47 crc kubenswrapper[4776]: I1125 10:48:47.821182 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"62f6bd2b248ceb776fb1c929efd14083d136607e3bcf021f3f4b235fbcacf0e1"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:48:47 crc kubenswrapper[4776]: I1125 10:48:47.821248 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://62f6bd2b248ceb776fb1c929efd14083d136607e3bcf021f3f4b235fbcacf0e1" gracePeriod=600 Nov 25 10:48:48 crc kubenswrapper[4776]: I1125 10:48:48.140265 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:48:48 crc kubenswrapper[4776]: I1125 10:48:48.566685 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="62f6bd2b248ceb776fb1c929efd14083d136607e3bcf021f3f4b235fbcacf0e1" exitCode=0 Nov 25 10:48:48 crc kubenswrapper[4776]: I1125 10:48:48.567015 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"62f6bd2b248ceb776fb1c929efd14083d136607e3bcf021f3f4b235fbcacf0e1"} Nov 25 10:48:48 crc kubenswrapper[4776]: I1125 10:48:48.567081 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40"} Nov 25 10:48:48 crc kubenswrapper[4776]: I1125 10:48:48.567104 4776 scope.go:117] "RemoveContainer" containerID="f5edfd4dcfe1ba26fdb152fd9eb70ec7fc477dea18dcf834e0f7e4182600d22a" Nov 25 10:48:52 crc kubenswrapper[4776]: I1125 10:48:52.210689 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1-default"] Nov 25 10:48:52 crc kubenswrapper[4776]: I1125 10:48:52.212361 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Nov 25 10:48:52 crc kubenswrapper[4776]: I1125 10:48:52.214737 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-rdbsh" Nov 25 10:48:52 crc kubenswrapper[4776]: I1125 10:48:52.218813 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Nov 25 10:48:52 crc kubenswrapper[4776]: I1125 10:48:52.377895 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhf8s\" (UniqueName: \"kubernetes.io/projected/87be33d4-2959-401d-971c-ec6cacdff3d7-kube-api-access-hhf8s\") pod \"mariadb-client-1-default\" (UID: \"87be33d4-2959-401d-971c-ec6cacdff3d7\") " pod="openstack/mariadb-client-1-default" Nov 25 10:48:52 crc kubenswrapper[4776]: I1125 10:48:52.479426 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhf8s\" (UniqueName: \"kubernetes.io/projected/87be33d4-2959-401d-971c-ec6cacdff3d7-kube-api-access-hhf8s\") pod \"mariadb-client-1-default\" (UID: \"87be33d4-2959-401d-971c-ec6cacdff3d7\") " pod="openstack/mariadb-client-1-default" Nov 25 10:48:52 crc kubenswrapper[4776]: I1125 10:48:52.502235 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhf8s\" (UniqueName: \"kubernetes.io/projected/87be33d4-2959-401d-971c-ec6cacdff3d7-kube-api-access-hhf8s\") pod \"mariadb-client-1-default\" (UID: \"87be33d4-2959-401d-971c-ec6cacdff3d7\") " pod="openstack/mariadb-client-1-default" Nov 25 10:48:52 crc kubenswrapper[4776]: I1125 10:48:52.530381 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Nov 25 10:48:53 crc kubenswrapper[4776]: I1125 10:48:53.072654 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Nov 25 10:48:53 crc kubenswrapper[4776]: I1125 10:48:53.080708 4776 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 10:48:53 crc kubenswrapper[4776]: I1125 10:48:53.641918 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"87be33d4-2959-401d-971c-ec6cacdff3d7","Type":"ContainerStarted","Data":"3b3584655fb08e5e588a033ddac0a457b4650c79f0e07d5d6c2e1e4c80756e1e"} Nov 25 10:48:54 crc kubenswrapper[4776]: I1125 10:48:54.656125 4776 generic.go:334] "Generic (PLEG): container finished" podID="87be33d4-2959-401d-971c-ec6cacdff3d7" containerID="f33011cd57579d5e58ba05f93cc43fc0a37b7edb4a30f34f09f4838c8957456c" exitCode=0 Nov 25 10:48:54 crc kubenswrapper[4776]: I1125 10:48:54.656308 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"87be33d4-2959-401d-971c-ec6cacdff3d7","Type":"ContainerDied","Data":"f33011cd57579d5e58ba05f93cc43fc0a37b7edb4a30f34f09f4838c8957456c"} Nov 25 10:48:56 crc kubenswrapper[4776]: I1125 10:48:56.018607 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Nov 25 10:48:56 crc kubenswrapper[4776]: I1125 10:48:56.045744 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1-default_87be33d4-2959-401d-971c-ec6cacdff3d7/mariadb-client-1-default/0.log" Nov 25 10:48:56 crc kubenswrapper[4776]: I1125 10:48:56.068498 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1-default"] Nov 25 10:48:56 crc kubenswrapper[4776]: I1125 10:48:56.077177 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1-default"] Nov 25 10:48:56 crc kubenswrapper[4776]: I1125 10:48:56.141337 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhf8s\" (UniqueName: \"kubernetes.io/projected/87be33d4-2959-401d-971c-ec6cacdff3d7-kube-api-access-hhf8s\") pod \"87be33d4-2959-401d-971c-ec6cacdff3d7\" (UID: \"87be33d4-2959-401d-971c-ec6cacdff3d7\") " Nov 25 10:48:56 crc kubenswrapper[4776]: I1125 10:48:56.147412 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87be33d4-2959-401d-971c-ec6cacdff3d7-kube-api-access-hhf8s" (OuterVolumeSpecName: "kube-api-access-hhf8s") pod "87be33d4-2959-401d-971c-ec6cacdff3d7" (UID: "87be33d4-2959-401d-971c-ec6cacdff3d7"). InnerVolumeSpecName "kube-api-access-hhf8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:48:56 crc kubenswrapper[4776]: I1125 10:48:56.243826 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhf8s\" (UniqueName: \"kubernetes.io/projected/87be33d4-2959-401d-971c-ec6cacdff3d7-kube-api-access-hhf8s\") on node \"crc\" DevicePath \"\"" Nov 25 10:48:56 crc kubenswrapper[4776]: I1125 10:48:56.467802 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2-default"] Nov 25 10:48:56 crc kubenswrapper[4776]: E1125 10:48:56.468278 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87be33d4-2959-401d-971c-ec6cacdff3d7" containerName="mariadb-client-1-default" Nov 25 10:48:56 crc kubenswrapper[4776]: I1125 10:48:56.468298 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="87be33d4-2959-401d-971c-ec6cacdff3d7" containerName="mariadb-client-1-default" Nov 25 10:48:56 crc kubenswrapper[4776]: I1125 10:48:56.468452 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="87be33d4-2959-401d-971c-ec6cacdff3d7" containerName="mariadb-client-1-default" Nov 25 10:48:56 crc kubenswrapper[4776]: I1125 10:48:56.469005 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Nov 25 10:48:56 crc kubenswrapper[4776]: I1125 10:48:56.477789 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Nov 25 10:48:56 crc kubenswrapper[4776]: I1125 10:48:56.648549 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-678tg\" (UniqueName: \"kubernetes.io/projected/ab912de2-b375-412b-a212-2c063f2ff622-kube-api-access-678tg\") pod \"mariadb-client-2-default\" (UID: \"ab912de2-b375-412b-a212-2c063f2ff622\") " pod="openstack/mariadb-client-2-default" Nov 25 10:48:56 crc kubenswrapper[4776]: I1125 10:48:56.671602 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b3584655fb08e5e588a033ddac0a457b4650c79f0e07d5d6c2e1e4c80756e1e" Nov 25 10:48:56 crc kubenswrapper[4776]: I1125 10:48:56.671646 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Nov 25 10:48:56 crc kubenswrapper[4776]: I1125 10:48:56.750413 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-678tg\" (UniqueName: \"kubernetes.io/projected/ab912de2-b375-412b-a212-2c063f2ff622-kube-api-access-678tg\") pod \"mariadb-client-2-default\" (UID: \"ab912de2-b375-412b-a212-2c063f2ff622\") " pod="openstack/mariadb-client-2-default" Nov 25 10:48:56 crc kubenswrapper[4776]: I1125 10:48:56.772325 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-678tg\" (UniqueName: \"kubernetes.io/projected/ab912de2-b375-412b-a212-2c063f2ff622-kube-api-access-678tg\") pod \"mariadb-client-2-default\" (UID: \"ab912de2-b375-412b-a212-2c063f2ff622\") " pod="openstack/mariadb-client-2-default" Nov 25 10:48:56 crc kubenswrapper[4776]: I1125 10:48:56.793566 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Nov 25 10:48:57 crc kubenswrapper[4776]: I1125 10:48:57.307815 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Nov 25 10:48:57 crc kubenswrapper[4776]: W1125 10:48:57.316619 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab912de2_b375_412b_a212_2c063f2ff622.slice/crio-016fa829f5eaa30a8a8f2b2f4e7cbc74e4f1c2dedddc5c6afbf62e1b22b38e5b WatchSource:0}: Error finding container 016fa829f5eaa30a8a8f2b2f4e7cbc74e4f1c2dedddc5c6afbf62e1b22b38e5b: Status 404 returned error can't find the container with id 016fa829f5eaa30a8a8f2b2f4e7cbc74e4f1c2dedddc5c6afbf62e1b22b38e5b Nov 25 10:48:57 crc kubenswrapper[4776]: I1125 10:48:57.676294 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87be33d4-2959-401d-971c-ec6cacdff3d7" path="/var/lib/kubelet/pods/87be33d4-2959-401d-971c-ec6cacdff3d7/volumes" Nov 25 10:48:57 crc kubenswrapper[4776]: I1125 10:48:57.686787 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"ab912de2-b375-412b-a212-2c063f2ff622","Type":"ContainerStarted","Data":"40264f8469d0bfd9c11717e506a93d399b948663dc75fd4618e5c4dc8388531a"} Nov 25 10:48:57 crc kubenswrapper[4776]: I1125 10:48:57.686832 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"ab912de2-b375-412b-a212-2c063f2ff622","Type":"ContainerStarted","Data":"016fa829f5eaa30a8a8f2b2f4e7cbc74e4f1c2dedddc5c6afbf62e1b22b38e5b"} Nov 25 10:48:57 crc kubenswrapper[4776]: I1125 10:48:57.703398 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-2-default" podStartSLOduration=1.703380434 podStartE2EDuration="1.703380434s" podCreationTimestamp="2025-11-25 10:48:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:48:57.703160559 +0000 UTC m=+5082.744220112" watchObservedRunningTime="2025-11-25 10:48:57.703380434 +0000 UTC m=+5082.744439997" Nov 25 10:48:58 crc kubenswrapper[4776]: I1125 10:48:58.694162 4776 generic.go:334] "Generic (PLEG): container finished" podID="ab912de2-b375-412b-a212-2c063f2ff622" containerID="40264f8469d0bfd9c11717e506a93d399b948663dc75fd4618e5c4dc8388531a" exitCode=1 Nov 25 10:48:58 crc kubenswrapper[4776]: I1125 10:48:58.694228 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"ab912de2-b375-412b-a212-2c063f2ff622","Type":"ContainerDied","Data":"40264f8469d0bfd9c11717e506a93d399b948663dc75fd4618e5c4dc8388531a"} Nov 25 10:49:00 crc kubenswrapper[4776]: I1125 10:49:00.029679 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Nov 25 10:49:00 crc kubenswrapper[4776]: I1125 10:49:00.074894 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2-default"] Nov 25 10:49:00 crc kubenswrapper[4776]: I1125 10:49:00.084926 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2-default"] Nov 25 10:49:00 crc kubenswrapper[4776]: I1125 10:49:00.098542 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-678tg\" (UniqueName: \"kubernetes.io/projected/ab912de2-b375-412b-a212-2c063f2ff622-kube-api-access-678tg\") pod \"ab912de2-b375-412b-a212-2c063f2ff622\" (UID: \"ab912de2-b375-412b-a212-2c063f2ff622\") " Nov 25 10:49:00 crc kubenswrapper[4776]: I1125 10:49:00.103402 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab912de2-b375-412b-a212-2c063f2ff622-kube-api-access-678tg" (OuterVolumeSpecName: "kube-api-access-678tg") pod "ab912de2-b375-412b-a212-2c063f2ff622" (UID: "ab912de2-b375-412b-a212-2c063f2ff622"). InnerVolumeSpecName "kube-api-access-678tg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:00 crc kubenswrapper[4776]: I1125 10:49:00.200567 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-678tg\" (UniqueName: \"kubernetes.io/projected/ab912de2-b375-412b-a212-2c063f2ff622-kube-api-access-678tg\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:00 crc kubenswrapper[4776]: I1125 10:49:00.711125 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="016fa829f5eaa30a8a8f2b2f4e7cbc74e4f1c2dedddc5c6afbf62e1b22b38e5b" Nov 25 10:49:00 crc kubenswrapper[4776]: I1125 10:49:00.711828 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Nov 25 10:49:01 crc kubenswrapper[4776]: I1125 10:49:01.299377 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1"] Nov 25 10:49:01 crc kubenswrapper[4776]: E1125 10:49:01.299786 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab912de2-b375-412b-a212-2c063f2ff622" containerName="mariadb-client-2-default" Nov 25 10:49:01 crc kubenswrapper[4776]: I1125 10:49:01.299802 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab912de2-b375-412b-a212-2c063f2ff622" containerName="mariadb-client-2-default" Nov 25 10:49:01 crc kubenswrapper[4776]: I1125 10:49:01.300099 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab912de2-b375-412b-a212-2c063f2ff622" containerName="mariadb-client-2-default" Nov 25 10:49:01 crc kubenswrapper[4776]: I1125 10:49:01.300775 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Nov 25 10:49:01 crc kubenswrapper[4776]: I1125 10:49:01.303226 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-rdbsh" Nov 25 10:49:01 crc kubenswrapper[4776]: I1125 10:49:01.319472 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Nov 25 10:49:01 crc kubenswrapper[4776]: I1125 10:49:01.418745 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvfjt\" (UniqueName: \"kubernetes.io/projected/254aa4f2-3b81-458d-8215-7966f4a7eeaf-kube-api-access-qvfjt\") pod \"mariadb-client-1\" (UID: \"254aa4f2-3b81-458d-8215-7966f4a7eeaf\") " pod="openstack/mariadb-client-1" Nov 25 10:49:01 crc kubenswrapper[4776]: I1125 10:49:01.520188 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvfjt\" (UniqueName: \"kubernetes.io/projected/254aa4f2-3b81-458d-8215-7966f4a7eeaf-kube-api-access-qvfjt\") pod \"mariadb-client-1\" (UID: \"254aa4f2-3b81-458d-8215-7966f4a7eeaf\") " pod="openstack/mariadb-client-1" Nov 25 10:49:01 crc kubenswrapper[4776]: I1125 10:49:01.552929 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvfjt\" (UniqueName: \"kubernetes.io/projected/254aa4f2-3b81-458d-8215-7966f4a7eeaf-kube-api-access-qvfjt\") pod \"mariadb-client-1\" (UID: \"254aa4f2-3b81-458d-8215-7966f4a7eeaf\") " pod="openstack/mariadb-client-1" Nov 25 10:49:01 crc kubenswrapper[4776]: I1125 10:49:01.623318 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Nov 25 10:49:01 crc kubenswrapper[4776]: I1125 10:49:01.673913 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab912de2-b375-412b-a212-2c063f2ff622" path="/var/lib/kubelet/pods/ab912de2-b375-412b-a212-2c063f2ff622/volumes" Nov 25 10:49:02 crc kubenswrapper[4776]: I1125 10:49:02.112283 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Nov 25 10:49:02 crc kubenswrapper[4776]: W1125 10:49:02.116447 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod254aa4f2_3b81_458d_8215_7966f4a7eeaf.slice/crio-6e26c41454e6adc0be133a54015c463f6b3a939659df92721f342cc1eac9cd24 WatchSource:0}: Error finding container 6e26c41454e6adc0be133a54015c463f6b3a939659df92721f342cc1eac9cd24: Status 404 returned error can't find the container with id 6e26c41454e6adc0be133a54015c463f6b3a939659df92721f342cc1eac9cd24 Nov 25 10:49:02 crc kubenswrapper[4776]: I1125 10:49:02.730693 4776 generic.go:334] "Generic (PLEG): container finished" podID="254aa4f2-3b81-458d-8215-7966f4a7eeaf" containerID="e0643220926571c88aea4d4c87aafdd6d7a4657df047f08405a0c8b31c6a3f2f" exitCode=0 Nov 25 10:49:02 crc kubenswrapper[4776]: I1125 10:49:02.730733 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"254aa4f2-3b81-458d-8215-7966f4a7eeaf","Type":"ContainerDied","Data":"e0643220926571c88aea4d4c87aafdd6d7a4657df047f08405a0c8b31c6a3f2f"} Nov 25 10:49:02 crc kubenswrapper[4776]: I1125 10:49:02.730759 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"254aa4f2-3b81-458d-8215-7966f4a7eeaf","Type":"ContainerStarted","Data":"6e26c41454e6adc0be133a54015c463f6b3a939659df92721f342cc1eac9cd24"} Nov 25 10:49:04 crc kubenswrapper[4776]: I1125 10:49:04.108564 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Nov 25 10:49:04 crc kubenswrapper[4776]: I1125 10:49:04.125888 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1_254aa4f2-3b81-458d-8215-7966f4a7eeaf/mariadb-client-1/0.log" Nov 25 10:49:04 crc kubenswrapper[4776]: I1125 10:49:04.153327 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1"] Nov 25 10:49:04 crc kubenswrapper[4776]: I1125 10:49:04.160882 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1"] Nov 25 10:49:04 crc kubenswrapper[4776]: I1125 10:49:04.264142 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvfjt\" (UniqueName: \"kubernetes.io/projected/254aa4f2-3b81-458d-8215-7966f4a7eeaf-kube-api-access-qvfjt\") pod \"254aa4f2-3b81-458d-8215-7966f4a7eeaf\" (UID: \"254aa4f2-3b81-458d-8215-7966f4a7eeaf\") " Nov 25 10:49:04 crc kubenswrapper[4776]: I1125 10:49:04.272314 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/254aa4f2-3b81-458d-8215-7966f4a7eeaf-kube-api-access-qvfjt" (OuterVolumeSpecName: "kube-api-access-qvfjt") pod "254aa4f2-3b81-458d-8215-7966f4a7eeaf" (UID: "254aa4f2-3b81-458d-8215-7966f4a7eeaf"). InnerVolumeSpecName "kube-api-access-qvfjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:04 crc kubenswrapper[4776]: I1125 10:49:04.366557 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvfjt\" (UniqueName: \"kubernetes.io/projected/254aa4f2-3b81-458d-8215-7966f4a7eeaf-kube-api-access-qvfjt\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:04 crc kubenswrapper[4776]: I1125 10:49:04.553313 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-4-default"] Nov 25 10:49:04 crc kubenswrapper[4776]: E1125 10:49:04.554043 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="254aa4f2-3b81-458d-8215-7966f4a7eeaf" containerName="mariadb-client-1" Nov 25 10:49:04 crc kubenswrapper[4776]: I1125 10:49:04.554086 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="254aa4f2-3b81-458d-8215-7966f4a7eeaf" containerName="mariadb-client-1" Nov 25 10:49:04 crc kubenswrapper[4776]: I1125 10:49:04.554280 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="254aa4f2-3b81-458d-8215-7966f4a7eeaf" containerName="mariadb-client-1" Nov 25 10:49:04 crc kubenswrapper[4776]: I1125 10:49:04.554949 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Nov 25 10:49:04 crc kubenswrapper[4776]: I1125 10:49:04.564591 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Nov 25 10:49:04 crc kubenswrapper[4776]: I1125 10:49:04.673714 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvbkt\" (UniqueName: \"kubernetes.io/projected/0590972b-9776-4bcc-9c23-968a709ed029-kube-api-access-gvbkt\") pod \"mariadb-client-4-default\" (UID: \"0590972b-9776-4bcc-9c23-968a709ed029\") " pod="openstack/mariadb-client-4-default" Nov 25 10:49:04 crc kubenswrapper[4776]: I1125 10:49:04.751500 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e26c41454e6adc0be133a54015c463f6b3a939659df92721f342cc1eac9cd24" Nov 25 10:49:04 crc kubenswrapper[4776]: I1125 10:49:04.751576 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Nov 25 10:49:04 crc kubenswrapper[4776]: I1125 10:49:04.775007 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvbkt\" (UniqueName: \"kubernetes.io/projected/0590972b-9776-4bcc-9c23-968a709ed029-kube-api-access-gvbkt\") pod \"mariadb-client-4-default\" (UID: \"0590972b-9776-4bcc-9c23-968a709ed029\") " pod="openstack/mariadb-client-4-default" Nov 25 10:49:04 crc kubenswrapper[4776]: I1125 10:49:04.796349 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvbkt\" (UniqueName: \"kubernetes.io/projected/0590972b-9776-4bcc-9c23-968a709ed029-kube-api-access-gvbkt\") pod \"mariadb-client-4-default\" (UID: \"0590972b-9776-4bcc-9c23-968a709ed029\") " pod="openstack/mariadb-client-4-default" Nov 25 10:49:04 crc kubenswrapper[4776]: I1125 10:49:04.871876 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Nov 25 10:49:05 crc kubenswrapper[4776]: I1125 10:49:05.382271 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Nov 25 10:49:05 crc kubenswrapper[4776]: I1125 10:49:05.672179 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="254aa4f2-3b81-458d-8215-7966f4a7eeaf" path="/var/lib/kubelet/pods/254aa4f2-3b81-458d-8215-7966f4a7eeaf/volumes" Nov 25 10:49:05 crc kubenswrapper[4776]: I1125 10:49:05.760293 4776 generic.go:334] "Generic (PLEG): container finished" podID="0590972b-9776-4bcc-9c23-968a709ed029" containerID="6e250d2f8d6c32ae1405e78f5748ee9e60693cdff5734fa804bf82a11975cc82" exitCode=0 Nov 25 10:49:05 crc kubenswrapper[4776]: I1125 10:49:05.760379 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"0590972b-9776-4bcc-9c23-968a709ed029","Type":"ContainerDied","Data":"6e250d2f8d6c32ae1405e78f5748ee9e60693cdff5734fa804bf82a11975cc82"} Nov 25 10:49:05 crc kubenswrapper[4776]: I1125 10:49:05.760415 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"0590972b-9776-4bcc-9c23-968a709ed029","Type":"ContainerStarted","Data":"aeb14b1d68774743116d4aac2109da5f8d94749689dcabaeb70b7bcbd87565d8"} Nov 25 10:49:07 crc kubenswrapper[4776]: I1125 10:49:07.105703 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Nov 25 10:49:07 crc kubenswrapper[4776]: I1125 10:49:07.131909 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-4-default_0590972b-9776-4bcc-9c23-968a709ed029/mariadb-client-4-default/0.log" Nov 25 10:49:07 crc kubenswrapper[4776]: I1125 10:49:07.158693 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-4-default"] Nov 25 10:49:07 crc kubenswrapper[4776]: I1125 10:49:07.163022 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-4-default"] Nov 25 10:49:07 crc kubenswrapper[4776]: I1125 10:49:07.222665 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvbkt\" (UniqueName: \"kubernetes.io/projected/0590972b-9776-4bcc-9c23-968a709ed029-kube-api-access-gvbkt\") pod \"0590972b-9776-4bcc-9c23-968a709ed029\" (UID: \"0590972b-9776-4bcc-9c23-968a709ed029\") " Nov 25 10:49:07 crc kubenswrapper[4776]: I1125 10:49:07.228272 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0590972b-9776-4bcc-9c23-968a709ed029-kube-api-access-gvbkt" (OuterVolumeSpecName: "kube-api-access-gvbkt") pod "0590972b-9776-4bcc-9c23-968a709ed029" (UID: "0590972b-9776-4bcc-9c23-968a709ed029"). InnerVolumeSpecName "kube-api-access-gvbkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:07 crc kubenswrapper[4776]: I1125 10:49:07.325046 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvbkt\" (UniqueName: \"kubernetes.io/projected/0590972b-9776-4bcc-9c23-968a709ed029-kube-api-access-gvbkt\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:07 crc kubenswrapper[4776]: I1125 10:49:07.671513 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0590972b-9776-4bcc-9c23-968a709ed029" path="/var/lib/kubelet/pods/0590972b-9776-4bcc-9c23-968a709ed029/volumes" Nov 25 10:49:07 crc kubenswrapper[4776]: I1125 10:49:07.776634 4776 scope.go:117] "RemoveContainer" containerID="6e250d2f8d6c32ae1405e78f5748ee9e60693cdff5734fa804bf82a11975cc82" Nov 25 10:49:07 crc kubenswrapper[4776]: I1125 10:49:07.776797 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Nov 25 10:49:11 crc kubenswrapper[4776]: I1125 10:49:11.328030 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-5-default"] Nov 25 10:49:11 crc kubenswrapper[4776]: E1125 10:49:11.328932 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0590972b-9776-4bcc-9c23-968a709ed029" containerName="mariadb-client-4-default" Nov 25 10:49:11 crc kubenswrapper[4776]: I1125 10:49:11.328948 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="0590972b-9776-4bcc-9c23-968a709ed029" containerName="mariadb-client-4-default" Nov 25 10:49:11 crc kubenswrapper[4776]: I1125 10:49:11.329143 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="0590972b-9776-4bcc-9c23-968a709ed029" containerName="mariadb-client-4-default" Nov 25 10:49:11 crc kubenswrapper[4776]: I1125 10:49:11.331214 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Nov 25 10:49:11 crc kubenswrapper[4776]: I1125 10:49:11.333757 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-rdbsh" Nov 25 10:49:11 crc kubenswrapper[4776]: I1125 10:49:11.334020 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Nov 25 10:49:11 crc kubenswrapper[4776]: I1125 10:49:11.491108 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2h2sk\" (UniqueName: \"kubernetes.io/projected/dafebc87-6fb2-4d38-aeb6-a84a609f1c3c-kube-api-access-2h2sk\") pod \"mariadb-client-5-default\" (UID: \"dafebc87-6fb2-4d38-aeb6-a84a609f1c3c\") " pod="openstack/mariadb-client-5-default" Nov 25 10:49:11 crc kubenswrapper[4776]: I1125 10:49:11.592132 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2h2sk\" (UniqueName: \"kubernetes.io/projected/dafebc87-6fb2-4d38-aeb6-a84a609f1c3c-kube-api-access-2h2sk\") pod \"mariadb-client-5-default\" (UID: \"dafebc87-6fb2-4d38-aeb6-a84a609f1c3c\") " pod="openstack/mariadb-client-5-default" Nov 25 10:49:11 crc kubenswrapper[4776]: I1125 10:49:11.628130 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2h2sk\" (UniqueName: \"kubernetes.io/projected/dafebc87-6fb2-4d38-aeb6-a84a609f1c3c-kube-api-access-2h2sk\") pod \"mariadb-client-5-default\" (UID: \"dafebc87-6fb2-4d38-aeb6-a84a609f1c3c\") " pod="openstack/mariadb-client-5-default" Nov 25 10:49:11 crc kubenswrapper[4776]: I1125 10:49:11.654223 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Nov 25 10:49:12 crc kubenswrapper[4776]: I1125 10:49:12.129326 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Nov 25 10:49:12 crc kubenswrapper[4776]: I1125 10:49:12.826243 4776 generic.go:334] "Generic (PLEG): container finished" podID="dafebc87-6fb2-4d38-aeb6-a84a609f1c3c" containerID="6fd5d1dd1a727cade74987e883f83a6b7b3e17be568974860f7b634243ef2aaa" exitCode=0 Nov 25 10:49:12 crc kubenswrapper[4776]: I1125 10:49:12.826329 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"dafebc87-6fb2-4d38-aeb6-a84a609f1c3c","Type":"ContainerDied","Data":"6fd5d1dd1a727cade74987e883f83a6b7b3e17be568974860f7b634243ef2aaa"} Nov 25 10:49:12 crc kubenswrapper[4776]: I1125 10:49:12.826599 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"dafebc87-6fb2-4d38-aeb6-a84a609f1c3c","Type":"ContainerStarted","Data":"0584b7d04271fe8780665dbee735ce4265a488df0b278ba918814890e389b610"} Nov 25 10:49:14 crc kubenswrapper[4776]: I1125 10:49:14.195623 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Nov 25 10:49:14 crc kubenswrapper[4776]: I1125 10:49:14.216625 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-5-default_dafebc87-6fb2-4d38-aeb6-a84a609f1c3c/mariadb-client-5-default/0.log" Nov 25 10:49:14 crc kubenswrapper[4776]: I1125 10:49:14.237945 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-5-default"] Nov 25 10:49:14 crc kubenswrapper[4776]: I1125 10:49:14.243691 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-5-default"] Nov 25 10:49:14 crc kubenswrapper[4776]: I1125 10:49:14.337459 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2h2sk\" (UniqueName: \"kubernetes.io/projected/dafebc87-6fb2-4d38-aeb6-a84a609f1c3c-kube-api-access-2h2sk\") pod \"dafebc87-6fb2-4d38-aeb6-a84a609f1c3c\" (UID: \"dafebc87-6fb2-4d38-aeb6-a84a609f1c3c\") " Nov 25 10:49:14 crc kubenswrapper[4776]: I1125 10:49:14.350388 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dafebc87-6fb2-4d38-aeb6-a84a609f1c3c-kube-api-access-2h2sk" (OuterVolumeSpecName: "kube-api-access-2h2sk") pod "dafebc87-6fb2-4d38-aeb6-a84a609f1c3c" (UID: "dafebc87-6fb2-4d38-aeb6-a84a609f1c3c"). InnerVolumeSpecName "kube-api-access-2h2sk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:14 crc kubenswrapper[4776]: I1125 10:49:14.372877 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-6-default"] Nov 25 10:49:14 crc kubenswrapper[4776]: E1125 10:49:14.373378 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dafebc87-6fb2-4d38-aeb6-a84a609f1c3c" containerName="mariadb-client-5-default" Nov 25 10:49:14 crc kubenswrapper[4776]: I1125 10:49:14.373413 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="dafebc87-6fb2-4d38-aeb6-a84a609f1c3c" containerName="mariadb-client-5-default" Nov 25 10:49:14 crc kubenswrapper[4776]: I1125 10:49:14.373671 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="dafebc87-6fb2-4d38-aeb6-a84a609f1c3c" containerName="mariadb-client-5-default" Nov 25 10:49:14 crc kubenswrapper[4776]: I1125 10:49:14.374384 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Nov 25 10:49:14 crc kubenswrapper[4776]: I1125 10:49:14.385560 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Nov 25 10:49:14 crc kubenswrapper[4776]: I1125 10:49:14.439301 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2h2sk\" (UniqueName: \"kubernetes.io/projected/dafebc87-6fb2-4d38-aeb6-a84a609f1c3c-kube-api-access-2h2sk\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:14 crc kubenswrapper[4776]: I1125 10:49:14.541261 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46mxm\" (UniqueName: \"kubernetes.io/projected/c0fba1e8-d6e6-4eb9-a464-d56353e252d3-kube-api-access-46mxm\") pod \"mariadb-client-6-default\" (UID: \"c0fba1e8-d6e6-4eb9-a464-d56353e252d3\") " pod="openstack/mariadb-client-6-default" Nov 25 10:49:14 crc kubenswrapper[4776]: I1125 10:49:14.642964 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46mxm\" (UniqueName: \"kubernetes.io/projected/c0fba1e8-d6e6-4eb9-a464-d56353e252d3-kube-api-access-46mxm\") pod \"mariadb-client-6-default\" (UID: \"c0fba1e8-d6e6-4eb9-a464-d56353e252d3\") " pod="openstack/mariadb-client-6-default" Nov 25 10:49:14 crc kubenswrapper[4776]: I1125 10:49:14.661619 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46mxm\" (UniqueName: \"kubernetes.io/projected/c0fba1e8-d6e6-4eb9-a464-d56353e252d3-kube-api-access-46mxm\") pod \"mariadb-client-6-default\" (UID: \"c0fba1e8-d6e6-4eb9-a464-d56353e252d3\") " pod="openstack/mariadb-client-6-default" Nov 25 10:49:14 crc kubenswrapper[4776]: I1125 10:49:14.707004 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Nov 25 10:49:14 crc kubenswrapper[4776]: I1125 10:49:14.850765 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0584b7d04271fe8780665dbee735ce4265a488df0b278ba918814890e389b610" Nov 25 10:49:14 crc kubenswrapper[4776]: I1125 10:49:14.850842 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Nov 25 10:49:15 crc kubenswrapper[4776]: I1125 10:49:15.186888 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Nov 25 10:49:15 crc kubenswrapper[4776]: I1125 10:49:15.672483 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dafebc87-6fb2-4d38-aeb6-a84a609f1c3c" path="/var/lib/kubelet/pods/dafebc87-6fb2-4d38-aeb6-a84a609f1c3c/volumes" Nov 25 10:49:15 crc kubenswrapper[4776]: I1125 10:49:15.859577 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"c0fba1e8-d6e6-4eb9-a464-d56353e252d3","Type":"ContainerStarted","Data":"4c2513c60a844ab3a406a3d220ad7622cc04585becd8919cdac99ada9d6b44e0"} Nov 25 10:49:15 crc kubenswrapper[4776]: I1125 10:49:15.859634 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"c0fba1e8-d6e6-4eb9-a464-d56353e252d3","Type":"ContainerStarted","Data":"32dfab45252f7c768e468b159c393ef73c9ffb3828140b11823505d130b29dfc"} Nov 25 10:49:15 crc kubenswrapper[4776]: I1125 10:49:15.882624 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-6-default" podStartSLOduration=1.882603424 podStartE2EDuration="1.882603424s" podCreationTimestamp="2025-11-25 10:49:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:49:15.875711471 +0000 UTC m=+5100.916771024" watchObservedRunningTime="2025-11-25 10:49:15.882603424 +0000 UTC m=+5100.923662977" Nov 25 10:49:16 crc kubenswrapper[4776]: I1125 10:49:16.871184 4776 generic.go:334] "Generic (PLEG): container finished" podID="c0fba1e8-d6e6-4eb9-a464-d56353e252d3" containerID="4c2513c60a844ab3a406a3d220ad7622cc04585becd8919cdac99ada9d6b44e0" exitCode=1 Nov 25 10:49:16 crc kubenswrapper[4776]: I1125 10:49:16.871252 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"c0fba1e8-d6e6-4eb9-a464-d56353e252d3","Type":"ContainerDied","Data":"4c2513c60a844ab3a406a3d220ad7622cc04585becd8919cdac99ada9d6b44e0"} Nov 25 10:49:18 crc kubenswrapper[4776]: I1125 10:49:18.248240 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Nov 25 10:49:18 crc kubenswrapper[4776]: I1125 10:49:18.278855 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-6-default"] Nov 25 10:49:18 crc kubenswrapper[4776]: I1125 10:49:18.284789 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-6-default"] Nov 25 10:49:18 crc kubenswrapper[4776]: I1125 10:49:18.410133 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46mxm\" (UniqueName: \"kubernetes.io/projected/c0fba1e8-d6e6-4eb9-a464-d56353e252d3-kube-api-access-46mxm\") pod \"c0fba1e8-d6e6-4eb9-a464-d56353e252d3\" (UID: \"c0fba1e8-d6e6-4eb9-a464-d56353e252d3\") " Nov 25 10:49:18 crc kubenswrapper[4776]: I1125 10:49:18.411358 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-7-default"] Nov 25 10:49:18 crc kubenswrapper[4776]: E1125 10:49:18.411697 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0fba1e8-d6e6-4eb9-a464-d56353e252d3" containerName="mariadb-client-6-default" Nov 25 10:49:18 crc kubenswrapper[4776]: I1125 10:49:18.411720 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0fba1e8-d6e6-4eb9-a464-d56353e252d3" containerName="mariadb-client-6-default" Nov 25 10:49:18 crc kubenswrapper[4776]: I1125 10:49:18.411937 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0fba1e8-d6e6-4eb9-a464-d56353e252d3" containerName="mariadb-client-6-default" Nov 25 10:49:18 crc kubenswrapper[4776]: I1125 10:49:18.412700 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Nov 25 10:49:18 crc kubenswrapper[4776]: I1125 10:49:18.415578 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0fba1e8-d6e6-4eb9-a464-d56353e252d3-kube-api-access-46mxm" (OuterVolumeSpecName: "kube-api-access-46mxm") pod "c0fba1e8-d6e6-4eb9-a464-d56353e252d3" (UID: "c0fba1e8-d6e6-4eb9-a464-d56353e252d3"). InnerVolumeSpecName "kube-api-access-46mxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:18 crc kubenswrapper[4776]: I1125 10:49:18.431418 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Nov 25 10:49:18 crc kubenswrapper[4776]: I1125 10:49:18.512143 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2r5x\" (UniqueName: \"kubernetes.io/projected/d92c6977-0181-4e3d-9b37-f3370ec0f974-kube-api-access-n2r5x\") pod \"mariadb-client-7-default\" (UID: \"d92c6977-0181-4e3d-9b37-f3370ec0f974\") " pod="openstack/mariadb-client-7-default" Nov 25 10:49:18 crc kubenswrapper[4776]: I1125 10:49:18.512369 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46mxm\" (UniqueName: \"kubernetes.io/projected/c0fba1e8-d6e6-4eb9-a464-d56353e252d3-kube-api-access-46mxm\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:18 crc kubenswrapper[4776]: I1125 10:49:18.613955 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2r5x\" (UniqueName: \"kubernetes.io/projected/d92c6977-0181-4e3d-9b37-f3370ec0f974-kube-api-access-n2r5x\") pod \"mariadb-client-7-default\" (UID: \"d92c6977-0181-4e3d-9b37-f3370ec0f974\") " pod="openstack/mariadb-client-7-default" Nov 25 10:49:18 crc kubenswrapper[4776]: I1125 10:49:18.631801 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2r5x\" (UniqueName: \"kubernetes.io/projected/d92c6977-0181-4e3d-9b37-f3370ec0f974-kube-api-access-n2r5x\") pod \"mariadb-client-7-default\" (UID: \"d92c6977-0181-4e3d-9b37-f3370ec0f974\") " pod="openstack/mariadb-client-7-default" Nov 25 10:49:18 crc kubenswrapper[4776]: I1125 10:49:18.758516 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Nov 25 10:49:18 crc kubenswrapper[4776]: I1125 10:49:18.921978 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32dfab45252f7c768e468b159c393ef73c9ffb3828140b11823505d130b29dfc" Nov 25 10:49:18 crc kubenswrapper[4776]: I1125 10:49:18.922305 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Nov 25 10:49:19 crc kubenswrapper[4776]: I1125 10:49:19.100894 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Nov 25 10:49:19 crc kubenswrapper[4776]: W1125 10:49:19.107871 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd92c6977_0181_4e3d_9b37_f3370ec0f974.slice/crio-60246b9731827167436a674b331019bfcdc4697139f2302563aebe8ad3e52534 WatchSource:0}: Error finding container 60246b9731827167436a674b331019bfcdc4697139f2302563aebe8ad3e52534: Status 404 returned error can't find the container with id 60246b9731827167436a674b331019bfcdc4697139f2302563aebe8ad3e52534 Nov 25 10:49:19 crc kubenswrapper[4776]: I1125 10:49:19.673023 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0fba1e8-d6e6-4eb9-a464-d56353e252d3" path="/var/lib/kubelet/pods/c0fba1e8-d6e6-4eb9-a464-d56353e252d3/volumes" Nov 25 10:49:19 crc kubenswrapper[4776]: I1125 10:49:19.931307 4776 generic.go:334] "Generic (PLEG): container finished" podID="d92c6977-0181-4e3d-9b37-f3370ec0f974" containerID="1c1f690721cc172ad8270fafa7e18a2d27e8974ebcaef52f9bda381c82baac88" exitCode=0 Nov 25 10:49:19 crc kubenswrapper[4776]: I1125 10:49:19.931364 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"d92c6977-0181-4e3d-9b37-f3370ec0f974","Type":"ContainerDied","Data":"1c1f690721cc172ad8270fafa7e18a2d27e8974ebcaef52f9bda381c82baac88"} Nov 25 10:49:19 crc kubenswrapper[4776]: I1125 10:49:19.931422 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"d92c6977-0181-4e3d-9b37-f3370ec0f974","Type":"ContainerStarted","Data":"60246b9731827167436a674b331019bfcdc4697139f2302563aebe8ad3e52534"} Nov 25 10:49:21 crc kubenswrapper[4776]: I1125 10:49:21.325241 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Nov 25 10:49:21 crc kubenswrapper[4776]: I1125 10:49:21.342614 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-7-default_d92c6977-0181-4e3d-9b37-f3370ec0f974/mariadb-client-7-default/0.log" Nov 25 10:49:21 crc kubenswrapper[4776]: I1125 10:49:21.374156 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-7-default"] Nov 25 10:49:21 crc kubenswrapper[4776]: I1125 10:49:21.379300 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-7-default"] Nov 25 10:49:21 crc kubenswrapper[4776]: I1125 10:49:21.456990 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2r5x\" (UniqueName: \"kubernetes.io/projected/d92c6977-0181-4e3d-9b37-f3370ec0f974-kube-api-access-n2r5x\") pod \"d92c6977-0181-4e3d-9b37-f3370ec0f974\" (UID: \"d92c6977-0181-4e3d-9b37-f3370ec0f974\") " Nov 25 10:49:21 crc kubenswrapper[4776]: I1125 10:49:21.465655 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d92c6977-0181-4e3d-9b37-f3370ec0f974-kube-api-access-n2r5x" (OuterVolumeSpecName: "kube-api-access-n2r5x") pod "d92c6977-0181-4e3d-9b37-f3370ec0f974" (UID: "d92c6977-0181-4e3d-9b37-f3370ec0f974"). InnerVolumeSpecName "kube-api-access-n2r5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:21 crc kubenswrapper[4776]: I1125 10:49:21.529253 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2"] Nov 25 10:49:21 crc kubenswrapper[4776]: E1125 10:49:21.529675 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d92c6977-0181-4e3d-9b37-f3370ec0f974" containerName="mariadb-client-7-default" Nov 25 10:49:21 crc kubenswrapper[4776]: I1125 10:49:21.529692 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="d92c6977-0181-4e3d-9b37-f3370ec0f974" containerName="mariadb-client-7-default" Nov 25 10:49:21 crc kubenswrapper[4776]: I1125 10:49:21.529909 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="d92c6977-0181-4e3d-9b37-f3370ec0f974" containerName="mariadb-client-7-default" Nov 25 10:49:21 crc kubenswrapper[4776]: I1125 10:49:21.530628 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Nov 25 10:49:21 crc kubenswrapper[4776]: I1125 10:49:21.536958 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Nov 25 10:49:21 crc kubenswrapper[4776]: I1125 10:49:21.559209 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2r5x\" (UniqueName: \"kubernetes.io/projected/d92c6977-0181-4e3d-9b37-f3370ec0f974-kube-api-access-n2r5x\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:21 crc kubenswrapper[4776]: I1125 10:49:21.660944 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56pw9\" (UniqueName: \"kubernetes.io/projected/2bb7a47f-21b4-4307-af34-35a27dcf0b64-kube-api-access-56pw9\") pod \"mariadb-client-2\" (UID: \"2bb7a47f-21b4-4307-af34-35a27dcf0b64\") " pod="openstack/mariadb-client-2" Nov 25 10:49:21 crc kubenswrapper[4776]: I1125 10:49:21.673821 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d92c6977-0181-4e3d-9b37-f3370ec0f974" path="/var/lib/kubelet/pods/d92c6977-0181-4e3d-9b37-f3370ec0f974/volumes" Nov 25 10:49:21 crc kubenswrapper[4776]: I1125 10:49:21.762517 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56pw9\" (UniqueName: \"kubernetes.io/projected/2bb7a47f-21b4-4307-af34-35a27dcf0b64-kube-api-access-56pw9\") pod \"mariadb-client-2\" (UID: \"2bb7a47f-21b4-4307-af34-35a27dcf0b64\") " pod="openstack/mariadb-client-2" Nov 25 10:49:21 crc kubenswrapper[4776]: I1125 10:49:21.783384 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56pw9\" (UniqueName: \"kubernetes.io/projected/2bb7a47f-21b4-4307-af34-35a27dcf0b64-kube-api-access-56pw9\") pod \"mariadb-client-2\" (UID: \"2bb7a47f-21b4-4307-af34-35a27dcf0b64\") " pod="openstack/mariadb-client-2" Nov 25 10:49:21 crc kubenswrapper[4776]: I1125 10:49:21.854365 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Nov 25 10:49:21 crc kubenswrapper[4776]: I1125 10:49:21.952628 4776 scope.go:117] "RemoveContainer" containerID="1c1f690721cc172ad8270fafa7e18a2d27e8974ebcaef52f9bda381c82baac88" Nov 25 10:49:21 crc kubenswrapper[4776]: I1125 10:49:21.952994 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Nov 25 10:49:22 crc kubenswrapper[4776]: I1125 10:49:22.120823 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Nov 25 10:49:22 crc kubenswrapper[4776]: W1125 10:49:22.125256 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bb7a47f_21b4_4307_af34_35a27dcf0b64.slice/crio-a084ffd98692e7a3f5d371079a9fd5655f41448b261302fb7ad9a613c6042e27 WatchSource:0}: Error finding container a084ffd98692e7a3f5d371079a9fd5655f41448b261302fb7ad9a613c6042e27: Status 404 returned error can't find the container with id a084ffd98692e7a3f5d371079a9fd5655f41448b261302fb7ad9a613c6042e27 Nov 25 10:49:22 crc kubenswrapper[4776]: I1125 10:49:22.961836 4776 generic.go:334] "Generic (PLEG): container finished" podID="2bb7a47f-21b4-4307-af34-35a27dcf0b64" containerID="3e6d9e176a10e478905aeab1fada0c8f3d49080f66a2255b66f0694bc0992e83" exitCode=0 Nov 25 10:49:22 crc kubenswrapper[4776]: I1125 10:49:22.961939 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"2bb7a47f-21b4-4307-af34-35a27dcf0b64","Type":"ContainerDied","Data":"3e6d9e176a10e478905aeab1fada0c8f3d49080f66a2255b66f0694bc0992e83"} Nov 25 10:49:22 crc kubenswrapper[4776]: I1125 10:49:22.962356 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"2bb7a47f-21b4-4307-af34-35a27dcf0b64","Type":"ContainerStarted","Data":"a084ffd98692e7a3f5d371079a9fd5655f41448b261302fb7ad9a613c6042e27"} Nov 25 10:49:24 crc kubenswrapper[4776]: I1125 10:49:24.315436 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Nov 25 10:49:24 crc kubenswrapper[4776]: I1125 10:49:24.336780 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2_2bb7a47f-21b4-4307-af34-35a27dcf0b64/mariadb-client-2/0.log" Nov 25 10:49:24 crc kubenswrapper[4776]: I1125 10:49:24.358930 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2"] Nov 25 10:49:24 crc kubenswrapper[4776]: I1125 10:49:24.366351 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2"] Nov 25 10:49:24 crc kubenswrapper[4776]: I1125 10:49:24.401415 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56pw9\" (UniqueName: \"kubernetes.io/projected/2bb7a47f-21b4-4307-af34-35a27dcf0b64-kube-api-access-56pw9\") pod \"2bb7a47f-21b4-4307-af34-35a27dcf0b64\" (UID: \"2bb7a47f-21b4-4307-af34-35a27dcf0b64\") " Nov 25 10:49:24 crc kubenswrapper[4776]: I1125 10:49:24.406609 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bb7a47f-21b4-4307-af34-35a27dcf0b64-kube-api-access-56pw9" (OuterVolumeSpecName: "kube-api-access-56pw9") pod "2bb7a47f-21b4-4307-af34-35a27dcf0b64" (UID: "2bb7a47f-21b4-4307-af34-35a27dcf0b64"). InnerVolumeSpecName "kube-api-access-56pw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:49:24 crc kubenswrapper[4776]: I1125 10:49:24.504172 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56pw9\" (UniqueName: \"kubernetes.io/projected/2bb7a47f-21b4-4307-af34-35a27dcf0b64-kube-api-access-56pw9\") on node \"crc\" DevicePath \"\"" Nov 25 10:49:24 crc kubenswrapper[4776]: I1125 10:49:24.984507 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a084ffd98692e7a3f5d371079a9fd5655f41448b261302fb7ad9a613c6042e27" Nov 25 10:49:24 crc kubenswrapper[4776]: I1125 10:49:24.984611 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Nov 25 10:49:25 crc kubenswrapper[4776]: I1125 10:49:25.671818 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bb7a47f-21b4-4307-af34-35a27dcf0b64" path="/var/lib/kubelet/pods/2bb7a47f-21b4-4307-af34-35a27dcf0b64/volumes" Nov 25 10:51:17 crc kubenswrapper[4776]: I1125 10:51:17.819346 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:51:17 crc kubenswrapper[4776]: I1125 10:51:17.820025 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:51:31 crc kubenswrapper[4776]: I1125 10:51:31.624165 4776 scope.go:117] "RemoveContainer" containerID="525ac307b87c2ac45fc15b62f0e9864a3c0ff4e0cb50536916647e40747636b2" Nov 25 10:51:47 crc kubenswrapper[4776]: I1125 10:51:47.818905 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:51:47 crc kubenswrapper[4776]: I1125 10:51:47.819714 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:51:58 crc kubenswrapper[4776]: I1125 10:51:58.508497 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2qkhz"] Nov 25 10:51:58 crc kubenswrapper[4776]: E1125 10:51:58.511786 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bb7a47f-21b4-4307-af34-35a27dcf0b64" containerName="mariadb-client-2" Nov 25 10:51:58 crc kubenswrapper[4776]: I1125 10:51:58.511852 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bb7a47f-21b4-4307-af34-35a27dcf0b64" containerName="mariadb-client-2" Nov 25 10:51:58 crc kubenswrapper[4776]: I1125 10:51:58.512339 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bb7a47f-21b4-4307-af34-35a27dcf0b64" containerName="mariadb-client-2" Nov 25 10:51:58 crc kubenswrapper[4776]: I1125 10:51:58.515211 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2qkhz" Nov 25 10:51:58 crc kubenswrapper[4776]: I1125 10:51:58.525208 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2qkhz"] Nov 25 10:51:58 crc kubenswrapper[4776]: I1125 10:51:58.660613 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c5ced81-2f00-4f4f-a7f6-aa25feb3721f-catalog-content\") pod \"certified-operators-2qkhz\" (UID: \"4c5ced81-2f00-4f4f-a7f6-aa25feb3721f\") " pod="openshift-marketplace/certified-operators-2qkhz" Nov 25 10:51:58 crc kubenswrapper[4776]: I1125 10:51:58.660808 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c5ced81-2f00-4f4f-a7f6-aa25feb3721f-utilities\") pod \"certified-operators-2qkhz\" (UID: \"4c5ced81-2f00-4f4f-a7f6-aa25feb3721f\") " pod="openshift-marketplace/certified-operators-2qkhz" Nov 25 10:51:58 crc kubenswrapper[4776]: I1125 10:51:58.660945 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8ndg\" (UniqueName: \"kubernetes.io/projected/4c5ced81-2f00-4f4f-a7f6-aa25feb3721f-kube-api-access-s8ndg\") pod \"certified-operators-2qkhz\" (UID: \"4c5ced81-2f00-4f4f-a7f6-aa25feb3721f\") " pod="openshift-marketplace/certified-operators-2qkhz" Nov 25 10:51:58 crc kubenswrapper[4776]: I1125 10:51:58.763165 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c5ced81-2f00-4f4f-a7f6-aa25feb3721f-catalog-content\") pod \"certified-operators-2qkhz\" (UID: \"4c5ced81-2f00-4f4f-a7f6-aa25feb3721f\") " pod="openshift-marketplace/certified-operators-2qkhz" Nov 25 10:51:58 crc kubenswrapper[4776]: I1125 10:51:58.763289 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c5ced81-2f00-4f4f-a7f6-aa25feb3721f-utilities\") pod \"certified-operators-2qkhz\" (UID: \"4c5ced81-2f00-4f4f-a7f6-aa25feb3721f\") " pod="openshift-marketplace/certified-operators-2qkhz" Nov 25 10:51:58 crc kubenswrapper[4776]: I1125 10:51:58.763351 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8ndg\" (UniqueName: \"kubernetes.io/projected/4c5ced81-2f00-4f4f-a7f6-aa25feb3721f-kube-api-access-s8ndg\") pod \"certified-operators-2qkhz\" (UID: \"4c5ced81-2f00-4f4f-a7f6-aa25feb3721f\") " pod="openshift-marketplace/certified-operators-2qkhz" Nov 25 10:51:58 crc kubenswrapper[4776]: I1125 10:51:58.763671 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c5ced81-2f00-4f4f-a7f6-aa25feb3721f-catalog-content\") pod \"certified-operators-2qkhz\" (UID: \"4c5ced81-2f00-4f4f-a7f6-aa25feb3721f\") " pod="openshift-marketplace/certified-operators-2qkhz" Nov 25 10:51:58 crc kubenswrapper[4776]: I1125 10:51:58.764006 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c5ced81-2f00-4f4f-a7f6-aa25feb3721f-utilities\") pod \"certified-operators-2qkhz\" (UID: \"4c5ced81-2f00-4f4f-a7f6-aa25feb3721f\") " pod="openshift-marketplace/certified-operators-2qkhz" Nov 25 10:51:58 crc kubenswrapper[4776]: I1125 10:51:58.787571 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8ndg\" (UniqueName: \"kubernetes.io/projected/4c5ced81-2f00-4f4f-a7f6-aa25feb3721f-kube-api-access-s8ndg\") pod \"certified-operators-2qkhz\" (UID: \"4c5ced81-2f00-4f4f-a7f6-aa25feb3721f\") " pod="openshift-marketplace/certified-operators-2qkhz" Nov 25 10:51:58 crc kubenswrapper[4776]: I1125 10:51:58.854670 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2qkhz" Nov 25 10:51:59 crc kubenswrapper[4776]: I1125 10:51:59.329260 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2qkhz"] Nov 25 10:51:59 crc kubenswrapper[4776]: W1125 10:51:59.339278 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c5ced81_2f00_4f4f_a7f6_aa25feb3721f.slice/crio-a72dfc3867d4a2fb7b0843bd33529c1765f3f001b8e5eab259ed4f23d779b296 WatchSource:0}: Error finding container a72dfc3867d4a2fb7b0843bd33529c1765f3f001b8e5eab259ed4f23d779b296: Status 404 returned error can't find the container with id a72dfc3867d4a2fb7b0843bd33529c1765f3f001b8e5eab259ed4f23d779b296 Nov 25 10:52:00 crc kubenswrapper[4776]: I1125 10:52:00.312760 4776 generic.go:334] "Generic (PLEG): container finished" podID="4c5ced81-2f00-4f4f-a7f6-aa25feb3721f" containerID="cb2469aed1090966c1081d6952160a7ffedecd66dbcbc8fa4a0bc1d811f4dc0c" exitCode=0 Nov 25 10:52:00 crc kubenswrapper[4776]: I1125 10:52:00.312901 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2qkhz" event={"ID":"4c5ced81-2f00-4f4f-a7f6-aa25feb3721f","Type":"ContainerDied","Data":"cb2469aed1090966c1081d6952160a7ffedecd66dbcbc8fa4a0bc1d811f4dc0c"} Nov 25 10:52:00 crc kubenswrapper[4776]: I1125 10:52:00.313170 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2qkhz" event={"ID":"4c5ced81-2f00-4f4f-a7f6-aa25feb3721f","Type":"ContainerStarted","Data":"a72dfc3867d4a2fb7b0843bd33529c1765f3f001b8e5eab259ed4f23d779b296"} Nov 25 10:52:04 crc kubenswrapper[4776]: I1125 10:52:04.348880 4776 generic.go:334] "Generic (PLEG): container finished" podID="4c5ced81-2f00-4f4f-a7f6-aa25feb3721f" containerID="033ce72c34374797185f2c00ac9302cc75ace40075f4f14a4bf0b0126c375ea5" exitCode=0 Nov 25 10:52:04 crc kubenswrapper[4776]: I1125 10:52:04.348945 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2qkhz" event={"ID":"4c5ced81-2f00-4f4f-a7f6-aa25feb3721f","Type":"ContainerDied","Data":"033ce72c34374797185f2c00ac9302cc75ace40075f4f14a4bf0b0126c375ea5"} Nov 25 10:52:07 crc kubenswrapper[4776]: I1125 10:52:07.375409 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2qkhz" event={"ID":"4c5ced81-2f00-4f4f-a7f6-aa25feb3721f","Type":"ContainerStarted","Data":"d36d86094b66ac0c5977500500f52890aea85ed6d130f2b54a872949f0ac4080"} Nov 25 10:52:07 crc kubenswrapper[4776]: I1125 10:52:07.392743 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2qkhz" podStartSLOduration=3.213646836 podStartE2EDuration="9.392728915s" podCreationTimestamp="2025-11-25 10:51:58 +0000 UTC" firstStartedPulling="2025-11-25 10:52:00.314827635 +0000 UTC m=+5265.355887188" lastFinishedPulling="2025-11-25 10:52:06.493909674 +0000 UTC m=+5271.534969267" observedRunningTime="2025-11-25 10:52:07.391399622 +0000 UTC m=+5272.432459175" watchObservedRunningTime="2025-11-25 10:52:07.392728915 +0000 UTC m=+5272.433788468" Nov 25 10:52:08 crc kubenswrapper[4776]: I1125 10:52:08.856364 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2qkhz" Nov 25 10:52:08 crc kubenswrapper[4776]: I1125 10:52:08.856461 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2qkhz" Nov 25 10:52:08 crc kubenswrapper[4776]: I1125 10:52:08.913461 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2qkhz" Nov 25 10:52:17 crc kubenswrapper[4776]: I1125 10:52:17.818178 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:52:17 crc kubenswrapper[4776]: I1125 10:52:17.820520 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:52:17 crc kubenswrapper[4776]: I1125 10:52:17.820801 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 10:52:17 crc kubenswrapper[4776]: I1125 10:52:17.821883 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:52:17 crc kubenswrapper[4776]: I1125 10:52:17.822139 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40" gracePeriod=600 Nov 25 10:52:17 crc kubenswrapper[4776]: E1125 10:52:17.947530 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:52:18 crc kubenswrapper[4776]: I1125 10:52:18.498171 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40" exitCode=0 Nov 25 10:52:18 crc kubenswrapper[4776]: I1125 10:52:18.498260 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40"} Nov 25 10:52:18 crc kubenswrapper[4776]: I1125 10:52:18.498375 4776 scope.go:117] "RemoveContainer" containerID="62f6bd2b248ceb776fb1c929efd14083d136607e3bcf021f3f4b235fbcacf0e1" Nov 25 10:52:18 crc kubenswrapper[4776]: I1125 10:52:18.499414 4776 scope.go:117] "RemoveContainer" containerID="f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40" Nov 25 10:52:18 crc kubenswrapper[4776]: E1125 10:52:18.499697 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:52:18 crc kubenswrapper[4776]: I1125 10:52:18.917683 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2qkhz" Nov 25 10:52:18 crc kubenswrapper[4776]: I1125 10:52:18.966645 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2qkhz"] Nov 25 10:52:19 crc kubenswrapper[4776]: I1125 10:52:19.516653 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2qkhz" podUID="4c5ced81-2f00-4f4f-a7f6-aa25feb3721f" containerName="registry-server" containerID="cri-o://d36d86094b66ac0c5977500500f52890aea85ed6d130f2b54a872949f0ac4080" gracePeriod=2 Nov 25 10:52:19 crc kubenswrapper[4776]: I1125 10:52:19.950906 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2qkhz" Nov 25 10:52:20 crc kubenswrapper[4776]: I1125 10:52:20.053303 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8ndg\" (UniqueName: \"kubernetes.io/projected/4c5ced81-2f00-4f4f-a7f6-aa25feb3721f-kube-api-access-s8ndg\") pod \"4c5ced81-2f00-4f4f-a7f6-aa25feb3721f\" (UID: \"4c5ced81-2f00-4f4f-a7f6-aa25feb3721f\") " Nov 25 10:52:20 crc kubenswrapper[4776]: I1125 10:52:20.053401 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c5ced81-2f00-4f4f-a7f6-aa25feb3721f-utilities\") pod \"4c5ced81-2f00-4f4f-a7f6-aa25feb3721f\" (UID: \"4c5ced81-2f00-4f4f-a7f6-aa25feb3721f\") " Nov 25 10:52:20 crc kubenswrapper[4776]: I1125 10:52:20.053434 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c5ced81-2f00-4f4f-a7f6-aa25feb3721f-catalog-content\") pod \"4c5ced81-2f00-4f4f-a7f6-aa25feb3721f\" (UID: \"4c5ced81-2f00-4f4f-a7f6-aa25feb3721f\") " Nov 25 10:52:20 crc kubenswrapper[4776]: I1125 10:52:20.055281 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c5ced81-2f00-4f4f-a7f6-aa25feb3721f-utilities" (OuterVolumeSpecName: "utilities") pod "4c5ced81-2f00-4f4f-a7f6-aa25feb3721f" (UID: "4c5ced81-2f00-4f4f-a7f6-aa25feb3721f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:52:20 crc kubenswrapper[4776]: I1125 10:52:20.061672 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c5ced81-2f00-4f4f-a7f6-aa25feb3721f-kube-api-access-s8ndg" (OuterVolumeSpecName: "kube-api-access-s8ndg") pod "4c5ced81-2f00-4f4f-a7f6-aa25feb3721f" (UID: "4c5ced81-2f00-4f4f-a7f6-aa25feb3721f"). InnerVolumeSpecName "kube-api-access-s8ndg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:52:20 crc kubenswrapper[4776]: I1125 10:52:20.101089 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c5ced81-2f00-4f4f-a7f6-aa25feb3721f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4c5ced81-2f00-4f4f-a7f6-aa25feb3721f" (UID: "4c5ced81-2f00-4f4f-a7f6-aa25feb3721f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:52:20 crc kubenswrapper[4776]: I1125 10:52:20.167739 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8ndg\" (UniqueName: \"kubernetes.io/projected/4c5ced81-2f00-4f4f-a7f6-aa25feb3721f-kube-api-access-s8ndg\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:20 crc kubenswrapper[4776]: I1125 10:52:20.168411 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c5ced81-2f00-4f4f-a7f6-aa25feb3721f-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:20 crc kubenswrapper[4776]: I1125 10:52:20.168428 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c5ced81-2f00-4f4f-a7f6-aa25feb3721f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:52:20 crc kubenswrapper[4776]: I1125 10:52:20.532565 4776 generic.go:334] "Generic (PLEG): container finished" podID="4c5ced81-2f00-4f4f-a7f6-aa25feb3721f" containerID="d36d86094b66ac0c5977500500f52890aea85ed6d130f2b54a872949f0ac4080" exitCode=0 Nov 25 10:52:20 crc kubenswrapper[4776]: I1125 10:52:20.532668 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2qkhz" event={"ID":"4c5ced81-2f00-4f4f-a7f6-aa25feb3721f","Type":"ContainerDied","Data":"d36d86094b66ac0c5977500500f52890aea85ed6d130f2b54a872949f0ac4080"} Nov 25 10:52:20 crc kubenswrapper[4776]: I1125 10:52:20.532747 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2qkhz" Nov 25 10:52:20 crc kubenswrapper[4776]: I1125 10:52:20.533062 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2qkhz" event={"ID":"4c5ced81-2f00-4f4f-a7f6-aa25feb3721f","Type":"ContainerDied","Data":"a72dfc3867d4a2fb7b0843bd33529c1765f3f001b8e5eab259ed4f23d779b296"} Nov 25 10:52:20 crc kubenswrapper[4776]: I1125 10:52:20.533117 4776 scope.go:117] "RemoveContainer" containerID="d36d86094b66ac0c5977500500f52890aea85ed6d130f2b54a872949f0ac4080" Nov 25 10:52:20 crc kubenswrapper[4776]: I1125 10:52:20.567838 4776 scope.go:117] "RemoveContainer" containerID="033ce72c34374797185f2c00ac9302cc75ace40075f4f14a4bf0b0126c375ea5" Nov 25 10:52:20 crc kubenswrapper[4776]: I1125 10:52:20.571814 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2qkhz"] Nov 25 10:52:20 crc kubenswrapper[4776]: I1125 10:52:20.581501 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2qkhz"] Nov 25 10:52:20 crc kubenswrapper[4776]: I1125 10:52:20.587754 4776 scope.go:117] "RemoveContainer" containerID="cb2469aed1090966c1081d6952160a7ffedecd66dbcbc8fa4a0bc1d811f4dc0c" Nov 25 10:52:20 crc kubenswrapper[4776]: I1125 10:52:20.622148 4776 scope.go:117] "RemoveContainer" containerID="d36d86094b66ac0c5977500500f52890aea85ed6d130f2b54a872949f0ac4080" Nov 25 10:52:20 crc kubenswrapper[4776]: E1125 10:52:20.622856 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d36d86094b66ac0c5977500500f52890aea85ed6d130f2b54a872949f0ac4080\": container with ID starting with d36d86094b66ac0c5977500500f52890aea85ed6d130f2b54a872949f0ac4080 not found: ID does not exist" containerID="d36d86094b66ac0c5977500500f52890aea85ed6d130f2b54a872949f0ac4080" Nov 25 10:52:20 crc kubenswrapper[4776]: I1125 10:52:20.623043 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d36d86094b66ac0c5977500500f52890aea85ed6d130f2b54a872949f0ac4080"} err="failed to get container status \"d36d86094b66ac0c5977500500f52890aea85ed6d130f2b54a872949f0ac4080\": rpc error: code = NotFound desc = could not find container \"d36d86094b66ac0c5977500500f52890aea85ed6d130f2b54a872949f0ac4080\": container with ID starting with d36d86094b66ac0c5977500500f52890aea85ed6d130f2b54a872949f0ac4080 not found: ID does not exist" Nov 25 10:52:20 crc kubenswrapper[4776]: I1125 10:52:20.623109 4776 scope.go:117] "RemoveContainer" containerID="033ce72c34374797185f2c00ac9302cc75ace40075f4f14a4bf0b0126c375ea5" Nov 25 10:52:20 crc kubenswrapper[4776]: E1125 10:52:20.623750 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"033ce72c34374797185f2c00ac9302cc75ace40075f4f14a4bf0b0126c375ea5\": container with ID starting with 033ce72c34374797185f2c00ac9302cc75ace40075f4f14a4bf0b0126c375ea5 not found: ID does not exist" containerID="033ce72c34374797185f2c00ac9302cc75ace40075f4f14a4bf0b0126c375ea5" Nov 25 10:52:20 crc kubenswrapper[4776]: I1125 10:52:20.623777 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"033ce72c34374797185f2c00ac9302cc75ace40075f4f14a4bf0b0126c375ea5"} err="failed to get container status \"033ce72c34374797185f2c00ac9302cc75ace40075f4f14a4bf0b0126c375ea5\": rpc error: code = NotFound desc = could not find container \"033ce72c34374797185f2c00ac9302cc75ace40075f4f14a4bf0b0126c375ea5\": container with ID starting with 033ce72c34374797185f2c00ac9302cc75ace40075f4f14a4bf0b0126c375ea5 not found: ID does not exist" Nov 25 10:52:20 crc kubenswrapper[4776]: I1125 10:52:20.623793 4776 scope.go:117] "RemoveContainer" containerID="cb2469aed1090966c1081d6952160a7ffedecd66dbcbc8fa4a0bc1d811f4dc0c" Nov 25 10:52:20 crc kubenswrapper[4776]: E1125 10:52:20.624280 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb2469aed1090966c1081d6952160a7ffedecd66dbcbc8fa4a0bc1d811f4dc0c\": container with ID starting with cb2469aed1090966c1081d6952160a7ffedecd66dbcbc8fa4a0bc1d811f4dc0c not found: ID does not exist" containerID="cb2469aed1090966c1081d6952160a7ffedecd66dbcbc8fa4a0bc1d811f4dc0c" Nov 25 10:52:20 crc kubenswrapper[4776]: I1125 10:52:20.624312 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb2469aed1090966c1081d6952160a7ffedecd66dbcbc8fa4a0bc1d811f4dc0c"} err="failed to get container status \"cb2469aed1090966c1081d6952160a7ffedecd66dbcbc8fa4a0bc1d811f4dc0c\": rpc error: code = NotFound desc = could not find container \"cb2469aed1090966c1081d6952160a7ffedecd66dbcbc8fa4a0bc1d811f4dc0c\": container with ID starting with cb2469aed1090966c1081d6952160a7ffedecd66dbcbc8fa4a0bc1d811f4dc0c not found: ID does not exist" Nov 25 10:52:21 crc kubenswrapper[4776]: I1125 10:52:21.673095 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c5ced81-2f00-4f4f-a7f6-aa25feb3721f" path="/var/lib/kubelet/pods/4c5ced81-2f00-4f4f-a7f6-aa25feb3721f/volumes" Nov 25 10:52:31 crc kubenswrapper[4776]: I1125 10:52:31.662235 4776 scope.go:117] "RemoveContainer" containerID="f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40" Nov 25 10:52:31 crc kubenswrapper[4776]: E1125 10:52:31.662755 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:52:44 crc kubenswrapper[4776]: I1125 10:52:44.662793 4776 scope.go:117] "RemoveContainer" containerID="f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40" Nov 25 10:52:44 crc kubenswrapper[4776]: E1125 10:52:44.663631 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:52:58 crc kubenswrapper[4776]: I1125 10:52:58.663503 4776 scope.go:117] "RemoveContainer" containerID="f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40" Nov 25 10:52:58 crc kubenswrapper[4776]: E1125 10:52:58.664873 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:53:11 crc kubenswrapper[4776]: I1125 10:53:11.662715 4776 scope.go:117] "RemoveContainer" containerID="f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40" Nov 25 10:53:11 crc kubenswrapper[4776]: E1125 10:53:11.664699 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:53:13 crc kubenswrapper[4776]: I1125 10:53:13.414455 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Nov 25 10:53:13 crc kubenswrapper[4776]: E1125 10:53:13.415957 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c5ced81-2f00-4f4f-a7f6-aa25feb3721f" containerName="registry-server" Nov 25 10:53:13 crc kubenswrapper[4776]: I1125 10:53:13.415995 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c5ced81-2f00-4f4f-a7f6-aa25feb3721f" containerName="registry-server" Nov 25 10:53:13 crc kubenswrapper[4776]: E1125 10:53:13.416056 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c5ced81-2f00-4f4f-a7f6-aa25feb3721f" containerName="extract-utilities" Nov 25 10:53:13 crc kubenswrapper[4776]: I1125 10:53:13.416187 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c5ced81-2f00-4f4f-a7f6-aa25feb3721f" containerName="extract-utilities" Nov 25 10:53:13 crc kubenswrapper[4776]: E1125 10:53:13.416229 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c5ced81-2f00-4f4f-a7f6-aa25feb3721f" containerName="extract-content" Nov 25 10:53:13 crc kubenswrapper[4776]: I1125 10:53:13.416246 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c5ced81-2f00-4f4f-a7f6-aa25feb3721f" containerName="extract-content" Nov 25 10:53:13 crc kubenswrapper[4776]: I1125 10:53:13.416740 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c5ced81-2f00-4f4f-a7f6-aa25feb3721f" containerName="registry-server" Nov 25 10:53:13 crc kubenswrapper[4776]: I1125 10:53:13.418692 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Nov 25 10:53:13 crc kubenswrapper[4776]: I1125 10:53:13.421227 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Nov 25 10:53:13 crc kubenswrapper[4776]: I1125 10:53:13.421525 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-rdbsh" Nov 25 10:53:13 crc kubenswrapper[4776]: I1125 10:53:13.565678 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-39e1751a-fb9e-4783-809a-8a50b438fb75\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-39e1751a-fb9e-4783-809a-8a50b438fb75\") pod \"mariadb-copy-data\" (UID: \"d4a1804d-1d0c-4976-840a-02335c611592\") " pod="openstack/mariadb-copy-data" Nov 25 10:53:13 crc kubenswrapper[4776]: I1125 10:53:13.565833 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f89bv\" (UniqueName: \"kubernetes.io/projected/d4a1804d-1d0c-4976-840a-02335c611592-kube-api-access-f89bv\") pod \"mariadb-copy-data\" (UID: \"d4a1804d-1d0c-4976-840a-02335c611592\") " pod="openstack/mariadb-copy-data" Nov 25 10:53:13 crc kubenswrapper[4776]: I1125 10:53:13.666940 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-39e1751a-fb9e-4783-809a-8a50b438fb75\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-39e1751a-fb9e-4783-809a-8a50b438fb75\") pod \"mariadb-copy-data\" (UID: \"d4a1804d-1d0c-4976-840a-02335c611592\") " pod="openstack/mariadb-copy-data" Nov 25 10:53:13 crc kubenswrapper[4776]: I1125 10:53:13.667263 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f89bv\" (UniqueName: \"kubernetes.io/projected/d4a1804d-1d0c-4976-840a-02335c611592-kube-api-access-f89bv\") pod \"mariadb-copy-data\" (UID: \"d4a1804d-1d0c-4976-840a-02335c611592\") " pod="openstack/mariadb-copy-data" Nov 25 10:53:13 crc kubenswrapper[4776]: I1125 10:53:13.670049 4776 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 10:53:13 crc kubenswrapper[4776]: I1125 10:53:13.670105 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-39e1751a-fb9e-4783-809a-8a50b438fb75\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-39e1751a-fb9e-4783-809a-8a50b438fb75\") pod \"mariadb-copy-data\" (UID: \"d4a1804d-1d0c-4976-840a-02335c611592\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/dcb215b378c5c518429fafa66a8a8a33a25795ca6d6094325011a84098e9a5e4/globalmount\"" pod="openstack/mariadb-copy-data" Nov 25 10:53:13 crc kubenswrapper[4776]: I1125 10:53:13.691250 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f89bv\" (UniqueName: \"kubernetes.io/projected/d4a1804d-1d0c-4976-840a-02335c611592-kube-api-access-f89bv\") pod \"mariadb-copy-data\" (UID: \"d4a1804d-1d0c-4976-840a-02335c611592\") " pod="openstack/mariadb-copy-data" Nov 25 10:53:13 crc kubenswrapper[4776]: I1125 10:53:13.708889 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-39e1751a-fb9e-4783-809a-8a50b438fb75\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-39e1751a-fb9e-4783-809a-8a50b438fb75\") pod \"mariadb-copy-data\" (UID: \"d4a1804d-1d0c-4976-840a-02335c611592\") " pod="openstack/mariadb-copy-data" Nov 25 10:53:13 crc kubenswrapper[4776]: I1125 10:53:13.741696 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Nov 25 10:53:14 crc kubenswrapper[4776]: I1125 10:53:14.279816 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Nov 25 10:53:14 crc kubenswrapper[4776]: I1125 10:53:14.967051 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"d4a1804d-1d0c-4976-840a-02335c611592","Type":"ContainerStarted","Data":"398d97eaeb03fef58b8fdf1ee14c3011b9e92d960abf3bfbc7e3c61f4f792c94"} Nov 25 10:53:14 crc kubenswrapper[4776]: I1125 10:53:14.967249 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"d4a1804d-1d0c-4976-840a-02335c611592","Type":"ContainerStarted","Data":"2304778b2312639657687cd5bcf5b78dc0faba9ba827ba060427e9a53c57d948"} Nov 25 10:53:14 crc kubenswrapper[4776]: I1125 10:53:14.994299 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=2.994272656 podStartE2EDuration="2.994272656s" podCreationTimestamp="2025-11-25 10:53:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:53:14.984417219 +0000 UTC m=+5340.025476772" watchObservedRunningTime="2025-11-25 10:53:14.994272656 +0000 UTC m=+5340.035332249" Nov 25 10:53:17 crc kubenswrapper[4776]: I1125 10:53:17.766711 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Nov 25 10:53:17 crc kubenswrapper[4776]: I1125 10:53:17.769592 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 25 10:53:17 crc kubenswrapper[4776]: I1125 10:53:17.777985 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Nov 25 10:53:17 crc kubenswrapper[4776]: I1125 10:53:17.942604 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgm87\" (UniqueName: \"kubernetes.io/projected/b6558bd5-9181-402a-affa-429c68574228-kube-api-access-dgm87\") pod \"mariadb-client\" (UID: \"b6558bd5-9181-402a-affa-429c68574228\") " pod="openstack/mariadb-client" Nov 25 10:53:18 crc kubenswrapper[4776]: I1125 10:53:18.043723 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgm87\" (UniqueName: \"kubernetes.io/projected/b6558bd5-9181-402a-affa-429c68574228-kube-api-access-dgm87\") pod \"mariadb-client\" (UID: \"b6558bd5-9181-402a-affa-429c68574228\") " pod="openstack/mariadb-client" Nov 25 10:53:18 crc kubenswrapper[4776]: I1125 10:53:18.066672 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgm87\" (UniqueName: \"kubernetes.io/projected/b6558bd5-9181-402a-affa-429c68574228-kube-api-access-dgm87\") pod \"mariadb-client\" (UID: \"b6558bd5-9181-402a-affa-429c68574228\") " pod="openstack/mariadb-client" Nov 25 10:53:18 crc kubenswrapper[4776]: I1125 10:53:18.091520 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 25 10:53:18 crc kubenswrapper[4776]: I1125 10:53:18.514344 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Nov 25 10:53:18 crc kubenswrapper[4776]: I1125 10:53:18.996202 4776 generic.go:334] "Generic (PLEG): container finished" podID="b6558bd5-9181-402a-affa-429c68574228" containerID="ef0b5866a0822f88197714a524be43881cd9ff6f5151bd4079365aff874aa509" exitCode=0 Nov 25 10:53:18 crc kubenswrapper[4776]: I1125 10:53:18.996260 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"b6558bd5-9181-402a-affa-429c68574228","Type":"ContainerDied","Data":"ef0b5866a0822f88197714a524be43881cd9ff6f5151bd4079365aff874aa509"} Nov 25 10:53:18 crc kubenswrapper[4776]: I1125 10:53:18.996296 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"b6558bd5-9181-402a-affa-429c68574228","Type":"ContainerStarted","Data":"88b44fe1c104a776b1e4018a4c37c7f59b3236af95b9eb0219609439a86bfbe3"} Nov 25 10:53:20 crc kubenswrapper[4776]: I1125 10:53:20.370512 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 25 10:53:20 crc kubenswrapper[4776]: I1125 10:53:20.397509 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_b6558bd5-9181-402a-affa-429c68574228/mariadb-client/0.log" Nov 25 10:53:20 crc kubenswrapper[4776]: I1125 10:53:20.426629 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Nov 25 10:53:20 crc kubenswrapper[4776]: I1125 10:53:20.433107 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Nov 25 10:53:20 crc kubenswrapper[4776]: I1125 10:53:20.495355 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgm87\" (UniqueName: \"kubernetes.io/projected/b6558bd5-9181-402a-affa-429c68574228-kube-api-access-dgm87\") pod \"b6558bd5-9181-402a-affa-429c68574228\" (UID: \"b6558bd5-9181-402a-affa-429c68574228\") " Nov 25 10:53:20 crc kubenswrapper[4776]: I1125 10:53:20.502442 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6558bd5-9181-402a-affa-429c68574228-kube-api-access-dgm87" (OuterVolumeSpecName: "kube-api-access-dgm87") pod "b6558bd5-9181-402a-affa-429c68574228" (UID: "b6558bd5-9181-402a-affa-429c68574228"). InnerVolumeSpecName "kube-api-access-dgm87". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:53:20 crc kubenswrapper[4776]: I1125 10:53:20.545286 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Nov 25 10:53:20 crc kubenswrapper[4776]: E1125 10:53:20.545697 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6558bd5-9181-402a-affa-429c68574228" containerName="mariadb-client" Nov 25 10:53:20 crc kubenswrapper[4776]: I1125 10:53:20.545715 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6558bd5-9181-402a-affa-429c68574228" containerName="mariadb-client" Nov 25 10:53:20 crc kubenswrapper[4776]: I1125 10:53:20.545906 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6558bd5-9181-402a-affa-429c68574228" containerName="mariadb-client" Nov 25 10:53:20 crc kubenswrapper[4776]: I1125 10:53:20.546657 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 25 10:53:20 crc kubenswrapper[4776]: I1125 10:53:20.553027 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Nov 25 10:53:20 crc kubenswrapper[4776]: I1125 10:53:20.597302 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgm87\" (UniqueName: \"kubernetes.io/projected/b6558bd5-9181-402a-affa-429c68574228-kube-api-access-dgm87\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:20 crc kubenswrapper[4776]: I1125 10:53:20.699376 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26nfr\" (UniqueName: \"kubernetes.io/projected/ff91dd93-dbe8-4c5c-b0e7-eaec3f15be27-kube-api-access-26nfr\") pod \"mariadb-client\" (UID: \"ff91dd93-dbe8-4c5c-b0e7-eaec3f15be27\") " pod="openstack/mariadb-client" Nov 25 10:53:20 crc kubenswrapper[4776]: I1125 10:53:20.801487 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26nfr\" (UniqueName: \"kubernetes.io/projected/ff91dd93-dbe8-4c5c-b0e7-eaec3f15be27-kube-api-access-26nfr\") pod \"mariadb-client\" (UID: \"ff91dd93-dbe8-4c5c-b0e7-eaec3f15be27\") " pod="openstack/mariadb-client" Nov 25 10:53:20 crc kubenswrapper[4776]: I1125 10:53:20.819375 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26nfr\" (UniqueName: \"kubernetes.io/projected/ff91dd93-dbe8-4c5c-b0e7-eaec3f15be27-kube-api-access-26nfr\") pod \"mariadb-client\" (UID: \"ff91dd93-dbe8-4c5c-b0e7-eaec3f15be27\") " pod="openstack/mariadb-client" Nov 25 10:53:20 crc kubenswrapper[4776]: I1125 10:53:20.866309 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 25 10:53:21 crc kubenswrapper[4776]: I1125 10:53:21.013200 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88b44fe1c104a776b1e4018a4c37c7f59b3236af95b9eb0219609439a86bfbe3" Nov 25 10:53:21 crc kubenswrapper[4776]: I1125 10:53:21.013238 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 25 10:53:21 crc kubenswrapper[4776]: I1125 10:53:21.035237 4776 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/mariadb-client" oldPodUID="b6558bd5-9181-402a-affa-429c68574228" podUID="ff91dd93-dbe8-4c5c-b0e7-eaec3f15be27" Nov 25 10:53:21 crc kubenswrapper[4776]: I1125 10:53:21.319486 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Nov 25 10:53:21 crc kubenswrapper[4776]: I1125 10:53:21.671865 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6558bd5-9181-402a-affa-429c68574228" path="/var/lib/kubelet/pods/b6558bd5-9181-402a-affa-429c68574228/volumes" Nov 25 10:53:22 crc kubenswrapper[4776]: I1125 10:53:22.027672 4776 generic.go:334] "Generic (PLEG): container finished" podID="ff91dd93-dbe8-4c5c-b0e7-eaec3f15be27" containerID="774ee73593d2eaba78e3c336f42ec751ccdef918d45b933cee975558c71ecac4" exitCode=0 Nov 25 10:53:22 crc kubenswrapper[4776]: I1125 10:53:22.027743 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"ff91dd93-dbe8-4c5c-b0e7-eaec3f15be27","Type":"ContainerDied","Data":"774ee73593d2eaba78e3c336f42ec751ccdef918d45b933cee975558c71ecac4"} Nov 25 10:53:22 crc kubenswrapper[4776]: I1125 10:53:22.027822 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"ff91dd93-dbe8-4c5c-b0e7-eaec3f15be27","Type":"ContainerStarted","Data":"63847d2dace8e968a64ed4919e44440ee642687ba967a3bbc6f7a8ad91be44a1"} Nov 25 10:53:23 crc kubenswrapper[4776]: I1125 10:53:23.390262 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 25 10:53:23 crc kubenswrapper[4776]: I1125 10:53:23.416490 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_ff91dd93-dbe8-4c5c-b0e7-eaec3f15be27/mariadb-client/0.log" Nov 25 10:53:23 crc kubenswrapper[4776]: I1125 10:53:23.444851 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Nov 25 10:53:23 crc kubenswrapper[4776]: I1125 10:53:23.452185 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Nov 25 10:53:23 crc kubenswrapper[4776]: I1125 10:53:23.561753 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26nfr\" (UniqueName: \"kubernetes.io/projected/ff91dd93-dbe8-4c5c-b0e7-eaec3f15be27-kube-api-access-26nfr\") pod \"ff91dd93-dbe8-4c5c-b0e7-eaec3f15be27\" (UID: \"ff91dd93-dbe8-4c5c-b0e7-eaec3f15be27\") " Nov 25 10:53:23 crc kubenswrapper[4776]: I1125 10:53:23.574391 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff91dd93-dbe8-4c5c-b0e7-eaec3f15be27-kube-api-access-26nfr" (OuterVolumeSpecName: "kube-api-access-26nfr") pod "ff91dd93-dbe8-4c5c-b0e7-eaec3f15be27" (UID: "ff91dd93-dbe8-4c5c-b0e7-eaec3f15be27"). InnerVolumeSpecName "kube-api-access-26nfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:53:23 crc kubenswrapper[4776]: I1125 10:53:23.663769 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26nfr\" (UniqueName: \"kubernetes.io/projected/ff91dd93-dbe8-4c5c-b0e7-eaec3f15be27-kube-api-access-26nfr\") on node \"crc\" DevicePath \"\"" Nov 25 10:53:23 crc kubenswrapper[4776]: I1125 10:53:23.676229 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff91dd93-dbe8-4c5c-b0e7-eaec3f15be27" path="/var/lib/kubelet/pods/ff91dd93-dbe8-4c5c-b0e7-eaec3f15be27/volumes" Nov 25 10:53:24 crc kubenswrapper[4776]: I1125 10:53:24.043899 4776 scope.go:117] "RemoveContainer" containerID="774ee73593d2eaba78e3c336f42ec751ccdef918d45b933cee975558c71ecac4" Nov 25 10:53:24 crc kubenswrapper[4776]: I1125 10:53:24.043966 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Nov 25 10:53:25 crc kubenswrapper[4776]: I1125 10:53:25.669864 4776 scope.go:117] "RemoveContainer" containerID="f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40" Nov 25 10:53:25 crc kubenswrapper[4776]: E1125 10:53:25.670259 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:53:31 crc kubenswrapper[4776]: I1125 10:53:31.721945 4776 scope.go:117] "RemoveContainer" containerID="75889533bb71117d5bf6b3d7ef073f97a4f09f60d30a706091fc8145e0f0bb1f" Nov 25 10:53:31 crc kubenswrapper[4776]: I1125 10:53:31.759865 4776 scope.go:117] "RemoveContainer" containerID="186b4dd940c3d859642772cffd1be431d22aa57aa04f371b7bce8b6627a54a91" Nov 25 10:53:38 crc kubenswrapper[4776]: I1125 10:53:38.663018 4776 scope.go:117] "RemoveContainer" containerID="f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40" Nov 25 10:53:38 crc kubenswrapper[4776]: E1125 10:53:38.663875 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:53:53 crc kubenswrapper[4776]: I1125 10:53:53.662172 4776 scope.go:117] "RemoveContainer" containerID="f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40" Nov 25 10:53:53 crc kubenswrapper[4776]: E1125 10:53:53.662965 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:53:58 crc kubenswrapper[4776]: I1125 10:53:58.698146 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6vlcg"] Nov 25 10:53:58 crc kubenswrapper[4776]: E1125 10:53:58.698905 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff91dd93-dbe8-4c5c-b0e7-eaec3f15be27" containerName="mariadb-client" Nov 25 10:53:58 crc kubenswrapper[4776]: I1125 10:53:58.698957 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff91dd93-dbe8-4c5c-b0e7-eaec3f15be27" containerName="mariadb-client" Nov 25 10:53:58 crc kubenswrapper[4776]: I1125 10:53:58.699207 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff91dd93-dbe8-4c5c-b0e7-eaec3f15be27" containerName="mariadb-client" Nov 25 10:53:58 crc kubenswrapper[4776]: I1125 10:53:58.700682 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6vlcg" Nov 25 10:53:58 crc kubenswrapper[4776]: I1125 10:53:58.754854 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6vlcg"] Nov 25 10:53:58 crc kubenswrapper[4776]: I1125 10:53:58.756577 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3-utilities\") pod \"community-operators-6vlcg\" (UID: \"6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3\") " pod="openshift-marketplace/community-operators-6vlcg" Nov 25 10:53:58 crc kubenswrapper[4776]: I1125 10:53:58.756626 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vtm9\" (UniqueName: \"kubernetes.io/projected/6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3-kube-api-access-4vtm9\") pod \"community-operators-6vlcg\" (UID: \"6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3\") " pod="openshift-marketplace/community-operators-6vlcg" Nov 25 10:53:58 crc kubenswrapper[4776]: I1125 10:53:58.756662 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3-catalog-content\") pod \"community-operators-6vlcg\" (UID: \"6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3\") " pod="openshift-marketplace/community-operators-6vlcg" Nov 25 10:53:58 crc kubenswrapper[4776]: I1125 10:53:58.858788 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3-utilities\") pod \"community-operators-6vlcg\" (UID: \"6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3\") " pod="openshift-marketplace/community-operators-6vlcg" Nov 25 10:53:58 crc kubenswrapper[4776]: I1125 10:53:58.859059 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vtm9\" (UniqueName: \"kubernetes.io/projected/6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3-kube-api-access-4vtm9\") pod \"community-operators-6vlcg\" (UID: \"6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3\") " pod="openshift-marketplace/community-operators-6vlcg" Nov 25 10:53:58 crc kubenswrapper[4776]: I1125 10:53:58.859116 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3-catalog-content\") pod \"community-operators-6vlcg\" (UID: \"6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3\") " pod="openshift-marketplace/community-operators-6vlcg" Nov 25 10:53:58 crc kubenswrapper[4776]: I1125 10:53:58.859513 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3-utilities\") pod \"community-operators-6vlcg\" (UID: \"6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3\") " pod="openshift-marketplace/community-operators-6vlcg" Nov 25 10:53:58 crc kubenswrapper[4776]: I1125 10:53:58.859567 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3-catalog-content\") pod \"community-operators-6vlcg\" (UID: \"6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3\") " pod="openshift-marketplace/community-operators-6vlcg" Nov 25 10:53:58 crc kubenswrapper[4776]: I1125 10:53:58.885519 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vtm9\" (UniqueName: \"kubernetes.io/projected/6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3-kube-api-access-4vtm9\") pod \"community-operators-6vlcg\" (UID: \"6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3\") " pod="openshift-marketplace/community-operators-6vlcg" Nov 25 10:53:59 crc kubenswrapper[4776]: I1125 10:53:59.085936 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6vlcg" Nov 25 10:53:59 crc kubenswrapper[4776]: I1125 10:53:59.564478 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6vlcg"] Nov 25 10:54:00 crc kubenswrapper[4776]: I1125 10:54:00.365275 4776 generic.go:334] "Generic (PLEG): container finished" podID="6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3" containerID="70a933b3ff6e4e88282229a4bc68c353347ef40229bcc36009a8f36b1f4ea8b0" exitCode=0 Nov 25 10:54:00 crc kubenswrapper[4776]: I1125 10:54:00.365332 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vlcg" event={"ID":"6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3","Type":"ContainerDied","Data":"70a933b3ff6e4e88282229a4bc68c353347ef40229bcc36009a8f36b1f4ea8b0"} Nov 25 10:54:00 crc kubenswrapper[4776]: I1125 10:54:00.365371 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vlcg" event={"ID":"6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3","Type":"ContainerStarted","Data":"d96309eb17ccf88e784b92865af330aad0221adad4f54efabe9de68f425daeaa"} Nov 25 10:54:00 crc kubenswrapper[4776]: I1125 10:54:00.367820 4776 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 10:54:02 crc kubenswrapper[4776]: I1125 10:54:02.383849 4776 generic.go:334] "Generic (PLEG): container finished" podID="6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3" containerID="83cf57c9f00865d38f8d72a34a1cd168267ccc529971bbb6599420d783c3c431" exitCode=0 Nov 25 10:54:02 crc kubenswrapper[4776]: I1125 10:54:02.383903 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vlcg" event={"ID":"6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3","Type":"ContainerDied","Data":"83cf57c9f00865d38f8d72a34a1cd168267ccc529971bbb6599420d783c3c431"} Nov 25 10:54:03 crc kubenswrapper[4776]: I1125 10:54:03.396885 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vlcg" event={"ID":"6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3","Type":"ContainerStarted","Data":"d3042da4bf2c571f150be94b720d2186a180299428a126f5b280e14d3070dbc4"} Nov 25 10:54:03 crc kubenswrapper[4776]: I1125 10:54:03.421410 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6vlcg" podStartSLOduration=3.006360868 podStartE2EDuration="5.421392746s" podCreationTimestamp="2025-11-25 10:53:58 +0000 UTC" firstStartedPulling="2025-11-25 10:54:00.367501002 +0000 UTC m=+5385.408560555" lastFinishedPulling="2025-11-25 10:54:02.78253287 +0000 UTC m=+5387.823592433" observedRunningTime="2025-11-25 10:54:03.414175265 +0000 UTC m=+5388.455234828" watchObservedRunningTime="2025-11-25 10:54:03.421392746 +0000 UTC m=+5388.462452299" Nov 25 10:54:04 crc kubenswrapper[4776]: I1125 10:54:04.662679 4776 scope.go:117] "RemoveContainer" containerID="f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40" Nov 25 10:54:04 crc kubenswrapper[4776]: E1125 10:54:04.663201 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:54:09 crc kubenswrapper[4776]: I1125 10:54:09.086110 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6vlcg" Nov 25 10:54:09 crc kubenswrapper[4776]: I1125 10:54:09.086439 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6vlcg" Nov 25 10:54:09 crc kubenswrapper[4776]: I1125 10:54:09.139844 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6vlcg" Nov 25 10:54:09 crc kubenswrapper[4776]: I1125 10:54:09.482756 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6vlcg" Nov 25 10:54:09 crc kubenswrapper[4776]: I1125 10:54:09.521989 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6vlcg"] Nov 25 10:54:11 crc kubenswrapper[4776]: I1125 10:54:11.453775 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6vlcg" podUID="6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3" containerName="registry-server" containerID="cri-o://d3042da4bf2c571f150be94b720d2186a180299428a126f5b280e14d3070dbc4" gracePeriod=2 Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.237023 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.238680 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.243131 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.243156 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.243453 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.243627 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-fvnnv" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.243709 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.251148 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.263928 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.265661 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.284237 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.285981 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.292680 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.304667 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.355756 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-283f1f3e-8133-41e2-8eb1-b6e0550bd384\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-283f1f3e-8133-41e2-8eb1-b6e0550bd384\") pod \"ovsdbserver-nb-2\" (UID: \"f5b7e9cb-0aa9-434f-b75c-783829b9ba56\") " pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.355835 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d16f6ea-879b-4b44-a9aa-187c59681215-config\") pod \"ovsdbserver-nb-0\" (UID: \"3d16f6ea-879b-4b44-a9aa-187c59681215\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.355867 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5b7e9cb-0aa9-434f-b75c-783829b9ba56-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"f5b7e9cb-0aa9-434f-b75c-783829b9ba56\") " pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.355896 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3d16f6ea-879b-4b44-a9aa-187c59681215-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"3d16f6ea-879b-4b44-a9aa-187c59681215\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.355925 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f8bd2df8-2364-4758-8c88-77c6b9f39e89\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f8bd2df8-2364-4758-8c88-77c6b9f39e89\") pod \"ovsdbserver-nb-0\" (UID: \"3d16f6ea-879b-4b44-a9aa-187c59681215\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.355962 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f5b7e9cb-0aa9-434f-b75c-783829b9ba56-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"f5b7e9cb-0aa9-434f-b75c-783829b9ba56\") " pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.355995 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d16f6ea-879b-4b44-a9aa-187c59681215-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3d16f6ea-879b-4b44-a9aa-187c59681215\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.356018 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5b7e9cb-0aa9-434f-b75c-783829b9ba56-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"f5b7e9cb-0aa9-434f-b75c-783829b9ba56\") " pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.356042 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d16f6ea-879b-4b44-a9aa-187c59681215-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3d16f6ea-879b-4b44-a9aa-187c59681215\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.356162 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjrfp\" (UniqueName: \"kubernetes.io/projected/f5b7e9cb-0aa9-434f-b75c-783829b9ba56-kube-api-access-sjrfp\") pod \"ovsdbserver-nb-2\" (UID: \"f5b7e9cb-0aa9-434f-b75c-783829b9ba56\") " pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.356257 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5b7e9cb-0aa9-434f-b75c-783829b9ba56-config\") pod \"ovsdbserver-nb-2\" (UID: \"f5b7e9cb-0aa9-434f-b75c-783829b9ba56\") " pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.356295 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d16f6ea-879b-4b44-a9aa-187c59681215-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"3d16f6ea-879b-4b44-a9aa-187c59681215\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.356320 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f5b7e9cb-0aa9-434f-b75c-783829b9ba56-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"f5b7e9cb-0aa9-434f-b75c-783829b9ba56\") " pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.356400 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4wwm\" (UniqueName: \"kubernetes.io/projected/3d16f6ea-879b-4b44-a9aa-187c59681215-kube-api-access-g4wwm\") pod \"ovsdbserver-nb-0\" (UID: \"3d16f6ea-879b-4b44-a9aa-187c59681215\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.356421 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5b7e9cb-0aa9-434f-b75c-783829b9ba56-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"f5b7e9cb-0aa9-434f-b75c-783829b9ba56\") " pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.356445 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d16f6ea-879b-4b44-a9aa-187c59681215-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"3d16f6ea-879b-4b44-a9aa-187c59681215\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.448332 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.449885 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.452232 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.458532 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.458847 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.459967 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d16f6ea-879b-4b44-a9aa-187c59681215-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3d16f6ea-879b-4b44-a9aa-187c59681215\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.459999 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5b7e9cb-0aa9-434f-b75c-783829b9ba56-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"f5b7e9cb-0aa9-434f-b75c-783829b9ba56\") " pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.460026 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d16f6ea-879b-4b44-a9aa-187c59681215-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3d16f6ea-879b-4b44-a9aa-187c59681215\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.460048 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjrfp\" (UniqueName: \"kubernetes.io/projected/f5b7e9cb-0aa9-434f-b75c-783829b9ba56-kube-api-access-sjrfp\") pod \"ovsdbserver-nb-2\" (UID: \"f5b7e9cb-0aa9-434f-b75c-783829b9ba56\") " pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.460116 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7a3aa625-3532-4d06-aa7e-e91af4df2f4a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7a3aa625-3532-4d06-aa7e-e91af4df2f4a\") pod \"ovsdbserver-nb-1\" (UID: \"bce8180f-fb70-4f9c-bc00-3a39e62da7a3\") " pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.460145 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5b7e9cb-0aa9-434f-b75c-783829b9ba56-config\") pod \"ovsdbserver-nb-2\" (UID: \"f5b7e9cb-0aa9-434f-b75c-783829b9ba56\") " pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.460167 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d16f6ea-879b-4b44-a9aa-187c59681215-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"3d16f6ea-879b-4b44-a9aa-187c59681215\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.460183 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f5b7e9cb-0aa9-434f-b75c-783829b9ba56-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"f5b7e9cb-0aa9-434f-b75c-783829b9ba56\") " pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.460207 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bce8180f-fb70-4f9c-bc00-3a39e62da7a3-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"bce8180f-fb70-4f9c-bc00-3a39e62da7a3\") " pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.460234 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4wwm\" (UniqueName: \"kubernetes.io/projected/3d16f6ea-879b-4b44-a9aa-187c59681215-kube-api-access-g4wwm\") pod \"ovsdbserver-nb-0\" (UID: \"3d16f6ea-879b-4b44-a9aa-187c59681215\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.460254 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5b7e9cb-0aa9-434f-b75c-783829b9ba56-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"f5b7e9cb-0aa9-434f-b75c-783829b9ba56\") " pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.460273 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bce8180f-fb70-4f9c-bc00-3a39e62da7a3-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"bce8180f-fb70-4f9c-bc00-3a39e62da7a3\") " pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.460294 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d16f6ea-879b-4b44-a9aa-187c59681215-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"3d16f6ea-879b-4b44-a9aa-187c59681215\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.460315 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bce8180f-fb70-4f9c-bc00-3a39e62da7a3-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"bce8180f-fb70-4f9c-bc00-3a39e62da7a3\") " pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.460344 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-283f1f3e-8133-41e2-8eb1-b6e0550bd384\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-283f1f3e-8133-41e2-8eb1-b6e0550bd384\") pod \"ovsdbserver-nb-2\" (UID: \"f5b7e9cb-0aa9-434f-b75c-783829b9ba56\") " pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.460393 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d16f6ea-879b-4b44-a9aa-187c59681215-config\") pod \"ovsdbserver-nb-0\" (UID: \"3d16f6ea-879b-4b44-a9aa-187c59681215\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.460413 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bce8180f-fb70-4f9c-bc00-3a39e62da7a3-config\") pod \"ovsdbserver-nb-1\" (UID: \"bce8180f-fb70-4f9c-bc00-3a39e62da7a3\") " pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.460435 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5b7e9cb-0aa9-434f-b75c-783829b9ba56-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"f5b7e9cb-0aa9-434f-b75c-783829b9ba56\") " pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.460452 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tc9ld\" (UniqueName: \"kubernetes.io/projected/bce8180f-fb70-4f9c-bc00-3a39e62da7a3-kube-api-access-tc9ld\") pod \"ovsdbserver-nb-1\" (UID: \"bce8180f-fb70-4f9c-bc00-3a39e62da7a3\") " pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.460470 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bce8180f-fb70-4f9c-bc00-3a39e62da7a3-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"bce8180f-fb70-4f9c-bc00-3a39e62da7a3\") " pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.460489 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3d16f6ea-879b-4b44-a9aa-187c59681215-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"3d16f6ea-879b-4b44-a9aa-187c59681215\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.460515 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f8bd2df8-2364-4758-8c88-77c6b9f39e89\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f8bd2df8-2364-4758-8c88-77c6b9f39e89\") pod \"ovsdbserver-nb-0\" (UID: \"3d16f6ea-879b-4b44-a9aa-187c59681215\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.460542 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f5b7e9cb-0aa9-434f-b75c-783829b9ba56-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"f5b7e9cb-0aa9-434f-b75c-783829b9ba56\") " pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.460566 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bce8180f-fb70-4f9c-bc00-3a39e62da7a3-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"bce8180f-fb70-4f9c-bc00-3a39e62da7a3\") " pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.461562 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-fvrmd" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.463794 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d16f6ea-879b-4b44-a9aa-187c59681215-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"3d16f6ea-879b-4b44-a9aa-187c59681215\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.466536 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3d16f6ea-879b-4b44-a9aa-187c59681215-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"3d16f6ea-879b-4b44-a9aa-187c59681215\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.470038 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d16f6ea-879b-4b44-a9aa-187c59681215-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"3d16f6ea-879b-4b44-a9aa-187c59681215\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.470371 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d16f6ea-879b-4b44-a9aa-187c59681215-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3d16f6ea-879b-4b44-a9aa-187c59681215\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.471124 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f5b7e9cb-0aa9-434f-b75c-783829b9ba56-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"f5b7e9cb-0aa9-434f-b75c-783829b9ba56\") " pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.471589 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d16f6ea-879b-4b44-a9aa-187c59681215-config\") pod \"ovsdbserver-nb-0\" (UID: \"3d16f6ea-879b-4b44-a9aa-187c59681215\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.472685 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f5b7e9cb-0aa9-434f-b75c-783829b9ba56-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"f5b7e9cb-0aa9-434f-b75c-783829b9ba56\") " pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.472825 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5b7e9cb-0aa9-434f-b75c-783829b9ba56-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"f5b7e9cb-0aa9-434f-b75c-783829b9ba56\") " pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.473881 4776 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.473914 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-283f1f3e-8133-41e2-8eb1-b6e0550bd384\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-283f1f3e-8133-41e2-8eb1-b6e0550bd384\") pod \"ovsdbserver-nb-2\" (UID: \"f5b7e9cb-0aa9-434f-b75c-783829b9ba56\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2432a05d53ca1b6b190765461c24347e4475c01dddc3d9388d4d49a06db29613/globalmount\"" pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.478127 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5b7e9cb-0aa9-434f-b75c-783829b9ba56-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"f5b7e9cb-0aa9-434f-b75c-783829b9ba56\") " pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.480946 4776 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.480957 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5b7e9cb-0aa9-434f-b75c-783829b9ba56-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"f5b7e9cb-0aa9-434f-b75c-783829b9ba56\") " pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.480984 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f8bd2df8-2364-4758-8c88-77c6b9f39e89\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f8bd2df8-2364-4758-8c88-77c6b9f39e89\") pod \"ovsdbserver-nb-0\" (UID: \"3d16f6ea-879b-4b44-a9aa-187c59681215\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0715c107060256b305641934e96ee8ca74b5fdc4f68759ff3495e052f02e0d42/globalmount\"" pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.485201 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d16f6ea-879b-4b44-a9aa-187c59681215-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3d16f6ea-879b-4b44-a9aa-187c59681215\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.489896 4776 generic.go:334] "Generic (PLEG): container finished" podID="6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3" containerID="d3042da4bf2c571f150be94b720d2186a180299428a126f5b280e14d3070dbc4" exitCode=0 Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.489971 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vlcg" event={"ID":"6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3","Type":"ContainerDied","Data":"d3042da4bf2c571f150be94b720d2186a180299428a126f5b280e14d3070dbc4"} Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.490842 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5b7e9cb-0aa9-434f-b75c-783829b9ba56-config\") pod \"ovsdbserver-nb-2\" (UID: \"f5b7e9cb-0aa9-434f-b75c-783829b9ba56\") " pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.491847 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4wwm\" (UniqueName: \"kubernetes.io/projected/3d16f6ea-879b-4b44-a9aa-187c59681215-kube-api-access-g4wwm\") pod \"ovsdbserver-nb-0\" (UID: \"3d16f6ea-879b-4b44-a9aa-187c59681215\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.498437 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.500325 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.501167 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjrfp\" (UniqueName: \"kubernetes.io/projected/f5b7e9cb-0aa9-434f-b75c-783829b9ba56-kube-api-access-sjrfp\") pod \"ovsdbserver-nb-2\" (UID: \"f5b7e9cb-0aa9-434f-b75c-783829b9ba56\") " pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.507536 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.509250 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.524150 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.543819 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-283f1f3e-8133-41e2-8eb1-b6e0550bd384\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-283f1f3e-8133-41e2-8eb1-b6e0550bd384\") pod \"ovsdbserver-nb-2\" (UID: \"f5b7e9cb-0aa9-434f-b75c-783829b9ba56\") " pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.545638 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.550401 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f8bd2df8-2364-4758-8c88-77c6b9f39e89\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f8bd2df8-2364-4758-8c88-77c6b9f39e89\") pod \"ovsdbserver-nb-0\" (UID: \"3d16f6ea-879b-4b44-a9aa-187c59681215\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.553040 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.557446 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.562221 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce32cbf1-d6a1-4361-8921-ce0abcda4667-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ce32cbf1-d6a1-4361-8921-ce0abcda4667\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.562359 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce32cbf1-d6a1-4361-8921-ce0abcda4667-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ce32cbf1-d6a1-4361-8921-ce0abcda4667\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.562447 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7a3aa625-3532-4d06-aa7e-e91af4df2f4a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7a3aa625-3532-4d06-aa7e-e91af4df2f4a\") pod \"ovsdbserver-nb-1\" (UID: \"bce8180f-fb70-4f9c-bc00-3a39e62da7a3\") " pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.562533 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce32cbf1-d6a1-4361-8921-ce0abcda4667-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ce32cbf1-d6a1-4361-8921-ce0abcda4667\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.562680 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bce8180f-fb70-4f9c-bc00-3a39e62da7a3-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"bce8180f-fb70-4f9c-bc00-3a39e62da7a3\") " pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.562763 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bce8180f-fb70-4f9c-bc00-3a39e62da7a3-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"bce8180f-fb70-4f9c-bc00-3a39e62da7a3\") " pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.562835 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bce8180f-fb70-4f9c-bc00-3a39e62da7a3-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"bce8180f-fb70-4f9c-bc00-3a39e62da7a3\") " pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.562924 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g746l\" (UniqueName: \"kubernetes.io/projected/ce32cbf1-d6a1-4361-8921-ce0abcda4667-kube-api-access-g746l\") pod \"ovsdbserver-sb-0\" (UID: \"ce32cbf1-d6a1-4361-8921-ce0abcda4667\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.563037 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce32cbf1-d6a1-4361-8921-ce0abcda4667-config\") pod \"ovsdbserver-sb-0\" (UID: \"ce32cbf1-d6a1-4361-8921-ce0abcda4667\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.563192 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ce32cbf1-d6a1-4361-8921-ce0abcda4667-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ce32cbf1-d6a1-4361-8921-ce0abcda4667\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.563288 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ce32cbf1-d6a1-4361-8921-ce0abcda4667-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ce32cbf1-d6a1-4361-8921-ce0abcda4667\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.563386 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-505e6906-c7aa-455d-a32e-fbbfe03e29a4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-505e6906-c7aa-455d-a32e-fbbfe03e29a4\") pod \"ovsdbserver-sb-0\" (UID: \"ce32cbf1-d6a1-4361-8921-ce0abcda4667\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.563521 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bce8180f-fb70-4f9c-bc00-3a39e62da7a3-config\") pod \"ovsdbserver-nb-1\" (UID: \"bce8180f-fb70-4f9c-bc00-3a39e62da7a3\") " pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.563640 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tc9ld\" (UniqueName: \"kubernetes.io/projected/bce8180f-fb70-4f9c-bc00-3a39e62da7a3-kube-api-access-tc9ld\") pod \"ovsdbserver-nb-1\" (UID: \"bce8180f-fb70-4f9c-bc00-3a39e62da7a3\") " pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.563719 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bce8180f-fb70-4f9c-bc00-3a39e62da7a3-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"bce8180f-fb70-4f9c-bc00-3a39e62da7a3\") " pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.563815 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bce8180f-fb70-4f9c-bc00-3a39e62da7a3-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"bce8180f-fb70-4f9c-bc00-3a39e62da7a3\") " pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.565230 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bce8180f-fb70-4f9c-bc00-3a39e62da7a3-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"bce8180f-fb70-4f9c-bc00-3a39e62da7a3\") " pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.565242 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bce8180f-fb70-4f9c-bc00-3a39e62da7a3-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"bce8180f-fb70-4f9c-bc00-3a39e62da7a3\") " pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.566383 4776 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.566445 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7a3aa625-3532-4d06-aa7e-e91af4df2f4a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7a3aa625-3532-4d06-aa7e-e91af4df2f4a\") pod \"ovsdbserver-nb-1\" (UID: \"bce8180f-fb70-4f9c-bc00-3a39e62da7a3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1fa36152be99da2d0f9b06e2b73316ce884dbd51a91178676bcb30960e44b5b0/globalmount\"" pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.566736 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bce8180f-fb70-4f9c-bc00-3a39e62da7a3-config\") pod \"ovsdbserver-nb-1\" (UID: \"bce8180f-fb70-4f9c-bc00-3a39e62da7a3\") " pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.567393 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bce8180f-fb70-4f9c-bc00-3a39e62da7a3-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"bce8180f-fb70-4f9c-bc00-3a39e62da7a3\") " pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.569289 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bce8180f-fb70-4f9c-bc00-3a39e62da7a3-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"bce8180f-fb70-4f9c-bc00-3a39e62da7a3\") " pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.569862 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bce8180f-fb70-4f9c-bc00-3a39e62da7a3-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"bce8180f-fb70-4f9c-bc00-3a39e62da7a3\") " pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.582944 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.584762 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tc9ld\" (UniqueName: \"kubernetes.io/projected/bce8180f-fb70-4f9c-bc00-3a39e62da7a3-kube-api-access-tc9ld\") pod \"ovsdbserver-nb-1\" (UID: \"bce8180f-fb70-4f9c-bc00-3a39e62da7a3\") " pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.603394 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7a3aa625-3532-4d06-aa7e-e91af4df2f4a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7a3aa625-3532-4d06-aa7e-e91af4df2f4a\") pod \"ovsdbserver-nb-1\" (UID: \"bce8180f-fb70-4f9c-bc00-3a39e62da7a3\") " pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.610869 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.664989 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/676f4247-9871-4bee-ad0f-3b42889be224-config\") pod \"ovsdbserver-sb-2\" (UID: \"676f4247-9871-4bee-ad0f-3b42889be224\") " pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.665133 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0d2f39ca-cd42-42c2-9d0f-51e31961a349-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"0d2f39ca-cd42-42c2-9d0f-51e31961a349\") " pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.665163 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/676f4247-9871-4bee-ad0f-3b42889be224-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"676f4247-9871-4bee-ad0f-3b42889be224\") " pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.665196 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d2f39ca-cd42-42c2-9d0f-51e31961a349-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"0d2f39ca-cd42-42c2-9d0f-51e31961a349\") " pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.665216 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d2f39ca-cd42-42c2-9d0f-51e31961a349-config\") pod \"ovsdbserver-sb-1\" (UID: \"0d2f39ca-cd42-42c2-9d0f-51e31961a349\") " pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.665244 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0d2f39ca-cd42-42c2-9d0f-51e31961a349-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"0d2f39ca-cd42-42c2-9d0f-51e31961a349\") " pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.665266 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d2f39ca-cd42-42c2-9d0f-51e31961a349-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"0d2f39ca-cd42-42c2-9d0f-51e31961a349\") " pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.665311 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/676f4247-9871-4bee-ad0f-3b42889be224-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"676f4247-9871-4bee-ad0f-3b42889be224\") " pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.665346 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g746l\" (UniqueName: \"kubernetes.io/projected/ce32cbf1-d6a1-4361-8921-ce0abcda4667-kube-api-access-g746l\") pod \"ovsdbserver-sb-0\" (UID: \"ce32cbf1-d6a1-4361-8921-ce0abcda4667\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.665379 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce32cbf1-d6a1-4361-8921-ce0abcda4667-config\") pod \"ovsdbserver-sb-0\" (UID: \"ce32cbf1-d6a1-4361-8921-ce0abcda4667\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.665400 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ce32cbf1-d6a1-4361-8921-ce0abcda4667-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ce32cbf1-d6a1-4361-8921-ce0abcda4667\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.665423 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-505e6906-c7aa-455d-a32e-fbbfe03e29a4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-505e6906-c7aa-455d-a32e-fbbfe03e29a4\") pod \"ovsdbserver-sb-0\" (UID: \"ce32cbf1-d6a1-4361-8921-ce0abcda4667\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.665446 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ce32cbf1-d6a1-4361-8921-ce0abcda4667-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ce32cbf1-d6a1-4361-8921-ce0abcda4667\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.665477 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ab217773-08e4-4587-a086-c974e5b2e782\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ab217773-08e4-4587-a086-c974e5b2e782\") pod \"ovsdbserver-sb-2\" (UID: \"676f4247-9871-4bee-ad0f-3b42889be224\") " pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.665521 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d2f39ca-cd42-42c2-9d0f-51e31961a349-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"0d2f39ca-cd42-42c2-9d0f-51e31961a349\") " pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.665547 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfcvv\" (UniqueName: \"kubernetes.io/projected/0d2f39ca-cd42-42c2-9d0f-51e31961a349-kube-api-access-lfcvv\") pod \"ovsdbserver-sb-1\" (UID: \"0d2f39ca-cd42-42c2-9d0f-51e31961a349\") " pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.665568 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/676f4247-9871-4bee-ad0f-3b42889be224-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"676f4247-9871-4bee-ad0f-3b42889be224\") " pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.665598 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/676f4247-9871-4bee-ad0f-3b42889be224-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"676f4247-9871-4bee-ad0f-3b42889be224\") " pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.665628 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwv9n\" (UniqueName: \"kubernetes.io/projected/676f4247-9871-4bee-ad0f-3b42889be224-kube-api-access-jwv9n\") pod \"ovsdbserver-sb-2\" (UID: \"676f4247-9871-4bee-ad0f-3b42889be224\") " pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.665657 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce32cbf1-d6a1-4361-8921-ce0abcda4667-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ce32cbf1-d6a1-4361-8921-ce0abcda4667\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.665682 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce32cbf1-d6a1-4361-8921-ce0abcda4667-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ce32cbf1-d6a1-4361-8921-ce0abcda4667\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.665705 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-49d063b4-82ae-43fb-bfc2-4538757f2986\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49d063b4-82ae-43fb-bfc2-4538757f2986\") pod \"ovsdbserver-sb-1\" (UID: \"0d2f39ca-cd42-42c2-9d0f-51e31961a349\") " pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.665733 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/676f4247-9871-4bee-ad0f-3b42889be224-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"676f4247-9871-4bee-ad0f-3b42889be224\") " pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.665767 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce32cbf1-d6a1-4361-8921-ce0abcda4667-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ce32cbf1-d6a1-4361-8921-ce0abcda4667\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.667431 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce32cbf1-d6a1-4361-8921-ce0abcda4667-config\") pod \"ovsdbserver-sb-0\" (UID: \"ce32cbf1-d6a1-4361-8921-ce0abcda4667\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.667558 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ce32cbf1-d6a1-4361-8921-ce0abcda4667-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ce32cbf1-d6a1-4361-8921-ce0abcda4667\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.668294 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ce32cbf1-d6a1-4361-8921-ce0abcda4667-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ce32cbf1-d6a1-4361-8921-ce0abcda4667\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.670220 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce32cbf1-d6a1-4361-8921-ce0abcda4667-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ce32cbf1-d6a1-4361-8921-ce0abcda4667\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.670562 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce32cbf1-d6a1-4361-8921-ce0abcda4667-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ce32cbf1-d6a1-4361-8921-ce0abcda4667\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.673957 4776 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.684925 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-505e6906-c7aa-455d-a32e-fbbfe03e29a4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-505e6906-c7aa-455d-a32e-fbbfe03e29a4\") pod \"ovsdbserver-sb-0\" (UID: \"ce32cbf1-d6a1-4361-8921-ce0abcda4667\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/7771e44e16797f1408dc92a10a8ba306243a5bb2a57268aa53eba28b7204bdf8/globalmount\"" pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.680174 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce32cbf1-d6a1-4361-8921-ce0abcda4667-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ce32cbf1-d6a1-4361-8921-ce0abcda4667\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.691170 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g746l\" (UniqueName: \"kubernetes.io/projected/ce32cbf1-d6a1-4361-8921-ce0abcda4667-kube-api-access-g746l\") pod \"ovsdbserver-sb-0\" (UID: \"ce32cbf1-d6a1-4361-8921-ce0abcda4667\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.723254 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-505e6906-c7aa-455d-a32e-fbbfe03e29a4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-505e6906-c7aa-455d-a32e-fbbfe03e29a4\") pod \"ovsdbserver-sb-0\" (UID: \"ce32cbf1-d6a1-4361-8921-ce0abcda4667\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.768389 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/676f4247-9871-4bee-ad0f-3b42889be224-config\") pod \"ovsdbserver-sb-2\" (UID: \"676f4247-9871-4bee-ad0f-3b42889be224\") " pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.770266 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0d2f39ca-cd42-42c2-9d0f-51e31961a349-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"0d2f39ca-cd42-42c2-9d0f-51e31961a349\") " pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.770287 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/676f4247-9871-4bee-ad0f-3b42889be224-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"676f4247-9871-4bee-ad0f-3b42889be224\") " pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.769870 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/676f4247-9871-4bee-ad0f-3b42889be224-config\") pod \"ovsdbserver-sb-2\" (UID: \"676f4247-9871-4bee-ad0f-3b42889be224\") " pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.771388 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/676f4247-9871-4bee-ad0f-3b42889be224-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"676f4247-9871-4bee-ad0f-3b42889be224\") " pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.771500 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0d2f39ca-cd42-42c2-9d0f-51e31961a349-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"0d2f39ca-cd42-42c2-9d0f-51e31961a349\") " pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.771525 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d2f39ca-cd42-42c2-9d0f-51e31961a349-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"0d2f39ca-cd42-42c2-9d0f-51e31961a349\") " pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.771561 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d2f39ca-cd42-42c2-9d0f-51e31961a349-config\") pod \"ovsdbserver-sb-1\" (UID: \"0d2f39ca-cd42-42c2-9d0f-51e31961a349\") " pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.771595 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0d2f39ca-cd42-42c2-9d0f-51e31961a349-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"0d2f39ca-cd42-42c2-9d0f-51e31961a349\") " pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.771616 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d2f39ca-cd42-42c2-9d0f-51e31961a349-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"0d2f39ca-cd42-42c2-9d0f-51e31961a349\") " pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.771633 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/676f4247-9871-4bee-ad0f-3b42889be224-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"676f4247-9871-4bee-ad0f-3b42889be224\") " pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.771699 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ab217773-08e4-4587-a086-c974e5b2e782\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ab217773-08e4-4587-a086-c974e5b2e782\") pod \"ovsdbserver-sb-2\" (UID: \"676f4247-9871-4bee-ad0f-3b42889be224\") " pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.771742 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d2f39ca-cd42-42c2-9d0f-51e31961a349-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"0d2f39ca-cd42-42c2-9d0f-51e31961a349\") " pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.771767 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/676f4247-9871-4bee-ad0f-3b42889be224-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"676f4247-9871-4bee-ad0f-3b42889be224\") " pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.771781 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfcvv\" (UniqueName: \"kubernetes.io/projected/0d2f39ca-cd42-42c2-9d0f-51e31961a349-kube-api-access-lfcvv\") pod \"ovsdbserver-sb-1\" (UID: \"0d2f39ca-cd42-42c2-9d0f-51e31961a349\") " pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.771802 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/676f4247-9871-4bee-ad0f-3b42889be224-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"676f4247-9871-4bee-ad0f-3b42889be224\") " pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.771832 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwv9n\" (UniqueName: \"kubernetes.io/projected/676f4247-9871-4bee-ad0f-3b42889be224-kube-api-access-jwv9n\") pod \"ovsdbserver-sb-2\" (UID: \"676f4247-9871-4bee-ad0f-3b42889be224\") " pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.773882 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/676f4247-9871-4bee-ad0f-3b42889be224-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"676f4247-9871-4bee-ad0f-3b42889be224\") " pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.775137 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0d2f39ca-cd42-42c2-9d0f-51e31961a349-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"0d2f39ca-cd42-42c2-9d0f-51e31961a349\") " pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.775758 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d2f39ca-cd42-42c2-9d0f-51e31961a349-config\") pod \"ovsdbserver-sb-1\" (UID: \"0d2f39ca-cd42-42c2-9d0f-51e31961a349\") " pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.775865 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-49d063b4-82ae-43fb-bfc2-4538757f2986\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49d063b4-82ae-43fb-bfc2-4538757f2986\") pod \"ovsdbserver-sb-1\" (UID: \"0d2f39ca-cd42-42c2-9d0f-51e31961a349\") " pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.776530 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d2f39ca-cd42-42c2-9d0f-51e31961a349-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"0d2f39ca-cd42-42c2-9d0f-51e31961a349\") " pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.776757 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/676f4247-9871-4bee-ad0f-3b42889be224-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"676f4247-9871-4bee-ad0f-3b42889be224\") " pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.777048 4776 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.777153 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ab217773-08e4-4587-a086-c974e5b2e782\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ab217773-08e4-4587-a086-c974e5b2e782\") pod \"ovsdbserver-sb-2\" (UID: \"676f4247-9871-4bee-ad0f-3b42889be224\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c2c1dc094054564f5f8729e736f1ffc5a677c70eb5d3b517799a3d5443e7e08c/globalmount\"" pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.782818 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d2f39ca-cd42-42c2-9d0f-51e31961a349-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"0d2f39ca-cd42-42c2-9d0f-51e31961a349\") " pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.784574 4776 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.784684 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-49d063b4-82ae-43fb-bfc2-4538757f2986\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49d063b4-82ae-43fb-bfc2-4538757f2986\") pod \"ovsdbserver-sb-1\" (UID: \"0d2f39ca-cd42-42c2-9d0f-51e31961a349\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/7bbe1e777e963eb3c958192207a9a8331d36f2a7d21bc85218f77616fa740dc9/globalmount\"" pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.791514 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwv9n\" (UniqueName: \"kubernetes.io/projected/676f4247-9871-4bee-ad0f-3b42889be224-kube-api-access-jwv9n\") pod \"ovsdbserver-sb-2\" (UID: \"676f4247-9871-4bee-ad0f-3b42889be224\") " pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.795176 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfcvv\" (UniqueName: \"kubernetes.io/projected/0d2f39ca-cd42-42c2-9d0f-51e31961a349-kube-api-access-lfcvv\") pod \"ovsdbserver-sb-1\" (UID: \"0d2f39ca-cd42-42c2-9d0f-51e31961a349\") " pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.809097 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/676f4247-9871-4bee-ad0f-3b42889be224-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"676f4247-9871-4bee-ad0f-3b42889be224\") " pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.820160 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/676f4247-9871-4bee-ad0f-3b42889be224-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"676f4247-9871-4bee-ad0f-3b42889be224\") " pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.822465 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/676f4247-9871-4bee-ad0f-3b42889be224-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"676f4247-9871-4bee-ad0f-3b42889be224\") " pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.824668 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d2f39ca-cd42-42c2-9d0f-51e31961a349-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"0d2f39ca-cd42-42c2-9d0f-51e31961a349\") " pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.843120 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ab217773-08e4-4587-a086-c974e5b2e782\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ab217773-08e4-4587-a086-c974e5b2e782\") pod \"ovsdbserver-sb-2\" (UID: \"676f4247-9871-4bee-ad0f-3b42889be224\") " pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.852772 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-49d063b4-82ae-43fb-bfc2-4538757f2986\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-49d063b4-82ae-43fb-bfc2-4538757f2986\") pod \"ovsdbserver-sb-1\" (UID: \"0d2f39ca-cd42-42c2-9d0f-51e31961a349\") " pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.917989 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.928407 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:12 crc kubenswrapper[4776]: I1125 10:54:12.934476 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:13 crc kubenswrapper[4776]: I1125 10:54:13.172210 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Nov 25 10:54:13 crc kubenswrapper[4776]: I1125 10:54:13.264911 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 25 10:54:13 crc kubenswrapper[4776]: I1125 10:54:13.404334 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 25 10:54:13 crc kubenswrapper[4776]: W1125 10:54:13.411598 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d16f6ea_879b_4b44_a9aa_187c59681215.slice/crio-f4bef96f764848977a3cfdf149aa16afde01aa393d4ffa2bebbf99b0c2f79031 WatchSource:0}: Error finding container f4bef96f764848977a3cfdf149aa16afde01aa393d4ffa2bebbf99b0c2f79031: Status 404 returned error can't find the container with id f4bef96f764848977a3cfdf149aa16afde01aa393d4ffa2bebbf99b0c2f79031 Nov 25 10:54:13 crc kubenswrapper[4776]: W1125 10:54:13.414631 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce32cbf1_d6a1_4361_8921_ce0abcda4667.slice/crio-32b80fa02a61eb8987554fce703ea41b2195f5dbaa315f20b738aa9982209af6 WatchSource:0}: Error finding container 32b80fa02a61eb8987554fce703ea41b2195f5dbaa315f20b738aa9982209af6: Status 404 returned error can't find the container with id 32b80fa02a61eb8987554fce703ea41b2195f5dbaa315f20b738aa9982209af6 Nov 25 10:54:13 crc kubenswrapper[4776]: I1125 10:54:13.500443 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ce32cbf1-d6a1-4361-8921-ce0abcda4667","Type":"ContainerStarted","Data":"32b80fa02a61eb8987554fce703ea41b2195f5dbaa315f20b738aa9982209af6"} Nov 25 10:54:13 crc kubenswrapper[4776]: I1125 10:54:13.505234 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"bce8180f-fb70-4f9c-bc00-3a39e62da7a3","Type":"ContainerStarted","Data":"730e154e61e6c17fba41ca11f86c17e4425c378bac001c149c7e75790b3b59d5"} Nov 25 10:54:13 crc kubenswrapper[4776]: I1125 10:54:13.506108 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"3d16f6ea-879b-4b44-a9aa-187c59681215","Type":"ContainerStarted","Data":"f4bef96f764848977a3cfdf149aa16afde01aa393d4ffa2bebbf99b0c2f79031"} Nov 25 10:54:13 crc kubenswrapper[4776]: I1125 10:54:13.520714 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Nov 25 10:54:13 crc kubenswrapper[4776]: I1125 10:54:13.617863 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Nov 25 10:54:13 crc kubenswrapper[4776]: W1125 10:54:13.667584 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod676f4247_9871_4bee_ad0f_3b42889be224.slice/crio-669a31281c406c02d343ffc119102481c7ba17df9e8d0bb9fad1b1fc6fb31047 WatchSource:0}: Error finding container 669a31281c406c02d343ffc119102481c7ba17df9e8d0bb9fad1b1fc6fb31047: Status 404 returned error can't find the container with id 669a31281c406c02d343ffc119102481c7ba17df9e8d0bb9fad1b1fc6fb31047 Nov 25 10:54:14 crc kubenswrapper[4776]: I1125 10:54:14.005798 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6vlcg" Nov 25 10:54:14 crc kubenswrapper[4776]: I1125 10:54:14.097637 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3-utilities\") pod \"6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3\" (UID: \"6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3\") " Nov 25 10:54:14 crc kubenswrapper[4776]: I1125 10:54:14.097891 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3-catalog-content\") pod \"6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3\" (UID: \"6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3\") " Nov 25 10:54:14 crc kubenswrapper[4776]: I1125 10:54:14.097970 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vtm9\" (UniqueName: \"kubernetes.io/projected/6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3-kube-api-access-4vtm9\") pod \"6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3\" (UID: \"6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3\") " Nov 25 10:54:14 crc kubenswrapper[4776]: I1125 10:54:14.099446 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3-utilities" (OuterVolumeSpecName: "utilities") pod "6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3" (UID: "6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:54:14 crc kubenswrapper[4776]: I1125 10:54:14.104047 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3-kube-api-access-4vtm9" (OuterVolumeSpecName: "kube-api-access-4vtm9") pod "6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3" (UID: "6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3"). InnerVolumeSpecName "kube-api-access-4vtm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:54:14 crc kubenswrapper[4776]: I1125 10:54:14.157360 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3" (UID: "6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:54:14 crc kubenswrapper[4776]: I1125 10:54:14.199800 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vtm9\" (UniqueName: \"kubernetes.io/projected/6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3-kube-api-access-4vtm9\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:14 crc kubenswrapper[4776]: I1125 10:54:14.199842 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:14 crc kubenswrapper[4776]: I1125 10:54:14.199867 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:14 crc kubenswrapper[4776]: I1125 10:54:14.338298 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Nov 25 10:54:14 crc kubenswrapper[4776]: I1125 10:54:14.536581 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"f5b7e9cb-0aa9-434f-b75c-783829b9ba56","Type":"ContainerStarted","Data":"f38b1a96c2e2da18bf9c78708ebd5dae71c461a1f1d5d02abab27d3f57c92942"} Nov 25 10:54:14 crc kubenswrapper[4776]: I1125 10:54:14.545703 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vlcg" event={"ID":"6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3","Type":"ContainerDied","Data":"d96309eb17ccf88e784b92865af330aad0221adad4f54efabe9de68f425daeaa"} Nov 25 10:54:14 crc kubenswrapper[4776]: I1125 10:54:14.545754 4776 scope.go:117] "RemoveContainer" containerID="d3042da4bf2c571f150be94b720d2186a180299428a126f5b280e14d3070dbc4" Nov 25 10:54:14 crc kubenswrapper[4776]: I1125 10:54:14.545869 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6vlcg" Nov 25 10:54:14 crc kubenswrapper[4776]: I1125 10:54:14.555832 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ce32cbf1-d6a1-4361-8921-ce0abcda4667","Type":"ContainerStarted","Data":"7bdd97d61f4ccbab4088c39e8b391465a2bb4e297d33f88b2173d35658074142"} Nov 25 10:54:14 crc kubenswrapper[4776]: I1125 10:54:14.559213 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"bce8180f-fb70-4f9c-bc00-3a39e62da7a3","Type":"ContainerStarted","Data":"ed16c004c50dcf6d2b1f50b282e80cbf4654fed59a48590183973900bb7d7947"} Nov 25 10:54:14 crc kubenswrapper[4776]: I1125 10:54:14.559269 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"bce8180f-fb70-4f9c-bc00-3a39e62da7a3","Type":"ContainerStarted","Data":"cd89fe4827241b08f4dbf080beebadd0c83abe4b11633a7b65ec6aad4fa00072"} Nov 25 10:54:14 crc kubenswrapper[4776]: I1125 10:54:14.562440 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"3d16f6ea-879b-4b44-a9aa-187c59681215","Type":"ContainerStarted","Data":"ac37c0fe336df079861ab17af0f816b9dcec6ddc87a6fb6223b50efebbe514e2"} Nov 25 10:54:14 crc kubenswrapper[4776]: I1125 10:54:14.568907 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"676f4247-9871-4bee-ad0f-3b42889be224","Type":"ContainerStarted","Data":"8a57118f1f74d1670fef1e59c59ee326fcc9d3b31163400d60c16e1cd76f004b"} Nov 25 10:54:14 crc kubenswrapper[4776]: I1125 10:54:14.569236 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"676f4247-9871-4bee-ad0f-3b42889be224","Type":"ContainerStarted","Data":"669a31281c406c02d343ffc119102481c7ba17df9e8d0bb9fad1b1fc6fb31047"} Nov 25 10:54:14 crc kubenswrapper[4776]: I1125 10:54:14.580740 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"0d2f39ca-cd42-42c2-9d0f-51e31961a349","Type":"ContainerStarted","Data":"aa8f25cd91a238a582dadab259f754e6747c3ee031b56755c4d47543581af17d"} Nov 25 10:54:14 crc kubenswrapper[4776]: I1125 10:54:14.580790 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"0d2f39ca-cd42-42c2-9d0f-51e31961a349","Type":"ContainerStarted","Data":"f536d0fcc12f3a3a28ef7254d2871740f48fe69e1a9c6b93a711535a2ac852b5"} Nov 25 10:54:14 crc kubenswrapper[4776]: I1125 10:54:14.584342 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6vlcg"] Nov 25 10:54:14 crc kubenswrapper[4776]: I1125 10:54:14.597153 4776 scope.go:117] "RemoveContainer" containerID="83cf57c9f00865d38f8d72a34a1cd168267ccc529971bbb6599420d783c3c431" Nov 25 10:54:14 crc kubenswrapper[4776]: I1125 10:54:14.608264 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=3.608242372 podStartE2EDuration="3.608242372s" podCreationTimestamp="2025-11-25 10:54:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:54:14.597498543 +0000 UTC m=+5399.638558096" watchObservedRunningTime="2025-11-25 10:54:14.608242372 +0000 UTC m=+5399.649301925" Nov 25 10:54:14 crc kubenswrapper[4776]: I1125 10:54:14.609762 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6vlcg"] Nov 25 10:54:14 crc kubenswrapper[4776]: I1125 10:54:14.620587 4776 scope.go:117] "RemoveContainer" containerID="70a933b3ff6e4e88282229a4bc68c353347ef40229bcc36009a8f36b1f4ea8b0" Nov 25 10:54:15 crc kubenswrapper[4776]: I1125 10:54:15.595164 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"676f4247-9871-4bee-ad0f-3b42889be224","Type":"ContainerStarted","Data":"87a1871da7afc91cf554692b20f58d54bff890408aa536b3a6911a0ef672784e"} Nov 25 10:54:15 crc kubenswrapper[4776]: I1125 10:54:15.598697 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"0d2f39ca-cd42-42c2-9d0f-51e31961a349","Type":"ContainerStarted","Data":"2a56def3439b037193235384e5ade7227eeca1dc35f20ce1b23026fa22741c41"} Nov 25 10:54:15 crc kubenswrapper[4776]: I1125 10:54:15.601495 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"f5b7e9cb-0aa9-434f-b75c-783829b9ba56","Type":"ContainerStarted","Data":"a5ef29a5a28fc119675d126203c9db32bd20381ae7a6adfbee0668a34bb52acf"} Nov 25 10:54:15 crc kubenswrapper[4776]: I1125 10:54:15.601616 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"f5b7e9cb-0aa9-434f-b75c-783829b9ba56","Type":"ContainerStarted","Data":"3c3dbe3ae62f0655ef367d76847b26e18049318c8afde93cbc04f1daa4763a6b"} Nov 25 10:54:15 crc kubenswrapper[4776]: I1125 10:54:15.607640 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ce32cbf1-d6a1-4361-8921-ce0abcda4667","Type":"ContainerStarted","Data":"71141b0c0a419cfef46987fb63c668cf2aac3c2c7c5e900018148a49bea4d2f4"} Nov 25 10:54:15 crc kubenswrapper[4776]: I1125 10:54:15.610301 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"3d16f6ea-879b-4b44-a9aa-187c59681215","Type":"ContainerStarted","Data":"668420cea42d43338535daec7a9800276191bdbaf1c56edb22abf6b5abd24909"} Nov 25 10:54:15 crc kubenswrapper[4776]: I1125 10:54:15.611113 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:15 crc kubenswrapper[4776]: I1125 10:54:15.625894 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=4.625871776 podStartE2EDuration="4.625871776s" podCreationTimestamp="2025-11-25 10:54:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:54:15.623768633 +0000 UTC m=+5400.664828206" watchObservedRunningTime="2025-11-25 10:54:15.625871776 +0000 UTC m=+5400.666931349" Nov 25 10:54:15 crc kubenswrapper[4776]: I1125 10:54:15.650198 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=4.650157976 podStartE2EDuration="4.650157976s" podCreationTimestamp="2025-11-25 10:54:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:54:15.64474271 +0000 UTC m=+5400.685802293" watchObservedRunningTime="2025-11-25 10:54:15.650157976 +0000 UTC m=+5400.691217529" Nov 25 10:54:15 crc kubenswrapper[4776]: I1125 10:54:15.668649 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=4.668629979 podStartE2EDuration="4.668629979s" podCreationTimestamp="2025-11-25 10:54:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:54:15.668158838 +0000 UTC m=+5400.709218401" watchObservedRunningTime="2025-11-25 10:54:15.668629979 +0000 UTC m=+5400.709689532" Nov 25 10:54:15 crc kubenswrapper[4776]: I1125 10:54:15.680333 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3" path="/var/lib/kubelet/pods/6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3/volumes" Nov 25 10:54:15 crc kubenswrapper[4776]: I1125 10:54:15.695036 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=4.695012702 podStartE2EDuration="4.695012702s" podCreationTimestamp="2025-11-25 10:54:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:54:15.689920304 +0000 UTC m=+5400.730979857" watchObservedRunningTime="2025-11-25 10:54:15.695012702 +0000 UTC m=+5400.736072255" Nov 25 10:54:15 crc kubenswrapper[4776]: I1125 10:54:15.718313 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=4.718293566 podStartE2EDuration="4.718293566s" podCreationTimestamp="2025-11-25 10:54:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:54:15.714785388 +0000 UTC m=+5400.755844951" watchObservedRunningTime="2025-11-25 10:54:15.718293566 +0000 UTC m=+5400.759353119" Nov 25 10:54:15 crc kubenswrapper[4776]: I1125 10:54:15.918202 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:15 crc kubenswrapper[4776]: I1125 10:54:15.928922 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:15 crc kubenswrapper[4776]: I1125 10:54:15.935738 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:17 crc kubenswrapper[4776]: I1125 10:54:17.558263 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:17 crc kubenswrapper[4776]: I1125 10:54:17.583436 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:17 crc kubenswrapper[4776]: I1125 10:54:17.611991 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:17 crc kubenswrapper[4776]: I1125 10:54:17.662984 4776 scope.go:117] "RemoveContainer" containerID="f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40" Nov 25 10:54:17 crc kubenswrapper[4776]: E1125 10:54:17.663268 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:54:17 crc kubenswrapper[4776]: I1125 10:54:17.919052 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:17 crc kubenswrapper[4776]: I1125 10:54:17.929451 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:17 crc kubenswrapper[4776]: I1125 10:54:17.935653 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:18 crc kubenswrapper[4776]: I1125 10:54:18.558493 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:18 crc kubenswrapper[4776]: I1125 10:54:18.583472 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:18 crc kubenswrapper[4776]: I1125 10:54:18.602331 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:18 crc kubenswrapper[4776]: I1125 10:54:18.638123 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:18 crc kubenswrapper[4776]: I1125 10:54:18.653114 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:18 crc kubenswrapper[4776]: I1125 10:54:18.707668 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Nov 25 10:54:18 crc kubenswrapper[4776]: I1125 10:54:18.970383 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bdcf644d5-mdztq"] Nov 25 10:54:18 crc kubenswrapper[4776]: E1125 10:54:18.970679 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3" containerName="extract-utilities" Nov 25 10:54:18 crc kubenswrapper[4776]: I1125 10:54:18.970690 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3" containerName="extract-utilities" Nov 25 10:54:18 crc kubenswrapper[4776]: E1125 10:54:18.970709 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3" containerName="extract-content" Nov 25 10:54:18 crc kubenswrapper[4776]: I1125 10:54:18.970715 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3" containerName="extract-content" Nov 25 10:54:18 crc kubenswrapper[4776]: E1125 10:54:18.970733 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3" containerName="registry-server" Nov 25 10:54:18 crc kubenswrapper[4776]: I1125 10:54:18.970740 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3" containerName="registry-server" Nov 25 10:54:18 crc kubenswrapper[4776]: I1125 10:54:18.970914 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d1a7c93-9cd6-4d3b-b94b-c2b323cc64b3" containerName="registry-server" Nov 25 10:54:18 crc kubenswrapper[4776]: I1125 10:54:18.971733 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdcf644d5-mdztq" Nov 25 10:54:18 crc kubenswrapper[4776]: I1125 10:54:18.977804 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 25 10:54:18 crc kubenswrapper[4776]: I1125 10:54:18.983901 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:18 crc kubenswrapper[4776]: I1125 10:54:18.995890 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:18 crc kubenswrapper[4776]: I1125 10:54:18.995940 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bdcf644d5-mdztq"] Nov 25 10:54:19 crc kubenswrapper[4776]: I1125 10:54:19.000568 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:19 crc kubenswrapper[4776]: I1125 10:54:19.105173 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzq74\" (UniqueName: \"kubernetes.io/projected/302e23a4-3978-4629-838e-94c5247d6cee-kube-api-access-jzq74\") pod \"dnsmasq-dns-7bdcf644d5-mdztq\" (UID: \"302e23a4-3978-4629-838e-94c5247d6cee\") " pod="openstack/dnsmasq-dns-7bdcf644d5-mdztq" Nov 25 10:54:19 crc kubenswrapper[4776]: I1125 10:54:19.105229 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/302e23a4-3978-4629-838e-94c5247d6cee-dns-svc\") pod \"dnsmasq-dns-7bdcf644d5-mdztq\" (UID: \"302e23a4-3978-4629-838e-94c5247d6cee\") " pod="openstack/dnsmasq-dns-7bdcf644d5-mdztq" Nov 25 10:54:19 crc kubenswrapper[4776]: I1125 10:54:19.105270 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/302e23a4-3978-4629-838e-94c5247d6cee-ovsdbserver-nb\") pod \"dnsmasq-dns-7bdcf644d5-mdztq\" (UID: \"302e23a4-3978-4629-838e-94c5247d6cee\") " pod="openstack/dnsmasq-dns-7bdcf644d5-mdztq" Nov 25 10:54:19 crc kubenswrapper[4776]: I1125 10:54:19.105315 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/302e23a4-3978-4629-838e-94c5247d6cee-config\") pod \"dnsmasq-dns-7bdcf644d5-mdztq\" (UID: \"302e23a4-3978-4629-838e-94c5247d6cee\") " pod="openstack/dnsmasq-dns-7bdcf644d5-mdztq" Nov 25 10:54:19 crc kubenswrapper[4776]: I1125 10:54:19.207388 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/302e23a4-3978-4629-838e-94c5247d6cee-ovsdbserver-nb\") pod \"dnsmasq-dns-7bdcf644d5-mdztq\" (UID: \"302e23a4-3978-4629-838e-94c5247d6cee\") " pod="openstack/dnsmasq-dns-7bdcf644d5-mdztq" Nov 25 10:54:19 crc kubenswrapper[4776]: I1125 10:54:19.207457 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/302e23a4-3978-4629-838e-94c5247d6cee-config\") pod \"dnsmasq-dns-7bdcf644d5-mdztq\" (UID: \"302e23a4-3978-4629-838e-94c5247d6cee\") " pod="openstack/dnsmasq-dns-7bdcf644d5-mdztq" Nov 25 10:54:19 crc kubenswrapper[4776]: I1125 10:54:19.207572 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzq74\" (UniqueName: \"kubernetes.io/projected/302e23a4-3978-4629-838e-94c5247d6cee-kube-api-access-jzq74\") pod \"dnsmasq-dns-7bdcf644d5-mdztq\" (UID: \"302e23a4-3978-4629-838e-94c5247d6cee\") " pod="openstack/dnsmasq-dns-7bdcf644d5-mdztq" Nov 25 10:54:19 crc kubenswrapper[4776]: I1125 10:54:19.207610 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/302e23a4-3978-4629-838e-94c5247d6cee-dns-svc\") pod \"dnsmasq-dns-7bdcf644d5-mdztq\" (UID: \"302e23a4-3978-4629-838e-94c5247d6cee\") " pod="openstack/dnsmasq-dns-7bdcf644d5-mdztq" Nov 25 10:54:19 crc kubenswrapper[4776]: I1125 10:54:19.208739 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/302e23a4-3978-4629-838e-94c5247d6cee-config\") pod \"dnsmasq-dns-7bdcf644d5-mdztq\" (UID: \"302e23a4-3978-4629-838e-94c5247d6cee\") " pod="openstack/dnsmasq-dns-7bdcf644d5-mdztq" Nov 25 10:54:19 crc kubenswrapper[4776]: I1125 10:54:19.208817 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/302e23a4-3978-4629-838e-94c5247d6cee-dns-svc\") pod \"dnsmasq-dns-7bdcf644d5-mdztq\" (UID: \"302e23a4-3978-4629-838e-94c5247d6cee\") " pod="openstack/dnsmasq-dns-7bdcf644d5-mdztq" Nov 25 10:54:19 crc kubenswrapper[4776]: I1125 10:54:19.209185 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/302e23a4-3978-4629-838e-94c5247d6cee-ovsdbserver-nb\") pod \"dnsmasq-dns-7bdcf644d5-mdztq\" (UID: \"302e23a4-3978-4629-838e-94c5247d6cee\") " pod="openstack/dnsmasq-dns-7bdcf644d5-mdztq" Nov 25 10:54:19 crc kubenswrapper[4776]: I1125 10:54:19.230413 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzq74\" (UniqueName: \"kubernetes.io/projected/302e23a4-3978-4629-838e-94c5247d6cee-kube-api-access-jzq74\") pod \"dnsmasq-dns-7bdcf644d5-mdztq\" (UID: \"302e23a4-3978-4629-838e-94c5247d6cee\") " pod="openstack/dnsmasq-dns-7bdcf644d5-mdztq" Nov 25 10:54:19 crc kubenswrapper[4776]: I1125 10:54:19.306831 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdcf644d5-mdztq" Nov 25 10:54:19 crc kubenswrapper[4776]: I1125 10:54:19.690266 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 25 10:54:19 crc kubenswrapper[4776]: I1125 10:54:19.695902 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Nov 25 10:54:19 crc kubenswrapper[4776]: I1125 10:54:19.696533 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Nov 25 10:54:19 crc kubenswrapper[4776]: I1125 10:54:19.697511 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Nov 25 10:54:19 crc kubenswrapper[4776]: I1125 10:54:19.697894 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 25 10:54:19 crc kubenswrapper[4776]: I1125 10:54:19.769388 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bdcf644d5-mdztq"] Nov 25 10:54:20 crc kubenswrapper[4776]: I1125 10:54:20.264945 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bdcf644d5-mdztq"] Nov 25 10:54:20 crc kubenswrapper[4776]: I1125 10:54:20.319092 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74767bbc9-lxk47"] Nov 25 10:54:20 crc kubenswrapper[4776]: I1125 10:54:20.320754 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74767bbc9-lxk47" Nov 25 10:54:20 crc kubenswrapper[4776]: I1125 10:54:20.323919 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 25 10:54:20 crc kubenswrapper[4776]: I1125 10:54:20.353207 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74767bbc9-lxk47"] Nov 25 10:54:20 crc kubenswrapper[4776]: I1125 10:54:20.429009 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32010807-df10-4d9e-b62c-b7e14706210b-dns-svc\") pod \"dnsmasq-dns-74767bbc9-lxk47\" (UID: \"32010807-df10-4d9e-b62c-b7e14706210b\") " pod="openstack/dnsmasq-dns-74767bbc9-lxk47" Nov 25 10:54:20 crc kubenswrapper[4776]: I1125 10:54:20.429152 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/32010807-df10-4d9e-b62c-b7e14706210b-ovsdbserver-sb\") pod \"dnsmasq-dns-74767bbc9-lxk47\" (UID: \"32010807-df10-4d9e-b62c-b7e14706210b\") " pod="openstack/dnsmasq-dns-74767bbc9-lxk47" Nov 25 10:54:20 crc kubenswrapper[4776]: I1125 10:54:20.429183 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/32010807-df10-4d9e-b62c-b7e14706210b-ovsdbserver-nb\") pod \"dnsmasq-dns-74767bbc9-lxk47\" (UID: \"32010807-df10-4d9e-b62c-b7e14706210b\") " pod="openstack/dnsmasq-dns-74767bbc9-lxk47" Nov 25 10:54:20 crc kubenswrapper[4776]: I1125 10:54:20.429342 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32010807-df10-4d9e-b62c-b7e14706210b-config\") pod \"dnsmasq-dns-74767bbc9-lxk47\" (UID: \"32010807-df10-4d9e-b62c-b7e14706210b\") " pod="openstack/dnsmasq-dns-74767bbc9-lxk47" Nov 25 10:54:20 crc kubenswrapper[4776]: I1125 10:54:20.429492 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbnsk\" (UniqueName: \"kubernetes.io/projected/32010807-df10-4d9e-b62c-b7e14706210b-kube-api-access-mbnsk\") pod \"dnsmasq-dns-74767bbc9-lxk47\" (UID: \"32010807-df10-4d9e-b62c-b7e14706210b\") " pod="openstack/dnsmasq-dns-74767bbc9-lxk47" Nov 25 10:54:20 crc kubenswrapper[4776]: I1125 10:54:20.530849 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32010807-df10-4d9e-b62c-b7e14706210b-config\") pod \"dnsmasq-dns-74767bbc9-lxk47\" (UID: \"32010807-df10-4d9e-b62c-b7e14706210b\") " pod="openstack/dnsmasq-dns-74767bbc9-lxk47" Nov 25 10:54:20 crc kubenswrapper[4776]: I1125 10:54:20.530956 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbnsk\" (UniqueName: \"kubernetes.io/projected/32010807-df10-4d9e-b62c-b7e14706210b-kube-api-access-mbnsk\") pod \"dnsmasq-dns-74767bbc9-lxk47\" (UID: \"32010807-df10-4d9e-b62c-b7e14706210b\") " pod="openstack/dnsmasq-dns-74767bbc9-lxk47" Nov 25 10:54:20 crc kubenswrapper[4776]: I1125 10:54:20.530984 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32010807-df10-4d9e-b62c-b7e14706210b-dns-svc\") pod \"dnsmasq-dns-74767bbc9-lxk47\" (UID: \"32010807-df10-4d9e-b62c-b7e14706210b\") " pod="openstack/dnsmasq-dns-74767bbc9-lxk47" Nov 25 10:54:20 crc kubenswrapper[4776]: I1125 10:54:20.531044 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/32010807-df10-4d9e-b62c-b7e14706210b-ovsdbserver-sb\") pod \"dnsmasq-dns-74767bbc9-lxk47\" (UID: \"32010807-df10-4d9e-b62c-b7e14706210b\") " pod="openstack/dnsmasq-dns-74767bbc9-lxk47" Nov 25 10:54:20 crc kubenswrapper[4776]: I1125 10:54:20.531085 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/32010807-df10-4d9e-b62c-b7e14706210b-ovsdbserver-nb\") pod \"dnsmasq-dns-74767bbc9-lxk47\" (UID: \"32010807-df10-4d9e-b62c-b7e14706210b\") " pod="openstack/dnsmasq-dns-74767bbc9-lxk47" Nov 25 10:54:20 crc kubenswrapper[4776]: I1125 10:54:20.531871 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32010807-df10-4d9e-b62c-b7e14706210b-config\") pod \"dnsmasq-dns-74767bbc9-lxk47\" (UID: \"32010807-df10-4d9e-b62c-b7e14706210b\") " pod="openstack/dnsmasq-dns-74767bbc9-lxk47" Nov 25 10:54:20 crc kubenswrapper[4776]: I1125 10:54:20.531881 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/32010807-df10-4d9e-b62c-b7e14706210b-ovsdbserver-nb\") pod \"dnsmasq-dns-74767bbc9-lxk47\" (UID: \"32010807-df10-4d9e-b62c-b7e14706210b\") " pod="openstack/dnsmasq-dns-74767bbc9-lxk47" Nov 25 10:54:20 crc kubenswrapper[4776]: I1125 10:54:20.532478 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32010807-df10-4d9e-b62c-b7e14706210b-dns-svc\") pod \"dnsmasq-dns-74767bbc9-lxk47\" (UID: \"32010807-df10-4d9e-b62c-b7e14706210b\") " pod="openstack/dnsmasq-dns-74767bbc9-lxk47" Nov 25 10:54:20 crc kubenswrapper[4776]: I1125 10:54:20.532704 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/32010807-df10-4d9e-b62c-b7e14706210b-ovsdbserver-sb\") pod \"dnsmasq-dns-74767bbc9-lxk47\" (UID: \"32010807-df10-4d9e-b62c-b7e14706210b\") " pod="openstack/dnsmasq-dns-74767bbc9-lxk47" Nov 25 10:54:20 crc kubenswrapper[4776]: I1125 10:54:20.553301 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbnsk\" (UniqueName: \"kubernetes.io/projected/32010807-df10-4d9e-b62c-b7e14706210b-kube-api-access-mbnsk\") pod \"dnsmasq-dns-74767bbc9-lxk47\" (UID: \"32010807-df10-4d9e-b62c-b7e14706210b\") " pod="openstack/dnsmasq-dns-74767bbc9-lxk47" Nov 25 10:54:20 crc kubenswrapper[4776]: I1125 10:54:20.651559 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74767bbc9-lxk47" Nov 25 10:54:20 crc kubenswrapper[4776]: I1125 10:54:20.660050 4776 generic.go:334] "Generic (PLEG): container finished" podID="302e23a4-3978-4629-838e-94c5247d6cee" containerID="d2a2deadef8e0b3ccb09b71976d1f28d83438f69ec0f13b675bb23b1b2097cba" exitCode=0 Nov 25 10:54:20 crc kubenswrapper[4776]: I1125 10:54:20.660330 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdcf644d5-mdztq" event={"ID":"302e23a4-3978-4629-838e-94c5247d6cee","Type":"ContainerDied","Data":"d2a2deadef8e0b3ccb09b71976d1f28d83438f69ec0f13b675bb23b1b2097cba"} Nov 25 10:54:20 crc kubenswrapper[4776]: I1125 10:54:20.660415 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdcf644d5-mdztq" event={"ID":"302e23a4-3978-4629-838e-94c5247d6cee","Type":"ContainerStarted","Data":"17e51aa08270c2f5ab7e113f4847899c785d773e5f716934b1b76035bb2b0978"} Nov 25 10:54:21 crc kubenswrapper[4776]: I1125 10:54:21.093276 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74767bbc9-lxk47"] Nov 25 10:54:21 crc kubenswrapper[4776]: W1125 10:54:21.094370 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32010807_df10_4d9e_b62c_b7e14706210b.slice/crio-94d236bbcb917ab79734d0b1640b13a58c95b364ee7fb894d825545a78b9eaa3 WatchSource:0}: Error finding container 94d236bbcb917ab79734d0b1640b13a58c95b364ee7fb894d825545a78b9eaa3: Status 404 returned error can't find the container with id 94d236bbcb917ab79734d0b1640b13a58c95b364ee7fb894d825545a78b9eaa3 Nov 25 10:54:21 crc kubenswrapper[4776]: I1125 10:54:21.671471 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7bdcf644d5-mdztq" podUID="302e23a4-3978-4629-838e-94c5247d6cee" containerName="dnsmasq-dns" containerID="cri-o://eb475cd597d3c22e6eb04c80a9fa27233987036c4864d1839c429bf9fcc381c9" gracePeriod=10 Nov 25 10:54:21 crc kubenswrapper[4776]: I1125 10:54:21.672699 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdcf644d5-mdztq" event={"ID":"302e23a4-3978-4629-838e-94c5247d6cee","Type":"ContainerStarted","Data":"eb475cd597d3c22e6eb04c80a9fa27233987036c4864d1839c429bf9fcc381c9"} Nov 25 10:54:21 crc kubenswrapper[4776]: I1125 10:54:21.672734 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7bdcf644d5-mdztq" Nov 25 10:54:21 crc kubenswrapper[4776]: I1125 10:54:21.673342 4776 generic.go:334] "Generic (PLEG): container finished" podID="32010807-df10-4d9e-b62c-b7e14706210b" containerID="fa22650a109f2356b9687b9bec7a6d93988d1cac25c975f4f83548d0996703a4" exitCode=0 Nov 25 10:54:21 crc kubenswrapper[4776]: I1125 10:54:21.673379 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74767bbc9-lxk47" event={"ID":"32010807-df10-4d9e-b62c-b7e14706210b","Type":"ContainerDied","Data":"fa22650a109f2356b9687b9bec7a6d93988d1cac25c975f4f83548d0996703a4"} Nov 25 10:54:21 crc kubenswrapper[4776]: I1125 10:54:21.673404 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74767bbc9-lxk47" event={"ID":"32010807-df10-4d9e-b62c-b7e14706210b","Type":"ContainerStarted","Data":"94d236bbcb917ab79734d0b1640b13a58c95b364ee7fb894d825545a78b9eaa3"} Nov 25 10:54:21 crc kubenswrapper[4776]: I1125 10:54:21.703429 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7bdcf644d5-mdztq" podStartSLOduration=3.703408375 podStartE2EDuration="3.703408375s" podCreationTimestamp="2025-11-25 10:54:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:54:21.691318672 +0000 UTC m=+5406.732378265" watchObservedRunningTime="2025-11-25 10:54:21.703408375 +0000 UTC m=+5406.744467928" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.036854 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdcf644d5-mdztq" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.162341 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/302e23a4-3978-4629-838e-94c5247d6cee-dns-svc\") pod \"302e23a4-3978-4629-838e-94c5247d6cee\" (UID: \"302e23a4-3978-4629-838e-94c5247d6cee\") " Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.162492 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/302e23a4-3978-4629-838e-94c5247d6cee-config\") pod \"302e23a4-3978-4629-838e-94c5247d6cee\" (UID: \"302e23a4-3978-4629-838e-94c5247d6cee\") " Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.162562 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzq74\" (UniqueName: \"kubernetes.io/projected/302e23a4-3978-4629-838e-94c5247d6cee-kube-api-access-jzq74\") pod \"302e23a4-3978-4629-838e-94c5247d6cee\" (UID: \"302e23a4-3978-4629-838e-94c5247d6cee\") " Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.162598 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/302e23a4-3978-4629-838e-94c5247d6cee-ovsdbserver-nb\") pod \"302e23a4-3978-4629-838e-94c5247d6cee\" (UID: \"302e23a4-3978-4629-838e-94c5247d6cee\") " Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.166717 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/302e23a4-3978-4629-838e-94c5247d6cee-kube-api-access-jzq74" (OuterVolumeSpecName: "kube-api-access-jzq74") pod "302e23a4-3978-4629-838e-94c5247d6cee" (UID: "302e23a4-3978-4629-838e-94c5247d6cee"). InnerVolumeSpecName "kube-api-access-jzq74". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.198257 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/302e23a4-3978-4629-838e-94c5247d6cee-config" (OuterVolumeSpecName: "config") pod "302e23a4-3978-4629-838e-94c5247d6cee" (UID: "302e23a4-3978-4629-838e-94c5247d6cee"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.201662 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/302e23a4-3978-4629-838e-94c5247d6cee-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "302e23a4-3978-4629-838e-94c5247d6cee" (UID: "302e23a4-3978-4629-838e-94c5247d6cee"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.201689 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/302e23a4-3978-4629-838e-94c5247d6cee-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "302e23a4-3978-4629-838e-94c5247d6cee" (UID: "302e23a4-3978-4629-838e-94c5247d6cee"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.264669 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzq74\" (UniqueName: \"kubernetes.io/projected/302e23a4-3978-4629-838e-94c5247d6cee-kube-api-access-jzq74\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.264716 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/302e23a4-3978-4629-838e-94c5247d6cee-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.264729 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/302e23a4-3978-4629-838e-94c5247d6cee-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.264742 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/302e23a4-3978-4629-838e-94c5247d6cee-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.663435 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Nov 25 10:54:22 crc kubenswrapper[4776]: E1125 10:54:22.663722 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="302e23a4-3978-4629-838e-94c5247d6cee" containerName="init" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.663738 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="302e23a4-3978-4629-838e-94c5247d6cee" containerName="init" Nov 25 10:54:22 crc kubenswrapper[4776]: E1125 10:54:22.663753 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="302e23a4-3978-4629-838e-94c5247d6cee" containerName="dnsmasq-dns" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.663759 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="302e23a4-3978-4629-838e-94c5247d6cee" containerName="dnsmasq-dns" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.663918 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="302e23a4-3978-4629-838e-94c5247d6cee" containerName="dnsmasq-dns" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.664457 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.665978 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.674439 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.683400 4776 generic.go:334] "Generic (PLEG): container finished" podID="302e23a4-3978-4629-838e-94c5247d6cee" containerID="eb475cd597d3c22e6eb04c80a9fa27233987036c4864d1839c429bf9fcc381c9" exitCode=0 Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.683444 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdcf644d5-mdztq" event={"ID":"302e23a4-3978-4629-838e-94c5247d6cee","Type":"ContainerDied","Data":"eb475cd597d3c22e6eb04c80a9fa27233987036c4864d1839c429bf9fcc381c9"} Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.683498 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bdcf644d5-mdztq" event={"ID":"302e23a4-3978-4629-838e-94c5247d6cee","Type":"ContainerDied","Data":"17e51aa08270c2f5ab7e113f4847899c785d773e5f716934b1b76035bb2b0978"} Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.683505 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bdcf644d5-mdztq" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.683515 4776 scope.go:117] "RemoveContainer" containerID="eb475cd597d3c22e6eb04c80a9fa27233987036c4864d1839c429bf9fcc381c9" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.685361 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74767bbc9-lxk47" event={"ID":"32010807-df10-4d9e-b62c-b7e14706210b","Type":"ContainerStarted","Data":"ba422d27a9918f591c7ba0af0fe8f866b7b0b1bb352015072379ae794b2b2e4a"} Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.685531 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74767bbc9-lxk47" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.715740 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74767bbc9-lxk47" podStartSLOduration=2.715717405 podStartE2EDuration="2.715717405s" podCreationTimestamp="2025-11-25 10:54:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:54:22.706211977 +0000 UTC m=+5407.747271530" watchObservedRunningTime="2025-11-25 10:54:22.715717405 +0000 UTC m=+5407.756776958" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.727267 4776 scope.go:117] "RemoveContainer" containerID="d2a2deadef8e0b3ccb09b71976d1f28d83438f69ec0f13b675bb23b1b2097cba" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.738233 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bdcf644d5-mdztq"] Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.742271 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bdcf644d5-mdztq"] Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.746494 4776 scope.go:117] "RemoveContainer" containerID="eb475cd597d3c22e6eb04c80a9fa27233987036c4864d1839c429bf9fcc381c9" Nov 25 10:54:22 crc kubenswrapper[4776]: E1125 10:54:22.747087 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb475cd597d3c22e6eb04c80a9fa27233987036c4864d1839c429bf9fcc381c9\": container with ID starting with eb475cd597d3c22e6eb04c80a9fa27233987036c4864d1839c429bf9fcc381c9 not found: ID does not exist" containerID="eb475cd597d3c22e6eb04c80a9fa27233987036c4864d1839c429bf9fcc381c9" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.747134 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb475cd597d3c22e6eb04c80a9fa27233987036c4864d1839c429bf9fcc381c9"} err="failed to get container status \"eb475cd597d3c22e6eb04c80a9fa27233987036c4864d1839c429bf9fcc381c9\": rpc error: code = NotFound desc = could not find container \"eb475cd597d3c22e6eb04c80a9fa27233987036c4864d1839c429bf9fcc381c9\": container with ID starting with eb475cd597d3c22e6eb04c80a9fa27233987036c4864d1839c429bf9fcc381c9 not found: ID does not exist" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.747163 4776 scope.go:117] "RemoveContainer" containerID="d2a2deadef8e0b3ccb09b71976d1f28d83438f69ec0f13b675bb23b1b2097cba" Nov 25 10:54:22 crc kubenswrapper[4776]: E1125 10:54:22.747505 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2a2deadef8e0b3ccb09b71976d1f28d83438f69ec0f13b675bb23b1b2097cba\": container with ID starting with d2a2deadef8e0b3ccb09b71976d1f28d83438f69ec0f13b675bb23b1b2097cba not found: ID does not exist" containerID="d2a2deadef8e0b3ccb09b71976d1f28d83438f69ec0f13b675bb23b1b2097cba" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.747555 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2a2deadef8e0b3ccb09b71976d1f28d83438f69ec0f13b675bb23b1b2097cba"} err="failed to get container status \"d2a2deadef8e0b3ccb09b71976d1f28d83438f69ec0f13b675bb23b1b2097cba\": rpc error: code = NotFound desc = could not find container \"d2a2deadef8e0b3ccb09b71976d1f28d83438f69ec0f13b675bb23b1b2097cba\": container with ID starting with d2a2deadef8e0b3ccb09b71976d1f28d83438f69ec0f13b675bb23b1b2097cba not found: ID does not exist" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.772449 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-69f0de1e-12d2-4c9e-a4b1-29b517947901\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-69f0de1e-12d2-4c9e-a4b1-29b517947901\") pod \"ovn-copy-data\" (UID: \"bb36b96d-a860-41e0-9cea-59479eb5ce28\") " pod="openstack/ovn-copy-data" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.772639 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/bb36b96d-a860-41e0-9cea-59479eb5ce28-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"bb36b96d-a860-41e0-9cea-59479eb5ce28\") " pod="openstack/ovn-copy-data" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.772678 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84s25\" (UniqueName: \"kubernetes.io/projected/bb36b96d-a860-41e0-9cea-59479eb5ce28-kube-api-access-84s25\") pod \"ovn-copy-data\" (UID: \"bb36b96d-a860-41e0-9cea-59479eb5ce28\") " pod="openstack/ovn-copy-data" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.874232 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/bb36b96d-a860-41e0-9cea-59479eb5ce28-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"bb36b96d-a860-41e0-9cea-59479eb5ce28\") " pod="openstack/ovn-copy-data" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.874582 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84s25\" (UniqueName: \"kubernetes.io/projected/bb36b96d-a860-41e0-9cea-59479eb5ce28-kube-api-access-84s25\") pod \"ovn-copy-data\" (UID: \"bb36b96d-a860-41e0-9cea-59479eb5ce28\") " pod="openstack/ovn-copy-data" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.874927 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-69f0de1e-12d2-4c9e-a4b1-29b517947901\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-69f0de1e-12d2-4c9e-a4b1-29b517947901\") pod \"ovn-copy-data\" (UID: \"bb36b96d-a860-41e0-9cea-59479eb5ce28\") " pod="openstack/ovn-copy-data" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.879026 4776 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.879095 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-69f0de1e-12d2-4c9e-a4b1-29b517947901\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-69f0de1e-12d2-4c9e-a4b1-29b517947901\") pod \"ovn-copy-data\" (UID: \"bb36b96d-a860-41e0-9cea-59479eb5ce28\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/79a4d6ce69c9d7346db8c4a79cadb5b1f18f220758af2bee31fa486d51659213/globalmount\"" pod="openstack/ovn-copy-data" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.880039 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/bb36b96d-a860-41e0-9cea-59479eb5ce28-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"bb36b96d-a860-41e0-9cea-59479eb5ce28\") " pod="openstack/ovn-copy-data" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.892666 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84s25\" (UniqueName: \"kubernetes.io/projected/bb36b96d-a860-41e0-9cea-59479eb5ce28-kube-api-access-84s25\") pod \"ovn-copy-data\" (UID: \"bb36b96d-a860-41e0-9cea-59479eb5ce28\") " pod="openstack/ovn-copy-data" Nov 25 10:54:22 crc kubenswrapper[4776]: I1125 10:54:22.910910 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-69f0de1e-12d2-4c9e-a4b1-29b517947901\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-69f0de1e-12d2-4c9e-a4b1-29b517947901\") pod \"ovn-copy-data\" (UID: \"bb36b96d-a860-41e0-9cea-59479eb5ce28\") " pod="openstack/ovn-copy-data" Nov 25 10:54:23 crc kubenswrapper[4776]: I1125 10:54:23.007787 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Nov 25 10:54:23 crc kubenswrapper[4776]: I1125 10:54:23.319900 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Nov 25 10:54:23 crc kubenswrapper[4776]: W1125 10:54:23.321426 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb36b96d_a860_41e0_9cea_59479eb5ce28.slice/crio-a814c0d7928e235b44086052c1869b9f5022c1e710d9aa180203f31305c76eb2 WatchSource:0}: Error finding container a814c0d7928e235b44086052c1869b9f5022c1e710d9aa180203f31305c76eb2: Status 404 returned error can't find the container with id a814c0d7928e235b44086052c1869b9f5022c1e710d9aa180203f31305c76eb2 Nov 25 10:54:23 crc kubenswrapper[4776]: I1125 10:54:23.676501 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="302e23a4-3978-4629-838e-94c5247d6cee" path="/var/lib/kubelet/pods/302e23a4-3978-4629-838e-94c5247d6cee/volumes" Nov 25 10:54:23 crc kubenswrapper[4776]: I1125 10:54:23.695900 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"bb36b96d-a860-41e0-9cea-59479eb5ce28","Type":"ContainerStarted","Data":"a814c0d7928e235b44086052c1869b9f5022c1e710d9aa180203f31305c76eb2"} Nov 25 10:54:24 crc kubenswrapper[4776]: I1125 10:54:24.705050 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"bb36b96d-a860-41e0-9cea-59479eb5ce28","Type":"ContainerStarted","Data":"ca02947c0f670b534f946ae0b95ced2d33b13e1cabb85697d6bd977eebd187d3"} Nov 25 10:54:24 crc kubenswrapper[4776]: I1125 10:54:24.728700 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=3.253094204 podStartE2EDuration="3.728676211s" podCreationTimestamp="2025-11-25 10:54:21 +0000 UTC" firstStartedPulling="2025-11-25 10:54:23.323407779 +0000 UTC m=+5408.364467342" lastFinishedPulling="2025-11-25 10:54:23.798989796 +0000 UTC m=+5408.840049349" observedRunningTime="2025-11-25 10:54:24.719638685 +0000 UTC m=+5409.760698238" watchObservedRunningTime="2025-11-25 10:54:24.728676211 +0000 UTC m=+5409.769735764" Nov 25 10:54:29 crc kubenswrapper[4776]: I1125 10:54:29.685059 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 25 10:54:29 crc kubenswrapper[4776]: I1125 10:54:29.694657 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 25 10:54:29 crc kubenswrapper[4776]: I1125 10:54:29.703904 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-rdkrp" Nov 25 10:54:29 crc kubenswrapper[4776]: I1125 10:54:29.704019 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 25 10:54:29 crc kubenswrapper[4776]: I1125 10:54:29.704058 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 25 10:54:29 crc kubenswrapper[4776]: I1125 10:54:29.704306 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 25 10:54:29 crc kubenswrapper[4776]: I1125 10:54:29.708438 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 25 10:54:29 crc kubenswrapper[4776]: I1125 10:54:29.789320 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f396897-96fa-422c-a2fd-92faadaff6e3-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"7f396897-96fa-422c-a2fd-92faadaff6e3\") " pod="openstack/ovn-northd-0" Nov 25 10:54:29 crc kubenswrapper[4776]: I1125 10:54:29.789412 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f396897-96fa-422c-a2fd-92faadaff6e3-scripts\") pod \"ovn-northd-0\" (UID: \"7f396897-96fa-422c-a2fd-92faadaff6e3\") " pod="openstack/ovn-northd-0" Nov 25 10:54:29 crc kubenswrapper[4776]: I1125 10:54:29.789450 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f396897-96fa-422c-a2fd-92faadaff6e3-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"7f396897-96fa-422c-a2fd-92faadaff6e3\") " pod="openstack/ovn-northd-0" Nov 25 10:54:29 crc kubenswrapper[4776]: I1125 10:54:29.789501 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7f396897-96fa-422c-a2fd-92faadaff6e3-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"7f396897-96fa-422c-a2fd-92faadaff6e3\") " pod="openstack/ovn-northd-0" Nov 25 10:54:29 crc kubenswrapper[4776]: I1125 10:54:29.789525 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jc5kz\" (UniqueName: \"kubernetes.io/projected/7f396897-96fa-422c-a2fd-92faadaff6e3-kube-api-access-jc5kz\") pod \"ovn-northd-0\" (UID: \"7f396897-96fa-422c-a2fd-92faadaff6e3\") " pod="openstack/ovn-northd-0" Nov 25 10:54:29 crc kubenswrapper[4776]: I1125 10:54:29.789582 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f396897-96fa-422c-a2fd-92faadaff6e3-config\") pod \"ovn-northd-0\" (UID: \"7f396897-96fa-422c-a2fd-92faadaff6e3\") " pod="openstack/ovn-northd-0" Nov 25 10:54:29 crc kubenswrapper[4776]: I1125 10:54:29.789612 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f396897-96fa-422c-a2fd-92faadaff6e3-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"7f396897-96fa-422c-a2fd-92faadaff6e3\") " pod="openstack/ovn-northd-0" Nov 25 10:54:29 crc kubenswrapper[4776]: I1125 10:54:29.890913 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7f396897-96fa-422c-a2fd-92faadaff6e3-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"7f396897-96fa-422c-a2fd-92faadaff6e3\") " pod="openstack/ovn-northd-0" Nov 25 10:54:29 crc kubenswrapper[4776]: I1125 10:54:29.890959 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jc5kz\" (UniqueName: \"kubernetes.io/projected/7f396897-96fa-422c-a2fd-92faadaff6e3-kube-api-access-jc5kz\") pod \"ovn-northd-0\" (UID: \"7f396897-96fa-422c-a2fd-92faadaff6e3\") " pod="openstack/ovn-northd-0" Nov 25 10:54:29 crc kubenswrapper[4776]: I1125 10:54:29.891009 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f396897-96fa-422c-a2fd-92faadaff6e3-config\") pod \"ovn-northd-0\" (UID: \"7f396897-96fa-422c-a2fd-92faadaff6e3\") " pod="openstack/ovn-northd-0" Nov 25 10:54:29 crc kubenswrapper[4776]: I1125 10:54:29.891027 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f396897-96fa-422c-a2fd-92faadaff6e3-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"7f396897-96fa-422c-a2fd-92faadaff6e3\") " pod="openstack/ovn-northd-0" Nov 25 10:54:29 crc kubenswrapper[4776]: I1125 10:54:29.891101 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f396897-96fa-422c-a2fd-92faadaff6e3-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"7f396897-96fa-422c-a2fd-92faadaff6e3\") " pod="openstack/ovn-northd-0" Nov 25 10:54:29 crc kubenswrapper[4776]: I1125 10:54:29.891138 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f396897-96fa-422c-a2fd-92faadaff6e3-scripts\") pod \"ovn-northd-0\" (UID: \"7f396897-96fa-422c-a2fd-92faadaff6e3\") " pod="openstack/ovn-northd-0" Nov 25 10:54:29 crc kubenswrapper[4776]: I1125 10:54:29.891166 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f396897-96fa-422c-a2fd-92faadaff6e3-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"7f396897-96fa-422c-a2fd-92faadaff6e3\") " pod="openstack/ovn-northd-0" Nov 25 10:54:29 crc kubenswrapper[4776]: I1125 10:54:29.892670 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7f396897-96fa-422c-a2fd-92faadaff6e3-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"7f396897-96fa-422c-a2fd-92faadaff6e3\") " pod="openstack/ovn-northd-0" Nov 25 10:54:29 crc kubenswrapper[4776]: I1125 10:54:29.893080 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f396897-96fa-422c-a2fd-92faadaff6e3-scripts\") pod \"ovn-northd-0\" (UID: \"7f396897-96fa-422c-a2fd-92faadaff6e3\") " pod="openstack/ovn-northd-0" Nov 25 10:54:29 crc kubenswrapper[4776]: I1125 10:54:29.893451 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f396897-96fa-422c-a2fd-92faadaff6e3-config\") pod \"ovn-northd-0\" (UID: \"7f396897-96fa-422c-a2fd-92faadaff6e3\") " pod="openstack/ovn-northd-0" Nov 25 10:54:29 crc kubenswrapper[4776]: I1125 10:54:29.903097 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f396897-96fa-422c-a2fd-92faadaff6e3-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"7f396897-96fa-422c-a2fd-92faadaff6e3\") " pod="openstack/ovn-northd-0" Nov 25 10:54:29 crc kubenswrapper[4776]: I1125 10:54:29.903204 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f396897-96fa-422c-a2fd-92faadaff6e3-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"7f396897-96fa-422c-a2fd-92faadaff6e3\") " pod="openstack/ovn-northd-0" Nov 25 10:54:29 crc kubenswrapper[4776]: I1125 10:54:29.904789 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f396897-96fa-422c-a2fd-92faadaff6e3-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"7f396897-96fa-422c-a2fd-92faadaff6e3\") " pod="openstack/ovn-northd-0" Nov 25 10:54:29 crc kubenswrapper[4776]: I1125 10:54:29.925803 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jc5kz\" (UniqueName: \"kubernetes.io/projected/7f396897-96fa-422c-a2fd-92faadaff6e3-kube-api-access-jc5kz\") pod \"ovn-northd-0\" (UID: \"7f396897-96fa-422c-a2fd-92faadaff6e3\") " pod="openstack/ovn-northd-0" Nov 25 10:54:30 crc kubenswrapper[4776]: I1125 10:54:30.017442 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 25 10:54:30 crc kubenswrapper[4776]: I1125 10:54:30.550972 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 25 10:54:30 crc kubenswrapper[4776]: W1125 10:54:30.562310 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f396897_96fa_422c_a2fd_92faadaff6e3.slice/crio-e8b67702c616b88f27a0952a0b3fe18b622018e1cc477c1a2aba1668a57fe707 WatchSource:0}: Error finding container e8b67702c616b88f27a0952a0b3fe18b622018e1cc477c1a2aba1668a57fe707: Status 404 returned error can't find the container with id e8b67702c616b88f27a0952a0b3fe18b622018e1cc477c1a2aba1668a57fe707 Nov 25 10:54:30 crc kubenswrapper[4776]: I1125 10:54:30.659307 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74767bbc9-lxk47" Nov 25 10:54:30 crc kubenswrapper[4776]: I1125 10:54:30.737751 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54564445dc-5wx2z"] Nov 25 10:54:30 crc kubenswrapper[4776]: I1125 10:54:30.737972 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-54564445dc-5wx2z" podUID="ac07dcc7-2c32-40b9-802b-8380d6c2ef3e" containerName="dnsmasq-dns" containerID="cri-o://e428da7d3af50e7504fbf05eed836a8b5431da41599a55c00486121248dd4b70" gracePeriod=10 Nov 25 10:54:30 crc kubenswrapper[4776]: I1125 10:54:30.767522 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7f396897-96fa-422c-a2fd-92faadaff6e3","Type":"ContainerStarted","Data":"e8b67702c616b88f27a0952a0b3fe18b622018e1cc477c1a2aba1668a57fe707"} Nov 25 10:54:31 crc kubenswrapper[4776]: I1125 10:54:31.273851 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54564445dc-5wx2z" Nov 25 10:54:31 crc kubenswrapper[4776]: I1125 10:54:31.331172 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac07dcc7-2c32-40b9-802b-8380d6c2ef3e-config\") pod \"ac07dcc7-2c32-40b9-802b-8380d6c2ef3e\" (UID: \"ac07dcc7-2c32-40b9-802b-8380d6c2ef3e\") " Nov 25 10:54:31 crc kubenswrapper[4776]: I1125 10:54:31.331311 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bx5pl\" (UniqueName: \"kubernetes.io/projected/ac07dcc7-2c32-40b9-802b-8380d6c2ef3e-kube-api-access-bx5pl\") pod \"ac07dcc7-2c32-40b9-802b-8380d6c2ef3e\" (UID: \"ac07dcc7-2c32-40b9-802b-8380d6c2ef3e\") " Nov 25 10:54:31 crc kubenswrapper[4776]: I1125 10:54:31.331355 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac07dcc7-2c32-40b9-802b-8380d6c2ef3e-dns-svc\") pod \"ac07dcc7-2c32-40b9-802b-8380d6c2ef3e\" (UID: \"ac07dcc7-2c32-40b9-802b-8380d6c2ef3e\") " Nov 25 10:54:31 crc kubenswrapper[4776]: I1125 10:54:31.335816 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac07dcc7-2c32-40b9-802b-8380d6c2ef3e-kube-api-access-bx5pl" (OuterVolumeSpecName: "kube-api-access-bx5pl") pod "ac07dcc7-2c32-40b9-802b-8380d6c2ef3e" (UID: "ac07dcc7-2c32-40b9-802b-8380d6c2ef3e"). InnerVolumeSpecName "kube-api-access-bx5pl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:54:31 crc kubenswrapper[4776]: I1125 10:54:31.376037 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac07dcc7-2c32-40b9-802b-8380d6c2ef3e-config" (OuterVolumeSpecName: "config") pod "ac07dcc7-2c32-40b9-802b-8380d6c2ef3e" (UID: "ac07dcc7-2c32-40b9-802b-8380d6c2ef3e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:31 crc kubenswrapper[4776]: I1125 10:54:31.376762 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac07dcc7-2c32-40b9-802b-8380d6c2ef3e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ac07dcc7-2c32-40b9-802b-8380d6c2ef3e" (UID: "ac07dcc7-2c32-40b9-802b-8380d6c2ef3e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:31 crc kubenswrapper[4776]: I1125 10:54:31.433265 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac07dcc7-2c32-40b9-802b-8380d6c2ef3e-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:31 crc kubenswrapper[4776]: I1125 10:54:31.433320 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bx5pl\" (UniqueName: \"kubernetes.io/projected/ac07dcc7-2c32-40b9-802b-8380d6c2ef3e-kube-api-access-bx5pl\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:31 crc kubenswrapper[4776]: I1125 10:54:31.433334 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac07dcc7-2c32-40b9-802b-8380d6c2ef3e-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:31 crc kubenswrapper[4776]: I1125 10:54:31.662528 4776 scope.go:117] "RemoveContainer" containerID="f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40" Nov 25 10:54:31 crc kubenswrapper[4776]: E1125 10:54:31.663240 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:54:31 crc kubenswrapper[4776]: I1125 10:54:31.775649 4776 generic.go:334] "Generic (PLEG): container finished" podID="ac07dcc7-2c32-40b9-802b-8380d6c2ef3e" containerID="e428da7d3af50e7504fbf05eed836a8b5431da41599a55c00486121248dd4b70" exitCode=0 Nov 25 10:54:31 crc kubenswrapper[4776]: I1125 10:54:31.775721 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54564445dc-5wx2z" Nov 25 10:54:31 crc kubenswrapper[4776]: I1125 10:54:31.775729 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54564445dc-5wx2z" event={"ID":"ac07dcc7-2c32-40b9-802b-8380d6c2ef3e","Type":"ContainerDied","Data":"e428da7d3af50e7504fbf05eed836a8b5431da41599a55c00486121248dd4b70"} Nov 25 10:54:31 crc kubenswrapper[4776]: I1125 10:54:31.775842 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54564445dc-5wx2z" event={"ID":"ac07dcc7-2c32-40b9-802b-8380d6c2ef3e","Type":"ContainerDied","Data":"15247c76da0d56702119c5f9ea2db2fda9961944ae06d1f7f6e72b7c349ede45"} Nov 25 10:54:31 crc kubenswrapper[4776]: I1125 10:54:31.775862 4776 scope.go:117] "RemoveContainer" containerID="e428da7d3af50e7504fbf05eed836a8b5431da41599a55c00486121248dd4b70" Nov 25 10:54:31 crc kubenswrapper[4776]: I1125 10:54:31.777701 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7f396897-96fa-422c-a2fd-92faadaff6e3","Type":"ContainerStarted","Data":"81f0d47badb0253174cf71e34c8bc9b25100a12be813d5c40237bc4c0c99dcbd"} Nov 25 10:54:31 crc kubenswrapper[4776]: I1125 10:54:31.777738 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7f396897-96fa-422c-a2fd-92faadaff6e3","Type":"ContainerStarted","Data":"fd03fefdc11ee093ab9045af3f44aec4500440d2cc0e12ef9de4c873d494b634"} Nov 25 10:54:31 crc kubenswrapper[4776]: I1125 10:54:31.777845 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 25 10:54:31 crc kubenswrapper[4776]: I1125 10:54:31.795344 4776 scope.go:117] "RemoveContainer" containerID="a9828407e3e3001e79a68285213e3c6a1a99a8ffe58d0b51a724488e05c7e7e6" Nov 25 10:54:31 crc kubenswrapper[4776]: I1125 10:54:31.802793 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.802768663 podStartE2EDuration="2.802768663s" podCreationTimestamp="2025-11-25 10:54:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:54:31.795972832 +0000 UTC m=+5416.837032385" watchObservedRunningTime="2025-11-25 10:54:31.802768663 +0000 UTC m=+5416.843828216" Nov 25 10:54:31 crc kubenswrapper[4776]: I1125 10:54:31.817187 4776 scope.go:117] "RemoveContainer" containerID="e428da7d3af50e7504fbf05eed836a8b5431da41599a55c00486121248dd4b70" Nov 25 10:54:31 crc kubenswrapper[4776]: E1125 10:54:31.818548 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e428da7d3af50e7504fbf05eed836a8b5431da41599a55c00486121248dd4b70\": container with ID starting with e428da7d3af50e7504fbf05eed836a8b5431da41599a55c00486121248dd4b70 not found: ID does not exist" containerID="e428da7d3af50e7504fbf05eed836a8b5431da41599a55c00486121248dd4b70" Nov 25 10:54:31 crc kubenswrapper[4776]: I1125 10:54:31.818579 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e428da7d3af50e7504fbf05eed836a8b5431da41599a55c00486121248dd4b70"} err="failed to get container status \"e428da7d3af50e7504fbf05eed836a8b5431da41599a55c00486121248dd4b70\": rpc error: code = NotFound desc = could not find container \"e428da7d3af50e7504fbf05eed836a8b5431da41599a55c00486121248dd4b70\": container with ID starting with e428da7d3af50e7504fbf05eed836a8b5431da41599a55c00486121248dd4b70 not found: ID does not exist" Nov 25 10:54:31 crc kubenswrapper[4776]: I1125 10:54:31.818597 4776 scope.go:117] "RemoveContainer" containerID="a9828407e3e3001e79a68285213e3c6a1a99a8ffe58d0b51a724488e05c7e7e6" Nov 25 10:54:31 crc kubenswrapper[4776]: E1125 10:54:31.822350 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9828407e3e3001e79a68285213e3c6a1a99a8ffe58d0b51a724488e05c7e7e6\": container with ID starting with a9828407e3e3001e79a68285213e3c6a1a99a8ffe58d0b51a724488e05c7e7e6 not found: ID does not exist" containerID="a9828407e3e3001e79a68285213e3c6a1a99a8ffe58d0b51a724488e05c7e7e6" Nov 25 10:54:31 crc kubenswrapper[4776]: I1125 10:54:31.822394 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9828407e3e3001e79a68285213e3c6a1a99a8ffe58d0b51a724488e05c7e7e6"} err="failed to get container status \"a9828407e3e3001e79a68285213e3c6a1a99a8ffe58d0b51a724488e05c7e7e6\": rpc error: code = NotFound desc = could not find container \"a9828407e3e3001e79a68285213e3c6a1a99a8ffe58d0b51a724488e05c7e7e6\": container with ID starting with a9828407e3e3001e79a68285213e3c6a1a99a8ffe58d0b51a724488e05c7e7e6 not found: ID does not exist" Nov 25 10:54:31 crc kubenswrapper[4776]: I1125 10:54:31.832671 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54564445dc-5wx2z"] Nov 25 10:54:31 crc kubenswrapper[4776]: I1125 10:54:31.838768 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54564445dc-5wx2z"] Nov 25 10:54:33 crc kubenswrapper[4776]: I1125 10:54:33.672836 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac07dcc7-2c32-40b9-802b-8380d6c2ef3e" path="/var/lib/kubelet/pods/ac07dcc7-2c32-40b9-802b-8380d6c2ef3e/volumes" Nov 25 10:54:34 crc kubenswrapper[4776]: I1125 10:54:34.969851 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-9zjdz"] Nov 25 10:54:34 crc kubenswrapper[4776]: E1125 10:54:34.970495 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac07dcc7-2c32-40b9-802b-8380d6c2ef3e" containerName="dnsmasq-dns" Nov 25 10:54:34 crc kubenswrapper[4776]: I1125 10:54:34.970509 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac07dcc7-2c32-40b9-802b-8380d6c2ef3e" containerName="dnsmasq-dns" Nov 25 10:54:34 crc kubenswrapper[4776]: E1125 10:54:34.970525 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac07dcc7-2c32-40b9-802b-8380d6c2ef3e" containerName="init" Nov 25 10:54:34 crc kubenswrapper[4776]: I1125 10:54:34.970534 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac07dcc7-2c32-40b9-802b-8380d6c2ef3e" containerName="init" Nov 25 10:54:34 crc kubenswrapper[4776]: I1125 10:54:34.970770 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac07dcc7-2c32-40b9-802b-8380d6c2ef3e" containerName="dnsmasq-dns" Nov 25 10:54:34 crc kubenswrapper[4776]: I1125 10:54:34.971395 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9zjdz" Nov 25 10:54:34 crc kubenswrapper[4776]: I1125 10:54:34.983141 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-9zjdz"] Nov 25 10:54:35 crc kubenswrapper[4776]: I1125 10:54:35.004097 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-3a30-account-create-rwgn5"] Nov 25 10:54:35 crc kubenswrapper[4776]: I1125 10:54:35.005131 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3a30-account-create-rwgn5" Nov 25 10:54:35 crc kubenswrapper[4776]: I1125 10:54:35.008494 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-3a30-account-create-rwgn5"] Nov 25 10:54:35 crc kubenswrapper[4776]: I1125 10:54:35.040305 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 25 10:54:35 crc kubenswrapper[4776]: I1125 10:54:35.097203 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjtxd\" (UniqueName: \"kubernetes.io/projected/0e9ca664-92bc-49cc-8bca-66d0fbf38343-kube-api-access-qjtxd\") pod \"keystone-3a30-account-create-rwgn5\" (UID: \"0e9ca664-92bc-49cc-8bca-66d0fbf38343\") " pod="openstack/keystone-3a30-account-create-rwgn5" Nov 25 10:54:35 crc kubenswrapper[4776]: I1125 10:54:35.097548 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8t59\" (UniqueName: \"kubernetes.io/projected/b3d99a94-8d26-4627-bd67-818fcca07854-kube-api-access-g8t59\") pod \"keystone-db-create-9zjdz\" (UID: \"b3d99a94-8d26-4627-bd67-818fcca07854\") " pod="openstack/keystone-db-create-9zjdz" Nov 25 10:54:35 crc kubenswrapper[4776]: I1125 10:54:35.097841 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e9ca664-92bc-49cc-8bca-66d0fbf38343-operator-scripts\") pod \"keystone-3a30-account-create-rwgn5\" (UID: \"0e9ca664-92bc-49cc-8bca-66d0fbf38343\") " pod="openstack/keystone-3a30-account-create-rwgn5" Nov 25 10:54:35 crc kubenswrapper[4776]: I1125 10:54:35.097885 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3d99a94-8d26-4627-bd67-818fcca07854-operator-scripts\") pod \"keystone-db-create-9zjdz\" (UID: \"b3d99a94-8d26-4627-bd67-818fcca07854\") " pod="openstack/keystone-db-create-9zjdz" Nov 25 10:54:35 crc kubenswrapper[4776]: I1125 10:54:35.199518 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e9ca664-92bc-49cc-8bca-66d0fbf38343-operator-scripts\") pod \"keystone-3a30-account-create-rwgn5\" (UID: \"0e9ca664-92bc-49cc-8bca-66d0fbf38343\") " pod="openstack/keystone-3a30-account-create-rwgn5" Nov 25 10:54:35 crc kubenswrapper[4776]: I1125 10:54:35.199562 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3d99a94-8d26-4627-bd67-818fcca07854-operator-scripts\") pod \"keystone-db-create-9zjdz\" (UID: \"b3d99a94-8d26-4627-bd67-818fcca07854\") " pod="openstack/keystone-db-create-9zjdz" Nov 25 10:54:35 crc kubenswrapper[4776]: I1125 10:54:35.199613 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjtxd\" (UniqueName: \"kubernetes.io/projected/0e9ca664-92bc-49cc-8bca-66d0fbf38343-kube-api-access-qjtxd\") pod \"keystone-3a30-account-create-rwgn5\" (UID: \"0e9ca664-92bc-49cc-8bca-66d0fbf38343\") " pod="openstack/keystone-3a30-account-create-rwgn5" Nov 25 10:54:35 crc kubenswrapper[4776]: I1125 10:54:35.199675 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8t59\" (UniqueName: \"kubernetes.io/projected/b3d99a94-8d26-4627-bd67-818fcca07854-kube-api-access-g8t59\") pod \"keystone-db-create-9zjdz\" (UID: \"b3d99a94-8d26-4627-bd67-818fcca07854\") " pod="openstack/keystone-db-create-9zjdz" Nov 25 10:54:35 crc kubenswrapper[4776]: I1125 10:54:35.200404 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e9ca664-92bc-49cc-8bca-66d0fbf38343-operator-scripts\") pod \"keystone-3a30-account-create-rwgn5\" (UID: \"0e9ca664-92bc-49cc-8bca-66d0fbf38343\") " pod="openstack/keystone-3a30-account-create-rwgn5" Nov 25 10:54:35 crc kubenswrapper[4776]: I1125 10:54:35.200516 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3d99a94-8d26-4627-bd67-818fcca07854-operator-scripts\") pod \"keystone-db-create-9zjdz\" (UID: \"b3d99a94-8d26-4627-bd67-818fcca07854\") " pod="openstack/keystone-db-create-9zjdz" Nov 25 10:54:35 crc kubenswrapper[4776]: I1125 10:54:35.230821 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjtxd\" (UniqueName: \"kubernetes.io/projected/0e9ca664-92bc-49cc-8bca-66d0fbf38343-kube-api-access-qjtxd\") pod \"keystone-3a30-account-create-rwgn5\" (UID: \"0e9ca664-92bc-49cc-8bca-66d0fbf38343\") " pod="openstack/keystone-3a30-account-create-rwgn5" Nov 25 10:54:35 crc kubenswrapper[4776]: I1125 10:54:35.239307 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8t59\" (UniqueName: \"kubernetes.io/projected/b3d99a94-8d26-4627-bd67-818fcca07854-kube-api-access-g8t59\") pod \"keystone-db-create-9zjdz\" (UID: \"b3d99a94-8d26-4627-bd67-818fcca07854\") " pod="openstack/keystone-db-create-9zjdz" Nov 25 10:54:35 crc kubenswrapper[4776]: I1125 10:54:35.290897 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9zjdz" Nov 25 10:54:35 crc kubenswrapper[4776]: I1125 10:54:35.354443 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3a30-account-create-rwgn5" Nov 25 10:54:35 crc kubenswrapper[4776]: I1125 10:54:35.764608 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-9zjdz"] Nov 25 10:54:35 crc kubenswrapper[4776]: I1125 10:54:35.809754 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9zjdz" event={"ID":"b3d99a94-8d26-4627-bd67-818fcca07854","Type":"ContainerStarted","Data":"02ef8c5fee56a38cda631d5a9302014f0560f0872efa1d7b0b8e4e1b8473d49d"} Nov 25 10:54:35 crc kubenswrapper[4776]: I1125 10:54:35.888575 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-3a30-account-create-rwgn5"] Nov 25 10:54:35 crc kubenswrapper[4776]: W1125 10:54:35.891516 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e9ca664_92bc_49cc_8bca_66d0fbf38343.slice/crio-80b26efd7b8be8a576c238186425825c0c8f6efad86e435aca17b615290a09f5 WatchSource:0}: Error finding container 80b26efd7b8be8a576c238186425825c0c8f6efad86e435aca17b615290a09f5: Status 404 returned error can't find the container with id 80b26efd7b8be8a576c238186425825c0c8f6efad86e435aca17b615290a09f5 Nov 25 10:54:36 crc kubenswrapper[4776]: I1125 10:54:36.823429 4776 generic.go:334] "Generic (PLEG): container finished" podID="b3d99a94-8d26-4627-bd67-818fcca07854" containerID="99918cbec2ca01810476b8024781dfaddbdd25714aa55a102117b6fa681ba895" exitCode=0 Nov 25 10:54:36 crc kubenswrapper[4776]: I1125 10:54:36.823486 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9zjdz" event={"ID":"b3d99a94-8d26-4627-bd67-818fcca07854","Type":"ContainerDied","Data":"99918cbec2ca01810476b8024781dfaddbdd25714aa55a102117b6fa681ba895"} Nov 25 10:54:36 crc kubenswrapper[4776]: I1125 10:54:36.825751 4776 generic.go:334] "Generic (PLEG): container finished" podID="0e9ca664-92bc-49cc-8bca-66d0fbf38343" containerID="2d918aa82f7a519fd0d43d5392ee5dfc86d0b00a061e1b3f1ddfa9b88d81c779" exitCode=0 Nov 25 10:54:36 crc kubenswrapper[4776]: I1125 10:54:36.825880 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3a30-account-create-rwgn5" event={"ID":"0e9ca664-92bc-49cc-8bca-66d0fbf38343","Type":"ContainerDied","Data":"2d918aa82f7a519fd0d43d5392ee5dfc86d0b00a061e1b3f1ddfa9b88d81c779"} Nov 25 10:54:36 crc kubenswrapper[4776]: I1125 10:54:36.825961 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3a30-account-create-rwgn5" event={"ID":"0e9ca664-92bc-49cc-8bca-66d0fbf38343","Type":"ContainerStarted","Data":"80b26efd7b8be8a576c238186425825c0c8f6efad86e435aca17b615290a09f5"} Nov 25 10:54:38 crc kubenswrapper[4776]: I1125 10:54:38.236915 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9zjdz" Nov 25 10:54:38 crc kubenswrapper[4776]: I1125 10:54:38.243279 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3a30-account-create-rwgn5" Nov 25 10:54:38 crc kubenswrapper[4776]: I1125 10:54:38.377792 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e9ca664-92bc-49cc-8bca-66d0fbf38343-operator-scripts\") pod \"0e9ca664-92bc-49cc-8bca-66d0fbf38343\" (UID: \"0e9ca664-92bc-49cc-8bca-66d0fbf38343\") " Nov 25 10:54:38 crc kubenswrapper[4776]: I1125 10:54:38.377944 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8t59\" (UniqueName: \"kubernetes.io/projected/b3d99a94-8d26-4627-bd67-818fcca07854-kube-api-access-g8t59\") pod \"b3d99a94-8d26-4627-bd67-818fcca07854\" (UID: \"b3d99a94-8d26-4627-bd67-818fcca07854\") " Nov 25 10:54:38 crc kubenswrapper[4776]: I1125 10:54:38.377979 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3d99a94-8d26-4627-bd67-818fcca07854-operator-scripts\") pod \"b3d99a94-8d26-4627-bd67-818fcca07854\" (UID: \"b3d99a94-8d26-4627-bd67-818fcca07854\") " Nov 25 10:54:38 crc kubenswrapper[4776]: I1125 10:54:38.378120 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjtxd\" (UniqueName: \"kubernetes.io/projected/0e9ca664-92bc-49cc-8bca-66d0fbf38343-kube-api-access-qjtxd\") pod \"0e9ca664-92bc-49cc-8bca-66d0fbf38343\" (UID: \"0e9ca664-92bc-49cc-8bca-66d0fbf38343\") " Nov 25 10:54:38 crc kubenswrapper[4776]: I1125 10:54:38.378625 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e9ca664-92bc-49cc-8bca-66d0fbf38343-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0e9ca664-92bc-49cc-8bca-66d0fbf38343" (UID: "0e9ca664-92bc-49cc-8bca-66d0fbf38343"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:38 crc kubenswrapper[4776]: I1125 10:54:38.378898 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3d99a94-8d26-4627-bd67-818fcca07854-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b3d99a94-8d26-4627-bd67-818fcca07854" (UID: "b3d99a94-8d26-4627-bd67-818fcca07854"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:38 crc kubenswrapper[4776]: I1125 10:54:38.383320 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e9ca664-92bc-49cc-8bca-66d0fbf38343-kube-api-access-qjtxd" (OuterVolumeSpecName: "kube-api-access-qjtxd") pod "0e9ca664-92bc-49cc-8bca-66d0fbf38343" (UID: "0e9ca664-92bc-49cc-8bca-66d0fbf38343"). InnerVolumeSpecName "kube-api-access-qjtxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:54:38 crc kubenswrapper[4776]: I1125 10:54:38.384822 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3d99a94-8d26-4627-bd67-818fcca07854-kube-api-access-g8t59" (OuterVolumeSpecName: "kube-api-access-g8t59") pod "b3d99a94-8d26-4627-bd67-818fcca07854" (UID: "b3d99a94-8d26-4627-bd67-818fcca07854"). InnerVolumeSpecName "kube-api-access-g8t59". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:54:38 crc kubenswrapper[4776]: I1125 10:54:38.479884 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e9ca664-92bc-49cc-8bca-66d0fbf38343-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:38 crc kubenswrapper[4776]: I1125 10:54:38.479921 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8t59\" (UniqueName: \"kubernetes.io/projected/b3d99a94-8d26-4627-bd67-818fcca07854-kube-api-access-g8t59\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:38 crc kubenswrapper[4776]: I1125 10:54:38.479933 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3d99a94-8d26-4627-bd67-818fcca07854-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:38 crc kubenswrapper[4776]: I1125 10:54:38.479941 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjtxd\" (UniqueName: \"kubernetes.io/projected/0e9ca664-92bc-49cc-8bca-66d0fbf38343-kube-api-access-qjtxd\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:38 crc kubenswrapper[4776]: I1125 10:54:38.844587 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3a30-account-create-rwgn5" event={"ID":"0e9ca664-92bc-49cc-8bca-66d0fbf38343","Type":"ContainerDied","Data":"80b26efd7b8be8a576c238186425825c0c8f6efad86e435aca17b615290a09f5"} Nov 25 10:54:38 crc kubenswrapper[4776]: I1125 10:54:38.844643 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80b26efd7b8be8a576c238186425825c0c8f6efad86e435aca17b615290a09f5" Nov 25 10:54:38 crc kubenswrapper[4776]: I1125 10:54:38.844712 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3a30-account-create-rwgn5" Nov 25 10:54:38 crc kubenswrapper[4776]: I1125 10:54:38.847366 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9zjdz" event={"ID":"b3d99a94-8d26-4627-bd67-818fcca07854","Type":"ContainerDied","Data":"02ef8c5fee56a38cda631d5a9302014f0560f0872efa1d7b0b8e4e1b8473d49d"} Nov 25 10:54:38 crc kubenswrapper[4776]: I1125 10:54:38.847417 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02ef8c5fee56a38cda631d5a9302014f0560f0872efa1d7b0b8e4e1b8473d49d" Nov 25 10:54:38 crc kubenswrapper[4776]: I1125 10:54:38.847464 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9zjdz" Nov 25 10:54:40 crc kubenswrapper[4776]: I1125 10:54:40.099578 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 25 10:54:40 crc kubenswrapper[4776]: I1125 10:54:40.553046 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-jg4k7"] Nov 25 10:54:40 crc kubenswrapper[4776]: E1125 10:54:40.553535 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3d99a94-8d26-4627-bd67-818fcca07854" containerName="mariadb-database-create" Nov 25 10:54:40 crc kubenswrapper[4776]: I1125 10:54:40.553558 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3d99a94-8d26-4627-bd67-818fcca07854" containerName="mariadb-database-create" Nov 25 10:54:40 crc kubenswrapper[4776]: E1125 10:54:40.553572 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e9ca664-92bc-49cc-8bca-66d0fbf38343" containerName="mariadb-account-create" Nov 25 10:54:40 crc kubenswrapper[4776]: I1125 10:54:40.553580 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e9ca664-92bc-49cc-8bca-66d0fbf38343" containerName="mariadb-account-create" Nov 25 10:54:40 crc kubenswrapper[4776]: I1125 10:54:40.553794 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e9ca664-92bc-49cc-8bca-66d0fbf38343" containerName="mariadb-account-create" Nov 25 10:54:40 crc kubenswrapper[4776]: I1125 10:54:40.553816 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3d99a94-8d26-4627-bd67-818fcca07854" containerName="mariadb-database-create" Nov 25 10:54:40 crc kubenswrapper[4776]: I1125 10:54:40.554515 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jg4k7" Nov 25 10:54:40 crc kubenswrapper[4776]: I1125 10:54:40.557404 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 10:54:40 crc kubenswrapper[4776]: I1125 10:54:40.557613 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-lnvfd" Nov 25 10:54:40 crc kubenswrapper[4776]: I1125 10:54:40.557761 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 10:54:40 crc kubenswrapper[4776]: I1125 10:54:40.563423 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-jg4k7"] Nov 25 10:54:40 crc kubenswrapper[4776]: I1125 10:54:40.565962 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 10:54:40 crc kubenswrapper[4776]: I1125 10:54:40.618765 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhpjd\" (UniqueName: \"kubernetes.io/projected/0da5ee40-85c5-4cf9-bf03-a1184d208541-kube-api-access-fhpjd\") pod \"keystone-db-sync-jg4k7\" (UID: \"0da5ee40-85c5-4cf9-bf03-a1184d208541\") " pod="openstack/keystone-db-sync-jg4k7" Nov 25 10:54:40 crc kubenswrapper[4776]: I1125 10:54:40.618819 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0da5ee40-85c5-4cf9-bf03-a1184d208541-config-data\") pod \"keystone-db-sync-jg4k7\" (UID: \"0da5ee40-85c5-4cf9-bf03-a1184d208541\") " pod="openstack/keystone-db-sync-jg4k7" Nov 25 10:54:40 crc kubenswrapper[4776]: I1125 10:54:40.618892 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0da5ee40-85c5-4cf9-bf03-a1184d208541-combined-ca-bundle\") pod \"keystone-db-sync-jg4k7\" (UID: \"0da5ee40-85c5-4cf9-bf03-a1184d208541\") " pod="openstack/keystone-db-sync-jg4k7" Nov 25 10:54:40 crc kubenswrapper[4776]: I1125 10:54:40.721232 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhpjd\" (UniqueName: \"kubernetes.io/projected/0da5ee40-85c5-4cf9-bf03-a1184d208541-kube-api-access-fhpjd\") pod \"keystone-db-sync-jg4k7\" (UID: \"0da5ee40-85c5-4cf9-bf03-a1184d208541\") " pod="openstack/keystone-db-sync-jg4k7" Nov 25 10:54:40 crc kubenswrapper[4776]: I1125 10:54:40.721383 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0da5ee40-85c5-4cf9-bf03-a1184d208541-config-data\") pod \"keystone-db-sync-jg4k7\" (UID: \"0da5ee40-85c5-4cf9-bf03-a1184d208541\") " pod="openstack/keystone-db-sync-jg4k7" Nov 25 10:54:40 crc kubenswrapper[4776]: I1125 10:54:40.721515 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0da5ee40-85c5-4cf9-bf03-a1184d208541-combined-ca-bundle\") pod \"keystone-db-sync-jg4k7\" (UID: \"0da5ee40-85c5-4cf9-bf03-a1184d208541\") " pod="openstack/keystone-db-sync-jg4k7" Nov 25 10:54:40 crc kubenswrapper[4776]: I1125 10:54:40.727908 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0da5ee40-85c5-4cf9-bf03-a1184d208541-combined-ca-bundle\") pod \"keystone-db-sync-jg4k7\" (UID: \"0da5ee40-85c5-4cf9-bf03-a1184d208541\") " pod="openstack/keystone-db-sync-jg4k7" Nov 25 10:54:40 crc kubenswrapper[4776]: I1125 10:54:40.728385 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0da5ee40-85c5-4cf9-bf03-a1184d208541-config-data\") pod \"keystone-db-sync-jg4k7\" (UID: \"0da5ee40-85c5-4cf9-bf03-a1184d208541\") " pod="openstack/keystone-db-sync-jg4k7" Nov 25 10:54:40 crc kubenswrapper[4776]: I1125 10:54:40.736958 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhpjd\" (UniqueName: \"kubernetes.io/projected/0da5ee40-85c5-4cf9-bf03-a1184d208541-kube-api-access-fhpjd\") pod \"keystone-db-sync-jg4k7\" (UID: \"0da5ee40-85c5-4cf9-bf03-a1184d208541\") " pod="openstack/keystone-db-sync-jg4k7" Nov 25 10:54:40 crc kubenswrapper[4776]: I1125 10:54:40.875811 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jg4k7" Nov 25 10:54:41 crc kubenswrapper[4776]: I1125 10:54:41.323370 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-jg4k7"] Nov 25 10:54:41 crc kubenswrapper[4776]: I1125 10:54:41.875132 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jg4k7" event={"ID":"0da5ee40-85c5-4cf9-bf03-a1184d208541","Type":"ContainerStarted","Data":"54822ea4c536a47287cc033d08783b1b390937d0ee1f7d8174f2653ca047f4df"} Nov 25 10:54:41 crc kubenswrapper[4776]: I1125 10:54:41.875697 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jg4k7" event={"ID":"0da5ee40-85c5-4cf9-bf03-a1184d208541","Type":"ContainerStarted","Data":"681011cad8d6b1872a6bf4b01a686d014a7d8f3605221e5d3172ec27a38e526f"} Nov 25 10:54:41 crc kubenswrapper[4776]: I1125 10:54:41.899144 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-jg4k7" podStartSLOduration=1.899118157 podStartE2EDuration="1.899118157s" podCreationTimestamp="2025-11-25 10:54:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:54:41.89763771 +0000 UTC m=+5426.938697263" watchObservedRunningTime="2025-11-25 10:54:41.899118157 +0000 UTC m=+5426.940177750" Nov 25 10:54:43 crc kubenswrapper[4776]: I1125 10:54:43.894000 4776 generic.go:334] "Generic (PLEG): container finished" podID="0da5ee40-85c5-4cf9-bf03-a1184d208541" containerID="54822ea4c536a47287cc033d08783b1b390937d0ee1f7d8174f2653ca047f4df" exitCode=0 Nov 25 10:54:43 crc kubenswrapper[4776]: I1125 10:54:43.894108 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jg4k7" event={"ID":"0da5ee40-85c5-4cf9-bf03-a1184d208541","Type":"ContainerDied","Data":"54822ea4c536a47287cc033d08783b1b390937d0ee1f7d8174f2653ca047f4df"} Nov 25 10:54:44 crc kubenswrapper[4776]: I1125 10:54:44.661925 4776 scope.go:117] "RemoveContainer" containerID="f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40" Nov 25 10:54:44 crc kubenswrapper[4776]: E1125 10:54:44.662524 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:54:45 crc kubenswrapper[4776]: I1125 10:54:45.211291 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jg4k7" Nov 25 10:54:45 crc kubenswrapper[4776]: I1125 10:54:45.299735 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0da5ee40-85c5-4cf9-bf03-a1184d208541-combined-ca-bundle\") pod \"0da5ee40-85c5-4cf9-bf03-a1184d208541\" (UID: \"0da5ee40-85c5-4cf9-bf03-a1184d208541\") " Nov 25 10:54:45 crc kubenswrapper[4776]: I1125 10:54:45.299879 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhpjd\" (UniqueName: \"kubernetes.io/projected/0da5ee40-85c5-4cf9-bf03-a1184d208541-kube-api-access-fhpjd\") pod \"0da5ee40-85c5-4cf9-bf03-a1184d208541\" (UID: \"0da5ee40-85c5-4cf9-bf03-a1184d208541\") " Nov 25 10:54:45 crc kubenswrapper[4776]: I1125 10:54:45.299980 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0da5ee40-85c5-4cf9-bf03-a1184d208541-config-data\") pod \"0da5ee40-85c5-4cf9-bf03-a1184d208541\" (UID: \"0da5ee40-85c5-4cf9-bf03-a1184d208541\") " Nov 25 10:54:45 crc kubenswrapper[4776]: I1125 10:54:45.305956 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0da5ee40-85c5-4cf9-bf03-a1184d208541-kube-api-access-fhpjd" (OuterVolumeSpecName: "kube-api-access-fhpjd") pod "0da5ee40-85c5-4cf9-bf03-a1184d208541" (UID: "0da5ee40-85c5-4cf9-bf03-a1184d208541"). InnerVolumeSpecName "kube-api-access-fhpjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:54:45 crc kubenswrapper[4776]: I1125 10:54:45.328313 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0da5ee40-85c5-4cf9-bf03-a1184d208541-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0da5ee40-85c5-4cf9-bf03-a1184d208541" (UID: "0da5ee40-85c5-4cf9-bf03-a1184d208541"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:45 crc kubenswrapper[4776]: I1125 10:54:45.347315 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0da5ee40-85c5-4cf9-bf03-a1184d208541-config-data" (OuterVolumeSpecName: "config-data") pod "0da5ee40-85c5-4cf9-bf03-a1184d208541" (UID: "0da5ee40-85c5-4cf9-bf03-a1184d208541"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:45 crc kubenswrapper[4776]: I1125 10:54:45.402346 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0da5ee40-85c5-4cf9-bf03-a1184d208541-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:45 crc kubenswrapper[4776]: I1125 10:54:45.402600 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhpjd\" (UniqueName: \"kubernetes.io/projected/0da5ee40-85c5-4cf9-bf03-a1184d208541-kube-api-access-fhpjd\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:45 crc kubenswrapper[4776]: I1125 10:54:45.402687 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0da5ee40-85c5-4cf9-bf03-a1184d208541-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:45 crc kubenswrapper[4776]: I1125 10:54:45.913040 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jg4k7" event={"ID":"0da5ee40-85c5-4cf9-bf03-a1184d208541","Type":"ContainerDied","Data":"681011cad8d6b1872a6bf4b01a686d014a7d8f3605221e5d3172ec27a38e526f"} Nov 25 10:54:45 crc kubenswrapper[4776]: I1125 10:54:45.913118 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jg4k7" Nov 25 10:54:45 crc kubenswrapper[4776]: I1125 10:54:45.913139 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="681011cad8d6b1872a6bf4b01a686d014a7d8f3605221e5d3172ec27a38e526f" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.151005 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-f7vrf"] Nov 25 10:54:46 crc kubenswrapper[4776]: E1125 10:54:46.151499 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0da5ee40-85c5-4cf9-bf03-a1184d208541" containerName="keystone-db-sync" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.151523 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="0da5ee40-85c5-4cf9-bf03-a1184d208541" containerName="keystone-db-sync" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.151739 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="0da5ee40-85c5-4cf9-bf03-a1184d208541" containerName="keystone-db-sync" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.152534 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-f7vrf" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.156860 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.157167 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-lnvfd" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.157356 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.157486 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.157625 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.164498 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-f7vrf"] Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.181828 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6cbbbb95c9-bng2s"] Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.183619 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cbbbb95c9-bng2s" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.205096 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6cbbbb95c9-bng2s"] Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.216579 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-combined-ca-bundle\") pod \"keystone-bootstrap-f7vrf\" (UID: \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\") " pod="openstack/keystone-bootstrap-f7vrf" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.216883 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-config-data\") pod \"keystone-bootstrap-f7vrf\" (UID: \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\") " pod="openstack/keystone-bootstrap-f7vrf" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.216940 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-scripts\") pod \"keystone-bootstrap-f7vrf\" (UID: \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\") " pod="openstack/keystone-bootstrap-f7vrf" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.216981 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-credential-keys\") pod \"keystone-bootstrap-f7vrf\" (UID: \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\") " pod="openstack/keystone-bootstrap-f7vrf" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.217004 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2djg\" (UniqueName: \"kubernetes.io/projected/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-kube-api-access-j2djg\") pod \"keystone-bootstrap-f7vrf\" (UID: \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\") " pod="openstack/keystone-bootstrap-f7vrf" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.217077 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-fernet-keys\") pod \"keystone-bootstrap-f7vrf\" (UID: \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\") " pod="openstack/keystone-bootstrap-f7vrf" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.318921 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-config-data\") pod \"keystone-bootstrap-f7vrf\" (UID: \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\") " pod="openstack/keystone-bootstrap-f7vrf" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.318983 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-scripts\") pod \"keystone-bootstrap-f7vrf\" (UID: \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\") " pod="openstack/keystone-bootstrap-f7vrf" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.319010 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-credential-keys\") pod \"keystone-bootstrap-f7vrf\" (UID: \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\") " pod="openstack/keystone-bootstrap-f7vrf" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.319029 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2djg\" (UniqueName: \"kubernetes.io/projected/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-kube-api-access-j2djg\") pod \"keystone-bootstrap-f7vrf\" (UID: \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\") " pod="openstack/keystone-bootstrap-f7vrf" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.319118 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2a467eb0-222b-4486-aa5c-07bba04b741c-ovsdbserver-sb\") pod \"dnsmasq-dns-6cbbbb95c9-bng2s\" (UID: \"2a467eb0-222b-4486-aa5c-07bba04b741c\") " pod="openstack/dnsmasq-dns-6cbbbb95c9-bng2s" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.319151 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a467eb0-222b-4486-aa5c-07bba04b741c-config\") pod \"dnsmasq-dns-6cbbbb95c9-bng2s\" (UID: \"2a467eb0-222b-4486-aa5c-07bba04b741c\") " pod="openstack/dnsmasq-dns-6cbbbb95c9-bng2s" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.319178 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2a467eb0-222b-4486-aa5c-07bba04b741c-ovsdbserver-nb\") pod \"dnsmasq-dns-6cbbbb95c9-bng2s\" (UID: \"2a467eb0-222b-4486-aa5c-07bba04b741c\") " pod="openstack/dnsmasq-dns-6cbbbb95c9-bng2s" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.319196 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-fernet-keys\") pod \"keystone-bootstrap-f7vrf\" (UID: \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\") " pod="openstack/keystone-bootstrap-f7vrf" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.319216 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwfwh\" (UniqueName: \"kubernetes.io/projected/2a467eb0-222b-4486-aa5c-07bba04b741c-kube-api-access-vwfwh\") pod \"dnsmasq-dns-6cbbbb95c9-bng2s\" (UID: \"2a467eb0-222b-4486-aa5c-07bba04b741c\") " pod="openstack/dnsmasq-dns-6cbbbb95c9-bng2s" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.319266 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-combined-ca-bundle\") pod \"keystone-bootstrap-f7vrf\" (UID: \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\") " pod="openstack/keystone-bootstrap-f7vrf" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.319291 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a467eb0-222b-4486-aa5c-07bba04b741c-dns-svc\") pod \"dnsmasq-dns-6cbbbb95c9-bng2s\" (UID: \"2a467eb0-222b-4486-aa5c-07bba04b741c\") " pod="openstack/dnsmasq-dns-6cbbbb95c9-bng2s" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.327131 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-combined-ca-bundle\") pod \"keystone-bootstrap-f7vrf\" (UID: \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\") " pod="openstack/keystone-bootstrap-f7vrf" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.329235 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-config-data\") pod \"keystone-bootstrap-f7vrf\" (UID: \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\") " pod="openstack/keystone-bootstrap-f7vrf" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.329656 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-scripts\") pod \"keystone-bootstrap-f7vrf\" (UID: \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\") " pod="openstack/keystone-bootstrap-f7vrf" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.330585 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-credential-keys\") pod \"keystone-bootstrap-f7vrf\" (UID: \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\") " pod="openstack/keystone-bootstrap-f7vrf" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.331699 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-fernet-keys\") pod \"keystone-bootstrap-f7vrf\" (UID: \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\") " pod="openstack/keystone-bootstrap-f7vrf" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.336588 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2djg\" (UniqueName: \"kubernetes.io/projected/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-kube-api-access-j2djg\") pod \"keystone-bootstrap-f7vrf\" (UID: \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\") " pod="openstack/keystone-bootstrap-f7vrf" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.422453 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2a467eb0-222b-4486-aa5c-07bba04b741c-ovsdbserver-sb\") pod \"dnsmasq-dns-6cbbbb95c9-bng2s\" (UID: \"2a467eb0-222b-4486-aa5c-07bba04b741c\") " pod="openstack/dnsmasq-dns-6cbbbb95c9-bng2s" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.422583 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a467eb0-222b-4486-aa5c-07bba04b741c-config\") pod \"dnsmasq-dns-6cbbbb95c9-bng2s\" (UID: \"2a467eb0-222b-4486-aa5c-07bba04b741c\") " pod="openstack/dnsmasq-dns-6cbbbb95c9-bng2s" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.422648 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2a467eb0-222b-4486-aa5c-07bba04b741c-ovsdbserver-nb\") pod \"dnsmasq-dns-6cbbbb95c9-bng2s\" (UID: \"2a467eb0-222b-4486-aa5c-07bba04b741c\") " pod="openstack/dnsmasq-dns-6cbbbb95c9-bng2s" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.422693 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwfwh\" (UniqueName: \"kubernetes.io/projected/2a467eb0-222b-4486-aa5c-07bba04b741c-kube-api-access-vwfwh\") pod \"dnsmasq-dns-6cbbbb95c9-bng2s\" (UID: \"2a467eb0-222b-4486-aa5c-07bba04b741c\") " pod="openstack/dnsmasq-dns-6cbbbb95c9-bng2s" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.422933 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a467eb0-222b-4486-aa5c-07bba04b741c-dns-svc\") pod \"dnsmasq-dns-6cbbbb95c9-bng2s\" (UID: \"2a467eb0-222b-4486-aa5c-07bba04b741c\") " pod="openstack/dnsmasq-dns-6cbbbb95c9-bng2s" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.423653 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2a467eb0-222b-4486-aa5c-07bba04b741c-ovsdbserver-sb\") pod \"dnsmasq-dns-6cbbbb95c9-bng2s\" (UID: \"2a467eb0-222b-4486-aa5c-07bba04b741c\") " pod="openstack/dnsmasq-dns-6cbbbb95c9-bng2s" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.423914 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a467eb0-222b-4486-aa5c-07bba04b741c-dns-svc\") pod \"dnsmasq-dns-6cbbbb95c9-bng2s\" (UID: \"2a467eb0-222b-4486-aa5c-07bba04b741c\") " pod="openstack/dnsmasq-dns-6cbbbb95c9-bng2s" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.424317 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2a467eb0-222b-4486-aa5c-07bba04b741c-ovsdbserver-nb\") pod \"dnsmasq-dns-6cbbbb95c9-bng2s\" (UID: \"2a467eb0-222b-4486-aa5c-07bba04b741c\") " pod="openstack/dnsmasq-dns-6cbbbb95c9-bng2s" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.424792 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a467eb0-222b-4486-aa5c-07bba04b741c-config\") pod \"dnsmasq-dns-6cbbbb95c9-bng2s\" (UID: \"2a467eb0-222b-4486-aa5c-07bba04b741c\") " pod="openstack/dnsmasq-dns-6cbbbb95c9-bng2s" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.442604 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwfwh\" (UniqueName: \"kubernetes.io/projected/2a467eb0-222b-4486-aa5c-07bba04b741c-kube-api-access-vwfwh\") pod \"dnsmasq-dns-6cbbbb95c9-bng2s\" (UID: \"2a467eb0-222b-4486-aa5c-07bba04b741c\") " pod="openstack/dnsmasq-dns-6cbbbb95c9-bng2s" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.475892 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-f7vrf" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.502780 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cbbbb95c9-bng2s" Nov 25 10:54:46 crc kubenswrapper[4776]: I1125 10:54:46.943562 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-f7vrf"] Nov 25 10:54:46 crc kubenswrapper[4776]: W1125 10:54:46.949348 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a73532e_2b0f_4a54_bbb5_d74714dfc6c5.slice/crio-844d8e835cc519540f43bb538b96ce9c11cb9c95a077ee4987c47795709ed7e2 WatchSource:0}: Error finding container 844d8e835cc519540f43bb538b96ce9c11cb9c95a077ee4987c47795709ed7e2: Status 404 returned error can't find the container with id 844d8e835cc519540f43bb538b96ce9c11cb9c95a077ee4987c47795709ed7e2 Nov 25 10:54:47 crc kubenswrapper[4776]: I1125 10:54:47.030918 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6cbbbb95c9-bng2s"] Nov 25 10:54:47 crc kubenswrapper[4776]: W1125 10:54:47.036761 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a467eb0_222b_4486_aa5c_07bba04b741c.slice/crio-fe775dc952678afd32fced32c17c6e0e297fc13fba3e27ba8f85f5b966d63e87 WatchSource:0}: Error finding container fe775dc952678afd32fced32c17c6e0e297fc13fba3e27ba8f85f5b966d63e87: Status 404 returned error can't find the container with id fe775dc952678afd32fced32c17c6e0e297fc13fba3e27ba8f85f5b966d63e87 Nov 25 10:54:47 crc kubenswrapper[4776]: I1125 10:54:47.928186 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-f7vrf" event={"ID":"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5","Type":"ContainerStarted","Data":"1bdc2fc4eeba1050b8a93b69e842c63b8967de131f1dfc0bf3dd2b51d11017eb"} Nov 25 10:54:47 crc kubenswrapper[4776]: I1125 10:54:47.929337 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-f7vrf" event={"ID":"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5","Type":"ContainerStarted","Data":"844d8e835cc519540f43bb538b96ce9c11cb9c95a077ee4987c47795709ed7e2"} Nov 25 10:54:47 crc kubenswrapper[4776]: I1125 10:54:47.932410 4776 generic.go:334] "Generic (PLEG): container finished" podID="2a467eb0-222b-4486-aa5c-07bba04b741c" containerID="54efdc3d5b3058f9372933646be9167aed12e3667fd4356cd3a4a0518ac7ea4d" exitCode=0 Nov 25 10:54:47 crc kubenswrapper[4776]: I1125 10:54:47.932487 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cbbbb95c9-bng2s" event={"ID":"2a467eb0-222b-4486-aa5c-07bba04b741c","Type":"ContainerDied","Data":"54efdc3d5b3058f9372933646be9167aed12e3667fd4356cd3a4a0518ac7ea4d"} Nov 25 10:54:47 crc kubenswrapper[4776]: I1125 10:54:47.932529 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cbbbb95c9-bng2s" event={"ID":"2a467eb0-222b-4486-aa5c-07bba04b741c","Type":"ContainerStarted","Data":"fe775dc952678afd32fced32c17c6e0e297fc13fba3e27ba8f85f5b966d63e87"} Nov 25 10:54:47 crc kubenswrapper[4776]: I1125 10:54:47.999700 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-f7vrf" podStartSLOduration=1.9996797769999999 podStartE2EDuration="1.999679777s" podCreationTimestamp="2025-11-25 10:54:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:54:47.96677043 +0000 UTC m=+5433.007829993" watchObservedRunningTime="2025-11-25 10:54:47.999679777 +0000 UTC m=+5433.040739340" Nov 25 10:54:48 crc kubenswrapper[4776]: I1125 10:54:48.948248 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cbbbb95c9-bng2s" event={"ID":"2a467eb0-222b-4486-aa5c-07bba04b741c","Type":"ContainerStarted","Data":"db9a42be803e0a46a6d076bb1f28a7b9fd5e0fc3ceb1a262457f73adffb30089"} Nov 25 10:54:48 crc kubenswrapper[4776]: I1125 10:54:48.948753 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6cbbbb95c9-bng2s" Nov 25 10:54:48 crc kubenswrapper[4776]: I1125 10:54:48.978029 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6cbbbb95c9-bng2s" podStartSLOduration=2.978006002 podStartE2EDuration="2.978006002s" podCreationTimestamp="2025-11-25 10:54:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:54:48.970762771 +0000 UTC m=+5434.011822324" watchObservedRunningTime="2025-11-25 10:54:48.978006002 +0000 UTC m=+5434.019065555" Nov 25 10:54:50 crc kubenswrapper[4776]: I1125 10:54:50.967304 4776 generic.go:334] "Generic (PLEG): container finished" podID="3a73532e-2b0f-4a54-bbb5-d74714dfc6c5" containerID="1bdc2fc4eeba1050b8a93b69e842c63b8967de131f1dfc0bf3dd2b51d11017eb" exitCode=0 Nov 25 10:54:50 crc kubenswrapper[4776]: I1125 10:54:50.967446 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-f7vrf" event={"ID":"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5","Type":"ContainerDied","Data":"1bdc2fc4eeba1050b8a93b69e842c63b8967de131f1dfc0bf3dd2b51d11017eb"} Nov 25 10:54:52 crc kubenswrapper[4776]: I1125 10:54:52.338360 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-f7vrf" Nov 25 10:54:52 crc kubenswrapper[4776]: I1125 10:54:52.444035 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-credential-keys\") pod \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\" (UID: \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\") " Nov 25 10:54:52 crc kubenswrapper[4776]: I1125 10:54:52.444109 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-config-data\") pod \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\" (UID: \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\") " Nov 25 10:54:52 crc kubenswrapper[4776]: I1125 10:54:52.444196 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-scripts\") pod \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\" (UID: \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\") " Nov 25 10:54:52 crc kubenswrapper[4776]: I1125 10:54:52.444260 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-combined-ca-bundle\") pod \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\" (UID: \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\") " Nov 25 10:54:52 crc kubenswrapper[4776]: I1125 10:54:52.444305 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2djg\" (UniqueName: \"kubernetes.io/projected/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-kube-api-access-j2djg\") pod \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\" (UID: \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\") " Nov 25 10:54:52 crc kubenswrapper[4776]: I1125 10:54:52.444395 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-fernet-keys\") pod \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\" (UID: \"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5\") " Nov 25 10:54:52 crc kubenswrapper[4776]: I1125 10:54:52.450527 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "3a73532e-2b0f-4a54-bbb5-d74714dfc6c5" (UID: "3a73532e-2b0f-4a54-bbb5-d74714dfc6c5"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:52 crc kubenswrapper[4776]: I1125 10:54:52.459345 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-scripts" (OuterVolumeSpecName: "scripts") pod "3a73532e-2b0f-4a54-bbb5-d74714dfc6c5" (UID: "3a73532e-2b0f-4a54-bbb5-d74714dfc6c5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:52 crc kubenswrapper[4776]: I1125 10:54:52.465859 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "3a73532e-2b0f-4a54-bbb5-d74714dfc6c5" (UID: "3a73532e-2b0f-4a54-bbb5-d74714dfc6c5"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:52 crc kubenswrapper[4776]: I1125 10:54:52.466234 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-kube-api-access-j2djg" (OuterVolumeSpecName: "kube-api-access-j2djg") pod "3a73532e-2b0f-4a54-bbb5-d74714dfc6c5" (UID: "3a73532e-2b0f-4a54-bbb5-d74714dfc6c5"). InnerVolumeSpecName "kube-api-access-j2djg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:54:52 crc kubenswrapper[4776]: I1125 10:54:52.480458 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-config-data" (OuterVolumeSpecName: "config-data") pod "3a73532e-2b0f-4a54-bbb5-d74714dfc6c5" (UID: "3a73532e-2b0f-4a54-bbb5-d74714dfc6c5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:52 crc kubenswrapper[4776]: I1125 10:54:52.495676 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a73532e-2b0f-4a54-bbb5-d74714dfc6c5" (UID: "3a73532e-2b0f-4a54-bbb5-d74714dfc6c5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:52 crc kubenswrapper[4776]: I1125 10:54:52.549664 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:52 crc kubenswrapper[4776]: I1125 10:54:52.549741 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:52 crc kubenswrapper[4776]: I1125 10:54:52.549761 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2djg\" (UniqueName: \"kubernetes.io/projected/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-kube-api-access-j2djg\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:52 crc kubenswrapper[4776]: I1125 10:54:52.549778 4776 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:52 crc kubenswrapper[4776]: I1125 10:54:52.549794 4776 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:52 crc kubenswrapper[4776]: I1125 10:54:52.549807 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:52 crc kubenswrapper[4776]: I1125 10:54:52.983653 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-f7vrf" event={"ID":"3a73532e-2b0f-4a54-bbb5-d74714dfc6c5","Type":"ContainerDied","Data":"844d8e835cc519540f43bb538b96ce9c11cb9c95a077ee4987c47795709ed7e2"} Nov 25 10:54:52 crc kubenswrapper[4776]: I1125 10:54:52.984038 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="844d8e835cc519540f43bb538b96ce9c11cb9c95a077ee4987c47795709ed7e2" Nov 25 10:54:52 crc kubenswrapper[4776]: I1125 10:54:52.983802 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-f7vrf" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.157926 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-f7vrf"] Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.164894 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-f7vrf"] Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.271329 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-wknmm"] Nov 25 10:54:53 crc kubenswrapper[4776]: E1125 10:54:53.271718 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a73532e-2b0f-4a54-bbb5-d74714dfc6c5" containerName="keystone-bootstrap" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.271740 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a73532e-2b0f-4a54-bbb5-d74714dfc6c5" containerName="keystone-bootstrap" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.271997 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a73532e-2b0f-4a54-bbb5-d74714dfc6c5" containerName="keystone-bootstrap" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.272625 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wknmm" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.274578 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.274991 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-lnvfd" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.275119 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.275661 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.276264 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.281508 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-wknmm"] Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.362347 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-scripts\") pod \"keystone-bootstrap-wknmm\" (UID: \"6a8513c8-6195-4ffe-8765-d3554db89257\") " pod="openstack/keystone-bootstrap-wknmm" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.362524 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86qkj\" (UniqueName: \"kubernetes.io/projected/6a8513c8-6195-4ffe-8765-d3554db89257-kube-api-access-86qkj\") pod \"keystone-bootstrap-wknmm\" (UID: \"6a8513c8-6195-4ffe-8765-d3554db89257\") " pod="openstack/keystone-bootstrap-wknmm" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.362652 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-fernet-keys\") pod \"keystone-bootstrap-wknmm\" (UID: \"6a8513c8-6195-4ffe-8765-d3554db89257\") " pod="openstack/keystone-bootstrap-wknmm" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.362748 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-config-data\") pod \"keystone-bootstrap-wknmm\" (UID: \"6a8513c8-6195-4ffe-8765-d3554db89257\") " pod="openstack/keystone-bootstrap-wknmm" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.362910 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-combined-ca-bundle\") pod \"keystone-bootstrap-wknmm\" (UID: \"6a8513c8-6195-4ffe-8765-d3554db89257\") " pod="openstack/keystone-bootstrap-wknmm" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.363332 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-credential-keys\") pod \"keystone-bootstrap-wknmm\" (UID: \"6a8513c8-6195-4ffe-8765-d3554db89257\") " pod="openstack/keystone-bootstrap-wknmm" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.464669 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-config-data\") pod \"keystone-bootstrap-wknmm\" (UID: \"6a8513c8-6195-4ffe-8765-d3554db89257\") " pod="openstack/keystone-bootstrap-wknmm" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.464949 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-combined-ca-bundle\") pod \"keystone-bootstrap-wknmm\" (UID: \"6a8513c8-6195-4ffe-8765-d3554db89257\") " pod="openstack/keystone-bootstrap-wknmm" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.464999 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-credential-keys\") pod \"keystone-bootstrap-wknmm\" (UID: \"6a8513c8-6195-4ffe-8765-d3554db89257\") " pod="openstack/keystone-bootstrap-wknmm" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.465042 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-scripts\") pod \"keystone-bootstrap-wknmm\" (UID: \"6a8513c8-6195-4ffe-8765-d3554db89257\") " pod="openstack/keystone-bootstrap-wknmm" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.465091 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86qkj\" (UniqueName: \"kubernetes.io/projected/6a8513c8-6195-4ffe-8765-d3554db89257-kube-api-access-86qkj\") pod \"keystone-bootstrap-wknmm\" (UID: \"6a8513c8-6195-4ffe-8765-d3554db89257\") " pod="openstack/keystone-bootstrap-wknmm" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.465123 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-fernet-keys\") pod \"keystone-bootstrap-wknmm\" (UID: \"6a8513c8-6195-4ffe-8765-d3554db89257\") " pod="openstack/keystone-bootstrap-wknmm" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.470370 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-credential-keys\") pod \"keystone-bootstrap-wknmm\" (UID: \"6a8513c8-6195-4ffe-8765-d3554db89257\") " pod="openstack/keystone-bootstrap-wknmm" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.470807 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-fernet-keys\") pod \"keystone-bootstrap-wknmm\" (UID: \"6a8513c8-6195-4ffe-8765-d3554db89257\") " pod="openstack/keystone-bootstrap-wknmm" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.471402 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-scripts\") pod \"keystone-bootstrap-wknmm\" (UID: \"6a8513c8-6195-4ffe-8765-d3554db89257\") " pod="openstack/keystone-bootstrap-wknmm" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.473184 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-config-data\") pod \"keystone-bootstrap-wknmm\" (UID: \"6a8513c8-6195-4ffe-8765-d3554db89257\") " pod="openstack/keystone-bootstrap-wknmm" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.473393 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-combined-ca-bundle\") pod \"keystone-bootstrap-wknmm\" (UID: \"6a8513c8-6195-4ffe-8765-d3554db89257\") " pod="openstack/keystone-bootstrap-wknmm" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.495886 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86qkj\" (UniqueName: \"kubernetes.io/projected/6a8513c8-6195-4ffe-8765-d3554db89257-kube-api-access-86qkj\") pod \"keystone-bootstrap-wknmm\" (UID: \"6a8513c8-6195-4ffe-8765-d3554db89257\") " pod="openstack/keystone-bootstrap-wknmm" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.599547 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wknmm" Nov 25 10:54:53 crc kubenswrapper[4776]: I1125 10:54:53.692859 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a73532e-2b0f-4a54-bbb5-d74714dfc6c5" path="/var/lib/kubelet/pods/3a73532e-2b0f-4a54-bbb5-d74714dfc6c5/volumes" Nov 25 10:54:54 crc kubenswrapper[4776]: I1125 10:54:54.164131 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-wknmm"] Nov 25 10:54:54 crc kubenswrapper[4776]: W1125 10:54:54.170126 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a8513c8_6195_4ffe_8765_d3554db89257.slice/crio-c8ed7c0ecb20b10fbabeca745621052ffe778cc4a28e9b69c84a9e1b833ee6c9 WatchSource:0}: Error finding container c8ed7c0ecb20b10fbabeca745621052ffe778cc4a28e9b69c84a9e1b833ee6c9: Status 404 returned error can't find the container with id c8ed7c0ecb20b10fbabeca745621052ffe778cc4a28e9b69c84a9e1b833ee6c9 Nov 25 10:54:55 crc kubenswrapper[4776]: I1125 10:54:55.001000 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wknmm" event={"ID":"6a8513c8-6195-4ffe-8765-d3554db89257","Type":"ContainerStarted","Data":"b281721f6ad7427a1e3f77d7eeaedc45a3a0ae1b688aa850dc6f343ee2a59668"} Nov 25 10:54:55 crc kubenswrapper[4776]: I1125 10:54:55.001433 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wknmm" event={"ID":"6a8513c8-6195-4ffe-8765-d3554db89257","Type":"ContainerStarted","Data":"c8ed7c0ecb20b10fbabeca745621052ffe778cc4a28e9b69c84a9e1b833ee6c9"} Nov 25 10:54:55 crc kubenswrapper[4776]: I1125 10:54:55.023711 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-wknmm" podStartSLOduration=2.023695124 podStartE2EDuration="2.023695124s" podCreationTimestamp="2025-11-25 10:54:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:54:55.018032682 +0000 UTC m=+5440.059092255" watchObservedRunningTime="2025-11-25 10:54:55.023695124 +0000 UTC m=+5440.064754697" Nov 25 10:54:56 crc kubenswrapper[4776]: I1125 10:54:56.505039 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6cbbbb95c9-bng2s" Nov 25 10:54:56 crc kubenswrapper[4776]: I1125 10:54:56.574026 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74767bbc9-lxk47"] Nov 25 10:54:56 crc kubenswrapper[4776]: I1125 10:54:56.574425 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74767bbc9-lxk47" podUID="32010807-df10-4d9e-b62c-b7e14706210b" containerName="dnsmasq-dns" containerID="cri-o://ba422d27a9918f591c7ba0af0fe8f866b7b0b1bb352015072379ae794b2b2e4a" gracePeriod=10 Nov 25 10:54:57 crc kubenswrapper[4776]: I1125 10:54:57.023354 4776 generic.go:334] "Generic (PLEG): container finished" podID="32010807-df10-4d9e-b62c-b7e14706210b" containerID="ba422d27a9918f591c7ba0af0fe8f866b7b0b1bb352015072379ae794b2b2e4a" exitCode=0 Nov 25 10:54:57 crc kubenswrapper[4776]: I1125 10:54:57.023392 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74767bbc9-lxk47" event={"ID":"32010807-df10-4d9e-b62c-b7e14706210b","Type":"ContainerDied","Data":"ba422d27a9918f591c7ba0af0fe8f866b7b0b1bb352015072379ae794b2b2e4a"} Nov 25 10:54:57 crc kubenswrapper[4776]: I1125 10:54:57.023416 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74767bbc9-lxk47" event={"ID":"32010807-df10-4d9e-b62c-b7e14706210b","Type":"ContainerDied","Data":"94d236bbcb917ab79734d0b1640b13a58c95b364ee7fb894d825545a78b9eaa3"} Nov 25 10:54:57 crc kubenswrapper[4776]: I1125 10:54:57.023426 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94d236bbcb917ab79734d0b1640b13a58c95b364ee7fb894d825545a78b9eaa3" Nov 25 10:54:57 crc kubenswrapper[4776]: I1125 10:54:57.056593 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74767bbc9-lxk47" Nov 25 10:54:57 crc kubenswrapper[4776]: I1125 10:54:57.126974 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32010807-df10-4d9e-b62c-b7e14706210b-dns-svc\") pod \"32010807-df10-4d9e-b62c-b7e14706210b\" (UID: \"32010807-df10-4d9e-b62c-b7e14706210b\") " Nov 25 10:54:57 crc kubenswrapper[4776]: I1125 10:54:57.127027 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/32010807-df10-4d9e-b62c-b7e14706210b-ovsdbserver-sb\") pod \"32010807-df10-4d9e-b62c-b7e14706210b\" (UID: \"32010807-df10-4d9e-b62c-b7e14706210b\") " Nov 25 10:54:57 crc kubenswrapper[4776]: I1125 10:54:57.127196 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbnsk\" (UniqueName: \"kubernetes.io/projected/32010807-df10-4d9e-b62c-b7e14706210b-kube-api-access-mbnsk\") pod \"32010807-df10-4d9e-b62c-b7e14706210b\" (UID: \"32010807-df10-4d9e-b62c-b7e14706210b\") " Nov 25 10:54:57 crc kubenswrapper[4776]: I1125 10:54:57.127271 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/32010807-df10-4d9e-b62c-b7e14706210b-ovsdbserver-nb\") pod \"32010807-df10-4d9e-b62c-b7e14706210b\" (UID: \"32010807-df10-4d9e-b62c-b7e14706210b\") " Nov 25 10:54:57 crc kubenswrapper[4776]: I1125 10:54:57.127319 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32010807-df10-4d9e-b62c-b7e14706210b-config\") pod \"32010807-df10-4d9e-b62c-b7e14706210b\" (UID: \"32010807-df10-4d9e-b62c-b7e14706210b\") " Nov 25 10:54:57 crc kubenswrapper[4776]: I1125 10:54:57.141924 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32010807-df10-4d9e-b62c-b7e14706210b-kube-api-access-mbnsk" (OuterVolumeSpecName: "kube-api-access-mbnsk") pod "32010807-df10-4d9e-b62c-b7e14706210b" (UID: "32010807-df10-4d9e-b62c-b7e14706210b"). InnerVolumeSpecName "kube-api-access-mbnsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:54:57 crc kubenswrapper[4776]: I1125 10:54:57.169465 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32010807-df10-4d9e-b62c-b7e14706210b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "32010807-df10-4d9e-b62c-b7e14706210b" (UID: "32010807-df10-4d9e-b62c-b7e14706210b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:57 crc kubenswrapper[4776]: I1125 10:54:57.172930 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32010807-df10-4d9e-b62c-b7e14706210b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "32010807-df10-4d9e-b62c-b7e14706210b" (UID: "32010807-df10-4d9e-b62c-b7e14706210b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:57 crc kubenswrapper[4776]: I1125 10:54:57.173874 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32010807-df10-4d9e-b62c-b7e14706210b-config" (OuterVolumeSpecName: "config") pod "32010807-df10-4d9e-b62c-b7e14706210b" (UID: "32010807-df10-4d9e-b62c-b7e14706210b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:57 crc kubenswrapper[4776]: I1125 10:54:57.188357 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32010807-df10-4d9e-b62c-b7e14706210b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "32010807-df10-4d9e-b62c-b7e14706210b" (UID: "32010807-df10-4d9e-b62c-b7e14706210b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:54:57 crc kubenswrapper[4776]: I1125 10:54:57.230528 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbnsk\" (UniqueName: \"kubernetes.io/projected/32010807-df10-4d9e-b62c-b7e14706210b-kube-api-access-mbnsk\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:57 crc kubenswrapper[4776]: I1125 10:54:57.230681 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/32010807-df10-4d9e-b62c-b7e14706210b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:57 crc kubenswrapper[4776]: I1125 10:54:57.230779 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32010807-df10-4d9e-b62c-b7e14706210b-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:57 crc kubenswrapper[4776]: I1125 10:54:57.230870 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/32010807-df10-4d9e-b62c-b7e14706210b-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:57 crc kubenswrapper[4776]: I1125 10:54:57.231155 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/32010807-df10-4d9e-b62c-b7e14706210b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:58 crc kubenswrapper[4776]: I1125 10:54:58.033999 4776 generic.go:334] "Generic (PLEG): container finished" podID="6a8513c8-6195-4ffe-8765-d3554db89257" containerID="b281721f6ad7427a1e3f77d7eeaedc45a3a0ae1b688aa850dc6f343ee2a59668" exitCode=0 Nov 25 10:54:58 crc kubenswrapper[4776]: I1125 10:54:58.034108 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wknmm" event={"ID":"6a8513c8-6195-4ffe-8765-d3554db89257","Type":"ContainerDied","Data":"b281721f6ad7427a1e3f77d7eeaedc45a3a0ae1b688aa850dc6f343ee2a59668"} Nov 25 10:54:58 crc kubenswrapper[4776]: I1125 10:54:58.034135 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74767bbc9-lxk47" Nov 25 10:54:58 crc kubenswrapper[4776]: I1125 10:54:58.084901 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74767bbc9-lxk47"] Nov 25 10:54:58 crc kubenswrapper[4776]: I1125 10:54:58.092429 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74767bbc9-lxk47"] Nov 25 10:54:58 crc kubenswrapper[4776]: I1125 10:54:58.663370 4776 scope.go:117] "RemoveContainer" containerID="f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40" Nov 25 10:54:58 crc kubenswrapper[4776]: E1125 10:54:58.663822 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:54:59 crc kubenswrapper[4776]: I1125 10:54:59.368184 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wknmm" Nov 25 10:54:59 crc kubenswrapper[4776]: I1125 10:54:59.473195 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-config-data\") pod \"6a8513c8-6195-4ffe-8765-d3554db89257\" (UID: \"6a8513c8-6195-4ffe-8765-d3554db89257\") " Nov 25 10:54:59 crc kubenswrapper[4776]: I1125 10:54:59.473318 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-combined-ca-bundle\") pod \"6a8513c8-6195-4ffe-8765-d3554db89257\" (UID: \"6a8513c8-6195-4ffe-8765-d3554db89257\") " Nov 25 10:54:59 crc kubenswrapper[4776]: I1125 10:54:59.473378 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86qkj\" (UniqueName: \"kubernetes.io/projected/6a8513c8-6195-4ffe-8765-d3554db89257-kube-api-access-86qkj\") pod \"6a8513c8-6195-4ffe-8765-d3554db89257\" (UID: \"6a8513c8-6195-4ffe-8765-d3554db89257\") " Nov 25 10:54:59 crc kubenswrapper[4776]: I1125 10:54:59.473430 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-fernet-keys\") pod \"6a8513c8-6195-4ffe-8765-d3554db89257\" (UID: \"6a8513c8-6195-4ffe-8765-d3554db89257\") " Nov 25 10:54:59 crc kubenswrapper[4776]: I1125 10:54:59.473529 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-credential-keys\") pod \"6a8513c8-6195-4ffe-8765-d3554db89257\" (UID: \"6a8513c8-6195-4ffe-8765-d3554db89257\") " Nov 25 10:54:59 crc kubenswrapper[4776]: I1125 10:54:59.473674 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-scripts\") pod \"6a8513c8-6195-4ffe-8765-d3554db89257\" (UID: \"6a8513c8-6195-4ffe-8765-d3554db89257\") " Nov 25 10:54:59 crc kubenswrapper[4776]: I1125 10:54:59.478514 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "6a8513c8-6195-4ffe-8765-d3554db89257" (UID: "6a8513c8-6195-4ffe-8765-d3554db89257"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:59 crc kubenswrapper[4776]: I1125 10:54:59.478567 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6a8513c8-6195-4ffe-8765-d3554db89257" (UID: "6a8513c8-6195-4ffe-8765-d3554db89257"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:59 crc kubenswrapper[4776]: I1125 10:54:59.478987 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-scripts" (OuterVolumeSpecName: "scripts") pod "6a8513c8-6195-4ffe-8765-d3554db89257" (UID: "6a8513c8-6195-4ffe-8765-d3554db89257"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:59 crc kubenswrapper[4776]: I1125 10:54:59.482941 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a8513c8-6195-4ffe-8765-d3554db89257-kube-api-access-86qkj" (OuterVolumeSpecName: "kube-api-access-86qkj") pod "6a8513c8-6195-4ffe-8765-d3554db89257" (UID: "6a8513c8-6195-4ffe-8765-d3554db89257"). InnerVolumeSpecName "kube-api-access-86qkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:54:59 crc kubenswrapper[4776]: I1125 10:54:59.497032 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a8513c8-6195-4ffe-8765-d3554db89257" (UID: "6a8513c8-6195-4ffe-8765-d3554db89257"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:59 crc kubenswrapper[4776]: I1125 10:54:59.516770 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-config-data" (OuterVolumeSpecName: "config-data") pod "6a8513c8-6195-4ffe-8765-d3554db89257" (UID: "6a8513c8-6195-4ffe-8765-d3554db89257"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:54:59 crc kubenswrapper[4776]: I1125 10:54:59.575839 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:59 crc kubenswrapper[4776]: I1125 10:54:59.575879 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86qkj\" (UniqueName: \"kubernetes.io/projected/6a8513c8-6195-4ffe-8765-d3554db89257-kube-api-access-86qkj\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:59 crc kubenswrapper[4776]: I1125 10:54:59.575895 4776 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:59 crc kubenswrapper[4776]: I1125 10:54:59.575909 4776 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:59 crc kubenswrapper[4776]: I1125 10:54:59.575920 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:59 crc kubenswrapper[4776]: I1125 10:54:59.575932 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a8513c8-6195-4ffe-8765-d3554db89257-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:54:59 crc kubenswrapper[4776]: I1125 10:54:59.672824 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32010807-df10-4d9e-b62c-b7e14706210b" path="/var/lib/kubelet/pods/32010807-df10-4d9e-b62c-b7e14706210b/volumes" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.052412 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wknmm" event={"ID":"6a8513c8-6195-4ffe-8765-d3554db89257","Type":"ContainerDied","Data":"c8ed7c0ecb20b10fbabeca745621052ffe778cc4a28e9b69c84a9e1b833ee6c9"} Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.052989 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8ed7c0ecb20b10fbabeca745621052ffe778cc4a28e9b69c84a9e1b833ee6c9" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.052464 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wknmm" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.235318 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-c5d66c966-5c6vj"] Nov 25 10:55:00 crc kubenswrapper[4776]: E1125 10:55:00.236050 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32010807-df10-4d9e-b62c-b7e14706210b" containerName="dnsmasq-dns" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.236119 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="32010807-df10-4d9e-b62c-b7e14706210b" containerName="dnsmasq-dns" Nov 25 10:55:00 crc kubenswrapper[4776]: E1125 10:55:00.236146 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a8513c8-6195-4ffe-8765-d3554db89257" containerName="keystone-bootstrap" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.236179 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a8513c8-6195-4ffe-8765-d3554db89257" containerName="keystone-bootstrap" Nov 25 10:55:00 crc kubenswrapper[4776]: E1125 10:55:00.236206 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32010807-df10-4d9e-b62c-b7e14706210b" containerName="init" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.236215 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="32010807-df10-4d9e-b62c-b7e14706210b" containerName="init" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.236657 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a8513c8-6195-4ffe-8765-d3554db89257" containerName="keystone-bootstrap" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.236681 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="32010807-df10-4d9e-b62c-b7e14706210b" containerName="dnsmasq-dns" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.239376 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c5d66c966-5c6vj" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.240827 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-c5d66c966-5c6vj"] Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.243053 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-lnvfd" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.243436 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.243979 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.244355 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.245580 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.248758 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.391751 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/974cf029-f130-4748-81c3-1dca65481a17-internal-tls-certs\") pod \"keystone-c5d66c966-5c6vj\" (UID: \"974cf029-f130-4748-81c3-1dca65481a17\") " pod="openstack/keystone-c5d66c966-5c6vj" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.391816 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f78x2\" (UniqueName: \"kubernetes.io/projected/974cf029-f130-4748-81c3-1dca65481a17-kube-api-access-f78x2\") pod \"keystone-c5d66c966-5c6vj\" (UID: \"974cf029-f130-4748-81c3-1dca65481a17\") " pod="openstack/keystone-c5d66c966-5c6vj" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.391884 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/974cf029-f130-4748-81c3-1dca65481a17-config-data\") pod \"keystone-c5d66c966-5c6vj\" (UID: \"974cf029-f130-4748-81c3-1dca65481a17\") " pod="openstack/keystone-c5d66c966-5c6vj" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.391915 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/974cf029-f130-4748-81c3-1dca65481a17-public-tls-certs\") pod \"keystone-c5d66c966-5c6vj\" (UID: \"974cf029-f130-4748-81c3-1dca65481a17\") " pod="openstack/keystone-c5d66c966-5c6vj" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.391979 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/974cf029-f130-4748-81c3-1dca65481a17-fernet-keys\") pod \"keystone-c5d66c966-5c6vj\" (UID: \"974cf029-f130-4748-81c3-1dca65481a17\") " pod="openstack/keystone-c5d66c966-5c6vj" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.392058 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/974cf029-f130-4748-81c3-1dca65481a17-combined-ca-bundle\") pod \"keystone-c5d66c966-5c6vj\" (UID: \"974cf029-f130-4748-81c3-1dca65481a17\") " pod="openstack/keystone-c5d66c966-5c6vj" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.392125 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/974cf029-f130-4748-81c3-1dca65481a17-scripts\") pod \"keystone-c5d66c966-5c6vj\" (UID: \"974cf029-f130-4748-81c3-1dca65481a17\") " pod="openstack/keystone-c5d66c966-5c6vj" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.392159 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/974cf029-f130-4748-81c3-1dca65481a17-credential-keys\") pod \"keystone-c5d66c966-5c6vj\" (UID: \"974cf029-f130-4748-81c3-1dca65481a17\") " pod="openstack/keystone-c5d66c966-5c6vj" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.493483 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f78x2\" (UniqueName: \"kubernetes.io/projected/974cf029-f130-4748-81c3-1dca65481a17-kube-api-access-f78x2\") pod \"keystone-c5d66c966-5c6vj\" (UID: \"974cf029-f130-4748-81c3-1dca65481a17\") " pod="openstack/keystone-c5d66c966-5c6vj" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.493559 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/974cf029-f130-4748-81c3-1dca65481a17-config-data\") pod \"keystone-c5d66c966-5c6vj\" (UID: \"974cf029-f130-4748-81c3-1dca65481a17\") " pod="openstack/keystone-c5d66c966-5c6vj" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.493585 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/974cf029-f130-4748-81c3-1dca65481a17-public-tls-certs\") pod \"keystone-c5d66c966-5c6vj\" (UID: \"974cf029-f130-4748-81c3-1dca65481a17\") " pod="openstack/keystone-c5d66c966-5c6vj" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.493652 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/974cf029-f130-4748-81c3-1dca65481a17-fernet-keys\") pod \"keystone-c5d66c966-5c6vj\" (UID: \"974cf029-f130-4748-81c3-1dca65481a17\") " pod="openstack/keystone-c5d66c966-5c6vj" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.494410 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/974cf029-f130-4748-81c3-1dca65481a17-combined-ca-bundle\") pod \"keystone-c5d66c966-5c6vj\" (UID: \"974cf029-f130-4748-81c3-1dca65481a17\") " pod="openstack/keystone-c5d66c966-5c6vj" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.494485 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/974cf029-f130-4748-81c3-1dca65481a17-scripts\") pod \"keystone-c5d66c966-5c6vj\" (UID: \"974cf029-f130-4748-81c3-1dca65481a17\") " pod="openstack/keystone-c5d66c966-5c6vj" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.494539 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/974cf029-f130-4748-81c3-1dca65481a17-credential-keys\") pod \"keystone-c5d66c966-5c6vj\" (UID: \"974cf029-f130-4748-81c3-1dca65481a17\") " pod="openstack/keystone-c5d66c966-5c6vj" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.494572 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/974cf029-f130-4748-81c3-1dca65481a17-internal-tls-certs\") pod \"keystone-c5d66c966-5c6vj\" (UID: \"974cf029-f130-4748-81c3-1dca65481a17\") " pod="openstack/keystone-c5d66c966-5c6vj" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.500008 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/974cf029-f130-4748-81c3-1dca65481a17-config-data\") pod \"keystone-c5d66c966-5c6vj\" (UID: \"974cf029-f130-4748-81c3-1dca65481a17\") " pod="openstack/keystone-c5d66c966-5c6vj" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.500417 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/974cf029-f130-4748-81c3-1dca65481a17-public-tls-certs\") pod \"keystone-c5d66c966-5c6vj\" (UID: \"974cf029-f130-4748-81c3-1dca65481a17\") " pod="openstack/keystone-c5d66c966-5c6vj" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.501056 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/974cf029-f130-4748-81c3-1dca65481a17-fernet-keys\") pod \"keystone-c5d66c966-5c6vj\" (UID: \"974cf029-f130-4748-81c3-1dca65481a17\") " pod="openstack/keystone-c5d66c966-5c6vj" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.501275 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/974cf029-f130-4748-81c3-1dca65481a17-combined-ca-bundle\") pod \"keystone-c5d66c966-5c6vj\" (UID: \"974cf029-f130-4748-81c3-1dca65481a17\") " pod="openstack/keystone-c5d66c966-5c6vj" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.502163 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/974cf029-f130-4748-81c3-1dca65481a17-credential-keys\") pod \"keystone-c5d66c966-5c6vj\" (UID: \"974cf029-f130-4748-81c3-1dca65481a17\") " pod="openstack/keystone-c5d66c966-5c6vj" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.502407 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/974cf029-f130-4748-81c3-1dca65481a17-internal-tls-certs\") pod \"keystone-c5d66c966-5c6vj\" (UID: \"974cf029-f130-4748-81c3-1dca65481a17\") " pod="openstack/keystone-c5d66c966-5c6vj" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.508437 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/974cf029-f130-4748-81c3-1dca65481a17-scripts\") pod \"keystone-c5d66c966-5c6vj\" (UID: \"974cf029-f130-4748-81c3-1dca65481a17\") " pod="openstack/keystone-c5d66c966-5c6vj" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.525832 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f78x2\" (UniqueName: \"kubernetes.io/projected/974cf029-f130-4748-81c3-1dca65481a17-kube-api-access-f78x2\") pod \"keystone-c5d66c966-5c6vj\" (UID: \"974cf029-f130-4748-81c3-1dca65481a17\") " pod="openstack/keystone-c5d66c966-5c6vj" Nov 25 10:55:00 crc kubenswrapper[4776]: I1125 10:55:00.577924 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c5d66c966-5c6vj" Nov 25 10:55:01 crc kubenswrapper[4776]: I1125 10:55:01.038228 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-c5d66c966-5c6vj"] Nov 25 10:55:01 crc kubenswrapper[4776]: I1125 10:55:01.060309 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c5d66c966-5c6vj" event={"ID":"974cf029-f130-4748-81c3-1dca65481a17","Type":"ContainerStarted","Data":"e7d6cabcebce840aad9e7d3eec5b0690af9ba50f2aecd61d96f114ef50276f12"} Nov 25 10:55:02 crc kubenswrapper[4776]: I1125 10:55:02.069193 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c5d66c966-5c6vj" event={"ID":"974cf029-f130-4748-81c3-1dca65481a17","Type":"ContainerStarted","Data":"d88710382e9a9f07cc779bef6ad4f6051ca5d554644c7c4b3f4d70cc4fb3a07b"} Nov 25 10:55:02 crc kubenswrapper[4776]: I1125 10:55:02.069707 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-c5d66c966-5c6vj" Nov 25 10:55:02 crc kubenswrapper[4776]: I1125 10:55:02.095609 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-c5d66c966-5c6vj" podStartSLOduration=2.095581743 podStartE2EDuration="2.095581743s" podCreationTimestamp="2025-11-25 10:55:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:55:02.090413803 +0000 UTC m=+5447.131473366" watchObservedRunningTime="2025-11-25 10:55:02.095581743 +0000 UTC m=+5447.136641296" Nov 25 10:55:09 crc kubenswrapper[4776]: I1125 10:55:09.662334 4776 scope.go:117] "RemoveContainer" containerID="f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40" Nov 25 10:55:09 crc kubenswrapper[4776]: E1125 10:55:09.667805 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:55:21 crc kubenswrapper[4776]: I1125 10:55:21.662454 4776 scope.go:117] "RemoveContainer" containerID="f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40" Nov 25 10:55:21 crc kubenswrapper[4776]: E1125 10:55:21.663197 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:55:31 crc kubenswrapper[4776]: I1125 10:55:31.912212 4776 scope.go:117] "RemoveContainer" containerID="6fd5d1dd1a727cade74987e883f83a6b7b3e17be568974860f7b634243ef2aaa" Nov 25 10:55:31 crc kubenswrapper[4776]: I1125 10:55:31.937833 4776 scope.go:117] "RemoveContainer" containerID="4c2513c60a844ab3a406a3d220ad7622cc04585becd8919cdac99ada9d6b44e0" Nov 25 10:55:31 crc kubenswrapper[4776]: I1125 10:55:31.994474 4776 scope.go:117] "RemoveContainer" containerID="f33011cd57579d5e58ba05f93cc43fc0a37b7edb4a30f34f09f4838c8957456c" Nov 25 10:55:32 crc kubenswrapper[4776]: I1125 10:55:32.043809 4776 scope.go:117] "RemoveContainer" containerID="3e6d9e176a10e478905aeab1fada0c8f3d49080f66a2255b66f0694bc0992e83" Nov 25 10:55:32 crc kubenswrapper[4776]: I1125 10:55:32.095551 4776 scope.go:117] "RemoveContainer" containerID="40264f8469d0bfd9c11717e506a93d399b948663dc75fd4618e5c4dc8388531a" Nov 25 10:55:32 crc kubenswrapper[4776]: I1125 10:55:32.109604 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-c5d66c966-5c6vj" Nov 25 10:55:32 crc kubenswrapper[4776]: I1125 10:55:32.136466 4776 scope.go:117] "RemoveContainer" containerID="e0643220926571c88aea4d4c87aafdd6d7a4657df047f08405a0c8b31c6a3f2f" Nov 25 10:55:33 crc kubenswrapper[4776]: I1125 10:55:33.662782 4776 scope.go:117] "RemoveContainer" containerID="f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40" Nov 25 10:55:33 crc kubenswrapper[4776]: E1125 10:55:33.663122 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:55:34 crc kubenswrapper[4776]: I1125 10:55:34.925998 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 25 10:55:34 crc kubenswrapper[4776]: I1125 10:55:34.927247 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 10:55:34 crc kubenswrapper[4776]: I1125 10:55:34.929696 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 25 10:55:34 crc kubenswrapper[4776]: I1125 10:55:34.929842 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 25 10:55:34 crc kubenswrapper[4776]: I1125 10:55:34.929888 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-pk4gt" Nov 25 10:55:34 crc kubenswrapper[4776]: I1125 10:55:34.940818 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 25 10:55:34 crc kubenswrapper[4776]: I1125 10:55:34.977109 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 25 10:55:34 crc kubenswrapper[4776]: E1125 10:55:34.977843 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle kube-api-access-btnp6 openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="afc941b5-127c-44d0-8269-ad572b31e1b2" Nov 25 10:55:34 crc kubenswrapper[4776]: I1125 10:55:34.985463 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 25 10:55:34 crc kubenswrapper[4776]: I1125 10:55:34.995043 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 25 10:55:34 crc kubenswrapper[4776]: I1125 10:55:34.996149 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.013477 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.050565 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afc941b5-127c-44d0-8269-ad572b31e1b2-combined-ca-bundle\") pod \"openstackclient\" (UID: \"afc941b5-127c-44d0-8269-ad572b31e1b2\") " pod="openstack/openstackclient" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.050700 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/afc941b5-127c-44d0-8269-ad572b31e1b2-openstack-config-secret\") pod \"openstackclient\" (UID: \"afc941b5-127c-44d0-8269-ad572b31e1b2\") " pod="openstack/openstackclient" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.050750 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/afc941b5-127c-44d0-8269-ad572b31e1b2-openstack-config\") pod \"openstackclient\" (UID: \"afc941b5-127c-44d0-8269-ad572b31e1b2\") " pod="openstack/openstackclient" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.050828 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btnp6\" (UniqueName: \"kubernetes.io/projected/afc941b5-127c-44d0-8269-ad572b31e1b2-kube-api-access-btnp6\") pod \"openstackclient\" (UID: \"afc941b5-127c-44d0-8269-ad572b31e1b2\") " pod="openstack/openstackclient" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.152048 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ee61762a-4aa0-4ae9-9107-fca99430ccfa-openstack-config-secret\") pod \"openstackclient\" (UID: \"ee61762a-4aa0-4ae9-9107-fca99430ccfa\") " pod="openstack/openstackclient" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.152132 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afc941b5-127c-44d0-8269-ad572b31e1b2-combined-ca-bundle\") pod \"openstackclient\" (UID: \"afc941b5-127c-44d0-8269-ad572b31e1b2\") " pod="openstack/openstackclient" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.152206 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/afc941b5-127c-44d0-8269-ad572b31e1b2-openstack-config-secret\") pod \"openstackclient\" (UID: \"afc941b5-127c-44d0-8269-ad572b31e1b2\") " pod="openstack/openstackclient" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.152240 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/afc941b5-127c-44d0-8269-ad572b31e1b2-openstack-config\") pod \"openstackclient\" (UID: \"afc941b5-127c-44d0-8269-ad572b31e1b2\") " pod="openstack/openstackclient" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.152274 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee61762a-4aa0-4ae9-9107-fca99430ccfa-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ee61762a-4aa0-4ae9-9107-fca99430ccfa\") " pod="openstack/openstackclient" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.152296 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ee61762a-4aa0-4ae9-9107-fca99430ccfa-openstack-config\") pod \"openstackclient\" (UID: \"ee61762a-4aa0-4ae9-9107-fca99430ccfa\") " pod="openstack/openstackclient" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.152332 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkz8f\" (UniqueName: \"kubernetes.io/projected/ee61762a-4aa0-4ae9-9107-fca99430ccfa-kube-api-access-lkz8f\") pod \"openstackclient\" (UID: \"ee61762a-4aa0-4ae9-9107-fca99430ccfa\") " pod="openstack/openstackclient" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.152351 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btnp6\" (UniqueName: \"kubernetes.io/projected/afc941b5-127c-44d0-8269-ad572b31e1b2-kube-api-access-btnp6\") pod \"openstackclient\" (UID: \"afc941b5-127c-44d0-8269-ad572b31e1b2\") " pod="openstack/openstackclient" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.153680 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/afc941b5-127c-44d0-8269-ad572b31e1b2-openstack-config\") pod \"openstackclient\" (UID: \"afc941b5-127c-44d0-8269-ad572b31e1b2\") " pod="openstack/openstackclient" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.159399 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/afc941b5-127c-44d0-8269-ad572b31e1b2-openstack-config-secret\") pod \"openstackclient\" (UID: \"afc941b5-127c-44d0-8269-ad572b31e1b2\") " pod="openstack/openstackclient" Nov 25 10:55:35 crc kubenswrapper[4776]: E1125 10:55:35.159722 4776 projected.go:194] Error preparing data for projected volume kube-api-access-btnp6 for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (afc941b5-127c-44d0-8269-ad572b31e1b2) does not match the UID in record. The object might have been deleted and then recreated Nov 25 10:55:35 crc kubenswrapper[4776]: E1125 10:55:35.159781 4776 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/afc941b5-127c-44d0-8269-ad572b31e1b2-kube-api-access-btnp6 podName:afc941b5-127c-44d0-8269-ad572b31e1b2 nodeName:}" failed. No retries permitted until 2025-11-25 10:55:35.659765502 +0000 UTC m=+5480.700825055 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-btnp6" (UniqueName: "kubernetes.io/projected/afc941b5-127c-44d0-8269-ad572b31e1b2-kube-api-access-btnp6") pod "openstackclient" (UID: "afc941b5-127c-44d0-8269-ad572b31e1b2") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (afc941b5-127c-44d0-8269-ad572b31e1b2) does not match the UID in record. The object might have been deleted and then recreated Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.162578 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afc941b5-127c-44d0-8269-ad572b31e1b2-combined-ca-bundle\") pod \"openstackclient\" (UID: \"afc941b5-127c-44d0-8269-ad572b31e1b2\") " pod="openstack/openstackclient" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.254260 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkz8f\" (UniqueName: \"kubernetes.io/projected/ee61762a-4aa0-4ae9-9107-fca99430ccfa-kube-api-access-lkz8f\") pod \"openstackclient\" (UID: \"ee61762a-4aa0-4ae9-9107-fca99430ccfa\") " pod="openstack/openstackclient" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.254355 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ee61762a-4aa0-4ae9-9107-fca99430ccfa-openstack-config-secret\") pod \"openstackclient\" (UID: \"ee61762a-4aa0-4ae9-9107-fca99430ccfa\") " pod="openstack/openstackclient" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.254486 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee61762a-4aa0-4ae9-9107-fca99430ccfa-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ee61762a-4aa0-4ae9-9107-fca99430ccfa\") " pod="openstack/openstackclient" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.254513 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ee61762a-4aa0-4ae9-9107-fca99430ccfa-openstack-config\") pod \"openstackclient\" (UID: \"ee61762a-4aa0-4ae9-9107-fca99430ccfa\") " pod="openstack/openstackclient" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.255616 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ee61762a-4aa0-4ae9-9107-fca99430ccfa-openstack-config\") pod \"openstackclient\" (UID: \"ee61762a-4aa0-4ae9-9107-fca99430ccfa\") " pod="openstack/openstackclient" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.258255 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee61762a-4aa0-4ae9-9107-fca99430ccfa-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ee61762a-4aa0-4ae9-9107-fca99430ccfa\") " pod="openstack/openstackclient" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.260035 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ee61762a-4aa0-4ae9-9107-fca99430ccfa-openstack-config-secret\") pod \"openstackclient\" (UID: \"ee61762a-4aa0-4ae9-9107-fca99430ccfa\") " pod="openstack/openstackclient" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.276031 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkz8f\" (UniqueName: \"kubernetes.io/projected/ee61762a-4aa0-4ae9-9107-fca99430ccfa-kube-api-access-lkz8f\") pod \"openstackclient\" (UID: \"ee61762a-4aa0-4ae9-9107-fca99430ccfa\") " pod="openstack/openstackclient" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.319417 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.362756 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.367890 4776 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="afc941b5-127c-44d0-8269-ad572b31e1b2" podUID="ee61762a-4aa0-4ae9-9107-fca99430ccfa" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.400969 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.559787 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afc941b5-127c-44d0-8269-ad572b31e1b2-combined-ca-bundle\") pod \"afc941b5-127c-44d0-8269-ad572b31e1b2\" (UID: \"afc941b5-127c-44d0-8269-ad572b31e1b2\") " Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.560141 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/afc941b5-127c-44d0-8269-ad572b31e1b2-openstack-config\") pod \"afc941b5-127c-44d0-8269-ad572b31e1b2\" (UID: \"afc941b5-127c-44d0-8269-ad572b31e1b2\") " Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.560164 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/afc941b5-127c-44d0-8269-ad572b31e1b2-openstack-config-secret\") pod \"afc941b5-127c-44d0-8269-ad572b31e1b2\" (UID: \"afc941b5-127c-44d0-8269-ad572b31e1b2\") " Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.560543 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btnp6\" (UniqueName: \"kubernetes.io/projected/afc941b5-127c-44d0-8269-ad572b31e1b2-kube-api-access-btnp6\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.560927 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afc941b5-127c-44d0-8269-ad572b31e1b2-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "afc941b5-127c-44d0-8269-ad572b31e1b2" (UID: "afc941b5-127c-44d0-8269-ad572b31e1b2"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.564577 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afc941b5-127c-44d0-8269-ad572b31e1b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "afc941b5-127c-44d0-8269-ad572b31e1b2" (UID: "afc941b5-127c-44d0-8269-ad572b31e1b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.565204 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afc941b5-127c-44d0-8269-ad572b31e1b2-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "afc941b5-127c-44d0-8269-ad572b31e1b2" (UID: "afc941b5-127c-44d0-8269-ad572b31e1b2"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.665822 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afc941b5-127c-44d0-8269-ad572b31e1b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.665852 4776 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/afc941b5-127c-44d0-8269-ad572b31e1b2-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.665862 4776 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/afc941b5-127c-44d0-8269-ad572b31e1b2-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.672161 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afc941b5-127c-44d0-8269-ad572b31e1b2" path="/var/lib/kubelet/pods/afc941b5-127c-44d0-8269-ad572b31e1b2/volumes" Nov 25 10:55:35 crc kubenswrapper[4776]: I1125 10:55:35.794533 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 25 10:55:36 crc kubenswrapper[4776]: I1125 10:55:36.373739 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 10:55:36 crc kubenswrapper[4776]: I1125 10:55:36.373713 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"ee61762a-4aa0-4ae9-9107-fca99430ccfa","Type":"ContainerStarted","Data":"b3605f75d8bd3af002c79d3388d36f17448d7027115745579f13ee502a0e5ded"} Nov 25 10:55:36 crc kubenswrapper[4776]: I1125 10:55:36.374550 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"ee61762a-4aa0-4ae9-9107-fca99430ccfa","Type":"ContainerStarted","Data":"f60872c6d9af051b827cf4a410bd5f7d78f08ee2ba430b7910a56b0c064290e3"} Nov 25 10:55:36 crc kubenswrapper[4776]: I1125 10:55:36.400946 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.400921265 podStartE2EDuration="2.400921265s" podCreationTimestamp="2025-11-25 10:55:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:55:36.40033176 +0000 UTC m=+5481.441391323" watchObservedRunningTime="2025-11-25 10:55:36.400921265 +0000 UTC m=+5481.441980858" Nov 25 10:55:36 crc kubenswrapper[4776]: I1125 10:55:36.403923 4776 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="afc941b5-127c-44d0-8269-ad572b31e1b2" podUID="ee61762a-4aa0-4ae9-9107-fca99430ccfa" Nov 25 10:55:44 crc kubenswrapper[4776]: I1125 10:55:44.663003 4776 scope.go:117] "RemoveContainer" containerID="f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40" Nov 25 10:55:44 crc kubenswrapper[4776]: E1125 10:55:44.664261 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:55:59 crc kubenswrapper[4776]: I1125 10:55:59.662707 4776 scope.go:117] "RemoveContainer" containerID="f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40" Nov 25 10:55:59 crc kubenswrapper[4776]: E1125 10:55:59.663509 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:56:13 crc kubenswrapper[4776]: I1125 10:56:13.663010 4776 scope.go:117] "RemoveContainer" containerID="f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40" Nov 25 10:56:13 crc kubenswrapper[4776]: E1125 10:56:13.663880 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:56:18 crc kubenswrapper[4776]: E1125 10:56:18.745367 4776 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.193:47804->38.102.83.193:36915: read tcp 38.102.83.193:47804->38.102.83.193:36915: read: connection reset by peer Nov 25 10:56:27 crc kubenswrapper[4776]: I1125 10:56:27.662355 4776 scope.go:117] "RemoveContainer" containerID="f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40" Nov 25 10:56:27 crc kubenswrapper[4776]: E1125 10:56:27.663096 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:56:39 crc kubenswrapper[4776]: I1125 10:56:39.662629 4776 scope.go:117] "RemoveContainer" containerID="f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40" Nov 25 10:56:39 crc kubenswrapper[4776]: E1125 10:56:39.663230 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:56:53 crc kubenswrapper[4776]: I1125 10:56:53.662768 4776 scope.go:117] "RemoveContainer" containerID="f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40" Nov 25 10:56:53 crc kubenswrapper[4776]: E1125 10:56:53.664139 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:57:04 crc kubenswrapper[4776]: I1125 10:57:04.662169 4776 scope.go:117] "RemoveContainer" containerID="f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40" Nov 25 10:57:04 crc kubenswrapper[4776]: E1125 10:57:04.664585 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:57:13 crc kubenswrapper[4776]: I1125 10:57:13.798416 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-x7zns"] Nov 25 10:57:13 crc kubenswrapper[4776]: I1125 10:57:13.800587 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-x7zns" Nov 25 10:57:13 crc kubenswrapper[4776]: I1125 10:57:13.816687 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtmhf\" (UniqueName: \"kubernetes.io/projected/3d2684fe-a46c-4bcc-9f5b-8d06e5966073-kube-api-access-rtmhf\") pod \"barbican-db-create-x7zns\" (UID: \"3d2684fe-a46c-4bcc-9f5b-8d06e5966073\") " pod="openstack/barbican-db-create-x7zns" Nov 25 10:57:13 crc kubenswrapper[4776]: I1125 10:57:13.816868 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d2684fe-a46c-4bcc-9f5b-8d06e5966073-operator-scripts\") pod \"barbican-db-create-x7zns\" (UID: \"3d2684fe-a46c-4bcc-9f5b-8d06e5966073\") " pod="openstack/barbican-db-create-x7zns" Nov 25 10:57:13 crc kubenswrapper[4776]: I1125 10:57:13.817727 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-60f3-account-create-2ffkm"] Nov 25 10:57:13 crc kubenswrapper[4776]: I1125 10:57:13.819199 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-60f3-account-create-2ffkm" Nov 25 10:57:13 crc kubenswrapper[4776]: I1125 10:57:13.822192 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 25 10:57:13 crc kubenswrapper[4776]: I1125 10:57:13.824979 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-x7zns"] Nov 25 10:57:13 crc kubenswrapper[4776]: I1125 10:57:13.839485 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-60f3-account-create-2ffkm"] Nov 25 10:57:13 crc kubenswrapper[4776]: I1125 10:57:13.918334 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d2684fe-a46c-4bcc-9f5b-8d06e5966073-operator-scripts\") pod \"barbican-db-create-x7zns\" (UID: \"3d2684fe-a46c-4bcc-9f5b-8d06e5966073\") " pod="openstack/barbican-db-create-x7zns" Nov 25 10:57:13 crc kubenswrapper[4776]: I1125 10:57:13.918405 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37973950-513b-427d-b640-cbd21bc1fd65-operator-scripts\") pod \"barbican-60f3-account-create-2ffkm\" (UID: \"37973950-513b-427d-b640-cbd21bc1fd65\") " pod="openstack/barbican-60f3-account-create-2ffkm" Nov 25 10:57:13 crc kubenswrapper[4776]: I1125 10:57:13.918438 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pk9h\" (UniqueName: \"kubernetes.io/projected/37973950-513b-427d-b640-cbd21bc1fd65-kube-api-access-2pk9h\") pod \"barbican-60f3-account-create-2ffkm\" (UID: \"37973950-513b-427d-b640-cbd21bc1fd65\") " pod="openstack/barbican-60f3-account-create-2ffkm" Nov 25 10:57:13 crc kubenswrapper[4776]: I1125 10:57:13.918772 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtmhf\" (UniqueName: \"kubernetes.io/projected/3d2684fe-a46c-4bcc-9f5b-8d06e5966073-kube-api-access-rtmhf\") pod \"barbican-db-create-x7zns\" (UID: \"3d2684fe-a46c-4bcc-9f5b-8d06e5966073\") " pod="openstack/barbican-db-create-x7zns" Nov 25 10:57:13 crc kubenswrapper[4776]: I1125 10:57:13.919165 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d2684fe-a46c-4bcc-9f5b-8d06e5966073-operator-scripts\") pod \"barbican-db-create-x7zns\" (UID: \"3d2684fe-a46c-4bcc-9f5b-8d06e5966073\") " pod="openstack/barbican-db-create-x7zns" Nov 25 10:57:13 crc kubenswrapper[4776]: I1125 10:57:13.940250 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtmhf\" (UniqueName: \"kubernetes.io/projected/3d2684fe-a46c-4bcc-9f5b-8d06e5966073-kube-api-access-rtmhf\") pod \"barbican-db-create-x7zns\" (UID: \"3d2684fe-a46c-4bcc-9f5b-8d06e5966073\") " pod="openstack/barbican-db-create-x7zns" Nov 25 10:57:14 crc kubenswrapper[4776]: I1125 10:57:14.019713 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37973950-513b-427d-b640-cbd21bc1fd65-operator-scripts\") pod \"barbican-60f3-account-create-2ffkm\" (UID: \"37973950-513b-427d-b640-cbd21bc1fd65\") " pod="openstack/barbican-60f3-account-create-2ffkm" Nov 25 10:57:14 crc kubenswrapper[4776]: I1125 10:57:14.019996 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pk9h\" (UniqueName: \"kubernetes.io/projected/37973950-513b-427d-b640-cbd21bc1fd65-kube-api-access-2pk9h\") pod \"barbican-60f3-account-create-2ffkm\" (UID: \"37973950-513b-427d-b640-cbd21bc1fd65\") " pod="openstack/barbican-60f3-account-create-2ffkm" Nov 25 10:57:14 crc kubenswrapper[4776]: I1125 10:57:14.021339 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37973950-513b-427d-b640-cbd21bc1fd65-operator-scripts\") pod \"barbican-60f3-account-create-2ffkm\" (UID: \"37973950-513b-427d-b640-cbd21bc1fd65\") " pod="openstack/barbican-60f3-account-create-2ffkm" Nov 25 10:57:14 crc kubenswrapper[4776]: I1125 10:57:14.042076 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pk9h\" (UniqueName: \"kubernetes.io/projected/37973950-513b-427d-b640-cbd21bc1fd65-kube-api-access-2pk9h\") pod \"barbican-60f3-account-create-2ffkm\" (UID: \"37973950-513b-427d-b640-cbd21bc1fd65\") " pod="openstack/barbican-60f3-account-create-2ffkm" Nov 25 10:57:14 crc kubenswrapper[4776]: I1125 10:57:14.144358 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-x7zns" Nov 25 10:57:14 crc kubenswrapper[4776]: I1125 10:57:14.151330 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-60f3-account-create-2ffkm" Nov 25 10:57:14 crc kubenswrapper[4776]: I1125 10:57:14.622857 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-x7zns"] Nov 25 10:57:14 crc kubenswrapper[4776]: W1125 10:57:14.698126 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37973950_513b_427d_b640_cbd21bc1fd65.slice/crio-2e76d294735c6db5856a1fd7809339c8f21f8d9da9877ced7ec947ea11c62863 WatchSource:0}: Error finding container 2e76d294735c6db5856a1fd7809339c8f21f8d9da9877ced7ec947ea11c62863: Status 404 returned error can't find the container with id 2e76d294735c6db5856a1fd7809339c8f21f8d9da9877ced7ec947ea11c62863 Nov 25 10:57:14 crc kubenswrapper[4776]: I1125 10:57:14.698997 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-60f3-account-create-2ffkm"] Nov 25 10:57:15 crc kubenswrapper[4776]: I1125 10:57:15.205085 4776 generic.go:334] "Generic (PLEG): container finished" podID="3d2684fe-a46c-4bcc-9f5b-8d06e5966073" containerID="f6e9f3a5d505b4c545dbbead2d55fa069dbc2b11537a386b96f16848ee212cfd" exitCode=0 Nov 25 10:57:15 crc kubenswrapper[4776]: I1125 10:57:15.205115 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-x7zns" event={"ID":"3d2684fe-a46c-4bcc-9f5b-8d06e5966073","Type":"ContainerDied","Data":"f6e9f3a5d505b4c545dbbead2d55fa069dbc2b11537a386b96f16848ee212cfd"} Nov 25 10:57:15 crc kubenswrapper[4776]: I1125 10:57:15.205155 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-x7zns" event={"ID":"3d2684fe-a46c-4bcc-9f5b-8d06e5966073","Type":"ContainerStarted","Data":"e73ff0274e39bf8cd6c45e6953a8f1349b15c8241e6e4ca05fb068c95236fb65"} Nov 25 10:57:15 crc kubenswrapper[4776]: I1125 10:57:15.207248 4776 generic.go:334] "Generic (PLEG): container finished" podID="37973950-513b-427d-b640-cbd21bc1fd65" containerID="fd039443c5df559bc8d525197566de07fa763f9c155c919c5dfb3352f577160a" exitCode=0 Nov 25 10:57:15 crc kubenswrapper[4776]: I1125 10:57:15.207280 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-60f3-account-create-2ffkm" event={"ID":"37973950-513b-427d-b640-cbd21bc1fd65","Type":"ContainerDied","Data":"fd039443c5df559bc8d525197566de07fa763f9c155c919c5dfb3352f577160a"} Nov 25 10:57:15 crc kubenswrapper[4776]: I1125 10:57:15.207299 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-60f3-account-create-2ffkm" event={"ID":"37973950-513b-427d-b640-cbd21bc1fd65","Type":"ContainerStarted","Data":"2e76d294735c6db5856a1fd7809339c8f21f8d9da9877ced7ec947ea11c62863"} Nov 25 10:57:16 crc kubenswrapper[4776]: I1125 10:57:16.604750 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-x7zns" Nov 25 10:57:16 crc kubenswrapper[4776]: I1125 10:57:16.615692 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-60f3-account-create-2ffkm" Nov 25 10:57:16 crc kubenswrapper[4776]: I1125 10:57:16.662573 4776 scope.go:117] "RemoveContainer" containerID="f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40" Nov 25 10:57:16 crc kubenswrapper[4776]: E1125 10:57:16.662795 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 10:57:16 crc kubenswrapper[4776]: I1125 10:57:16.762853 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pk9h\" (UniqueName: \"kubernetes.io/projected/37973950-513b-427d-b640-cbd21bc1fd65-kube-api-access-2pk9h\") pod \"37973950-513b-427d-b640-cbd21bc1fd65\" (UID: \"37973950-513b-427d-b640-cbd21bc1fd65\") " Nov 25 10:57:16 crc kubenswrapper[4776]: I1125 10:57:16.762989 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtmhf\" (UniqueName: \"kubernetes.io/projected/3d2684fe-a46c-4bcc-9f5b-8d06e5966073-kube-api-access-rtmhf\") pod \"3d2684fe-a46c-4bcc-9f5b-8d06e5966073\" (UID: \"3d2684fe-a46c-4bcc-9f5b-8d06e5966073\") " Nov 25 10:57:16 crc kubenswrapper[4776]: I1125 10:57:16.763147 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37973950-513b-427d-b640-cbd21bc1fd65-operator-scripts\") pod \"37973950-513b-427d-b640-cbd21bc1fd65\" (UID: \"37973950-513b-427d-b640-cbd21bc1fd65\") " Nov 25 10:57:16 crc kubenswrapper[4776]: I1125 10:57:16.763376 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d2684fe-a46c-4bcc-9f5b-8d06e5966073-operator-scripts\") pod \"3d2684fe-a46c-4bcc-9f5b-8d06e5966073\" (UID: \"3d2684fe-a46c-4bcc-9f5b-8d06e5966073\") " Nov 25 10:57:16 crc kubenswrapper[4776]: I1125 10:57:16.763984 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d2684fe-a46c-4bcc-9f5b-8d06e5966073-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3d2684fe-a46c-4bcc-9f5b-8d06e5966073" (UID: "3d2684fe-a46c-4bcc-9f5b-8d06e5966073"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:57:16 crc kubenswrapper[4776]: I1125 10:57:16.764045 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37973950-513b-427d-b640-cbd21bc1fd65-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "37973950-513b-427d-b640-cbd21bc1fd65" (UID: "37973950-513b-427d-b640-cbd21bc1fd65"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:57:16 crc kubenswrapper[4776]: I1125 10:57:16.765325 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37973950-513b-427d-b640-cbd21bc1fd65-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:16 crc kubenswrapper[4776]: I1125 10:57:16.765362 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d2684fe-a46c-4bcc-9f5b-8d06e5966073-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:16 crc kubenswrapper[4776]: I1125 10:57:16.771912 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d2684fe-a46c-4bcc-9f5b-8d06e5966073-kube-api-access-rtmhf" (OuterVolumeSpecName: "kube-api-access-rtmhf") pod "3d2684fe-a46c-4bcc-9f5b-8d06e5966073" (UID: "3d2684fe-a46c-4bcc-9f5b-8d06e5966073"). InnerVolumeSpecName "kube-api-access-rtmhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:57:16 crc kubenswrapper[4776]: I1125 10:57:16.772745 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37973950-513b-427d-b640-cbd21bc1fd65-kube-api-access-2pk9h" (OuterVolumeSpecName: "kube-api-access-2pk9h") pod "37973950-513b-427d-b640-cbd21bc1fd65" (UID: "37973950-513b-427d-b640-cbd21bc1fd65"). InnerVolumeSpecName "kube-api-access-2pk9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:57:16 crc kubenswrapper[4776]: I1125 10:57:16.867790 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pk9h\" (UniqueName: \"kubernetes.io/projected/37973950-513b-427d-b640-cbd21bc1fd65-kube-api-access-2pk9h\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:16 crc kubenswrapper[4776]: I1125 10:57:16.867848 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtmhf\" (UniqueName: \"kubernetes.io/projected/3d2684fe-a46c-4bcc-9f5b-8d06e5966073-kube-api-access-rtmhf\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:17 crc kubenswrapper[4776]: I1125 10:57:17.229978 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-60f3-account-create-2ffkm" Nov 25 10:57:17 crc kubenswrapper[4776]: I1125 10:57:17.229988 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-60f3-account-create-2ffkm" event={"ID":"37973950-513b-427d-b640-cbd21bc1fd65","Type":"ContainerDied","Data":"2e76d294735c6db5856a1fd7809339c8f21f8d9da9877ced7ec947ea11c62863"} Nov 25 10:57:17 crc kubenswrapper[4776]: I1125 10:57:17.230186 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e76d294735c6db5856a1fd7809339c8f21f8d9da9877ced7ec947ea11c62863" Nov 25 10:57:17 crc kubenswrapper[4776]: I1125 10:57:17.233150 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-x7zns" event={"ID":"3d2684fe-a46c-4bcc-9f5b-8d06e5966073","Type":"ContainerDied","Data":"e73ff0274e39bf8cd6c45e6953a8f1349b15c8241e6e4ca05fb068c95236fb65"} Nov 25 10:57:17 crc kubenswrapper[4776]: I1125 10:57:17.233213 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e73ff0274e39bf8cd6c45e6953a8f1349b15c8241e6e4ca05fb068c95236fb65" Nov 25 10:57:17 crc kubenswrapper[4776]: I1125 10:57:17.233421 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-x7zns" Nov 25 10:57:19 crc kubenswrapper[4776]: I1125 10:57:19.182200 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-5ppz8"] Nov 25 10:57:19 crc kubenswrapper[4776]: E1125 10:57:19.183311 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d2684fe-a46c-4bcc-9f5b-8d06e5966073" containerName="mariadb-database-create" Nov 25 10:57:19 crc kubenswrapper[4776]: I1125 10:57:19.183330 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d2684fe-a46c-4bcc-9f5b-8d06e5966073" containerName="mariadb-database-create" Nov 25 10:57:19 crc kubenswrapper[4776]: E1125 10:57:19.183374 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37973950-513b-427d-b640-cbd21bc1fd65" containerName="mariadb-account-create" Nov 25 10:57:19 crc kubenswrapper[4776]: I1125 10:57:19.183382 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="37973950-513b-427d-b640-cbd21bc1fd65" containerName="mariadb-account-create" Nov 25 10:57:19 crc kubenswrapper[4776]: I1125 10:57:19.183610 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d2684fe-a46c-4bcc-9f5b-8d06e5966073" containerName="mariadb-database-create" Nov 25 10:57:19 crc kubenswrapper[4776]: I1125 10:57:19.183626 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="37973950-513b-427d-b640-cbd21bc1fd65" containerName="mariadb-account-create" Nov 25 10:57:19 crc kubenswrapper[4776]: I1125 10:57:19.184575 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5ppz8" Nov 25 10:57:19 crc kubenswrapper[4776]: I1125 10:57:19.185418 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-5ppz8"] Nov 25 10:57:19 crc kubenswrapper[4776]: I1125 10:57:19.191141 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-n77jw" Nov 25 10:57:19 crc kubenswrapper[4776]: I1125 10:57:19.191599 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 25 10:57:19 crc kubenswrapper[4776]: I1125 10:57:19.317303 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5pdm\" (UniqueName: \"kubernetes.io/projected/68963dc0-5607-450a-8bf3-b2a92abdefc9-kube-api-access-d5pdm\") pod \"barbican-db-sync-5ppz8\" (UID: \"68963dc0-5607-450a-8bf3-b2a92abdefc9\") " pod="openstack/barbican-db-sync-5ppz8" Nov 25 10:57:19 crc kubenswrapper[4776]: I1125 10:57:19.317471 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/68963dc0-5607-450a-8bf3-b2a92abdefc9-db-sync-config-data\") pod \"barbican-db-sync-5ppz8\" (UID: \"68963dc0-5607-450a-8bf3-b2a92abdefc9\") " pod="openstack/barbican-db-sync-5ppz8" Nov 25 10:57:19 crc kubenswrapper[4776]: I1125 10:57:19.317558 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68963dc0-5607-450a-8bf3-b2a92abdefc9-combined-ca-bundle\") pod \"barbican-db-sync-5ppz8\" (UID: \"68963dc0-5607-450a-8bf3-b2a92abdefc9\") " pod="openstack/barbican-db-sync-5ppz8" Nov 25 10:57:19 crc kubenswrapper[4776]: I1125 10:57:19.419368 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5pdm\" (UniqueName: \"kubernetes.io/projected/68963dc0-5607-450a-8bf3-b2a92abdefc9-kube-api-access-d5pdm\") pod \"barbican-db-sync-5ppz8\" (UID: \"68963dc0-5607-450a-8bf3-b2a92abdefc9\") " pod="openstack/barbican-db-sync-5ppz8" Nov 25 10:57:19 crc kubenswrapper[4776]: I1125 10:57:19.419805 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/68963dc0-5607-450a-8bf3-b2a92abdefc9-db-sync-config-data\") pod \"barbican-db-sync-5ppz8\" (UID: \"68963dc0-5607-450a-8bf3-b2a92abdefc9\") " pod="openstack/barbican-db-sync-5ppz8" Nov 25 10:57:19 crc kubenswrapper[4776]: I1125 10:57:19.419902 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68963dc0-5607-450a-8bf3-b2a92abdefc9-combined-ca-bundle\") pod \"barbican-db-sync-5ppz8\" (UID: \"68963dc0-5607-450a-8bf3-b2a92abdefc9\") " pod="openstack/barbican-db-sync-5ppz8" Nov 25 10:57:19 crc kubenswrapper[4776]: I1125 10:57:19.425443 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68963dc0-5607-450a-8bf3-b2a92abdefc9-combined-ca-bundle\") pod \"barbican-db-sync-5ppz8\" (UID: \"68963dc0-5607-450a-8bf3-b2a92abdefc9\") " pod="openstack/barbican-db-sync-5ppz8" Nov 25 10:57:19 crc kubenswrapper[4776]: I1125 10:57:19.427083 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/68963dc0-5607-450a-8bf3-b2a92abdefc9-db-sync-config-data\") pod \"barbican-db-sync-5ppz8\" (UID: \"68963dc0-5607-450a-8bf3-b2a92abdefc9\") " pod="openstack/barbican-db-sync-5ppz8" Nov 25 10:57:19 crc kubenswrapper[4776]: I1125 10:57:19.434744 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5pdm\" (UniqueName: \"kubernetes.io/projected/68963dc0-5607-450a-8bf3-b2a92abdefc9-kube-api-access-d5pdm\") pod \"barbican-db-sync-5ppz8\" (UID: \"68963dc0-5607-450a-8bf3-b2a92abdefc9\") " pod="openstack/barbican-db-sync-5ppz8" Nov 25 10:57:19 crc kubenswrapper[4776]: I1125 10:57:19.508255 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5ppz8" Nov 25 10:57:19 crc kubenswrapper[4776]: I1125 10:57:19.989940 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-5ppz8"] Nov 25 10:57:20 crc kubenswrapper[4776]: I1125 10:57:20.261765 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5ppz8" event={"ID":"68963dc0-5607-450a-8bf3-b2a92abdefc9","Type":"ContainerStarted","Data":"d94e4d0d2d389030273930dc6b70c474012aa26f933c8c8c8ede8945ce5e2d05"} Nov 25 10:57:20 crc kubenswrapper[4776]: I1125 10:57:20.262195 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5ppz8" event={"ID":"68963dc0-5607-450a-8bf3-b2a92abdefc9","Type":"ContainerStarted","Data":"e00f9d957672870d7d9ee0800e14189d63fdb66b7493b5d9ba38a4e178c40bf8"} Nov 25 10:57:20 crc kubenswrapper[4776]: I1125 10:57:20.281876 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-5ppz8" podStartSLOduration=1.281852669 podStartE2EDuration="1.281852669s" podCreationTimestamp="2025-11-25 10:57:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:57:20.275644733 +0000 UTC m=+5585.316704306" watchObservedRunningTime="2025-11-25 10:57:20.281852669 +0000 UTC m=+5585.322912222" Nov 25 10:57:21 crc kubenswrapper[4776]: I1125 10:57:21.128740 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9drrs"] Nov 25 10:57:21 crc kubenswrapper[4776]: I1125 10:57:21.130762 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9drrs" Nov 25 10:57:21 crc kubenswrapper[4776]: I1125 10:57:21.140731 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9drrs"] Nov 25 10:57:21 crc kubenswrapper[4776]: I1125 10:57:21.249903 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f85nw\" (UniqueName: \"kubernetes.io/projected/e48cd85d-fdd6-4c2f-95d9-01e1132f89a6-kube-api-access-f85nw\") pod \"redhat-operators-9drrs\" (UID: \"e48cd85d-fdd6-4c2f-95d9-01e1132f89a6\") " pod="openshift-marketplace/redhat-operators-9drrs" Nov 25 10:57:21 crc kubenswrapper[4776]: I1125 10:57:21.250257 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e48cd85d-fdd6-4c2f-95d9-01e1132f89a6-utilities\") pod \"redhat-operators-9drrs\" (UID: \"e48cd85d-fdd6-4c2f-95d9-01e1132f89a6\") " pod="openshift-marketplace/redhat-operators-9drrs" Nov 25 10:57:21 crc kubenswrapper[4776]: I1125 10:57:21.250412 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e48cd85d-fdd6-4c2f-95d9-01e1132f89a6-catalog-content\") pod \"redhat-operators-9drrs\" (UID: \"e48cd85d-fdd6-4c2f-95d9-01e1132f89a6\") " pod="openshift-marketplace/redhat-operators-9drrs" Nov 25 10:57:21 crc kubenswrapper[4776]: I1125 10:57:21.353142 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e48cd85d-fdd6-4c2f-95d9-01e1132f89a6-utilities\") pod \"redhat-operators-9drrs\" (UID: \"e48cd85d-fdd6-4c2f-95d9-01e1132f89a6\") " pod="openshift-marketplace/redhat-operators-9drrs" Nov 25 10:57:21 crc kubenswrapper[4776]: I1125 10:57:21.353208 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e48cd85d-fdd6-4c2f-95d9-01e1132f89a6-catalog-content\") pod \"redhat-operators-9drrs\" (UID: \"e48cd85d-fdd6-4c2f-95d9-01e1132f89a6\") " pod="openshift-marketplace/redhat-operators-9drrs" Nov 25 10:57:21 crc kubenswrapper[4776]: I1125 10:57:21.353506 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f85nw\" (UniqueName: \"kubernetes.io/projected/e48cd85d-fdd6-4c2f-95d9-01e1132f89a6-kube-api-access-f85nw\") pod \"redhat-operators-9drrs\" (UID: \"e48cd85d-fdd6-4c2f-95d9-01e1132f89a6\") " pod="openshift-marketplace/redhat-operators-9drrs" Nov 25 10:57:21 crc kubenswrapper[4776]: I1125 10:57:21.353676 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e48cd85d-fdd6-4c2f-95d9-01e1132f89a6-catalog-content\") pod \"redhat-operators-9drrs\" (UID: \"e48cd85d-fdd6-4c2f-95d9-01e1132f89a6\") " pod="openshift-marketplace/redhat-operators-9drrs" Nov 25 10:57:21 crc kubenswrapper[4776]: I1125 10:57:21.353801 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e48cd85d-fdd6-4c2f-95d9-01e1132f89a6-utilities\") pod \"redhat-operators-9drrs\" (UID: \"e48cd85d-fdd6-4c2f-95d9-01e1132f89a6\") " pod="openshift-marketplace/redhat-operators-9drrs" Nov 25 10:57:21 crc kubenswrapper[4776]: I1125 10:57:21.379658 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f85nw\" (UniqueName: \"kubernetes.io/projected/e48cd85d-fdd6-4c2f-95d9-01e1132f89a6-kube-api-access-f85nw\") pod \"redhat-operators-9drrs\" (UID: \"e48cd85d-fdd6-4c2f-95d9-01e1132f89a6\") " pod="openshift-marketplace/redhat-operators-9drrs" Nov 25 10:57:21 crc kubenswrapper[4776]: I1125 10:57:21.470589 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9drrs" Nov 25 10:57:21 crc kubenswrapper[4776]: I1125 10:57:21.933596 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9drrs"] Nov 25 10:57:21 crc kubenswrapper[4776]: W1125 10:57:21.938030 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode48cd85d_fdd6_4c2f_95d9_01e1132f89a6.slice/crio-310988a89e39841fddbff10b933febf82195efb52e3660f9432d25a8cc466630 WatchSource:0}: Error finding container 310988a89e39841fddbff10b933febf82195efb52e3660f9432d25a8cc466630: Status 404 returned error can't find the container with id 310988a89e39841fddbff10b933febf82195efb52e3660f9432d25a8cc466630 Nov 25 10:57:22 crc kubenswrapper[4776]: I1125 10:57:22.277629 4776 generic.go:334] "Generic (PLEG): container finished" podID="e48cd85d-fdd6-4c2f-95d9-01e1132f89a6" containerID="82bbc19d07dfab861e2f1ed7dda9c277badf862e81eedadd3b63710308b02265" exitCode=0 Nov 25 10:57:22 crc kubenswrapper[4776]: I1125 10:57:22.277689 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9drrs" event={"ID":"e48cd85d-fdd6-4c2f-95d9-01e1132f89a6","Type":"ContainerDied","Data":"82bbc19d07dfab861e2f1ed7dda9c277badf862e81eedadd3b63710308b02265"} Nov 25 10:57:22 crc kubenswrapper[4776]: I1125 10:57:22.277717 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9drrs" event={"ID":"e48cd85d-fdd6-4c2f-95d9-01e1132f89a6","Type":"ContainerStarted","Data":"310988a89e39841fddbff10b933febf82195efb52e3660f9432d25a8cc466630"} Nov 25 10:57:23 crc kubenswrapper[4776]: I1125 10:57:23.288573 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9drrs" event={"ID":"e48cd85d-fdd6-4c2f-95d9-01e1132f89a6","Type":"ContainerStarted","Data":"e4e46f7f33323064245856dd9bda671691008a9546817c2c15d6284184a97b81"} Nov 25 10:57:24 crc kubenswrapper[4776]: I1125 10:57:24.301349 4776 generic.go:334] "Generic (PLEG): container finished" podID="68963dc0-5607-450a-8bf3-b2a92abdefc9" containerID="d94e4d0d2d389030273930dc6b70c474012aa26f933c8c8c8ede8945ce5e2d05" exitCode=0 Nov 25 10:57:24 crc kubenswrapper[4776]: I1125 10:57:24.301448 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5ppz8" event={"ID":"68963dc0-5607-450a-8bf3-b2a92abdefc9","Type":"ContainerDied","Data":"d94e4d0d2d389030273930dc6b70c474012aa26f933c8c8c8ede8945ce5e2d05"} Nov 25 10:57:24 crc kubenswrapper[4776]: I1125 10:57:24.305047 4776 generic.go:334] "Generic (PLEG): container finished" podID="e48cd85d-fdd6-4c2f-95d9-01e1132f89a6" containerID="e4e46f7f33323064245856dd9bda671691008a9546817c2c15d6284184a97b81" exitCode=0 Nov 25 10:57:24 crc kubenswrapper[4776]: I1125 10:57:24.305104 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9drrs" event={"ID":"e48cd85d-fdd6-4c2f-95d9-01e1132f89a6","Type":"ContainerDied","Data":"e4e46f7f33323064245856dd9bda671691008a9546817c2c15d6284184a97b81"} Nov 25 10:57:25 crc kubenswrapper[4776]: I1125 10:57:25.651283 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5ppz8" Nov 25 10:57:25 crc kubenswrapper[4776]: I1125 10:57:25.684586 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/68963dc0-5607-450a-8bf3-b2a92abdefc9-db-sync-config-data\") pod \"68963dc0-5607-450a-8bf3-b2a92abdefc9\" (UID: \"68963dc0-5607-450a-8bf3-b2a92abdefc9\") " Nov 25 10:57:25 crc kubenswrapper[4776]: I1125 10:57:25.684649 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5pdm\" (UniqueName: \"kubernetes.io/projected/68963dc0-5607-450a-8bf3-b2a92abdefc9-kube-api-access-d5pdm\") pod \"68963dc0-5607-450a-8bf3-b2a92abdefc9\" (UID: \"68963dc0-5607-450a-8bf3-b2a92abdefc9\") " Nov 25 10:57:25 crc kubenswrapper[4776]: I1125 10:57:25.684677 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68963dc0-5607-450a-8bf3-b2a92abdefc9-combined-ca-bundle\") pod \"68963dc0-5607-450a-8bf3-b2a92abdefc9\" (UID: \"68963dc0-5607-450a-8bf3-b2a92abdefc9\") " Nov 25 10:57:25 crc kubenswrapper[4776]: I1125 10:57:25.689636 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68963dc0-5607-450a-8bf3-b2a92abdefc9-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "68963dc0-5607-450a-8bf3-b2a92abdefc9" (UID: "68963dc0-5607-450a-8bf3-b2a92abdefc9"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:25 crc kubenswrapper[4776]: I1125 10:57:25.692357 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68963dc0-5607-450a-8bf3-b2a92abdefc9-kube-api-access-d5pdm" (OuterVolumeSpecName: "kube-api-access-d5pdm") pod "68963dc0-5607-450a-8bf3-b2a92abdefc9" (UID: "68963dc0-5607-450a-8bf3-b2a92abdefc9"). InnerVolumeSpecName "kube-api-access-d5pdm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:57:25 crc kubenswrapper[4776]: I1125 10:57:25.716367 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68963dc0-5607-450a-8bf3-b2a92abdefc9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "68963dc0-5607-450a-8bf3-b2a92abdefc9" (UID: "68963dc0-5607-450a-8bf3-b2a92abdefc9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:25 crc kubenswrapper[4776]: I1125 10:57:25.786751 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5pdm\" (UniqueName: \"kubernetes.io/projected/68963dc0-5607-450a-8bf3-b2a92abdefc9-kube-api-access-d5pdm\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:25 crc kubenswrapper[4776]: I1125 10:57:25.786782 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68963dc0-5607-450a-8bf3-b2a92abdefc9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:25 crc kubenswrapper[4776]: I1125 10:57:25.786792 4776 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/68963dc0-5607-450a-8bf3-b2a92abdefc9-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.324186 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5ppz8" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.324196 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5ppz8" event={"ID":"68963dc0-5607-450a-8bf3-b2a92abdefc9","Type":"ContainerDied","Data":"e00f9d957672870d7d9ee0800e14189d63fdb66b7493b5d9ba38a4e178c40bf8"} Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.324265 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e00f9d957672870d7d9ee0800e14189d63fdb66b7493b5d9ba38a4e178c40bf8" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.326995 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9drrs" event={"ID":"e48cd85d-fdd6-4c2f-95d9-01e1132f89a6","Type":"ContainerStarted","Data":"8f71cb47e2f7ecf851a76cb48c0708548e30e8e6f9ade51a839b756190b9fac1"} Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.368053 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9drrs" podStartSLOduration=2.149722115 podStartE2EDuration="5.368033457s" podCreationTimestamp="2025-11-25 10:57:21 +0000 UTC" firstStartedPulling="2025-11-25 10:57:22.280377244 +0000 UTC m=+5587.321436797" lastFinishedPulling="2025-11-25 10:57:25.498688586 +0000 UTC m=+5590.539748139" observedRunningTime="2025-11-25 10:57:26.351781559 +0000 UTC m=+5591.392841122" watchObservedRunningTime="2025-11-25 10:57:26.368033457 +0000 UTC m=+5591.409093020" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.611100 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5647554789-tq4dw"] Nov 25 10:57:26 crc kubenswrapper[4776]: E1125 10:57:26.612942 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68963dc0-5607-450a-8bf3-b2a92abdefc9" containerName="barbican-db-sync" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.612958 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="68963dc0-5607-450a-8bf3-b2a92abdefc9" containerName="barbican-db-sync" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.613157 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="68963dc0-5607-450a-8bf3-b2a92abdefc9" containerName="barbican-db-sync" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.614104 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5647554789-tq4dw" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.628014 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-n77jw" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.629628 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.634100 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.642669 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5d87dd74f4-rzv2q"] Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.644368 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5d87dd74f4-rzv2q" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.646799 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.656781 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5647554789-tq4dw"] Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.684282 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5d87dd74f4-rzv2q"] Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.727324 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea1d5f90-b587-4071-a507-c5869e7fc650-combined-ca-bundle\") pod \"barbican-worker-5647554789-tq4dw\" (UID: \"ea1d5f90-b587-4071-a507-c5869e7fc650\") " pod="openstack/barbican-worker-5647554789-tq4dw" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.727411 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qbdg\" (UniqueName: \"kubernetes.io/projected/cd7d445b-c6b1-4479-8158-1be3fd658211-kube-api-access-6qbdg\") pod \"barbican-keystone-listener-5d87dd74f4-rzv2q\" (UID: \"cd7d445b-c6b1-4479-8158-1be3fd658211\") " pod="openstack/barbican-keystone-listener-5d87dd74f4-rzv2q" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.727451 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea1d5f90-b587-4071-a507-c5869e7fc650-logs\") pod \"barbican-worker-5647554789-tq4dw\" (UID: \"ea1d5f90-b587-4071-a507-c5869e7fc650\") " pod="openstack/barbican-worker-5647554789-tq4dw" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.727475 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea1d5f90-b587-4071-a507-c5869e7fc650-config-data\") pod \"barbican-worker-5647554789-tq4dw\" (UID: \"ea1d5f90-b587-4071-a507-c5869e7fc650\") " pod="openstack/barbican-worker-5647554789-tq4dw" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.727551 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd7d445b-c6b1-4479-8158-1be3fd658211-logs\") pod \"barbican-keystone-listener-5d87dd74f4-rzv2q\" (UID: \"cd7d445b-c6b1-4479-8158-1be3fd658211\") " pod="openstack/barbican-keystone-listener-5d87dd74f4-rzv2q" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.727723 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd7d445b-c6b1-4479-8158-1be3fd658211-config-data\") pod \"barbican-keystone-listener-5d87dd74f4-rzv2q\" (UID: \"cd7d445b-c6b1-4479-8158-1be3fd658211\") " pod="openstack/barbican-keystone-listener-5d87dd74f4-rzv2q" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.727758 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ea1d5f90-b587-4071-a507-c5869e7fc650-config-data-custom\") pod \"barbican-worker-5647554789-tq4dw\" (UID: \"ea1d5f90-b587-4071-a507-c5869e7fc650\") " pod="openstack/barbican-worker-5647554789-tq4dw" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.727784 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd7d445b-c6b1-4479-8158-1be3fd658211-combined-ca-bundle\") pod \"barbican-keystone-listener-5d87dd74f4-rzv2q\" (UID: \"cd7d445b-c6b1-4479-8158-1be3fd658211\") " pod="openstack/barbican-keystone-listener-5d87dd74f4-rzv2q" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.727811 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmmcq\" (UniqueName: \"kubernetes.io/projected/ea1d5f90-b587-4071-a507-c5869e7fc650-kube-api-access-tmmcq\") pod \"barbican-worker-5647554789-tq4dw\" (UID: \"ea1d5f90-b587-4071-a507-c5869e7fc650\") " pod="openstack/barbican-worker-5647554789-tq4dw" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.727865 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd7d445b-c6b1-4479-8158-1be3fd658211-config-data-custom\") pod \"barbican-keystone-listener-5d87dd74f4-rzv2q\" (UID: \"cd7d445b-c6b1-4479-8158-1be3fd658211\") " pod="openstack/barbican-keystone-listener-5d87dd74f4-rzv2q" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.758013 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-658f7f865f-fkgks"] Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.759750 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-658f7f865f-fkgks" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.779734 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-658f7f865f-fkgks"] Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.829694 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5161e133-bd4f-436d-8d36-cbcc7faccf9a-config\") pod \"dnsmasq-dns-658f7f865f-fkgks\" (UID: \"5161e133-bd4f-436d-8d36-cbcc7faccf9a\") " pod="openstack/dnsmasq-dns-658f7f865f-fkgks" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.829771 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd7d445b-c6b1-4479-8158-1be3fd658211-config-data\") pod \"barbican-keystone-listener-5d87dd74f4-rzv2q\" (UID: \"cd7d445b-c6b1-4479-8158-1be3fd658211\") " pod="openstack/barbican-keystone-listener-5d87dd74f4-rzv2q" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.829797 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd7d445b-c6b1-4479-8158-1be3fd658211-combined-ca-bundle\") pod \"barbican-keystone-listener-5d87dd74f4-rzv2q\" (UID: \"cd7d445b-c6b1-4479-8158-1be3fd658211\") " pod="openstack/barbican-keystone-listener-5d87dd74f4-rzv2q" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.829829 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ea1d5f90-b587-4071-a507-c5869e7fc650-config-data-custom\") pod \"barbican-worker-5647554789-tq4dw\" (UID: \"ea1d5f90-b587-4071-a507-c5869e7fc650\") " pod="openstack/barbican-worker-5647554789-tq4dw" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.829849 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmmcq\" (UniqueName: \"kubernetes.io/projected/ea1d5f90-b587-4071-a507-c5869e7fc650-kube-api-access-tmmcq\") pod \"barbican-worker-5647554789-tq4dw\" (UID: \"ea1d5f90-b587-4071-a507-c5869e7fc650\") " pod="openstack/barbican-worker-5647554789-tq4dw" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.829881 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd7d445b-c6b1-4479-8158-1be3fd658211-config-data-custom\") pod \"barbican-keystone-listener-5d87dd74f4-rzv2q\" (UID: \"cd7d445b-c6b1-4479-8158-1be3fd658211\") " pod="openstack/barbican-keystone-listener-5d87dd74f4-rzv2q" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.829940 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsmxl\" (UniqueName: \"kubernetes.io/projected/5161e133-bd4f-436d-8d36-cbcc7faccf9a-kube-api-access-wsmxl\") pod \"dnsmasq-dns-658f7f865f-fkgks\" (UID: \"5161e133-bd4f-436d-8d36-cbcc7faccf9a\") " pod="openstack/dnsmasq-dns-658f7f865f-fkgks" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.829963 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5161e133-bd4f-436d-8d36-cbcc7faccf9a-dns-svc\") pod \"dnsmasq-dns-658f7f865f-fkgks\" (UID: \"5161e133-bd4f-436d-8d36-cbcc7faccf9a\") " pod="openstack/dnsmasq-dns-658f7f865f-fkgks" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.829999 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5161e133-bd4f-436d-8d36-cbcc7faccf9a-ovsdbserver-sb\") pod \"dnsmasq-dns-658f7f865f-fkgks\" (UID: \"5161e133-bd4f-436d-8d36-cbcc7faccf9a\") " pod="openstack/dnsmasq-dns-658f7f865f-fkgks" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.830016 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5161e133-bd4f-436d-8d36-cbcc7faccf9a-ovsdbserver-nb\") pod \"dnsmasq-dns-658f7f865f-fkgks\" (UID: \"5161e133-bd4f-436d-8d36-cbcc7faccf9a\") " pod="openstack/dnsmasq-dns-658f7f865f-fkgks" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.830042 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea1d5f90-b587-4071-a507-c5869e7fc650-combined-ca-bundle\") pod \"barbican-worker-5647554789-tq4dw\" (UID: \"ea1d5f90-b587-4071-a507-c5869e7fc650\") " pod="openstack/barbican-worker-5647554789-tq4dw" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.830182 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qbdg\" (UniqueName: \"kubernetes.io/projected/cd7d445b-c6b1-4479-8158-1be3fd658211-kube-api-access-6qbdg\") pod \"barbican-keystone-listener-5d87dd74f4-rzv2q\" (UID: \"cd7d445b-c6b1-4479-8158-1be3fd658211\") " pod="openstack/barbican-keystone-listener-5d87dd74f4-rzv2q" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.830207 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea1d5f90-b587-4071-a507-c5869e7fc650-logs\") pod \"barbican-worker-5647554789-tq4dw\" (UID: \"ea1d5f90-b587-4071-a507-c5869e7fc650\") " pod="openstack/barbican-worker-5647554789-tq4dw" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.830274 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea1d5f90-b587-4071-a507-c5869e7fc650-config-data\") pod \"barbican-worker-5647554789-tq4dw\" (UID: \"ea1d5f90-b587-4071-a507-c5869e7fc650\") " pod="openstack/barbican-worker-5647554789-tq4dw" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.830312 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd7d445b-c6b1-4479-8158-1be3fd658211-logs\") pod \"barbican-keystone-listener-5d87dd74f4-rzv2q\" (UID: \"cd7d445b-c6b1-4479-8158-1be3fd658211\") " pod="openstack/barbican-keystone-listener-5d87dd74f4-rzv2q" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.831791 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea1d5f90-b587-4071-a507-c5869e7fc650-logs\") pod \"barbican-worker-5647554789-tq4dw\" (UID: \"ea1d5f90-b587-4071-a507-c5869e7fc650\") " pod="openstack/barbican-worker-5647554789-tq4dw" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.833124 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd7d445b-c6b1-4479-8158-1be3fd658211-logs\") pod \"barbican-keystone-listener-5d87dd74f4-rzv2q\" (UID: \"cd7d445b-c6b1-4479-8158-1be3fd658211\") " pod="openstack/barbican-keystone-listener-5d87dd74f4-rzv2q" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.842598 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd7d445b-c6b1-4479-8158-1be3fd658211-combined-ca-bundle\") pod \"barbican-keystone-listener-5d87dd74f4-rzv2q\" (UID: \"cd7d445b-c6b1-4479-8158-1be3fd658211\") " pod="openstack/barbican-keystone-listener-5d87dd74f4-rzv2q" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.853740 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea1d5f90-b587-4071-a507-c5869e7fc650-combined-ca-bundle\") pod \"barbican-worker-5647554789-tq4dw\" (UID: \"ea1d5f90-b587-4071-a507-c5869e7fc650\") " pod="openstack/barbican-worker-5647554789-tq4dw" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.854783 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea1d5f90-b587-4071-a507-c5869e7fc650-config-data\") pod \"barbican-worker-5647554789-tq4dw\" (UID: \"ea1d5f90-b587-4071-a507-c5869e7fc650\") " pod="openstack/barbican-worker-5647554789-tq4dw" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.855262 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd7d445b-c6b1-4479-8158-1be3fd658211-config-data-custom\") pod \"barbican-keystone-listener-5d87dd74f4-rzv2q\" (UID: \"cd7d445b-c6b1-4479-8158-1be3fd658211\") " pod="openstack/barbican-keystone-listener-5d87dd74f4-rzv2q" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.856247 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd7d445b-c6b1-4479-8158-1be3fd658211-config-data\") pod \"barbican-keystone-listener-5d87dd74f4-rzv2q\" (UID: \"cd7d445b-c6b1-4479-8158-1be3fd658211\") " pod="openstack/barbican-keystone-listener-5d87dd74f4-rzv2q" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.860362 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qbdg\" (UniqueName: \"kubernetes.io/projected/cd7d445b-c6b1-4479-8158-1be3fd658211-kube-api-access-6qbdg\") pod \"barbican-keystone-listener-5d87dd74f4-rzv2q\" (UID: \"cd7d445b-c6b1-4479-8158-1be3fd658211\") " pod="openstack/barbican-keystone-listener-5d87dd74f4-rzv2q" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.866685 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmmcq\" (UniqueName: \"kubernetes.io/projected/ea1d5f90-b587-4071-a507-c5869e7fc650-kube-api-access-tmmcq\") pod \"barbican-worker-5647554789-tq4dw\" (UID: \"ea1d5f90-b587-4071-a507-c5869e7fc650\") " pod="openstack/barbican-worker-5647554789-tq4dw" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.867150 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ea1d5f90-b587-4071-a507-c5869e7fc650-config-data-custom\") pod \"barbican-worker-5647554789-tq4dw\" (UID: \"ea1d5f90-b587-4071-a507-c5869e7fc650\") " pod="openstack/barbican-worker-5647554789-tq4dw" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.905842 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5f45c84864-9dvhh"] Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.909887 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f45c84864-9dvhh" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.917410 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.931884 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5161e133-bd4f-436d-8d36-cbcc7faccf9a-ovsdbserver-nb\") pod \"dnsmasq-dns-658f7f865f-fkgks\" (UID: \"5161e133-bd4f-436d-8d36-cbcc7faccf9a\") " pod="openstack/dnsmasq-dns-658f7f865f-fkgks" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.932021 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5161e133-bd4f-436d-8d36-cbcc7faccf9a-config\") pod \"dnsmasq-dns-658f7f865f-fkgks\" (UID: \"5161e133-bd4f-436d-8d36-cbcc7faccf9a\") " pod="openstack/dnsmasq-dns-658f7f865f-fkgks" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.932119 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsmxl\" (UniqueName: \"kubernetes.io/projected/5161e133-bd4f-436d-8d36-cbcc7faccf9a-kube-api-access-wsmxl\") pod \"dnsmasq-dns-658f7f865f-fkgks\" (UID: \"5161e133-bd4f-436d-8d36-cbcc7faccf9a\") " pod="openstack/dnsmasq-dns-658f7f865f-fkgks" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.932144 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5161e133-bd4f-436d-8d36-cbcc7faccf9a-dns-svc\") pod \"dnsmasq-dns-658f7f865f-fkgks\" (UID: \"5161e133-bd4f-436d-8d36-cbcc7faccf9a\") " pod="openstack/dnsmasq-dns-658f7f865f-fkgks" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.932166 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5161e133-bd4f-436d-8d36-cbcc7faccf9a-ovsdbserver-sb\") pod \"dnsmasq-dns-658f7f865f-fkgks\" (UID: \"5161e133-bd4f-436d-8d36-cbcc7faccf9a\") " pod="openstack/dnsmasq-dns-658f7f865f-fkgks" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.933164 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5161e133-bd4f-436d-8d36-cbcc7faccf9a-ovsdbserver-sb\") pod \"dnsmasq-dns-658f7f865f-fkgks\" (UID: \"5161e133-bd4f-436d-8d36-cbcc7faccf9a\") " pod="openstack/dnsmasq-dns-658f7f865f-fkgks" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.934014 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5161e133-bd4f-436d-8d36-cbcc7faccf9a-dns-svc\") pod \"dnsmasq-dns-658f7f865f-fkgks\" (UID: \"5161e133-bd4f-436d-8d36-cbcc7faccf9a\") " pod="openstack/dnsmasq-dns-658f7f865f-fkgks" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.934350 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5161e133-bd4f-436d-8d36-cbcc7faccf9a-config\") pod \"dnsmasq-dns-658f7f865f-fkgks\" (UID: \"5161e133-bd4f-436d-8d36-cbcc7faccf9a\") " pod="openstack/dnsmasq-dns-658f7f865f-fkgks" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.935528 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5f45c84864-9dvhh"] Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.935790 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5161e133-bd4f-436d-8d36-cbcc7faccf9a-ovsdbserver-nb\") pod \"dnsmasq-dns-658f7f865f-fkgks\" (UID: \"5161e133-bd4f-436d-8d36-cbcc7faccf9a\") " pod="openstack/dnsmasq-dns-658f7f865f-fkgks" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.951949 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5647554789-tq4dw" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.969827 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5d87dd74f4-rzv2q" Nov 25 10:57:26 crc kubenswrapper[4776]: I1125 10:57:26.975117 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsmxl\" (UniqueName: \"kubernetes.io/projected/5161e133-bd4f-436d-8d36-cbcc7faccf9a-kube-api-access-wsmxl\") pod \"dnsmasq-dns-658f7f865f-fkgks\" (UID: \"5161e133-bd4f-436d-8d36-cbcc7faccf9a\") " pod="openstack/dnsmasq-dns-658f7f865f-fkgks" Nov 25 10:57:27 crc kubenswrapper[4776]: I1125 10:57:27.033258 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-config-data\") pod \"barbican-api-5f45c84864-9dvhh\" (UID: \"67cbdd15-6d67-42ce-a67b-8ff0849a2bba\") " pod="openstack/barbican-api-5f45c84864-9dvhh" Nov 25 10:57:27 crc kubenswrapper[4776]: I1125 10:57:27.033308 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-config-data-custom\") pod \"barbican-api-5f45c84864-9dvhh\" (UID: \"67cbdd15-6d67-42ce-a67b-8ff0849a2bba\") " pod="openstack/barbican-api-5f45c84864-9dvhh" Nov 25 10:57:27 crc kubenswrapper[4776]: I1125 10:57:27.033343 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4t8gp\" (UniqueName: \"kubernetes.io/projected/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-kube-api-access-4t8gp\") pod \"barbican-api-5f45c84864-9dvhh\" (UID: \"67cbdd15-6d67-42ce-a67b-8ff0849a2bba\") " pod="openstack/barbican-api-5f45c84864-9dvhh" Nov 25 10:57:27 crc kubenswrapper[4776]: I1125 10:57:27.033444 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-logs\") pod \"barbican-api-5f45c84864-9dvhh\" (UID: \"67cbdd15-6d67-42ce-a67b-8ff0849a2bba\") " pod="openstack/barbican-api-5f45c84864-9dvhh" Nov 25 10:57:27 crc kubenswrapper[4776]: I1125 10:57:27.033662 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-combined-ca-bundle\") pod \"barbican-api-5f45c84864-9dvhh\" (UID: \"67cbdd15-6d67-42ce-a67b-8ff0849a2bba\") " pod="openstack/barbican-api-5f45c84864-9dvhh" Nov 25 10:57:27 crc kubenswrapper[4776]: I1125 10:57:27.090602 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-658f7f865f-fkgks" Nov 25 10:57:27 crc kubenswrapper[4776]: I1125 10:57:27.136350 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-combined-ca-bundle\") pod \"barbican-api-5f45c84864-9dvhh\" (UID: \"67cbdd15-6d67-42ce-a67b-8ff0849a2bba\") " pod="openstack/barbican-api-5f45c84864-9dvhh" Nov 25 10:57:27 crc kubenswrapper[4776]: I1125 10:57:27.136456 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-config-data\") pod \"barbican-api-5f45c84864-9dvhh\" (UID: \"67cbdd15-6d67-42ce-a67b-8ff0849a2bba\") " pod="openstack/barbican-api-5f45c84864-9dvhh" Nov 25 10:57:27 crc kubenswrapper[4776]: I1125 10:57:27.136485 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-config-data-custom\") pod \"barbican-api-5f45c84864-9dvhh\" (UID: \"67cbdd15-6d67-42ce-a67b-8ff0849a2bba\") " pod="openstack/barbican-api-5f45c84864-9dvhh" Nov 25 10:57:27 crc kubenswrapper[4776]: I1125 10:57:27.136514 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4t8gp\" (UniqueName: \"kubernetes.io/projected/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-kube-api-access-4t8gp\") pod \"barbican-api-5f45c84864-9dvhh\" (UID: \"67cbdd15-6d67-42ce-a67b-8ff0849a2bba\") " pod="openstack/barbican-api-5f45c84864-9dvhh" Nov 25 10:57:27 crc kubenswrapper[4776]: I1125 10:57:27.136532 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-logs\") pod \"barbican-api-5f45c84864-9dvhh\" (UID: \"67cbdd15-6d67-42ce-a67b-8ff0849a2bba\") " pod="openstack/barbican-api-5f45c84864-9dvhh" Nov 25 10:57:27 crc kubenswrapper[4776]: I1125 10:57:27.136904 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-logs\") pod \"barbican-api-5f45c84864-9dvhh\" (UID: \"67cbdd15-6d67-42ce-a67b-8ff0849a2bba\") " pod="openstack/barbican-api-5f45c84864-9dvhh" Nov 25 10:57:27 crc kubenswrapper[4776]: I1125 10:57:27.142432 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-config-data-custom\") pod \"barbican-api-5f45c84864-9dvhh\" (UID: \"67cbdd15-6d67-42ce-a67b-8ff0849a2bba\") " pod="openstack/barbican-api-5f45c84864-9dvhh" Nov 25 10:57:27 crc kubenswrapper[4776]: I1125 10:57:27.154588 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-config-data\") pod \"barbican-api-5f45c84864-9dvhh\" (UID: \"67cbdd15-6d67-42ce-a67b-8ff0849a2bba\") " pod="openstack/barbican-api-5f45c84864-9dvhh" Nov 25 10:57:27 crc kubenswrapper[4776]: I1125 10:57:27.156631 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-combined-ca-bundle\") pod \"barbican-api-5f45c84864-9dvhh\" (UID: \"67cbdd15-6d67-42ce-a67b-8ff0849a2bba\") " pod="openstack/barbican-api-5f45c84864-9dvhh" Nov 25 10:57:27 crc kubenswrapper[4776]: I1125 10:57:27.170157 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4t8gp\" (UniqueName: \"kubernetes.io/projected/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-kube-api-access-4t8gp\") pod \"barbican-api-5f45c84864-9dvhh\" (UID: \"67cbdd15-6d67-42ce-a67b-8ff0849a2bba\") " pod="openstack/barbican-api-5f45c84864-9dvhh" Nov 25 10:57:27 crc kubenswrapper[4776]: I1125 10:57:27.271718 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f45c84864-9dvhh" Nov 25 10:57:27 crc kubenswrapper[4776]: I1125 10:57:27.475363 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5d87dd74f4-rzv2q"] Nov 25 10:57:27 crc kubenswrapper[4776]: I1125 10:57:27.541552 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5647554789-tq4dw"] Nov 25 10:57:27 crc kubenswrapper[4776]: W1125 10:57:27.558134 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea1d5f90_b587_4071_a507_c5869e7fc650.slice/crio-31ee8e6a7f4517e257d61c130163479f03a94bc8ffe63e10a0c9502935261786 WatchSource:0}: Error finding container 31ee8e6a7f4517e257d61c130163479f03a94bc8ffe63e10a0c9502935261786: Status 404 returned error can't find the container with id 31ee8e6a7f4517e257d61c130163479f03a94bc8ffe63e10a0c9502935261786 Nov 25 10:57:27 crc kubenswrapper[4776]: I1125 10:57:27.659323 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-658f7f865f-fkgks"] Nov 25 10:57:27 crc kubenswrapper[4776]: W1125 10:57:27.668244 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5161e133_bd4f_436d_8d36_cbcc7faccf9a.slice/crio-dcd2e47687f705e9afa7f9fee5c824d9c3535b1daba652fa126699841123e770 WatchSource:0}: Error finding container dcd2e47687f705e9afa7f9fee5c824d9c3535b1daba652fa126699841123e770: Status 404 returned error can't find the container with id dcd2e47687f705e9afa7f9fee5c824d9c3535b1daba652fa126699841123e770 Nov 25 10:57:27 crc kubenswrapper[4776]: I1125 10:57:27.749871 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5f45c84864-9dvhh"] Nov 25 10:57:28 crc kubenswrapper[4776]: I1125 10:57:28.348179 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5d87dd74f4-rzv2q" event={"ID":"cd7d445b-c6b1-4479-8158-1be3fd658211","Type":"ContainerStarted","Data":"95dac4f17f307765fbfbede02e29c0d7fd9f7d95f1a71b4b57b308a45b7bb2e1"} Nov 25 10:57:28 crc kubenswrapper[4776]: I1125 10:57:28.353471 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5d87dd74f4-rzv2q" event={"ID":"cd7d445b-c6b1-4479-8158-1be3fd658211","Type":"ContainerStarted","Data":"a7cf2722d0defecec6df56492b79c1fd79e420d79a8e6cf565b7e0982432d996"} Nov 25 10:57:28 crc kubenswrapper[4776]: I1125 10:57:28.353492 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5d87dd74f4-rzv2q" event={"ID":"cd7d445b-c6b1-4479-8158-1be3fd658211","Type":"ContainerStarted","Data":"98577fd60bd5a6e877f60d0f34667b08f12a803586641ed311e448263e13b579"} Nov 25 10:57:28 crc kubenswrapper[4776]: I1125 10:57:28.353504 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5647554789-tq4dw" event={"ID":"ea1d5f90-b587-4071-a507-c5869e7fc650","Type":"ContainerStarted","Data":"c7d325f0b747533de5d011324265d57693b4feb7b38319296189c97f3476dca2"} Nov 25 10:57:28 crc kubenswrapper[4776]: I1125 10:57:28.353516 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5647554789-tq4dw" event={"ID":"ea1d5f90-b587-4071-a507-c5869e7fc650","Type":"ContainerStarted","Data":"d45a59a63c900fa0f4d16e38bb3d32af966bdcc5f24421f6e79c373fa9babdee"} Nov 25 10:57:28 crc kubenswrapper[4776]: I1125 10:57:28.353535 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5647554789-tq4dw" event={"ID":"ea1d5f90-b587-4071-a507-c5869e7fc650","Type":"ContainerStarted","Data":"31ee8e6a7f4517e257d61c130163479f03a94bc8ffe63e10a0c9502935261786"} Nov 25 10:57:28 crc kubenswrapper[4776]: I1125 10:57:28.354397 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f45c84864-9dvhh" event={"ID":"67cbdd15-6d67-42ce-a67b-8ff0849a2bba","Type":"ContainerStarted","Data":"66b80cc77fd76390e264638fb4f548f552fc078853f8098fcef6c174d8442210"} Nov 25 10:57:28 crc kubenswrapper[4776]: I1125 10:57:28.354449 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f45c84864-9dvhh" event={"ID":"67cbdd15-6d67-42ce-a67b-8ff0849a2bba","Type":"ContainerStarted","Data":"bf0fe4035a64450ea15cb43a3bad5e224ac9c0e69f50c23762a03f9404830647"} Nov 25 10:57:28 crc kubenswrapper[4776]: I1125 10:57:28.354460 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f45c84864-9dvhh" event={"ID":"67cbdd15-6d67-42ce-a67b-8ff0849a2bba","Type":"ContainerStarted","Data":"d94980a82e1054ad18cc5437ec6a7e097ccb31ddde9517b2908fd9d6035c4b50"} Nov 25 10:57:28 crc kubenswrapper[4776]: I1125 10:57:28.354769 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5f45c84864-9dvhh" Nov 25 10:57:28 crc kubenswrapper[4776]: I1125 10:57:28.355491 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5f45c84864-9dvhh" Nov 25 10:57:28 crc kubenswrapper[4776]: I1125 10:57:28.359527 4776 generic.go:334] "Generic (PLEG): container finished" podID="5161e133-bd4f-436d-8d36-cbcc7faccf9a" containerID="2c8dbdb4d3af73e3946b030d348bf16e708d07c3eba781cced54fe765b0839b9" exitCode=0 Nov 25 10:57:28 crc kubenswrapper[4776]: I1125 10:57:28.359570 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-658f7f865f-fkgks" event={"ID":"5161e133-bd4f-436d-8d36-cbcc7faccf9a","Type":"ContainerDied","Data":"2c8dbdb4d3af73e3946b030d348bf16e708d07c3eba781cced54fe765b0839b9"} Nov 25 10:57:28 crc kubenswrapper[4776]: I1125 10:57:28.359595 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-658f7f865f-fkgks" event={"ID":"5161e133-bd4f-436d-8d36-cbcc7faccf9a","Type":"ContainerStarted","Data":"dcd2e47687f705e9afa7f9fee5c824d9c3535b1daba652fa126699841123e770"} Nov 25 10:57:28 crc kubenswrapper[4776]: I1125 10:57:28.401857 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5d87dd74f4-rzv2q" podStartSLOduration=2.401830446 podStartE2EDuration="2.401830446s" podCreationTimestamp="2025-11-25 10:57:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:57:28.375912676 +0000 UTC m=+5593.416972249" watchObservedRunningTime="2025-11-25 10:57:28.401830446 +0000 UTC m=+5593.442889999" Nov 25 10:57:28 crc kubenswrapper[4776]: I1125 10:57:28.411496 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5647554789-tq4dw" podStartSLOduration=2.411468578 podStartE2EDuration="2.411468578s" podCreationTimestamp="2025-11-25 10:57:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:57:28.398171774 +0000 UTC m=+5593.439231327" watchObservedRunningTime="2025-11-25 10:57:28.411468578 +0000 UTC m=+5593.452528131" Nov 25 10:57:28 crc kubenswrapper[4776]: I1125 10:57:28.451028 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5f45c84864-9dvhh" podStartSLOduration=2.451011711 podStartE2EDuration="2.451011711s" podCreationTimestamp="2025-11-25 10:57:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:57:28.437148523 +0000 UTC m=+5593.478208076" watchObservedRunningTime="2025-11-25 10:57:28.451011711 +0000 UTC m=+5593.492071274" Nov 25 10:57:29 crc kubenswrapper[4776]: I1125 10:57:29.372385 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-658f7f865f-fkgks" event={"ID":"5161e133-bd4f-436d-8d36-cbcc7faccf9a","Type":"ContainerStarted","Data":"c2625123e156e2f6f7330db6834dc55ed01bc5ba1d5281595e746092ee6edb11"} Nov 25 10:57:29 crc kubenswrapper[4776]: I1125 10:57:29.393407 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-658f7f865f-fkgks" podStartSLOduration=3.393392615 podStartE2EDuration="3.393392615s" podCreationTimestamp="2025-11-25 10:57:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:57:29.389567099 +0000 UTC m=+5594.430626652" watchObservedRunningTime="2025-11-25 10:57:29.393392615 +0000 UTC m=+5594.434452168" Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.383676 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-658f7f865f-fkgks" Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.521388 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-645d68dd56-v4nsw"] Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.526557 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-645d68dd56-v4nsw" Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.529135 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.531147 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.533990 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-645d68dd56-v4nsw"] Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.599790 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xhx2\" (UniqueName: \"kubernetes.io/projected/90cec8c4-dd53-4430-839e-ed26c40b10ed-kube-api-access-7xhx2\") pod \"barbican-api-645d68dd56-v4nsw\" (UID: \"90cec8c4-dd53-4430-839e-ed26c40b10ed\") " pod="openstack/barbican-api-645d68dd56-v4nsw" Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.599885 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90cec8c4-dd53-4430-839e-ed26c40b10ed-combined-ca-bundle\") pod \"barbican-api-645d68dd56-v4nsw\" (UID: \"90cec8c4-dd53-4430-839e-ed26c40b10ed\") " pod="openstack/barbican-api-645d68dd56-v4nsw" Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.599914 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/90cec8c4-dd53-4430-839e-ed26c40b10ed-public-tls-certs\") pod \"barbican-api-645d68dd56-v4nsw\" (UID: \"90cec8c4-dd53-4430-839e-ed26c40b10ed\") " pod="openstack/barbican-api-645d68dd56-v4nsw" Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.599960 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90cec8c4-dd53-4430-839e-ed26c40b10ed-logs\") pod \"barbican-api-645d68dd56-v4nsw\" (UID: \"90cec8c4-dd53-4430-839e-ed26c40b10ed\") " pod="openstack/barbican-api-645d68dd56-v4nsw" Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.599997 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/90cec8c4-dd53-4430-839e-ed26c40b10ed-internal-tls-certs\") pod \"barbican-api-645d68dd56-v4nsw\" (UID: \"90cec8c4-dd53-4430-839e-ed26c40b10ed\") " pod="openstack/barbican-api-645d68dd56-v4nsw" Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.600033 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90cec8c4-dd53-4430-839e-ed26c40b10ed-config-data\") pod \"barbican-api-645d68dd56-v4nsw\" (UID: \"90cec8c4-dd53-4430-839e-ed26c40b10ed\") " pod="openstack/barbican-api-645d68dd56-v4nsw" Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.600050 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/90cec8c4-dd53-4430-839e-ed26c40b10ed-config-data-custom\") pod \"barbican-api-645d68dd56-v4nsw\" (UID: \"90cec8c4-dd53-4430-839e-ed26c40b10ed\") " pod="openstack/barbican-api-645d68dd56-v4nsw" Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.662513 4776 scope.go:117] "RemoveContainer" containerID="f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40" Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.702566 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xhx2\" (UniqueName: \"kubernetes.io/projected/90cec8c4-dd53-4430-839e-ed26c40b10ed-kube-api-access-7xhx2\") pod \"barbican-api-645d68dd56-v4nsw\" (UID: \"90cec8c4-dd53-4430-839e-ed26c40b10ed\") " pod="openstack/barbican-api-645d68dd56-v4nsw" Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.702996 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90cec8c4-dd53-4430-839e-ed26c40b10ed-combined-ca-bundle\") pod \"barbican-api-645d68dd56-v4nsw\" (UID: \"90cec8c4-dd53-4430-839e-ed26c40b10ed\") " pod="openstack/barbican-api-645d68dd56-v4nsw" Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.703043 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/90cec8c4-dd53-4430-839e-ed26c40b10ed-public-tls-certs\") pod \"barbican-api-645d68dd56-v4nsw\" (UID: \"90cec8c4-dd53-4430-839e-ed26c40b10ed\") " pod="openstack/barbican-api-645d68dd56-v4nsw" Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.703086 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90cec8c4-dd53-4430-839e-ed26c40b10ed-logs\") pod \"barbican-api-645d68dd56-v4nsw\" (UID: \"90cec8c4-dd53-4430-839e-ed26c40b10ed\") " pod="openstack/barbican-api-645d68dd56-v4nsw" Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.703142 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/90cec8c4-dd53-4430-839e-ed26c40b10ed-internal-tls-certs\") pod \"barbican-api-645d68dd56-v4nsw\" (UID: \"90cec8c4-dd53-4430-839e-ed26c40b10ed\") " pod="openstack/barbican-api-645d68dd56-v4nsw" Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.703184 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90cec8c4-dd53-4430-839e-ed26c40b10ed-config-data\") pod \"barbican-api-645d68dd56-v4nsw\" (UID: \"90cec8c4-dd53-4430-839e-ed26c40b10ed\") " pod="openstack/barbican-api-645d68dd56-v4nsw" Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.703208 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/90cec8c4-dd53-4430-839e-ed26c40b10ed-config-data-custom\") pod \"barbican-api-645d68dd56-v4nsw\" (UID: \"90cec8c4-dd53-4430-839e-ed26c40b10ed\") " pod="openstack/barbican-api-645d68dd56-v4nsw" Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.703782 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90cec8c4-dd53-4430-839e-ed26c40b10ed-logs\") pod \"barbican-api-645d68dd56-v4nsw\" (UID: \"90cec8c4-dd53-4430-839e-ed26c40b10ed\") " pod="openstack/barbican-api-645d68dd56-v4nsw" Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.709567 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90cec8c4-dd53-4430-839e-ed26c40b10ed-combined-ca-bundle\") pod \"barbican-api-645d68dd56-v4nsw\" (UID: \"90cec8c4-dd53-4430-839e-ed26c40b10ed\") " pod="openstack/barbican-api-645d68dd56-v4nsw" Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.709978 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90cec8c4-dd53-4430-839e-ed26c40b10ed-config-data\") pod \"barbican-api-645d68dd56-v4nsw\" (UID: \"90cec8c4-dd53-4430-839e-ed26c40b10ed\") " pod="openstack/barbican-api-645d68dd56-v4nsw" Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.710105 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/90cec8c4-dd53-4430-839e-ed26c40b10ed-internal-tls-certs\") pod \"barbican-api-645d68dd56-v4nsw\" (UID: \"90cec8c4-dd53-4430-839e-ed26c40b10ed\") " pod="openstack/barbican-api-645d68dd56-v4nsw" Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.722955 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/90cec8c4-dd53-4430-839e-ed26c40b10ed-public-tls-certs\") pod \"barbican-api-645d68dd56-v4nsw\" (UID: \"90cec8c4-dd53-4430-839e-ed26c40b10ed\") " pod="openstack/barbican-api-645d68dd56-v4nsw" Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.723347 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/90cec8c4-dd53-4430-839e-ed26c40b10ed-config-data-custom\") pod \"barbican-api-645d68dd56-v4nsw\" (UID: \"90cec8c4-dd53-4430-839e-ed26c40b10ed\") " pod="openstack/barbican-api-645d68dd56-v4nsw" Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.726901 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xhx2\" (UniqueName: \"kubernetes.io/projected/90cec8c4-dd53-4430-839e-ed26c40b10ed-kube-api-access-7xhx2\") pod \"barbican-api-645d68dd56-v4nsw\" (UID: \"90cec8c4-dd53-4430-839e-ed26c40b10ed\") " pod="openstack/barbican-api-645d68dd56-v4nsw" Nov 25 10:57:30 crc kubenswrapper[4776]: I1125 10:57:30.844690 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-645d68dd56-v4nsw" Nov 25 10:57:31 crc kubenswrapper[4776]: I1125 10:57:31.294263 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-645d68dd56-v4nsw"] Nov 25 10:57:31 crc kubenswrapper[4776]: W1125 10:57:31.296483 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90cec8c4_dd53_4430_839e_ed26c40b10ed.slice/crio-7957422224e8349f59b37553ab588f27012f4d6a5c6a90d014e29bff8c46214a WatchSource:0}: Error finding container 7957422224e8349f59b37553ab588f27012f4d6a5c6a90d014e29bff8c46214a: Status 404 returned error can't find the container with id 7957422224e8349f59b37553ab588f27012f4d6a5c6a90d014e29bff8c46214a Nov 25 10:57:31 crc kubenswrapper[4776]: I1125 10:57:31.392796 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-645d68dd56-v4nsw" event={"ID":"90cec8c4-dd53-4430-839e-ed26c40b10ed","Type":"ContainerStarted","Data":"7957422224e8349f59b37553ab588f27012f4d6a5c6a90d014e29bff8c46214a"} Nov 25 10:57:31 crc kubenswrapper[4776]: I1125 10:57:31.395997 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"8fb1377358b27002d683fed0c035aa47a3bd9a7c30a5ee14f1b728159b2160e7"} Nov 25 10:57:31 crc kubenswrapper[4776]: I1125 10:57:31.470899 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9drrs" Nov 25 10:57:31 crc kubenswrapper[4776]: I1125 10:57:31.470966 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9drrs" Nov 25 10:57:31 crc kubenswrapper[4776]: I1125 10:57:31.518433 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9drrs" Nov 25 10:57:32 crc kubenswrapper[4776]: I1125 10:57:32.405673 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-645d68dd56-v4nsw" event={"ID":"90cec8c4-dd53-4430-839e-ed26c40b10ed","Type":"ContainerStarted","Data":"97fccd9ad7505d4fd778d04d91c50cf3bbc1ecd44cf927411e3b113a740b0322"} Nov 25 10:57:32 crc kubenswrapper[4776]: I1125 10:57:32.406408 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-645d68dd56-v4nsw" event={"ID":"90cec8c4-dd53-4430-839e-ed26c40b10ed","Type":"ContainerStarted","Data":"7ce4b0dcabf23b2b1d5a2d70d4b12976ce8be2cb1f04e98a95b5e0dfd0b796e5"} Nov 25 10:57:32 crc kubenswrapper[4776]: I1125 10:57:32.407953 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-645d68dd56-v4nsw" Nov 25 10:57:32 crc kubenswrapper[4776]: I1125 10:57:32.408020 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-645d68dd56-v4nsw" Nov 25 10:57:32 crc kubenswrapper[4776]: I1125 10:57:32.442972 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-645d68dd56-v4nsw" podStartSLOduration=2.442951881 podStartE2EDuration="2.442951881s" podCreationTimestamp="2025-11-25 10:57:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:57:32.431386821 +0000 UTC m=+5597.472446374" watchObservedRunningTime="2025-11-25 10:57:32.442951881 +0000 UTC m=+5597.484011424" Nov 25 10:57:32 crc kubenswrapper[4776]: I1125 10:57:32.461595 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9drrs" Nov 25 10:57:32 crc kubenswrapper[4776]: I1125 10:57:32.506866 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9drrs"] Nov 25 10:57:33 crc kubenswrapper[4776]: I1125 10:57:33.966843 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5f45c84864-9dvhh" Nov 25 10:57:34 crc kubenswrapper[4776]: I1125 10:57:34.420592 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9drrs" podUID="e48cd85d-fdd6-4c2f-95d9-01e1132f89a6" containerName="registry-server" containerID="cri-o://8f71cb47e2f7ecf851a76cb48c0708548e30e8e6f9ade51a839b756190b9fac1" gracePeriod=2 Nov 25 10:57:35 crc kubenswrapper[4776]: I1125 10:57:35.391153 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5f45c84864-9dvhh" Nov 25 10:57:35 crc kubenswrapper[4776]: I1125 10:57:35.431294 4776 generic.go:334] "Generic (PLEG): container finished" podID="e48cd85d-fdd6-4c2f-95d9-01e1132f89a6" containerID="8f71cb47e2f7ecf851a76cb48c0708548e30e8e6f9ade51a839b756190b9fac1" exitCode=0 Nov 25 10:57:35 crc kubenswrapper[4776]: I1125 10:57:35.431357 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9drrs" event={"ID":"e48cd85d-fdd6-4c2f-95d9-01e1132f89a6","Type":"ContainerDied","Data":"8f71cb47e2f7ecf851a76cb48c0708548e30e8e6f9ade51a839b756190b9fac1"} Nov 25 10:57:35 crc kubenswrapper[4776]: I1125 10:57:35.841189 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9drrs" Nov 25 10:57:35 crc kubenswrapper[4776]: I1125 10:57:35.903809 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e48cd85d-fdd6-4c2f-95d9-01e1132f89a6-utilities\") pod \"e48cd85d-fdd6-4c2f-95d9-01e1132f89a6\" (UID: \"e48cd85d-fdd6-4c2f-95d9-01e1132f89a6\") " Nov 25 10:57:35 crc kubenswrapper[4776]: I1125 10:57:35.904208 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f85nw\" (UniqueName: \"kubernetes.io/projected/e48cd85d-fdd6-4c2f-95d9-01e1132f89a6-kube-api-access-f85nw\") pod \"e48cd85d-fdd6-4c2f-95d9-01e1132f89a6\" (UID: \"e48cd85d-fdd6-4c2f-95d9-01e1132f89a6\") " Nov 25 10:57:35 crc kubenswrapper[4776]: I1125 10:57:35.904245 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e48cd85d-fdd6-4c2f-95d9-01e1132f89a6-catalog-content\") pod \"e48cd85d-fdd6-4c2f-95d9-01e1132f89a6\" (UID: \"e48cd85d-fdd6-4c2f-95d9-01e1132f89a6\") " Nov 25 10:57:35 crc kubenswrapper[4776]: I1125 10:57:35.904909 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e48cd85d-fdd6-4c2f-95d9-01e1132f89a6-utilities" (OuterVolumeSpecName: "utilities") pod "e48cd85d-fdd6-4c2f-95d9-01e1132f89a6" (UID: "e48cd85d-fdd6-4c2f-95d9-01e1132f89a6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:57:35 crc kubenswrapper[4776]: I1125 10:57:35.912945 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e48cd85d-fdd6-4c2f-95d9-01e1132f89a6-kube-api-access-f85nw" (OuterVolumeSpecName: "kube-api-access-f85nw") pod "e48cd85d-fdd6-4c2f-95d9-01e1132f89a6" (UID: "e48cd85d-fdd6-4c2f-95d9-01e1132f89a6"). InnerVolumeSpecName "kube-api-access-f85nw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:57:36 crc kubenswrapper[4776]: I1125 10:57:36.000247 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e48cd85d-fdd6-4c2f-95d9-01e1132f89a6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e48cd85d-fdd6-4c2f-95d9-01e1132f89a6" (UID: "e48cd85d-fdd6-4c2f-95d9-01e1132f89a6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:57:36 crc kubenswrapper[4776]: I1125 10:57:36.006177 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f85nw\" (UniqueName: \"kubernetes.io/projected/e48cd85d-fdd6-4c2f-95d9-01e1132f89a6-kube-api-access-f85nw\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:36 crc kubenswrapper[4776]: I1125 10:57:36.006215 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e48cd85d-fdd6-4c2f-95d9-01e1132f89a6-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:36 crc kubenswrapper[4776]: I1125 10:57:36.006227 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e48cd85d-fdd6-4c2f-95d9-01e1132f89a6-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:36 crc kubenswrapper[4776]: I1125 10:57:36.462931 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9drrs" Nov 25 10:57:36 crc kubenswrapper[4776]: I1125 10:57:36.462815 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9drrs" event={"ID":"e48cd85d-fdd6-4c2f-95d9-01e1132f89a6","Type":"ContainerDied","Data":"310988a89e39841fddbff10b933febf82195efb52e3660f9432d25a8cc466630"} Nov 25 10:57:36 crc kubenswrapper[4776]: I1125 10:57:36.464227 4776 scope.go:117] "RemoveContainer" containerID="8f71cb47e2f7ecf851a76cb48c0708548e30e8e6f9ade51a839b756190b9fac1" Nov 25 10:57:36 crc kubenswrapper[4776]: I1125 10:57:36.486612 4776 scope.go:117] "RemoveContainer" containerID="e4e46f7f33323064245856dd9bda671691008a9546817c2c15d6284184a97b81" Nov 25 10:57:36 crc kubenswrapper[4776]: I1125 10:57:36.509841 4776 scope.go:117] "RemoveContainer" containerID="82bbc19d07dfab861e2f1ed7dda9c277badf862e81eedadd3b63710308b02265" Nov 25 10:57:36 crc kubenswrapper[4776]: I1125 10:57:36.598441 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9drrs"] Nov 25 10:57:36 crc kubenswrapper[4776]: I1125 10:57:36.608343 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9drrs"] Nov 25 10:57:37 crc kubenswrapper[4776]: I1125 10:57:37.093309 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-658f7f865f-fkgks" Nov 25 10:57:37 crc kubenswrapper[4776]: I1125 10:57:37.173740 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6cbbbb95c9-bng2s"] Nov 25 10:57:37 crc kubenswrapper[4776]: I1125 10:57:37.174047 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6cbbbb95c9-bng2s" podUID="2a467eb0-222b-4486-aa5c-07bba04b741c" containerName="dnsmasq-dns" containerID="cri-o://db9a42be803e0a46a6d076bb1f28a7b9fd5e0fc3ceb1a262457f73adffb30089" gracePeriod=10 Nov 25 10:57:37 crc kubenswrapper[4776]: I1125 10:57:37.501502 4776 generic.go:334] "Generic (PLEG): container finished" podID="2a467eb0-222b-4486-aa5c-07bba04b741c" containerID="db9a42be803e0a46a6d076bb1f28a7b9fd5e0fc3ceb1a262457f73adffb30089" exitCode=0 Nov 25 10:57:37 crc kubenswrapper[4776]: I1125 10:57:37.501613 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cbbbb95c9-bng2s" event={"ID":"2a467eb0-222b-4486-aa5c-07bba04b741c","Type":"ContainerDied","Data":"db9a42be803e0a46a6d076bb1f28a7b9fd5e0fc3ceb1a262457f73adffb30089"} Nov 25 10:57:37 crc kubenswrapper[4776]: I1125 10:57:37.682976 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e48cd85d-fdd6-4c2f-95d9-01e1132f89a6" path="/var/lib/kubelet/pods/e48cd85d-fdd6-4c2f-95d9-01e1132f89a6/volumes" Nov 25 10:57:37 crc kubenswrapper[4776]: I1125 10:57:37.741869 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cbbbb95c9-bng2s" Nov 25 10:57:37 crc kubenswrapper[4776]: I1125 10:57:37.847149 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a467eb0-222b-4486-aa5c-07bba04b741c-config\") pod \"2a467eb0-222b-4486-aa5c-07bba04b741c\" (UID: \"2a467eb0-222b-4486-aa5c-07bba04b741c\") " Nov 25 10:57:37 crc kubenswrapper[4776]: I1125 10:57:37.847271 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a467eb0-222b-4486-aa5c-07bba04b741c-dns-svc\") pod \"2a467eb0-222b-4486-aa5c-07bba04b741c\" (UID: \"2a467eb0-222b-4486-aa5c-07bba04b741c\") " Nov 25 10:57:37 crc kubenswrapper[4776]: I1125 10:57:37.847324 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwfwh\" (UniqueName: \"kubernetes.io/projected/2a467eb0-222b-4486-aa5c-07bba04b741c-kube-api-access-vwfwh\") pod \"2a467eb0-222b-4486-aa5c-07bba04b741c\" (UID: \"2a467eb0-222b-4486-aa5c-07bba04b741c\") " Nov 25 10:57:37 crc kubenswrapper[4776]: I1125 10:57:37.847381 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2a467eb0-222b-4486-aa5c-07bba04b741c-ovsdbserver-sb\") pod \"2a467eb0-222b-4486-aa5c-07bba04b741c\" (UID: \"2a467eb0-222b-4486-aa5c-07bba04b741c\") " Nov 25 10:57:37 crc kubenswrapper[4776]: I1125 10:57:37.847485 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2a467eb0-222b-4486-aa5c-07bba04b741c-ovsdbserver-nb\") pod \"2a467eb0-222b-4486-aa5c-07bba04b741c\" (UID: \"2a467eb0-222b-4486-aa5c-07bba04b741c\") " Nov 25 10:57:37 crc kubenswrapper[4776]: I1125 10:57:37.855467 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a467eb0-222b-4486-aa5c-07bba04b741c-kube-api-access-vwfwh" (OuterVolumeSpecName: "kube-api-access-vwfwh") pod "2a467eb0-222b-4486-aa5c-07bba04b741c" (UID: "2a467eb0-222b-4486-aa5c-07bba04b741c"). InnerVolumeSpecName "kube-api-access-vwfwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:57:37 crc kubenswrapper[4776]: I1125 10:57:37.894554 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a467eb0-222b-4486-aa5c-07bba04b741c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2a467eb0-222b-4486-aa5c-07bba04b741c" (UID: "2a467eb0-222b-4486-aa5c-07bba04b741c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:57:37 crc kubenswrapper[4776]: I1125 10:57:37.895052 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a467eb0-222b-4486-aa5c-07bba04b741c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2a467eb0-222b-4486-aa5c-07bba04b741c" (UID: "2a467eb0-222b-4486-aa5c-07bba04b741c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:57:37 crc kubenswrapper[4776]: I1125 10:57:37.895982 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a467eb0-222b-4486-aa5c-07bba04b741c-config" (OuterVolumeSpecName: "config") pod "2a467eb0-222b-4486-aa5c-07bba04b741c" (UID: "2a467eb0-222b-4486-aa5c-07bba04b741c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:57:37 crc kubenswrapper[4776]: I1125 10:57:37.909942 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a467eb0-222b-4486-aa5c-07bba04b741c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2a467eb0-222b-4486-aa5c-07bba04b741c" (UID: "2a467eb0-222b-4486-aa5c-07bba04b741c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:57:37 crc kubenswrapper[4776]: I1125 10:57:37.949347 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2a467eb0-222b-4486-aa5c-07bba04b741c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:37 crc kubenswrapper[4776]: I1125 10:57:37.949375 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a467eb0-222b-4486-aa5c-07bba04b741c-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:37 crc kubenswrapper[4776]: I1125 10:57:37.949384 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a467eb0-222b-4486-aa5c-07bba04b741c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:37 crc kubenswrapper[4776]: I1125 10:57:37.949393 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwfwh\" (UniqueName: \"kubernetes.io/projected/2a467eb0-222b-4486-aa5c-07bba04b741c-kube-api-access-vwfwh\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:37 crc kubenswrapper[4776]: I1125 10:57:37.949405 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2a467eb0-222b-4486-aa5c-07bba04b741c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:38 crc kubenswrapper[4776]: I1125 10:57:38.515692 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cbbbb95c9-bng2s" event={"ID":"2a467eb0-222b-4486-aa5c-07bba04b741c","Type":"ContainerDied","Data":"fe775dc952678afd32fced32c17c6e0e297fc13fba3e27ba8f85f5b966d63e87"} Nov 25 10:57:38 crc kubenswrapper[4776]: I1125 10:57:38.515748 4776 scope.go:117] "RemoveContainer" containerID="db9a42be803e0a46a6d076bb1f28a7b9fd5e0fc3ceb1a262457f73adffb30089" Nov 25 10:57:38 crc kubenswrapper[4776]: I1125 10:57:38.516741 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cbbbb95c9-bng2s" Nov 25 10:57:38 crc kubenswrapper[4776]: I1125 10:57:38.549059 4776 scope.go:117] "RemoveContainer" containerID="54efdc3d5b3058f9372933646be9167aed12e3667fd4356cd3a4a0518ac7ea4d" Nov 25 10:57:38 crc kubenswrapper[4776]: I1125 10:57:38.554542 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6cbbbb95c9-bng2s"] Nov 25 10:57:38 crc kubenswrapper[4776]: I1125 10:57:38.561809 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6cbbbb95c9-bng2s"] Nov 25 10:57:38 crc kubenswrapper[4776]: E1125 10:57:38.623161 4776 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a467eb0_222b_4486_aa5c_07bba04b741c.slice\": RecentStats: unable to find data in memory cache]" Nov 25 10:57:39 crc kubenswrapper[4776]: I1125 10:57:39.674765 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a467eb0-222b-4486-aa5c-07bba04b741c" path="/var/lib/kubelet/pods/2a467eb0-222b-4486-aa5c-07bba04b741c/volumes" Nov 25 10:57:42 crc kubenswrapper[4776]: I1125 10:57:42.261972 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-645d68dd56-v4nsw" Nov 25 10:57:42 crc kubenswrapper[4776]: I1125 10:57:42.317328 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-645d68dd56-v4nsw" Nov 25 10:57:42 crc kubenswrapper[4776]: I1125 10:57:42.379141 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5f45c84864-9dvhh"] Nov 25 10:57:42 crc kubenswrapper[4776]: I1125 10:57:42.379415 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5f45c84864-9dvhh" podUID="67cbdd15-6d67-42ce-a67b-8ff0849a2bba" containerName="barbican-api-log" containerID="cri-o://bf0fe4035a64450ea15cb43a3bad5e224ac9c0e69f50c23762a03f9404830647" gracePeriod=30 Nov 25 10:57:42 crc kubenswrapper[4776]: I1125 10:57:42.379551 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5f45c84864-9dvhh" podUID="67cbdd15-6d67-42ce-a67b-8ff0849a2bba" containerName="barbican-api" containerID="cri-o://66b80cc77fd76390e264638fb4f548f552fc078853f8098fcef6c174d8442210" gracePeriod=30 Nov 25 10:57:42 crc kubenswrapper[4776]: I1125 10:57:42.569297 4776 generic.go:334] "Generic (PLEG): container finished" podID="67cbdd15-6d67-42ce-a67b-8ff0849a2bba" containerID="bf0fe4035a64450ea15cb43a3bad5e224ac9c0e69f50c23762a03f9404830647" exitCode=143 Nov 25 10:57:42 crc kubenswrapper[4776]: I1125 10:57:42.570295 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f45c84864-9dvhh" event={"ID":"67cbdd15-6d67-42ce-a67b-8ff0849a2bba","Type":"ContainerDied","Data":"bf0fe4035a64450ea15cb43a3bad5e224ac9c0e69f50c23762a03f9404830647"} Nov 25 10:57:45 crc kubenswrapper[4776]: I1125 10:57:45.593550 4776 generic.go:334] "Generic (PLEG): container finished" podID="67cbdd15-6d67-42ce-a67b-8ff0849a2bba" containerID="66b80cc77fd76390e264638fb4f548f552fc078853f8098fcef6c174d8442210" exitCode=0 Nov 25 10:57:45 crc kubenswrapper[4776]: I1125 10:57:45.593606 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f45c84864-9dvhh" event={"ID":"67cbdd15-6d67-42ce-a67b-8ff0849a2bba","Type":"ContainerDied","Data":"66b80cc77fd76390e264638fb4f548f552fc078853f8098fcef6c174d8442210"} Nov 25 10:57:46 crc kubenswrapper[4776]: I1125 10:57:46.325145 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f45c84864-9dvhh" Nov 25 10:57:46 crc kubenswrapper[4776]: I1125 10:57:46.394324 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-config-data\") pod \"67cbdd15-6d67-42ce-a67b-8ff0849a2bba\" (UID: \"67cbdd15-6d67-42ce-a67b-8ff0849a2bba\") " Nov 25 10:57:46 crc kubenswrapper[4776]: I1125 10:57:46.394397 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4t8gp\" (UniqueName: \"kubernetes.io/projected/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-kube-api-access-4t8gp\") pod \"67cbdd15-6d67-42ce-a67b-8ff0849a2bba\" (UID: \"67cbdd15-6d67-42ce-a67b-8ff0849a2bba\") " Nov 25 10:57:46 crc kubenswrapper[4776]: I1125 10:57:46.394509 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-config-data-custom\") pod \"67cbdd15-6d67-42ce-a67b-8ff0849a2bba\" (UID: \"67cbdd15-6d67-42ce-a67b-8ff0849a2bba\") " Nov 25 10:57:46 crc kubenswrapper[4776]: I1125 10:57:46.394551 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-combined-ca-bundle\") pod \"67cbdd15-6d67-42ce-a67b-8ff0849a2bba\" (UID: \"67cbdd15-6d67-42ce-a67b-8ff0849a2bba\") " Nov 25 10:57:46 crc kubenswrapper[4776]: I1125 10:57:46.394633 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-logs\") pod \"67cbdd15-6d67-42ce-a67b-8ff0849a2bba\" (UID: \"67cbdd15-6d67-42ce-a67b-8ff0849a2bba\") " Nov 25 10:57:46 crc kubenswrapper[4776]: I1125 10:57:46.395844 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-logs" (OuterVolumeSpecName: "logs") pod "67cbdd15-6d67-42ce-a67b-8ff0849a2bba" (UID: "67cbdd15-6d67-42ce-a67b-8ff0849a2bba"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:57:46 crc kubenswrapper[4776]: I1125 10:57:46.400283 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-kube-api-access-4t8gp" (OuterVolumeSpecName: "kube-api-access-4t8gp") pod "67cbdd15-6d67-42ce-a67b-8ff0849a2bba" (UID: "67cbdd15-6d67-42ce-a67b-8ff0849a2bba"). InnerVolumeSpecName "kube-api-access-4t8gp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:57:46 crc kubenswrapper[4776]: I1125 10:57:46.400301 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "67cbdd15-6d67-42ce-a67b-8ff0849a2bba" (UID: "67cbdd15-6d67-42ce-a67b-8ff0849a2bba"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:46 crc kubenswrapper[4776]: I1125 10:57:46.427214 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67cbdd15-6d67-42ce-a67b-8ff0849a2bba" (UID: "67cbdd15-6d67-42ce-a67b-8ff0849a2bba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:46 crc kubenswrapper[4776]: I1125 10:57:46.438502 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-config-data" (OuterVolumeSpecName: "config-data") pod "67cbdd15-6d67-42ce-a67b-8ff0849a2bba" (UID: "67cbdd15-6d67-42ce-a67b-8ff0849a2bba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:57:46 crc kubenswrapper[4776]: I1125 10:57:46.496809 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:46 crc kubenswrapper[4776]: I1125 10:57:46.496847 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:46 crc kubenswrapper[4776]: I1125 10:57:46.496856 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:46 crc kubenswrapper[4776]: I1125 10:57:46.496865 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4t8gp\" (UniqueName: \"kubernetes.io/projected/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-kube-api-access-4t8gp\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:46 crc kubenswrapper[4776]: I1125 10:57:46.496878 4776 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67cbdd15-6d67-42ce-a67b-8ff0849a2bba-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:46 crc kubenswrapper[4776]: I1125 10:57:46.605864 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f45c84864-9dvhh" event={"ID":"67cbdd15-6d67-42ce-a67b-8ff0849a2bba","Type":"ContainerDied","Data":"d94980a82e1054ad18cc5437ec6a7e097ccb31ddde9517b2908fd9d6035c4b50"} Nov 25 10:57:46 crc kubenswrapper[4776]: I1125 10:57:46.605924 4776 scope.go:117] "RemoveContainer" containerID="66b80cc77fd76390e264638fb4f548f552fc078853f8098fcef6c174d8442210" Nov 25 10:57:46 crc kubenswrapper[4776]: I1125 10:57:46.606084 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f45c84864-9dvhh" Nov 25 10:57:46 crc kubenswrapper[4776]: I1125 10:57:46.634380 4776 scope.go:117] "RemoveContainer" containerID="bf0fe4035a64450ea15cb43a3bad5e224ac9c0e69f50c23762a03f9404830647" Nov 25 10:57:46 crc kubenswrapper[4776]: I1125 10:57:46.638594 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5f45c84864-9dvhh"] Nov 25 10:57:46 crc kubenswrapper[4776]: I1125 10:57:46.650390 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5f45c84864-9dvhh"] Nov 25 10:57:47 crc kubenswrapper[4776]: I1125 10:57:47.673256 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67cbdd15-6d67-42ce-a67b-8ff0849a2bba" path="/var/lib/kubelet/pods/67cbdd15-6d67-42ce-a67b-8ff0849a2bba/volumes" Nov 25 10:57:47 crc kubenswrapper[4776]: I1125 10:57:47.928734 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-6t7lg"] Nov 25 10:57:47 crc kubenswrapper[4776]: E1125 10:57:47.929381 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67cbdd15-6d67-42ce-a67b-8ff0849a2bba" containerName="barbican-api" Nov 25 10:57:47 crc kubenswrapper[4776]: I1125 10:57:47.929401 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="67cbdd15-6d67-42ce-a67b-8ff0849a2bba" containerName="barbican-api" Nov 25 10:57:47 crc kubenswrapper[4776]: E1125 10:57:47.929423 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e48cd85d-fdd6-4c2f-95d9-01e1132f89a6" containerName="extract-content" Nov 25 10:57:47 crc kubenswrapper[4776]: I1125 10:57:47.929435 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="e48cd85d-fdd6-4c2f-95d9-01e1132f89a6" containerName="extract-content" Nov 25 10:57:47 crc kubenswrapper[4776]: E1125 10:57:47.929457 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a467eb0-222b-4486-aa5c-07bba04b741c" containerName="init" Nov 25 10:57:47 crc kubenswrapper[4776]: I1125 10:57:47.929466 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a467eb0-222b-4486-aa5c-07bba04b741c" containerName="init" Nov 25 10:57:47 crc kubenswrapper[4776]: E1125 10:57:47.929482 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67cbdd15-6d67-42ce-a67b-8ff0849a2bba" containerName="barbican-api-log" Nov 25 10:57:47 crc kubenswrapper[4776]: I1125 10:57:47.929489 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="67cbdd15-6d67-42ce-a67b-8ff0849a2bba" containerName="barbican-api-log" Nov 25 10:57:47 crc kubenswrapper[4776]: E1125 10:57:47.929513 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a467eb0-222b-4486-aa5c-07bba04b741c" containerName="dnsmasq-dns" Nov 25 10:57:47 crc kubenswrapper[4776]: I1125 10:57:47.929519 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a467eb0-222b-4486-aa5c-07bba04b741c" containerName="dnsmasq-dns" Nov 25 10:57:47 crc kubenswrapper[4776]: E1125 10:57:47.929534 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e48cd85d-fdd6-4c2f-95d9-01e1132f89a6" containerName="extract-utilities" Nov 25 10:57:47 crc kubenswrapper[4776]: I1125 10:57:47.929542 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="e48cd85d-fdd6-4c2f-95d9-01e1132f89a6" containerName="extract-utilities" Nov 25 10:57:47 crc kubenswrapper[4776]: E1125 10:57:47.929554 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e48cd85d-fdd6-4c2f-95d9-01e1132f89a6" containerName="registry-server" Nov 25 10:57:47 crc kubenswrapper[4776]: I1125 10:57:47.929565 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="e48cd85d-fdd6-4c2f-95d9-01e1132f89a6" containerName="registry-server" Nov 25 10:57:47 crc kubenswrapper[4776]: I1125 10:57:47.929720 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="e48cd85d-fdd6-4c2f-95d9-01e1132f89a6" containerName="registry-server" Nov 25 10:57:47 crc kubenswrapper[4776]: I1125 10:57:47.929738 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a467eb0-222b-4486-aa5c-07bba04b741c" containerName="dnsmasq-dns" Nov 25 10:57:47 crc kubenswrapper[4776]: I1125 10:57:47.929750 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="67cbdd15-6d67-42ce-a67b-8ff0849a2bba" containerName="barbican-api-log" Nov 25 10:57:47 crc kubenswrapper[4776]: I1125 10:57:47.929760 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="67cbdd15-6d67-42ce-a67b-8ff0849a2bba" containerName="barbican-api" Nov 25 10:57:47 crc kubenswrapper[4776]: I1125 10:57:47.930374 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-6t7lg" Nov 25 10:57:47 crc kubenswrapper[4776]: I1125 10:57:47.936941 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-6t7lg"] Nov 25 10:57:48 crc kubenswrapper[4776]: I1125 10:57:48.032748 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7eac8e5-4d91-4316-981c-14c58d3a261a-operator-scripts\") pod \"neutron-db-create-6t7lg\" (UID: \"b7eac8e5-4d91-4316-981c-14c58d3a261a\") " pod="openstack/neutron-db-create-6t7lg" Nov 25 10:57:48 crc kubenswrapper[4776]: I1125 10:57:48.033166 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqchk\" (UniqueName: \"kubernetes.io/projected/b7eac8e5-4d91-4316-981c-14c58d3a261a-kube-api-access-cqchk\") pod \"neutron-db-create-6t7lg\" (UID: \"b7eac8e5-4d91-4316-981c-14c58d3a261a\") " pod="openstack/neutron-db-create-6t7lg" Nov 25 10:57:48 crc kubenswrapper[4776]: I1125 10:57:48.036285 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-2c0f-account-create-pn6jc"] Nov 25 10:57:48 crc kubenswrapper[4776]: I1125 10:57:48.037685 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2c0f-account-create-pn6jc" Nov 25 10:57:48 crc kubenswrapper[4776]: I1125 10:57:48.040735 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 25 10:57:48 crc kubenswrapper[4776]: I1125 10:57:48.059390 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-2c0f-account-create-pn6jc"] Nov 25 10:57:48 crc kubenswrapper[4776]: I1125 10:57:48.135053 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7eac8e5-4d91-4316-981c-14c58d3a261a-operator-scripts\") pod \"neutron-db-create-6t7lg\" (UID: \"b7eac8e5-4d91-4316-981c-14c58d3a261a\") " pod="openstack/neutron-db-create-6t7lg" Nov 25 10:57:48 crc kubenswrapper[4776]: I1125 10:57:48.135167 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df4x5\" (UniqueName: \"kubernetes.io/projected/410928cc-80b1-415a-a02c-83ca27897dca-kube-api-access-df4x5\") pod \"neutron-2c0f-account-create-pn6jc\" (UID: \"410928cc-80b1-415a-a02c-83ca27897dca\") " pod="openstack/neutron-2c0f-account-create-pn6jc" Nov 25 10:57:48 crc kubenswrapper[4776]: I1125 10:57:48.135293 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/410928cc-80b1-415a-a02c-83ca27897dca-operator-scripts\") pod \"neutron-2c0f-account-create-pn6jc\" (UID: \"410928cc-80b1-415a-a02c-83ca27897dca\") " pod="openstack/neutron-2c0f-account-create-pn6jc" Nov 25 10:57:48 crc kubenswrapper[4776]: I1125 10:57:48.135394 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqchk\" (UniqueName: \"kubernetes.io/projected/b7eac8e5-4d91-4316-981c-14c58d3a261a-kube-api-access-cqchk\") pod \"neutron-db-create-6t7lg\" (UID: \"b7eac8e5-4d91-4316-981c-14c58d3a261a\") " pod="openstack/neutron-db-create-6t7lg" Nov 25 10:57:48 crc kubenswrapper[4776]: I1125 10:57:48.135938 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7eac8e5-4d91-4316-981c-14c58d3a261a-operator-scripts\") pod \"neutron-db-create-6t7lg\" (UID: \"b7eac8e5-4d91-4316-981c-14c58d3a261a\") " pod="openstack/neutron-db-create-6t7lg" Nov 25 10:57:48 crc kubenswrapper[4776]: I1125 10:57:48.158140 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqchk\" (UniqueName: \"kubernetes.io/projected/b7eac8e5-4d91-4316-981c-14c58d3a261a-kube-api-access-cqchk\") pod \"neutron-db-create-6t7lg\" (UID: \"b7eac8e5-4d91-4316-981c-14c58d3a261a\") " pod="openstack/neutron-db-create-6t7lg" Nov 25 10:57:48 crc kubenswrapper[4776]: I1125 10:57:48.237426 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/410928cc-80b1-415a-a02c-83ca27897dca-operator-scripts\") pod \"neutron-2c0f-account-create-pn6jc\" (UID: \"410928cc-80b1-415a-a02c-83ca27897dca\") " pod="openstack/neutron-2c0f-account-create-pn6jc" Nov 25 10:57:48 crc kubenswrapper[4776]: I1125 10:57:48.237585 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df4x5\" (UniqueName: \"kubernetes.io/projected/410928cc-80b1-415a-a02c-83ca27897dca-kube-api-access-df4x5\") pod \"neutron-2c0f-account-create-pn6jc\" (UID: \"410928cc-80b1-415a-a02c-83ca27897dca\") " pod="openstack/neutron-2c0f-account-create-pn6jc" Nov 25 10:57:48 crc kubenswrapper[4776]: I1125 10:57:48.238234 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/410928cc-80b1-415a-a02c-83ca27897dca-operator-scripts\") pod \"neutron-2c0f-account-create-pn6jc\" (UID: \"410928cc-80b1-415a-a02c-83ca27897dca\") " pod="openstack/neutron-2c0f-account-create-pn6jc" Nov 25 10:57:48 crc kubenswrapper[4776]: I1125 10:57:48.252337 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-6t7lg" Nov 25 10:57:48 crc kubenswrapper[4776]: I1125 10:57:48.252381 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df4x5\" (UniqueName: \"kubernetes.io/projected/410928cc-80b1-415a-a02c-83ca27897dca-kube-api-access-df4x5\") pod \"neutron-2c0f-account-create-pn6jc\" (UID: \"410928cc-80b1-415a-a02c-83ca27897dca\") " pod="openstack/neutron-2c0f-account-create-pn6jc" Nov 25 10:57:48 crc kubenswrapper[4776]: I1125 10:57:48.356267 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2c0f-account-create-pn6jc" Nov 25 10:57:48 crc kubenswrapper[4776]: I1125 10:57:48.680289 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-6t7lg"] Nov 25 10:57:48 crc kubenswrapper[4776]: W1125 10:57:48.685472 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7eac8e5_4d91_4316_981c_14c58d3a261a.slice/crio-a72ec7ddae2258dfbf7d1ca390a757b73c0627ff3ead1368f8937497523d9f5c WatchSource:0}: Error finding container a72ec7ddae2258dfbf7d1ca390a757b73c0627ff3ead1368f8937497523d9f5c: Status 404 returned error can't find the container with id a72ec7ddae2258dfbf7d1ca390a757b73c0627ff3ead1368f8937497523d9f5c Nov 25 10:57:48 crc kubenswrapper[4776]: I1125 10:57:48.814942 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-2c0f-account-create-pn6jc"] Nov 25 10:57:49 crc kubenswrapper[4776]: I1125 10:57:49.629906 4776 generic.go:334] "Generic (PLEG): container finished" podID="410928cc-80b1-415a-a02c-83ca27897dca" containerID="82259d19d7e1f4c13b2c491ddc71c89165b933f1c30a6993ba43a9831cd7b6a6" exitCode=0 Nov 25 10:57:49 crc kubenswrapper[4776]: I1125 10:57:49.630023 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2c0f-account-create-pn6jc" event={"ID":"410928cc-80b1-415a-a02c-83ca27897dca","Type":"ContainerDied","Data":"82259d19d7e1f4c13b2c491ddc71c89165b933f1c30a6993ba43a9831cd7b6a6"} Nov 25 10:57:49 crc kubenswrapper[4776]: I1125 10:57:49.630291 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2c0f-account-create-pn6jc" event={"ID":"410928cc-80b1-415a-a02c-83ca27897dca","Type":"ContainerStarted","Data":"704d1d4bb4d9a4fa2dcf63144e39ce19c0bad4abab226af3c05dcc1038afa265"} Nov 25 10:57:49 crc kubenswrapper[4776]: I1125 10:57:49.632769 4776 generic.go:334] "Generic (PLEG): container finished" podID="b7eac8e5-4d91-4316-981c-14c58d3a261a" containerID="5be9efcdeacb408f3fd872d2ba344432f9c962a46ad4a8ab56fa7d01586c114c" exitCode=0 Nov 25 10:57:49 crc kubenswrapper[4776]: I1125 10:57:49.632810 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-6t7lg" event={"ID":"b7eac8e5-4d91-4316-981c-14c58d3a261a","Type":"ContainerDied","Data":"5be9efcdeacb408f3fd872d2ba344432f9c962a46ad4a8ab56fa7d01586c114c"} Nov 25 10:57:49 crc kubenswrapper[4776]: I1125 10:57:49.632837 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-6t7lg" event={"ID":"b7eac8e5-4d91-4316-981c-14c58d3a261a","Type":"ContainerStarted","Data":"a72ec7ddae2258dfbf7d1ca390a757b73c0627ff3ead1368f8937497523d9f5c"} Nov 25 10:57:51 crc kubenswrapper[4776]: I1125 10:57:51.018738 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-6t7lg" Nov 25 10:57:51 crc kubenswrapper[4776]: I1125 10:57:51.027050 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2c0f-account-create-pn6jc" Nov 25 10:57:51 crc kubenswrapper[4776]: I1125 10:57:51.089982 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/410928cc-80b1-415a-a02c-83ca27897dca-operator-scripts\") pod \"410928cc-80b1-415a-a02c-83ca27897dca\" (UID: \"410928cc-80b1-415a-a02c-83ca27897dca\") " Nov 25 10:57:51 crc kubenswrapper[4776]: I1125 10:57:51.090147 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-df4x5\" (UniqueName: \"kubernetes.io/projected/410928cc-80b1-415a-a02c-83ca27897dca-kube-api-access-df4x5\") pod \"410928cc-80b1-415a-a02c-83ca27897dca\" (UID: \"410928cc-80b1-415a-a02c-83ca27897dca\") " Nov 25 10:57:51 crc kubenswrapper[4776]: I1125 10:57:51.090378 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqchk\" (UniqueName: \"kubernetes.io/projected/b7eac8e5-4d91-4316-981c-14c58d3a261a-kube-api-access-cqchk\") pod \"b7eac8e5-4d91-4316-981c-14c58d3a261a\" (UID: \"b7eac8e5-4d91-4316-981c-14c58d3a261a\") " Nov 25 10:57:51 crc kubenswrapper[4776]: I1125 10:57:51.090461 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7eac8e5-4d91-4316-981c-14c58d3a261a-operator-scripts\") pod \"b7eac8e5-4d91-4316-981c-14c58d3a261a\" (UID: \"b7eac8e5-4d91-4316-981c-14c58d3a261a\") " Nov 25 10:57:51 crc kubenswrapper[4776]: I1125 10:57:51.091381 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7eac8e5-4d91-4316-981c-14c58d3a261a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b7eac8e5-4d91-4316-981c-14c58d3a261a" (UID: "b7eac8e5-4d91-4316-981c-14c58d3a261a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:57:51 crc kubenswrapper[4776]: I1125 10:57:51.091509 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/410928cc-80b1-415a-a02c-83ca27897dca-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "410928cc-80b1-415a-a02c-83ca27897dca" (UID: "410928cc-80b1-415a-a02c-83ca27897dca"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:57:51 crc kubenswrapper[4776]: I1125 10:57:51.099000 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7eac8e5-4d91-4316-981c-14c58d3a261a-kube-api-access-cqchk" (OuterVolumeSpecName: "kube-api-access-cqchk") pod "b7eac8e5-4d91-4316-981c-14c58d3a261a" (UID: "b7eac8e5-4d91-4316-981c-14c58d3a261a"). InnerVolumeSpecName "kube-api-access-cqchk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:57:51 crc kubenswrapper[4776]: I1125 10:57:51.100681 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/410928cc-80b1-415a-a02c-83ca27897dca-kube-api-access-df4x5" (OuterVolumeSpecName: "kube-api-access-df4x5") pod "410928cc-80b1-415a-a02c-83ca27897dca" (UID: "410928cc-80b1-415a-a02c-83ca27897dca"). InnerVolumeSpecName "kube-api-access-df4x5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:57:51 crc kubenswrapper[4776]: I1125 10:57:51.192209 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqchk\" (UniqueName: \"kubernetes.io/projected/b7eac8e5-4d91-4316-981c-14c58d3a261a-kube-api-access-cqchk\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:51 crc kubenswrapper[4776]: I1125 10:57:51.192240 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7eac8e5-4d91-4316-981c-14c58d3a261a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:51 crc kubenswrapper[4776]: I1125 10:57:51.192250 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/410928cc-80b1-415a-a02c-83ca27897dca-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:51 crc kubenswrapper[4776]: I1125 10:57:51.192259 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-df4x5\" (UniqueName: \"kubernetes.io/projected/410928cc-80b1-415a-a02c-83ca27897dca-kube-api-access-df4x5\") on node \"crc\" DevicePath \"\"" Nov 25 10:57:51 crc kubenswrapper[4776]: I1125 10:57:51.652938 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2c0f-account-create-pn6jc" event={"ID":"410928cc-80b1-415a-a02c-83ca27897dca","Type":"ContainerDied","Data":"704d1d4bb4d9a4fa2dcf63144e39ce19c0bad4abab226af3c05dcc1038afa265"} Nov 25 10:57:51 crc kubenswrapper[4776]: I1125 10:57:51.653002 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2c0f-account-create-pn6jc" Nov 25 10:57:51 crc kubenswrapper[4776]: I1125 10:57:51.653361 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="704d1d4bb4d9a4fa2dcf63144e39ce19c0bad4abab226af3c05dcc1038afa265" Nov 25 10:57:51 crc kubenswrapper[4776]: I1125 10:57:51.654720 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-6t7lg" event={"ID":"b7eac8e5-4d91-4316-981c-14c58d3a261a","Type":"ContainerDied","Data":"a72ec7ddae2258dfbf7d1ca390a757b73c0627ff3ead1368f8937497523d9f5c"} Nov 25 10:57:51 crc kubenswrapper[4776]: I1125 10:57:51.654772 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a72ec7ddae2258dfbf7d1ca390a757b73c0627ff3ead1368f8937497523d9f5c" Nov 25 10:57:51 crc kubenswrapper[4776]: I1125 10:57:51.654837 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-6t7lg" Nov 25 10:57:53 crc kubenswrapper[4776]: I1125 10:57:53.281535 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-gwjl8"] Nov 25 10:57:53 crc kubenswrapper[4776]: E1125 10:57:53.281943 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7eac8e5-4d91-4316-981c-14c58d3a261a" containerName="mariadb-database-create" Nov 25 10:57:53 crc kubenswrapper[4776]: I1125 10:57:53.281958 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7eac8e5-4d91-4316-981c-14c58d3a261a" containerName="mariadb-database-create" Nov 25 10:57:53 crc kubenswrapper[4776]: E1125 10:57:53.281991 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="410928cc-80b1-415a-a02c-83ca27897dca" containerName="mariadb-account-create" Nov 25 10:57:53 crc kubenswrapper[4776]: I1125 10:57:53.281999 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="410928cc-80b1-415a-a02c-83ca27897dca" containerName="mariadb-account-create" Nov 25 10:57:53 crc kubenswrapper[4776]: I1125 10:57:53.282204 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="410928cc-80b1-415a-a02c-83ca27897dca" containerName="mariadb-account-create" Nov 25 10:57:53 crc kubenswrapper[4776]: I1125 10:57:53.282235 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7eac8e5-4d91-4316-981c-14c58d3a261a" containerName="mariadb-database-create" Nov 25 10:57:53 crc kubenswrapper[4776]: I1125 10:57:53.282935 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-gwjl8" Nov 25 10:57:53 crc kubenswrapper[4776]: I1125 10:57:53.285223 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 25 10:57:53 crc kubenswrapper[4776]: I1125 10:57:53.285294 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 25 10:57:53 crc kubenswrapper[4776]: I1125 10:57:53.285691 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-fhgrd" Nov 25 10:57:53 crc kubenswrapper[4776]: I1125 10:57:53.296404 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-gwjl8"] Nov 25 10:57:53 crc kubenswrapper[4776]: I1125 10:57:53.332715 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c027d749-e028-4f8e-9b29-f70d6608c04f-config\") pod \"neutron-db-sync-gwjl8\" (UID: \"c027d749-e028-4f8e-9b29-f70d6608c04f\") " pod="openstack/neutron-db-sync-gwjl8" Nov 25 10:57:53 crc kubenswrapper[4776]: I1125 10:57:53.332792 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c027d749-e028-4f8e-9b29-f70d6608c04f-combined-ca-bundle\") pod \"neutron-db-sync-gwjl8\" (UID: \"c027d749-e028-4f8e-9b29-f70d6608c04f\") " pod="openstack/neutron-db-sync-gwjl8" Nov 25 10:57:53 crc kubenswrapper[4776]: I1125 10:57:53.332873 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4rkk\" (UniqueName: \"kubernetes.io/projected/c027d749-e028-4f8e-9b29-f70d6608c04f-kube-api-access-v4rkk\") pod \"neutron-db-sync-gwjl8\" (UID: \"c027d749-e028-4f8e-9b29-f70d6608c04f\") " pod="openstack/neutron-db-sync-gwjl8" Nov 25 10:57:53 crc kubenswrapper[4776]: I1125 10:57:53.434308 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c027d749-e028-4f8e-9b29-f70d6608c04f-combined-ca-bundle\") pod \"neutron-db-sync-gwjl8\" (UID: \"c027d749-e028-4f8e-9b29-f70d6608c04f\") " pod="openstack/neutron-db-sync-gwjl8" Nov 25 10:57:53 crc kubenswrapper[4776]: I1125 10:57:53.434404 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4rkk\" (UniqueName: \"kubernetes.io/projected/c027d749-e028-4f8e-9b29-f70d6608c04f-kube-api-access-v4rkk\") pod \"neutron-db-sync-gwjl8\" (UID: \"c027d749-e028-4f8e-9b29-f70d6608c04f\") " pod="openstack/neutron-db-sync-gwjl8" Nov 25 10:57:53 crc kubenswrapper[4776]: I1125 10:57:53.434473 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c027d749-e028-4f8e-9b29-f70d6608c04f-config\") pod \"neutron-db-sync-gwjl8\" (UID: \"c027d749-e028-4f8e-9b29-f70d6608c04f\") " pod="openstack/neutron-db-sync-gwjl8" Nov 25 10:57:53 crc kubenswrapper[4776]: I1125 10:57:53.440377 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c027d749-e028-4f8e-9b29-f70d6608c04f-config\") pod \"neutron-db-sync-gwjl8\" (UID: \"c027d749-e028-4f8e-9b29-f70d6608c04f\") " pod="openstack/neutron-db-sync-gwjl8" Nov 25 10:57:53 crc kubenswrapper[4776]: I1125 10:57:53.441549 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c027d749-e028-4f8e-9b29-f70d6608c04f-combined-ca-bundle\") pod \"neutron-db-sync-gwjl8\" (UID: \"c027d749-e028-4f8e-9b29-f70d6608c04f\") " pod="openstack/neutron-db-sync-gwjl8" Nov 25 10:57:53 crc kubenswrapper[4776]: I1125 10:57:53.451541 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4rkk\" (UniqueName: \"kubernetes.io/projected/c027d749-e028-4f8e-9b29-f70d6608c04f-kube-api-access-v4rkk\") pod \"neutron-db-sync-gwjl8\" (UID: \"c027d749-e028-4f8e-9b29-f70d6608c04f\") " pod="openstack/neutron-db-sync-gwjl8" Nov 25 10:57:53 crc kubenswrapper[4776]: I1125 10:57:53.601217 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-gwjl8" Nov 25 10:57:54 crc kubenswrapper[4776]: I1125 10:57:54.080540 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-gwjl8"] Nov 25 10:57:54 crc kubenswrapper[4776]: I1125 10:57:54.680127 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-gwjl8" event={"ID":"c027d749-e028-4f8e-9b29-f70d6608c04f","Type":"ContainerStarted","Data":"7fa743ab4a5cc8af2b03d5467b695a058f40063b22924ac92defa7c75c4ad680"} Nov 25 10:57:54 crc kubenswrapper[4776]: I1125 10:57:54.680171 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-gwjl8" event={"ID":"c027d749-e028-4f8e-9b29-f70d6608c04f","Type":"ContainerStarted","Data":"e182f6a57d093b8874cbd5005b46713c836e5da85f832128c43ec39b96406758"} Nov 25 10:57:54 crc kubenswrapper[4776]: I1125 10:57:54.696453 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-gwjl8" podStartSLOduration=1.696430898 podStartE2EDuration="1.696430898s" podCreationTimestamp="2025-11-25 10:57:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:57:54.695604607 +0000 UTC m=+5619.736664180" watchObservedRunningTime="2025-11-25 10:57:54.696430898 +0000 UTC m=+5619.737490441" Nov 25 10:58:00 crc kubenswrapper[4776]: I1125 10:58:00.772676 4776 generic.go:334] "Generic (PLEG): container finished" podID="c027d749-e028-4f8e-9b29-f70d6608c04f" containerID="7fa743ab4a5cc8af2b03d5467b695a058f40063b22924ac92defa7c75c4ad680" exitCode=0 Nov 25 10:58:00 crc kubenswrapper[4776]: I1125 10:58:00.772777 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-gwjl8" event={"ID":"c027d749-e028-4f8e-9b29-f70d6608c04f","Type":"ContainerDied","Data":"7fa743ab4a5cc8af2b03d5467b695a058f40063b22924ac92defa7c75c4ad680"} Nov 25 10:58:02 crc kubenswrapper[4776]: I1125 10:58:02.113132 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-gwjl8" Nov 25 10:58:02 crc kubenswrapper[4776]: I1125 10:58:02.227334 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c027d749-e028-4f8e-9b29-f70d6608c04f-config\") pod \"c027d749-e028-4f8e-9b29-f70d6608c04f\" (UID: \"c027d749-e028-4f8e-9b29-f70d6608c04f\") " Nov 25 10:58:02 crc kubenswrapper[4776]: I1125 10:58:02.227505 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4rkk\" (UniqueName: \"kubernetes.io/projected/c027d749-e028-4f8e-9b29-f70d6608c04f-kube-api-access-v4rkk\") pod \"c027d749-e028-4f8e-9b29-f70d6608c04f\" (UID: \"c027d749-e028-4f8e-9b29-f70d6608c04f\") " Nov 25 10:58:02 crc kubenswrapper[4776]: I1125 10:58:02.227533 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c027d749-e028-4f8e-9b29-f70d6608c04f-combined-ca-bundle\") pod \"c027d749-e028-4f8e-9b29-f70d6608c04f\" (UID: \"c027d749-e028-4f8e-9b29-f70d6608c04f\") " Nov 25 10:58:02 crc kubenswrapper[4776]: I1125 10:58:02.234228 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c027d749-e028-4f8e-9b29-f70d6608c04f-kube-api-access-v4rkk" (OuterVolumeSpecName: "kube-api-access-v4rkk") pod "c027d749-e028-4f8e-9b29-f70d6608c04f" (UID: "c027d749-e028-4f8e-9b29-f70d6608c04f"). InnerVolumeSpecName "kube-api-access-v4rkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:58:02 crc kubenswrapper[4776]: I1125 10:58:02.260132 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c027d749-e028-4f8e-9b29-f70d6608c04f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c027d749-e028-4f8e-9b29-f70d6608c04f" (UID: "c027d749-e028-4f8e-9b29-f70d6608c04f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:02 crc kubenswrapper[4776]: I1125 10:58:02.261411 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c027d749-e028-4f8e-9b29-f70d6608c04f-config" (OuterVolumeSpecName: "config") pod "c027d749-e028-4f8e-9b29-f70d6608c04f" (UID: "c027d749-e028-4f8e-9b29-f70d6608c04f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:02 crc kubenswrapper[4776]: I1125 10:58:02.329674 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c027d749-e028-4f8e-9b29-f70d6608c04f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:02 crc kubenswrapper[4776]: I1125 10:58:02.329729 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/c027d749-e028-4f8e-9b29-f70d6608c04f-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:02 crc kubenswrapper[4776]: I1125 10:58:02.329751 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4rkk\" (UniqueName: \"kubernetes.io/projected/c027d749-e028-4f8e-9b29-f70d6608c04f-kube-api-access-v4rkk\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:02 crc kubenswrapper[4776]: I1125 10:58:02.794797 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-gwjl8" event={"ID":"c027d749-e028-4f8e-9b29-f70d6608c04f","Type":"ContainerDied","Data":"e182f6a57d093b8874cbd5005b46713c836e5da85f832128c43ec39b96406758"} Nov 25 10:58:02 crc kubenswrapper[4776]: I1125 10:58:02.794843 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e182f6a57d093b8874cbd5005b46713c836e5da85f832128c43ec39b96406758" Nov 25 10:58:02 crc kubenswrapper[4776]: I1125 10:58:02.794897 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-gwjl8" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.033839 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d5c556d87-vf246"] Nov 25 10:58:03 crc kubenswrapper[4776]: E1125 10:58:03.034267 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c027d749-e028-4f8e-9b29-f70d6608c04f" containerName="neutron-db-sync" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.034291 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="c027d749-e028-4f8e-9b29-f70d6608c04f" containerName="neutron-db-sync" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.034484 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="c027d749-e028-4f8e-9b29-f70d6608c04f" containerName="neutron-db-sync" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.037161 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d5c556d87-vf246" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.058046 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d5c556d87-vf246"] Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.146192 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-dns-svc\") pod \"dnsmasq-dns-5d5c556d87-vf246\" (UID: \"7fdc4742-4288-4756-8dfd-3f3f0d2174e2\") " pod="openstack/dnsmasq-dns-5d5c556d87-vf246" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.146239 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-ovsdbserver-sb\") pod \"dnsmasq-dns-5d5c556d87-vf246\" (UID: \"7fdc4742-4288-4756-8dfd-3f3f0d2174e2\") " pod="openstack/dnsmasq-dns-5d5c556d87-vf246" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.146279 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-ovsdbserver-nb\") pod \"dnsmasq-dns-5d5c556d87-vf246\" (UID: \"7fdc4742-4288-4756-8dfd-3f3f0d2174e2\") " pod="openstack/dnsmasq-dns-5d5c556d87-vf246" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.146325 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6zsc\" (UniqueName: \"kubernetes.io/projected/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-kube-api-access-l6zsc\") pod \"dnsmasq-dns-5d5c556d87-vf246\" (UID: \"7fdc4742-4288-4756-8dfd-3f3f0d2174e2\") " pod="openstack/dnsmasq-dns-5d5c556d87-vf246" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.146346 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-config\") pod \"dnsmasq-dns-5d5c556d87-vf246\" (UID: \"7fdc4742-4288-4756-8dfd-3f3f0d2174e2\") " pod="openstack/dnsmasq-dns-5d5c556d87-vf246" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.190550 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-8757674c4-gwcws"] Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.192466 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8757674c4-gwcws" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.194753 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-fhgrd" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.194939 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.195123 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.195854 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.212390 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8757674c4-gwcws"] Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.247747 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-dns-svc\") pod \"dnsmasq-dns-5d5c556d87-vf246\" (UID: \"7fdc4742-4288-4756-8dfd-3f3f0d2174e2\") " pod="openstack/dnsmasq-dns-5d5c556d87-vf246" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.247793 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-ovsdbserver-sb\") pod \"dnsmasq-dns-5d5c556d87-vf246\" (UID: \"7fdc4742-4288-4756-8dfd-3f3f0d2174e2\") " pod="openstack/dnsmasq-dns-5d5c556d87-vf246" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.247823 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-ovsdbserver-nb\") pod \"dnsmasq-dns-5d5c556d87-vf246\" (UID: \"7fdc4742-4288-4756-8dfd-3f3f0d2174e2\") " pod="openstack/dnsmasq-dns-5d5c556d87-vf246" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.247868 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6zsc\" (UniqueName: \"kubernetes.io/projected/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-kube-api-access-l6zsc\") pod \"dnsmasq-dns-5d5c556d87-vf246\" (UID: \"7fdc4742-4288-4756-8dfd-3f3f0d2174e2\") " pod="openstack/dnsmasq-dns-5d5c556d87-vf246" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.247892 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-config\") pod \"dnsmasq-dns-5d5c556d87-vf246\" (UID: \"7fdc4742-4288-4756-8dfd-3f3f0d2174e2\") " pod="openstack/dnsmasq-dns-5d5c556d87-vf246" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.248949 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-config\") pod \"dnsmasq-dns-5d5c556d87-vf246\" (UID: \"7fdc4742-4288-4756-8dfd-3f3f0d2174e2\") " pod="openstack/dnsmasq-dns-5d5c556d87-vf246" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.249646 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-ovsdbserver-nb\") pod \"dnsmasq-dns-5d5c556d87-vf246\" (UID: \"7fdc4742-4288-4756-8dfd-3f3f0d2174e2\") " pod="openstack/dnsmasq-dns-5d5c556d87-vf246" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.249672 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-dns-svc\") pod \"dnsmasq-dns-5d5c556d87-vf246\" (UID: \"7fdc4742-4288-4756-8dfd-3f3f0d2174e2\") " pod="openstack/dnsmasq-dns-5d5c556d87-vf246" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.250192 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-ovsdbserver-sb\") pod \"dnsmasq-dns-5d5c556d87-vf246\" (UID: \"7fdc4742-4288-4756-8dfd-3f3f0d2174e2\") " pod="openstack/dnsmasq-dns-5d5c556d87-vf246" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.274864 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6zsc\" (UniqueName: \"kubernetes.io/projected/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-kube-api-access-l6zsc\") pod \"dnsmasq-dns-5d5c556d87-vf246\" (UID: \"7fdc4742-4288-4756-8dfd-3f3f0d2174e2\") " pod="openstack/dnsmasq-dns-5d5c556d87-vf246" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.349039 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwvnx\" (UniqueName: \"kubernetes.io/projected/1bcfb866-9492-4db9-b843-404306f8fb01-kube-api-access-pwvnx\") pod \"neutron-8757674c4-gwcws\" (UID: \"1bcfb866-9492-4db9-b843-404306f8fb01\") " pod="openstack/neutron-8757674c4-gwcws" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.349114 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1bcfb866-9492-4db9-b843-404306f8fb01-httpd-config\") pod \"neutron-8757674c4-gwcws\" (UID: \"1bcfb866-9492-4db9-b843-404306f8fb01\") " pod="openstack/neutron-8757674c4-gwcws" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.349142 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bcfb866-9492-4db9-b843-404306f8fb01-ovndb-tls-certs\") pod \"neutron-8757674c4-gwcws\" (UID: \"1bcfb866-9492-4db9-b843-404306f8fb01\") " pod="openstack/neutron-8757674c4-gwcws" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.349165 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1bcfb866-9492-4db9-b843-404306f8fb01-config\") pod \"neutron-8757674c4-gwcws\" (UID: \"1bcfb866-9492-4db9-b843-404306f8fb01\") " pod="openstack/neutron-8757674c4-gwcws" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.349418 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bcfb866-9492-4db9-b843-404306f8fb01-combined-ca-bundle\") pod \"neutron-8757674c4-gwcws\" (UID: \"1bcfb866-9492-4db9-b843-404306f8fb01\") " pod="openstack/neutron-8757674c4-gwcws" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.359336 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d5c556d87-vf246" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.453087 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1bcfb866-9492-4db9-b843-404306f8fb01-httpd-config\") pod \"neutron-8757674c4-gwcws\" (UID: \"1bcfb866-9492-4db9-b843-404306f8fb01\") " pod="openstack/neutron-8757674c4-gwcws" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.453136 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bcfb866-9492-4db9-b843-404306f8fb01-ovndb-tls-certs\") pod \"neutron-8757674c4-gwcws\" (UID: \"1bcfb866-9492-4db9-b843-404306f8fb01\") " pod="openstack/neutron-8757674c4-gwcws" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.453167 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1bcfb866-9492-4db9-b843-404306f8fb01-config\") pod \"neutron-8757674c4-gwcws\" (UID: \"1bcfb866-9492-4db9-b843-404306f8fb01\") " pod="openstack/neutron-8757674c4-gwcws" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.453292 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bcfb866-9492-4db9-b843-404306f8fb01-combined-ca-bundle\") pod \"neutron-8757674c4-gwcws\" (UID: \"1bcfb866-9492-4db9-b843-404306f8fb01\") " pod="openstack/neutron-8757674c4-gwcws" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.453394 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwvnx\" (UniqueName: \"kubernetes.io/projected/1bcfb866-9492-4db9-b843-404306f8fb01-kube-api-access-pwvnx\") pod \"neutron-8757674c4-gwcws\" (UID: \"1bcfb866-9492-4db9-b843-404306f8fb01\") " pod="openstack/neutron-8757674c4-gwcws" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.467016 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bcfb866-9492-4db9-b843-404306f8fb01-ovndb-tls-certs\") pod \"neutron-8757674c4-gwcws\" (UID: \"1bcfb866-9492-4db9-b843-404306f8fb01\") " pod="openstack/neutron-8757674c4-gwcws" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.468525 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/1bcfb866-9492-4db9-b843-404306f8fb01-config\") pod \"neutron-8757674c4-gwcws\" (UID: \"1bcfb866-9492-4db9-b843-404306f8fb01\") " pod="openstack/neutron-8757674c4-gwcws" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.469966 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bcfb866-9492-4db9-b843-404306f8fb01-combined-ca-bundle\") pod \"neutron-8757674c4-gwcws\" (UID: \"1bcfb866-9492-4db9-b843-404306f8fb01\") " pod="openstack/neutron-8757674c4-gwcws" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.470606 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1bcfb866-9492-4db9-b843-404306f8fb01-httpd-config\") pod \"neutron-8757674c4-gwcws\" (UID: \"1bcfb866-9492-4db9-b843-404306f8fb01\") " pod="openstack/neutron-8757674c4-gwcws" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.472019 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwvnx\" (UniqueName: \"kubernetes.io/projected/1bcfb866-9492-4db9-b843-404306f8fb01-kube-api-access-pwvnx\") pod \"neutron-8757674c4-gwcws\" (UID: \"1bcfb866-9492-4db9-b843-404306f8fb01\") " pod="openstack/neutron-8757674c4-gwcws" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.510608 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8757674c4-gwcws" Nov 25 10:58:03 crc kubenswrapper[4776]: I1125 10:58:03.873581 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d5c556d87-vf246"] Nov 25 10:58:04 crc kubenswrapper[4776]: I1125 10:58:04.087902 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8757674c4-gwcws"] Nov 25 10:58:04 crc kubenswrapper[4776]: W1125 10:58:04.098724 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1bcfb866_9492_4db9_b843_404306f8fb01.slice/crio-26c65ad02b16f0b3ace827528de36885b590949ff2c0412373572719419c31cb WatchSource:0}: Error finding container 26c65ad02b16f0b3ace827528de36885b590949ff2c0412373572719419c31cb: Status 404 returned error can't find the container with id 26c65ad02b16f0b3ace827528de36885b590949ff2c0412373572719419c31cb Nov 25 10:58:04 crc kubenswrapper[4776]: I1125 10:58:04.809720 4776 generic.go:334] "Generic (PLEG): container finished" podID="7fdc4742-4288-4756-8dfd-3f3f0d2174e2" containerID="0fec0a1c6a9f05a8bee620a8b81d7e77e7f5245e6ec480c3bb1babf947c2091c" exitCode=0 Nov 25 10:58:04 crc kubenswrapper[4776]: I1125 10:58:04.809817 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d5c556d87-vf246" event={"ID":"7fdc4742-4288-4756-8dfd-3f3f0d2174e2","Type":"ContainerDied","Data":"0fec0a1c6a9f05a8bee620a8b81d7e77e7f5245e6ec480c3bb1babf947c2091c"} Nov 25 10:58:04 crc kubenswrapper[4776]: I1125 10:58:04.810035 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d5c556d87-vf246" event={"ID":"7fdc4742-4288-4756-8dfd-3f3f0d2174e2","Type":"ContainerStarted","Data":"aafebdf2e101ad2a8a6a5a09e9273cb8e9da432f29abc8692d0d0b5cf699d1a6"} Nov 25 10:58:04 crc kubenswrapper[4776]: I1125 10:58:04.818627 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8757674c4-gwcws" event={"ID":"1bcfb866-9492-4db9-b843-404306f8fb01","Type":"ContainerStarted","Data":"ccf1c362e5da0b9a7574b6bfb35a482f32cd9a4fd7d7027b84e6f6662603ef54"} Nov 25 10:58:04 crc kubenswrapper[4776]: I1125 10:58:04.818675 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8757674c4-gwcws" event={"ID":"1bcfb866-9492-4db9-b843-404306f8fb01","Type":"ContainerStarted","Data":"a2c3be559b1055faa120b087e39b0ef9ff0642f312abe7091eeaeb7c77711a49"} Nov 25 10:58:04 crc kubenswrapper[4776]: I1125 10:58:04.818688 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8757674c4-gwcws" event={"ID":"1bcfb866-9492-4db9-b843-404306f8fb01","Type":"ContainerStarted","Data":"26c65ad02b16f0b3ace827528de36885b590949ff2c0412373572719419c31cb"} Nov 25 10:58:04 crc kubenswrapper[4776]: I1125 10:58:04.819574 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-8757674c4-gwcws" Nov 25 10:58:04 crc kubenswrapper[4776]: I1125 10:58:04.863833 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-8757674c4-gwcws" podStartSLOduration=1.863809175 podStartE2EDuration="1.863809175s" podCreationTimestamp="2025-11-25 10:58:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:58:04.859520217 +0000 UTC m=+5629.900579760" watchObservedRunningTime="2025-11-25 10:58:04.863809175 +0000 UTC m=+5629.904868738" Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.324345 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7db9d4c697-g5n67"] Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.326475 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7db9d4c697-g5n67" Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.331751 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.331846 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.398732 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xj7xv\" (UniqueName: \"kubernetes.io/projected/79fed1e6-a706-4827-85e1-1d46cd4fe514-kube-api-access-xj7xv\") pod \"neutron-7db9d4c697-g5n67\" (UID: \"79fed1e6-a706-4827-85e1-1d46cd4fe514\") " pod="openstack/neutron-7db9d4c697-g5n67" Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.398799 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/79fed1e6-a706-4827-85e1-1d46cd4fe514-config\") pod \"neutron-7db9d4c697-g5n67\" (UID: \"79fed1e6-a706-4827-85e1-1d46cd4fe514\") " pod="openstack/neutron-7db9d4c697-g5n67" Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.398849 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/79fed1e6-a706-4827-85e1-1d46cd4fe514-ovndb-tls-certs\") pod \"neutron-7db9d4c697-g5n67\" (UID: \"79fed1e6-a706-4827-85e1-1d46cd4fe514\") " pod="openstack/neutron-7db9d4c697-g5n67" Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.398878 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/79fed1e6-a706-4827-85e1-1d46cd4fe514-internal-tls-certs\") pod \"neutron-7db9d4c697-g5n67\" (UID: \"79fed1e6-a706-4827-85e1-1d46cd4fe514\") " pod="openstack/neutron-7db9d4c697-g5n67" Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.398903 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/79fed1e6-a706-4827-85e1-1d46cd4fe514-httpd-config\") pod \"neutron-7db9d4c697-g5n67\" (UID: \"79fed1e6-a706-4827-85e1-1d46cd4fe514\") " pod="openstack/neutron-7db9d4c697-g5n67" Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.398925 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/79fed1e6-a706-4827-85e1-1d46cd4fe514-public-tls-certs\") pod \"neutron-7db9d4c697-g5n67\" (UID: \"79fed1e6-a706-4827-85e1-1d46cd4fe514\") " pod="openstack/neutron-7db9d4c697-g5n67" Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.398967 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79fed1e6-a706-4827-85e1-1d46cd4fe514-combined-ca-bundle\") pod \"neutron-7db9d4c697-g5n67\" (UID: \"79fed1e6-a706-4827-85e1-1d46cd4fe514\") " pod="openstack/neutron-7db9d4c697-g5n67" Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.402674 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7db9d4c697-g5n67"] Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.499734 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/79fed1e6-a706-4827-85e1-1d46cd4fe514-ovndb-tls-certs\") pod \"neutron-7db9d4c697-g5n67\" (UID: \"79fed1e6-a706-4827-85e1-1d46cd4fe514\") " pod="openstack/neutron-7db9d4c697-g5n67" Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.499803 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/79fed1e6-a706-4827-85e1-1d46cd4fe514-internal-tls-certs\") pod \"neutron-7db9d4c697-g5n67\" (UID: \"79fed1e6-a706-4827-85e1-1d46cd4fe514\") " pod="openstack/neutron-7db9d4c697-g5n67" Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.499839 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/79fed1e6-a706-4827-85e1-1d46cd4fe514-httpd-config\") pod \"neutron-7db9d4c697-g5n67\" (UID: \"79fed1e6-a706-4827-85e1-1d46cd4fe514\") " pod="openstack/neutron-7db9d4c697-g5n67" Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.499868 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/79fed1e6-a706-4827-85e1-1d46cd4fe514-public-tls-certs\") pod \"neutron-7db9d4c697-g5n67\" (UID: \"79fed1e6-a706-4827-85e1-1d46cd4fe514\") " pod="openstack/neutron-7db9d4c697-g5n67" Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.499923 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79fed1e6-a706-4827-85e1-1d46cd4fe514-combined-ca-bundle\") pod \"neutron-7db9d4c697-g5n67\" (UID: \"79fed1e6-a706-4827-85e1-1d46cd4fe514\") " pod="openstack/neutron-7db9d4c697-g5n67" Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.499981 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xj7xv\" (UniqueName: \"kubernetes.io/projected/79fed1e6-a706-4827-85e1-1d46cd4fe514-kube-api-access-xj7xv\") pod \"neutron-7db9d4c697-g5n67\" (UID: \"79fed1e6-a706-4827-85e1-1d46cd4fe514\") " pod="openstack/neutron-7db9d4c697-g5n67" Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.500012 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/79fed1e6-a706-4827-85e1-1d46cd4fe514-config\") pod \"neutron-7db9d4c697-g5n67\" (UID: \"79fed1e6-a706-4827-85e1-1d46cd4fe514\") " pod="openstack/neutron-7db9d4c697-g5n67" Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.505972 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79fed1e6-a706-4827-85e1-1d46cd4fe514-combined-ca-bundle\") pod \"neutron-7db9d4c697-g5n67\" (UID: \"79fed1e6-a706-4827-85e1-1d46cd4fe514\") " pod="openstack/neutron-7db9d4c697-g5n67" Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.506551 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/79fed1e6-a706-4827-85e1-1d46cd4fe514-ovndb-tls-certs\") pod \"neutron-7db9d4c697-g5n67\" (UID: \"79fed1e6-a706-4827-85e1-1d46cd4fe514\") " pod="openstack/neutron-7db9d4c697-g5n67" Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.510141 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/79fed1e6-a706-4827-85e1-1d46cd4fe514-config\") pod \"neutron-7db9d4c697-g5n67\" (UID: \"79fed1e6-a706-4827-85e1-1d46cd4fe514\") " pod="openstack/neutron-7db9d4c697-g5n67" Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.510824 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/79fed1e6-a706-4827-85e1-1d46cd4fe514-internal-tls-certs\") pod \"neutron-7db9d4c697-g5n67\" (UID: \"79fed1e6-a706-4827-85e1-1d46cd4fe514\") " pod="openstack/neutron-7db9d4c697-g5n67" Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.512308 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/79fed1e6-a706-4827-85e1-1d46cd4fe514-public-tls-certs\") pod \"neutron-7db9d4c697-g5n67\" (UID: \"79fed1e6-a706-4827-85e1-1d46cd4fe514\") " pod="openstack/neutron-7db9d4c697-g5n67" Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.512793 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/79fed1e6-a706-4827-85e1-1d46cd4fe514-httpd-config\") pod \"neutron-7db9d4c697-g5n67\" (UID: \"79fed1e6-a706-4827-85e1-1d46cd4fe514\") " pod="openstack/neutron-7db9d4c697-g5n67" Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.519314 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xj7xv\" (UniqueName: \"kubernetes.io/projected/79fed1e6-a706-4827-85e1-1d46cd4fe514-kube-api-access-xj7xv\") pod \"neutron-7db9d4c697-g5n67\" (UID: \"79fed1e6-a706-4827-85e1-1d46cd4fe514\") " pod="openstack/neutron-7db9d4c697-g5n67" Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.659418 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7db9d4c697-g5n67" Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.832279 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d5c556d87-vf246" event={"ID":"7fdc4742-4288-4756-8dfd-3f3f0d2174e2","Type":"ContainerStarted","Data":"f46bb4d4a6a370a5680b4d71418cbdc0fa0aed30d3492279f265d8dfb0aa84f1"} Nov 25 10:58:05 crc kubenswrapper[4776]: I1125 10:58:05.854693 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d5c556d87-vf246" podStartSLOduration=2.854678676 podStartE2EDuration="2.854678676s" podCreationTimestamp="2025-11-25 10:58:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:58:05.849868825 +0000 UTC m=+5630.890928378" watchObservedRunningTime="2025-11-25 10:58:05.854678676 +0000 UTC m=+5630.895738229" Nov 25 10:58:06 crc kubenswrapper[4776]: I1125 10:58:06.028437 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7db9d4c697-g5n67"] Nov 25 10:58:06 crc kubenswrapper[4776]: W1125 10:58:06.033152 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79fed1e6_a706_4827_85e1_1d46cd4fe514.slice/crio-9a5eeb1954eaf08c8cc7ddd7bf3896c1d1ee341cdfbf758f8e755b510377b165 WatchSource:0}: Error finding container 9a5eeb1954eaf08c8cc7ddd7bf3896c1d1ee341cdfbf758f8e755b510377b165: Status 404 returned error can't find the container with id 9a5eeb1954eaf08c8cc7ddd7bf3896c1d1ee341cdfbf758f8e755b510377b165 Nov 25 10:58:06 crc kubenswrapper[4776]: I1125 10:58:06.841462 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7db9d4c697-g5n67" event={"ID":"79fed1e6-a706-4827-85e1-1d46cd4fe514","Type":"ContainerStarted","Data":"75cef86d1161195369701d853de9a77d5a4a1abff34f61b1d0d6bb5f5b3b12ec"} Nov 25 10:58:06 crc kubenswrapper[4776]: I1125 10:58:06.841768 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d5c556d87-vf246" Nov 25 10:58:06 crc kubenswrapper[4776]: I1125 10:58:06.841782 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7db9d4c697-g5n67" Nov 25 10:58:06 crc kubenswrapper[4776]: I1125 10:58:06.841791 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7db9d4c697-g5n67" event={"ID":"79fed1e6-a706-4827-85e1-1d46cd4fe514","Type":"ContainerStarted","Data":"abd6b53c610685d82b54b888c2a74e88a75f767a705fbfe81b956fe97d60316c"} Nov 25 10:58:06 crc kubenswrapper[4776]: I1125 10:58:06.841800 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7db9d4c697-g5n67" event={"ID":"79fed1e6-a706-4827-85e1-1d46cd4fe514","Type":"ContainerStarted","Data":"9a5eeb1954eaf08c8cc7ddd7bf3896c1d1ee341cdfbf758f8e755b510377b165"} Nov 25 10:58:06 crc kubenswrapper[4776]: I1125 10:58:06.870765 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7db9d4c697-g5n67" podStartSLOduration=1.87074154 podStartE2EDuration="1.87074154s" podCreationTimestamp="2025-11-25 10:58:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:58:06.862842422 +0000 UTC m=+5631.903901975" watchObservedRunningTime="2025-11-25 10:58:06.87074154 +0000 UTC m=+5631.911801093" Nov 25 10:58:13 crc kubenswrapper[4776]: I1125 10:58:13.361187 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d5c556d87-vf246" Nov 25 10:58:13 crc kubenswrapper[4776]: I1125 10:58:13.418350 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-658f7f865f-fkgks"] Nov 25 10:58:13 crc kubenswrapper[4776]: I1125 10:58:13.418655 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-658f7f865f-fkgks" podUID="5161e133-bd4f-436d-8d36-cbcc7faccf9a" containerName="dnsmasq-dns" containerID="cri-o://c2625123e156e2f6f7330db6834dc55ed01bc5ba1d5281595e746092ee6edb11" gracePeriod=10 Nov 25 10:58:13 crc kubenswrapper[4776]: I1125 10:58:13.900403 4776 generic.go:334] "Generic (PLEG): container finished" podID="5161e133-bd4f-436d-8d36-cbcc7faccf9a" containerID="c2625123e156e2f6f7330db6834dc55ed01bc5ba1d5281595e746092ee6edb11" exitCode=0 Nov 25 10:58:13 crc kubenswrapper[4776]: I1125 10:58:13.900586 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-658f7f865f-fkgks" event={"ID":"5161e133-bd4f-436d-8d36-cbcc7faccf9a","Type":"ContainerDied","Data":"c2625123e156e2f6f7330db6834dc55ed01bc5ba1d5281595e746092ee6edb11"} Nov 25 10:58:13 crc kubenswrapper[4776]: I1125 10:58:13.900751 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-658f7f865f-fkgks" event={"ID":"5161e133-bd4f-436d-8d36-cbcc7faccf9a","Type":"ContainerDied","Data":"dcd2e47687f705e9afa7f9fee5c824d9c3535b1daba652fa126699841123e770"} Nov 25 10:58:13 crc kubenswrapper[4776]: I1125 10:58:13.900767 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dcd2e47687f705e9afa7f9fee5c824d9c3535b1daba652fa126699841123e770" Nov 25 10:58:13 crc kubenswrapper[4776]: I1125 10:58:13.931352 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-658f7f865f-fkgks" Nov 25 10:58:14 crc kubenswrapper[4776]: I1125 10:58:14.098431 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5161e133-bd4f-436d-8d36-cbcc7faccf9a-ovsdbserver-nb\") pod \"5161e133-bd4f-436d-8d36-cbcc7faccf9a\" (UID: \"5161e133-bd4f-436d-8d36-cbcc7faccf9a\") " Nov 25 10:58:14 crc kubenswrapper[4776]: I1125 10:58:14.098511 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5161e133-bd4f-436d-8d36-cbcc7faccf9a-ovsdbserver-sb\") pod \"5161e133-bd4f-436d-8d36-cbcc7faccf9a\" (UID: \"5161e133-bd4f-436d-8d36-cbcc7faccf9a\") " Nov 25 10:58:14 crc kubenswrapper[4776]: I1125 10:58:14.098547 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5161e133-bd4f-436d-8d36-cbcc7faccf9a-dns-svc\") pod \"5161e133-bd4f-436d-8d36-cbcc7faccf9a\" (UID: \"5161e133-bd4f-436d-8d36-cbcc7faccf9a\") " Nov 25 10:58:14 crc kubenswrapper[4776]: I1125 10:58:14.098649 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5161e133-bd4f-436d-8d36-cbcc7faccf9a-config\") pod \"5161e133-bd4f-436d-8d36-cbcc7faccf9a\" (UID: \"5161e133-bd4f-436d-8d36-cbcc7faccf9a\") " Nov 25 10:58:14 crc kubenswrapper[4776]: I1125 10:58:14.098703 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsmxl\" (UniqueName: \"kubernetes.io/projected/5161e133-bd4f-436d-8d36-cbcc7faccf9a-kube-api-access-wsmxl\") pod \"5161e133-bd4f-436d-8d36-cbcc7faccf9a\" (UID: \"5161e133-bd4f-436d-8d36-cbcc7faccf9a\") " Nov 25 10:58:14 crc kubenswrapper[4776]: I1125 10:58:14.110378 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5161e133-bd4f-436d-8d36-cbcc7faccf9a-kube-api-access-wsmxl" (OuterVolumeSpecName: "kube-api-access-wsmxl") pod "5161e133-bd4f-436d-8d36-cbcc7faccf9a" (UID: "5161e133-bd4f-436d-8d36-cbcc7faccf9a"). InnerVolumeSpecName "kube-api-access-wsmxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:58:14 crc kubenswrapper[4776]: I1125 10:58:14.144466 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5161e133-bd4f-436d-8d36-cbcc7faccf9a-config" (OuterVolumeSpecName: "config") pod "5161e133-bd4f-436d-8d36-cbcc7faccf9a" (UID: "5161e133-bd4f-436d-8d36-cbcc7faccf9a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:58:14 crc kubenswrapper[4776]: I1125 10:58:14.145232 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5161e133-bd4f-436d-8d36-cbcc7faccf9a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5161e133-bd4f-436d-8d36-cbcc7faccf9a" (UID: "5161e133-bd4f-436d-8d36-cbcc7faccf9a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:58:14 crc kubenswrapper[4776]: I1125 10:58:14.147786 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5161e133-bd4f-436d-8d36-cbcc7faccf9a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5161e133-bd4f-436d-8d36-cbcc7faccf9a" (UID: "5161e133-bd4f-436d-8d36-cbcc7faccf9a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:58:14 crc kubenswrapper[4776]: I1125 10:58:14.174608 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5161e133-bd4f-436d-8d36-cbcc7faccf9a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5161e133-bd4f-436d-8d36-cbcc7faccf9a" (UID: "5161e133-bd4f-436d-8d36-cbcc7faccf9a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:58:14 crc kubenswrapper[4776]: I1125 10:58:14.201120 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5161e133-bd4f-436d-8d36-cbcc7faccf9a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:14 crc kubenswrapper[4776]: I1125 10:58:14.201334 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5161e133-bd4f-436d-8d36-cbcc7faccf9a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:14 crc kubenswrapper[4776]: I1125 10:58:14.201396 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5161e133-bd4f-436d-8d36-cbcc7faccf9a-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:14 crc kubenswrapper[4776]: I1125 10:58:14.201464 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5161e133-bd4f-436d-8d36-cbcc7faccf9a-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:14 crc kubenswrapper[4776]: I1125 10:58:14.201524 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsmxl\" (UniqueName: \"kubernetes.io/projected/5161e133-bd4f-436d-8d36-cbcc7faccf9a-kube-api-access-wsmxl\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:14 crc kubenswrapper[4776]: I1125 10:58:14.908238 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-658f7f865f-fkgks" Nov 25 10:58:14 crc kubenswrapper[4776]: I1125 10:58:14.939400 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-658f7f865f-fkgks"] Nov 25 10:58:14 crc kubenswrapper[4776]: I1125 10:58:14.948494 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-658f7f865f-fkgks"] Nov 25 10:58:15 crc kubenswrapper[4776]: I1125 10:58:15.676829 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5161e133-bd4f-436d-8d36-cbcc7faccf9a" path="/var/lib/kubelet/pods/5161e133-bd4f-436d-8d36-cbcc7faccf9a/volumes" Nov 25 10:58:33 crc kubenswrapper[4776]: I1125 10:58:33.519122 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-8757674c4-gwcws" Nov 25 10:58:35 crc kubenswrapper[4776]: I1125 10:58:35.672178 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7db9d4c697-g5n67" Nov 25 10:58:35 crc kubenswrapper[4776]: I1125 10:58:35.731009 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-8757674c4-gwcws"] Nov 25 10:58:35 crc kubenswrapper[4776]: I1125 10:58:35.731883 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-8757674c4-gwcws" podUID="1bcfb866-9492-4db9-b843-404306f8fb01" containerName="neutron-api" containerID="cri-o://a2c3be559b1055faa120b087e39b0ef9ff0642f312abe7091eeaeb7c77711a49" gracePeriod=30 Nov 25 10:58:35 crc kubenswrapper[4776]: I1125 10:58:35.731951 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-8757674c4-gwcws" podUID="1bcfb866-9492-4db9-b843-404306f8fb01" containerName="neutron-httpd" containerID="cri-o://ccf1c362e5da0b9a7574b6bfb35a482f32cd9a4fd7d7027b84e6f6662603ef54" gracePeriod=30 Nov 25 10:58:36 crc kubenswrapper[4776]: I1125 10:58:36.092357 4776 generic.go:334] "Generic (PLEG): container finished" podID="1bcfb866-9492-4db9-b843-404306f8fb01" containerID="ccf1c362e5da0b9a7574b6bfb35a482f32cd9a4fd7d7027b84e6f6662603ef54" exitCode=0 Nov 25 10:58:36 crc kubenswrapper[4776]: I1125 10:58:36.092435 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8757674c4-gwcws" event={"ID":"1bcfb866-9492-4db9-b843-404306f8fb01","Type":"ContainerDied","Data":"ccf1c362e5da0b9a7574b6bfb35a482f32cd9a4fd7d7027b84e6f6662603ef54"} Nov 25 10:58:39 crc kubenswrapper[4776]: I1125 10:58:39.701287 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8757674c4-gwcws" Nov 25 10:58:39 crc kubenswrapper[4776]: I1125 10:58:39.780724 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bcfb866-9492-4db9-b843-404306f8fb01-combined-ca-bundle\") pod \"1bcfb866-9492-4db9-b843-404306f8fb01\" (UID: \"1bcfb866-9492-4db9-b843-404306f8fb01\") " Nov 25 10:58:39 crc kubenswrapper[4776]: I1125 10:58:39.781166 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1bcfb866-9492-4db9-b843-404306f8fb01-httpd-config\") pod \"1bcfb866-9492-4db9-b843-404306f8fb01\" (UID: \"1bcfb866-9492-4db9-b843-404306f8fb01\") " Nov 25 10:58:39 crc kubenswrapper[4776]: I1125 10:58:39.781195 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bcfb866-9492-4db9-b843-404306f8fb01-ovndb-tls-certs\") pod \"1bcfb866-9492-4db9-b843-404306f8fb01\" (UID: \"1bcfb866-9492-4db9-b843-404306f8fb01\") " Nov 25 10:58:39 crc kubenswrapper[4776]: I1125 10:58:39.781269 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwvnx\" (UniqueName: \"kubernetes.io/projected/1bcfb866-9492-4db9-b843-404306f8fb01-kube-api-access-pwvnx\") pod \"1bcfb866-9492-4db9-b843-404306f8fb01\" (UID: \"1bcfb866-9492-4db9-b843-404306f8fb01\") " Nov 25 10:58:39 crc kubenswrapper[4776]: I1125 10:58:39.781368 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1bcfb866-9492-4db9-b843-404306f8fb01-config\") pod \"1bcfb866-9492-4db9-b843-404306f8fb01\" (UID: \"1bcfb866-9492-4db9-b843-404306f8fb01\") " Nov 25 10:58:39 crc kubenswrapper[4776]: I1125 10:58:39.786763 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bcfb866-9492-4db9-b843-404306f8fb01-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "1bcfb866-9492-4db9-b843-404306f8fb01" (UID: "1bcfb866-9492-4db9-b843-404306f8fb01"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:39 crc kubenswrapper[4776]: I1125 10:58:39.786958 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bcfb866-9492-4db9-b843-404306f8fb01-kube-api-access-pwvnx" (OuterVolumeSpecName: "kube-api-access-pwvnx") pod "1bcfb866-9492-4db9-b843-404306f8fb01" (UID: "1bcfb866-9492-4db9-b843-404306f8fb01"). InnerVolumeSpecName "kube-api-access-pwvnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:58:39 crc kubenswrapper[4776]: I1125 10:58:39.830015 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bcfb866-9492-4db9-b843-404306f8fb01-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1bcfb866-9492-4db9-b843-404306f8fb01" (UID: "1bcfb866-9492-4db9-b843-404306f8fb01"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:39 crc kubenswrapper[4776]: I1125 10:58:39.830403 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bcfb866-9492-4db9-b843-404306f8fb01-config" (OuterVolumeSpecName: "config") pod "1bcfb866-9492-4db9-b843-404306f8fb01" (UID: "1bcfb866-9492-4db9-b843-404306f8fb01"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:39 crc kubenswrapper[4776]: I1125 10:58:39.853327 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bcfb866-9492-4db9-b843-404306f8fb01-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "1bcfb866-9492-4db9-b843-404306f8fb01" (UID: "1bcfb866-9492-4db9-b843-404306f8fb01"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:58:39 crc kubenswrapper[4776]: I1125 10:58:39.882768 4776 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1bcfb866-9492-4db9-b843-404306f8fb01-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:39 crc kubenswrapper[4776]: I1125 10:58:39.882816 4776 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bcfb866-9492-4db9-b843-404306f8fb01-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:39 crc kubenswrapper[4776]: I1125 10:58:39.882827 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwvnx\" (UniqueName: \"kubernetes.io/projected/1bcfb866-9492-4db9-b843-404306f8fb01-kube-api-access-pwvnx\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:39 crc kubenswrapper[4776]: I1125 10:58:39.882837 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/1bcfb866-9492-4db9-b843-404306f8fb01-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:39 crc kubenswrapper[4776]: I1125 10:58:39.882847 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bcfb866-9492-4db9-b843-404306f8fb01-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:58:40 crc kubenswrapper[4776]: I1125 10:58:40.129031 4776 generic.go:334] "Generic (PLEG): container finished" podID="1bcfb866-9492-4db9-b843-404306f8fb01" containerID="a2c3be559b1055faa120b087e39b0ef9ff0642f312abe7091eeaeb7c77711a49" exitCode=0 Nov 25 10:58:40 crc kubenswrapper[4776]: I1125 10:58:40.129091 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8757674c4-gwcws" Nov 25 10:58:40 crc kubenswrapper[4776]: I1125 10:58:40.129092 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8757674c4-gwcws" event={"ID":"1bcfb866-9492-4db9-b843-404306f8fb01","Type":"ContainerDied","Data":"a2c3be559b1055faa120b087e39b0ef9ff0642f312abe7091eeaeb7c77711a49"} Nov 25 10:58:40 crc kubenswrapper[4776]: I1125 10:58:40.129164 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8757674c4-gwcws" event={"ID":"1bcfb866-9492-4db9-b843-404306f8fb01","Type":"ContainerDied","Data":"26c65ad02b16f0b3ace827528de36885b590949ff2c0412373572719419c31cb"} Nov 25 10:58:40 crc kubenswrapper[4776]: I1125 10:58:40.129182 4776 scope.go:117] "RemoveContainer" containerID="ccf1c362e5da0b9a7574b6bfb35a482f32cd9a4fd7d7027b84e6f6662603ef54" Nov 25 10:58:40 crc kubenswrapper[4776]: I1125 10:58:40.161273 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-8757674c4-gwcws"] Nov 25 10:58:40 crc kubenswrapper[4776]: I1125 10:58:40.162847 4776 scope.go:117] "RemoveContainer" containerID="a2c3be559b1055faa120b087e39b0ef9ff0642f312abe7091eeaeb7c77711a49" Nov 25 10:58:40 crc kubenswrapper[4776]: I1125 10:58:40.166883 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-8757674c4-gwcws"] Nov 25 10:58:40 crc kubenswrapper[4776]: I1125 10:58:40.180730 4776 scope.go:117] "RemoveContainer" containerID="ccf1c362e5da0b9a7574b6bfb35a482f32cd9a4fd7d7027b84e6f6662603ef54" Nov 25 10:58:40 crc kubenswrapper[4776]: E1125 10:58:40.181135 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccf1c362e5da0b9a7574b6bfb35a482f32cd9a4fd7d7027b84e6f6662603ef54\": container with ID starting with ccf1c362e5da0b9a7574b6bfb35a482f32cd9a4fd7d7027b84e6f6662603ef54 not found: ID does not exist" containerID="ccf1c362e5da0b9a7574b6bfb35a482f32cd9a4fd7d7027b84e6f6662603ef54" Nov 25 10:58:40 crc kubenswrapper[4776]: I1125 10:58:40.181166 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccf1c362e5da0b9a7574b6bfb35a482f32cd9a4fd7d7027b84e6f6662603ef54"} err="failed to get container status \"ccf1c362e5da0b9a7574b6bfb35a482f32cd9a4fd7d7027b84e6f6662603ef54\": rpc error: code = NotFound desc = could not find container \"ccf1c362e5da0b9a7574b6bfb35a482f32cd9a4fd7d7027b84e6f6662603ef54\": container with ID starting with ccf1c362e5da0b9a7574b6bfb35a482f32cd9a4fd7d7027b84e6f6662603ef54 not found: ID does not exist" Nov 25 10:58:40 crc kubenswrapper[4776]: I1125 10:58:40.181185 4776 scope.go:117] "RemoveContainer" containerID="a2c3be559b1055faa120b087e39b0ef9ff0642f312abe7091eeaeb7c77711a49" Nov 25 10:58:40 crc kubenswrapper[4776]: E1125 10:58:40.181541 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2c3be559b1055faa120b087e39b0ef9ff0642f312abe7091eeaeb7c77711a49\": container with ID starting with a2c3be559b1055faa120b087e39b0ef9ff0642f312abe7091eeaeb7c77711a49 not found: ID does not exist" containerID="a2c3be559b1055faa120b087e39b0ef9ff0642f312abe7091eeaeb7c77711a49" Nov 25 10:58:40 crc kubenswrapper[4776]: I1125 10:58:40.181586 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2c3be559b1055faa120b087e39b0ef9ff0642f312abe7091eeaeb7c77711a49"} err="failed to get container status \"a2c3be559b1055faa120b087e39b0ef9ff0642f312abe7091eeaeb7c77711a49\": rpc error: code = NotFound desc = could not find container \"a2c3be559b1055faa120b087e39b0ef9ff0642f312abe7091eeaeb7c77711a49\": container with ID starting with a2c3be559b1055faa120b087e39b0ef9ff0642f312abe7091eeaeb7c77711a49 not found: ID does not exist" Nov 25 10:58:41 crc kubenswrapper[4776]: I1125 10:58:41.675630 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bcfb866-9492-4db9-b843-404306f8fb01" path="/var/lib/kubelet/pods/1bcfb866-9492-4db9-b843-404306f8fb01/volumes" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.398714 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-dqc6q"] Nov 25 10:59:03 crc kubenswrapper[4776]: E1125 10:59:03.401704 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bcfb866-9492-4db9-b843-404306f8fb01" containerName="neutron-httpd" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.401878 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bcfb866-9492-4db9-b843-404306f8fb01" containerName="neutron-httpd" Nov 25 10:59:03 crc kubenswrapper[4776]: E1125 10:59:03.402039 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5161e133-bd4f-436d-8d36-cbcc7faccf9a" containerName="dnsmasq-dns" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.402191 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="5161e133-bd4f-436d-8d36-cbcc7faccf9a" containerName="dnsmasq-dns" Nov 25 10:59:03 crc kubenswrapper[4776]: E1125 10:59:03.402353 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bcfb866-9492-4db9-b843-404306f8fb01" containerName="neutron-api" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.402461 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bcfb866-9492-4db9-b843-404306f8fb01" containerName="neutron-api" Nov 25 10:59:03 crc kubenswrapper[4776]: E1125 10:59:03.402589 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5161e133-bd4f-436d-8d36-cbcc7faccf9a" containerName="init" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.402689 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="5161e133-bd4f-436d-8d36-cbcc7faccf9a" containerName="init" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.403199 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bcfb866-9492-4db9-b843-404306f8fb01" containerName="neutron-httpd" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.403360 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="5161e133-bd4f-436d-8d36-cbcc7faccf9a" containerName="dnsmasq-dns" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.403504 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bcfb866-9492-4db9-b843-404306f8fb01" containerName="neutron-api" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.404933 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-dqc6q" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.409224 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.409971 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.410346 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-8mb9k" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.410581 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.410697 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.432216 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-dqc6q"] Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.443663 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-dqc6q"] Nov 25 10:59:03 crc kubenswrapper[4776]: E1125 10:59:03.451208 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-zh4cw ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-zh4cw ring-data-devices scripts swiftconf]: context canceled" pod="openstack/swift-ring-rebalance-dqc6q" podUID="b602405c-162d-4b6f-99d3-9c0422e3d84d" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.488126 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67d94b9bc9-5cqlc"] Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.489537 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67d94b9bc9-5cqlc" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.491097 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b602405c-162d-4b6f-99d3-9c0422e3d84d-combined-ca-bundle\") pod \"swift-ring-rebalance-dqc6q\" (UID: \"b602405c-162d-4b6f-99d3-9c0422e3d84d\") " pod="openstack/swift-ring-rebalance-dqc6q" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.491155 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b602405c-162d-4b6f-99d3-9c0422e3d84d-swiftconf\") pod \"swift-ring-rebalance-dqc6q\" (UID: \"b602405c-162d-4b6f-99d3-9c0422e3d84d\") " pod="openstack/swift-ring-rebalance-dqc6q" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.491218 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b602405c-162d-4b6f-99d3-9c0422e3d84d-ring-data-devices\") pod \"swift-ring-rebalance-dqc6q\" (UID: \"b602405c-162d-4b6f-99d3-9c0422e3d84d\") " pod="openstack/swift-ring-rebalance-dqc6q" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.491275 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zh4cw\" (UniqueName: \"kubernetes.io/projected/b602405c-162d-4b6f-99d3-9c0422e3d84d-kube-api-access-zh4cw\") pod \"swift-ring-rebalance-dqc6q\" (UID: \"b602405c-162d-4b6f-99d3-9c0422e3d84d\") " pod="openstack/swift-ring-rebalance-dqc6q" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.491308 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b602405c-162d-4b6f-99d3-9c0422e3d84d-etc-swift\") pod \"swift-ring-rebalance-dqc6q\" (UID: \"b602405c-162d-4b6f-99d3-9c0422e3d84d\") " pod="openstack/swift-ring-rebalance-dqc6q" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.491331 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b602405c-162d-4b6f-99d3-9c0422e3d84d-scripts\") pod \"swift-ring-rebalance-dqc6q\" (UID: \"b602405c-162d-4b6f-99d3-9c0422e3d84d\") " pod="openstack/swift-ring-rebalance-dqc6q" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.491361 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b602405c-162d-4b6f-99d3-9c0422e3d84d-dispersionconf\") pod \"swift-ring-rebalance-dqc6q\" (UID: \"b602405c-162d-4b6f-99d3-9c0422e3d84d\") " pod="openstack/swift-ring-rebalance-dqc6q" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.508820 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67d94b9bc9-5cqlc"] Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.592667 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b602405c-162d-4b6f-99d3-9c0422e3d84d-combined-ca-bundle\") pod \"swift-ring-rebalance-dqc6q\" (UID: \"b602405c-162d-4b6f-99d3-9c0422e3d84d\") " pod="openstack/swift-ring-rebalance-dqc6q" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.592743 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b602405c-162d-4b6f-99d3-9c0422e3d84d-swiftconf\") pod \"swift-ring-rebalance-dqc6q\" (UID: \"b602405c-162d-4b6f-99d3-9c0422e3d84d\") " pod="openstack/swift-ring-rebalance-dqc6q" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.592803 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26290360-835e-4f34-a1ad-1803bce72437-ovsdbserver-sb\") pod \"dnsmasq-dns-67d94b9bc9-5cqlc\" (UID: \"26290360-835e-4f34-a1ad-1803bce72437\") " pod="openstack/dnsmasq-dns-67d94b9bc9-5cqlc" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.592855 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klfk4\" (UniqueName: \"kubernetes.io/projected/26290360-835e-4f34-a1ad-1803bce72437-kube-api-access-klfk4\") pod \"dnsmasq-dns-67d94b9bc9-5cqlc\" (UID: \"26290360-835e-4f34-a1ad-1803bce72437\") " pod="openstack/dnsmasq-dns-67d94b9bc9-5cqlc" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.592889 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b602405c-162d-4b6f-99d3-9c0422e3d84d-ring-data-devices\") pod \"swift-ring-rebalance-dqc6q\" (UID: \"b602405c-162d-4b6f-99d3-9c0422e3d84d\") " pod="openstack/swift-ring-rebalance-dqc6q" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.592951 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zh4cw\" (UniqueName: \"kubernetes.io/projected/b602405c-162d-4b6f-99d3-9c0422e3d84d-kube-api-access-zh4cw\") pod \"swift-ring-rebalance-dqc6q\" (UID: \"b602405c-162d-4b6f-99d3-9c0422e3d84d\") " pod="openstack/swift-ring-rebalance-dqc6q" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.592993 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b602405c-162d-4b6f-99d3-9c0422e3d84d-etc-swift\") pod \"swift-ring-rebalance-dqc6q\" (UID: \"b602405c-162d-4b6f-99d3-9c0422e3d84d\") " pod="openstack/swift-ring-rebalance-dqc6q" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.593022 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26290360-835e-4f34-a1ad-1803bce72437-dns-svc\") pod \"dnsmasq-dns-67d94b9bc9-5cqlc\" (UID: \"26290360-835e-4f34-a1ad-1803bce72437\") " pod="openstack/dnsmasq-dns-67d94b9bc9-5cqlc" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.593049 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b602405c-162d-4b6f-99d3-9c0422e3d84d-scripts\") pod \"swift-ring-rebalance-dqc6q\" (UID: \"b602405c-162d-4b6f-99d3-9c0422e3d84d\") " pod="openstack/swift-ring-rebalance-dqc6q" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.593097 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26290360-835e-4f34-a1ad-1803bce72437-config\") pod \"dnsmasq-dns-67d94b9bc9-5cqlc\" (UID: \"26290360-835e-4f34-a1ad-1803bce72437\") " pod="openstack/dnsmasq-dns-67d94b9bc9-5cqlc" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.593135 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26290360-835e-4f34-a1ad-1803bce72437-ovsdbserver-nb\") pod \"dnsmasq-dns-67d94b9bc9-5cqlc\" (UID: \"26290360-835e-4f34-a1ad-1803bce72437\") " pod="openstack/dnsmasq-dns-67d94b9bc9-5cqlc" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.593161 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b602405c-162d-4b6f-99d3-9c0422e3d84d-dispersionconf\") pod \"swift-ring-rebalance-dqc6q\" (UID: \"b602405c-162d-4b6f-99d3-9c0422e3d84d\") " pod="openstack/swift-ring-rebalance-dqc6q" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.593733 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b602405c-162d-4b6f-99d3-9c0422e3d84d-scripts\") pod \"swift-ring-rebalance-dqc6q\" (UID: \"b602405c-162d-4b6f-99d3-9c0422e3d84d\") " pod="openstack/swift-ring-rebalance-dqc6q" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.593954 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b602405c-162d-4b6f-99d3-9c0422e3d84d-etc-swift\") pod \"swift-ring-rebalance-dqc6q\" (UID: \"b602405c-162d-4b6f-99d3-9c0422e3d84d\") " pod="openstack/swift-ring-rebalance-dqc6q" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.594368 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b602405c-162d-4b6f-99d3-9c0422e3d84d-ring-data-devices\") pod \"swift-ring-rebalance-dqc6q\" (UID: \"b602405c-162d-4b6f-99d3-9c0422e3d84d\") " pod="openstack/swift-ring-rebalance-dqc6q" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.598571 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b602405c-162d-4b6f-99d3-9c0422e3d84d-dispersionconf\") pod \"swift-ring-rebalance-dqc6q\" (UID: \"b602405c-162d-4b6f-99d3-9c0422e3d84d\") " pod="openstack/swift-ring-rebalance-dqc6q" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.598834 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b602405c-162d-4b6f-99d3-9c0422e3d84d-combined-ca-bundle\") pod \"swift-ring-rebalance-dqc6q\" (UID: \"b602405c-162d-4b6f-99d3-9c0422e3d84d\") " pod="openstack/swift-ring-rebalance-dqc6q" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.599004 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b602405c-162d-4b6f-99d3-9c0422e3d84d-swiftconf\") pod \"swift-ring-rebalance-dqc6q\" (UID: \"b602405c-162d-4b6f-99d3-9c0422e3d84d\") " pod="openstack/swift-ring-rebalance-dqc6q" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.620643 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zh4cw\" (UniqueName: \"kubernetes.io/projected/b602405c-162d-4b6f-99d3-9c0422e3d84d-kube-api-access-zh4cw\") pod \"swift-ring-rebalance-dqc6q\" (UID: \"b602405c-162d-4b6f-99d3-9c0422e3d84d\") " pod="openstack/swift-ring-rebalance-dqc6q" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.694020 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26290360-835e-4f34-a1ad-1803bce72437-dns-svc\") pod \"dnsmasq-dns-67d94b9bc9-5cqlc\" (UID: \"26290360-835e-4f34-a1ad-1803bce72437\") " pod="openstack/dnsmasq-dns-67d94b9bc9-5cqlc" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.694399 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26290360-835e-4f34-a1ad-1803bce72437-config\") pod \"dnsmasq-dns-67d94b9bc9-5cqlc\" (UID: \"26290360-835e-4f34-a1ad-1803bce72437\") " pod="openstack/dnsmasq-dns-67d94b9bc9-5cqlc" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.695129 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26290360-835e-4f34-a1ad-1803bce72437-dns-svc\") pod \"dnsmasq-dns-67d94b9bc9-5cqlc\" (UID: \"26290360-835e-4f34-a1ad-1803bce72437\") " pod="openstack/dnsmasq-dns-67d94b9bc9-5cqlc" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.695290 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26290360-835e-4f34-a1ad-1803bce72437-config\") pod \"dnsmasq-dns-67d94b9bc9-5cqlc\" (UID: \"26290360-835e-4f34-a1ad-1803bce72437\") " pod="openstack/dnsmasq-dns-67d94b9bc9-5cqlc" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.695433 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26290360-835e-4f34-a1ad-1803bce72437-ovsdbserver-nb\") pod \"dnsmasq-dns-67d94b9bc9-5cqlc\" (UID: \"26290360-835e-4f34-a1ad-1803bce72437\") " pod="openstack/dnsmasq-dns-67d94b9bc9-5cqlc" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.695591 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26290360-835e-4f34-a1ad-1803bce72437-ovsdbserver-sb\") pod \"dnsmasq-dns-67d94b9bc9-5cqlc\" (UID: \"26290360-835e-4f34-a1ad-1803bce72437\") " pod="openstack/dnsmasq-dns-67d94b9bc9-5cqlc" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.695714 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klfk4\" (UniqueName: \"kubernetes.io/projected/26290360-835e-4f34-a1ad-1803bce72437-kube-api-access-klfk4\") pod \"dnsmasq-dns-67d94b9bc9-5cqlc\" (UID: \"26290360-835e-4f34-a1ad-1803bce72437\") " pod="openstack/dnsmasq-dns-67d94b9bc9-5cqlc" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.696131 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26290360-835e-4f34-a1ad-1803bce72437-ovsdbserver-nb\") pod \"dnsmasq-dns-67d94b9bc9-5cqlc\" (UID: \"26290360-835e-4f34-a1ad-1803bce72437\") " pod="openstack/dnsmasq-dns-67d94b9bc9-5cqlc" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.696734 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26290360-835e-4f34-a1ad-1803bce72437-ovsdbserver-sb\") pod \"dnsmasq-dns-67d94b9bc9-5cqlc\" (UID: \"26290360-835e-4f34-a1ad-1803bce72437\") " pod="openstack/dnsmasq-dns-67d94b9bc9-5cqlc" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.714713 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klfk4\" (UniqueName: \"kubernetes.io/projected/26290360-835e-4f34-a1ad-1803bce72437-kube-api-access-klfk4\") pod \"dnsmasq-dns-67d94b9bc9-5cqlc\" (UID: \"26290360-835e-4f34-a1ad-1803bce72437\") " pod="openstack/dnsmasq-dns-67d94b9bc9-5cqlc" Nov 25 10:59:03 crc kubenswrapper[4776]: I1125 10:59:03.814682 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67d94b9bc9-5cqlc" Nov 25 10:59:04 crc kubenswrapper[4776]: I1125 10:59:04.299847 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67d94b9bc9-5cqlc"] Nov 25 10:59:04 crc kubenswrapper[4776]: I1125 10:59:04.327781 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-dqc6q" Nov 25 10:59:04 crc kubenswrapper[4776]: I1125 10:59:04.328550 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67d94b9bc9-5cqlc" event={"ID":"26290360-835e-4f34-a1ad-1803bce72437","Type":"ContainerStarted","Data":"69a4797ee8efbc57a0adc79a051c65ba0cbab45a0d1738400eab53719af11e80"} Nov 25 10:59:04 crc kubenswrapper[4776]: I1125 10:59:04.342894 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-dqc6q" Nov 25 10:59:04 crc kubenswrapper[4776]: I1125 10:59:04.406986 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b602405c-162d-4b6f-99d3-9c0422e3d84d-ring-data-devices\") pod \"b602405c-162d-4b6f-99d3-9c0422e3d84d\" (UID: \"b602405c-162d-4b6f-99d3-9c0422e3d84d\") " Nov 25 10:59:04 crc kubenswrapper[4776]: I1125 10:59:04.407147 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b602405c-162d-4b6f-99d3-9c0422e3d84d-dispersionconf\") pod \"b602405c-162d-4b6f-99d3-9c0422e3d84d\" (UID: \"b602405c-162d-4b6f-99d3-9c0422e3d84d\") " Nov 25 10:59:04 crc kubenswrapper[4776]: I1125 10:59:04.407190 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b602405c-162d-4b6f-99d3-9c0422e3d84d-swiftconf\") pod \"b602405c-162d-4b6f-99d3-9c0422e3d84d\" (UID: \"b602405c-162d-4b6f-99d3-9c0422e3d84d\") " Nov 25 10:59:04 crc kubenswrapper[4776]: I1125 10:59:04.407209 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zh4cw\" (UniqueName: \"kubernetes.io/projected/b602405c-162d-4b6f-99d3-9c0422e3d84d-kube-api-access-zh4cw\") pod \"b602405c-162d-4b6f-99d3-9c0422e3d84d\" (UID: \"b602405c-162d-4b6f-99d3-9c0422e3d84d\") " Nov 25 10:59:04 crc kubenswrapper[4776]: I1125 10:59:04.407318 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b602405c-162d-4b6f-99d3-9c0422e3d84d-scripts\") pod \"b602405c-162d-4b6f-99d3-9c0422e3d84d\" (UID: \"b602405c-162d-4b6f-99d3-9c0422e3d84d\") " Nov 25 10:59:04 crc kubenswrapper[4776]: I1125 10:59:04.407350 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b602405c-162d-4b6f-99d3-9c0422e3d84d-etc-swift\") pod \"b602405c-162d-4b6f-99d3-9c0422e3d84d\" (UID: \"b602405c-162d-4b6f-99d3-9c0422e3d84d\") " Nov 25 10:59:04 crc kubenswrapper[4776]: I1125 10:59:04.407400 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b602405c-162d-4b6f-99d3-9c0422e3d84d-combined-ca-bundle\") pod \"b602405c-162d-4b6f-99d3-9c0422e3d84d\" (UID: \"b602405c-162d-4b6f-99d3-9c0422e3d84d\") " Nov 25 10:59:04 crc kubenswrapper[4776]: I1125 10:59:04.408561 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b602405c-162d-4b6f-99d3-9c0422e3d84d-scripts" (OuterVolumeSpecName: "scripts") pod "b602405c-162d-4b6f-99d3-9c0422e3d84d" (UID: "b602405c-162d-4b6f-99d3-9c0422e3d84d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:04 crc kubenswrapper[4776]: I1125 10:59:04.410570 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b602405c-162d-4b6f-99d3-9c0422e3d84d-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "b602405c-162d-4b6f-99d3-9c0422e3d84d" (UID: "b602405c-162d-4b6f-99d3-9c0422e3d84d"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:59:04 crc kubenswrapper[4776]: I1125 10:59:04.410920 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b602405c-162d-4b6f-99d3-9c0422e3d84d-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "b602405c-162d-4b6f-99d3-9c0422e3d84d" (UID: "b602405c-162d-4b6f-99d3-9c0422e3d84d"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:04 crc kubenswrapper[4776]: I1125 10:59:04.414471 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b602405c-162d-4b6f-99d3-9c0422e3d84d-kube-api-access-zh4cw" (OuterVolumeSpecName: "kube-api-access-zh4cw") pod "b602405c-162d-4b6f-99d3-9c0422e3d84d" (UID: "b602405c-162d-4b6f-99d3-9c0422e3d84d"). InnerVolumeSpecName "kube-api-access-zh4cw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:59:04 crc kubenswrapper[4776]: I1125 10:59:04.414811 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b602405c-162d-4b6f-99d3-9c0422e3d84d-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "b602405c-162d-4b6f-99d3-9c0422e3d84d" (UID: "b602405c-162d-4b6f-99d3-9c0422e3d84d"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:59:04 crc kubenswrapper[4776]: I1125 10:59:04.416177 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b602405c-162d-4b6f-99d3-9c0422e3d84d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b602405c-162d-4b6f-99d3-9c0422e3d84d" (UID: "b602405c-162d-4b6f-99d3-9c0422e3d84d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:59:04 crc kubenswrapper[4776]: I1125 10:59:04.416211 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b602405c-162d-4b6f-99d3-9c0422e3d84d-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "b602405c-162d-4b6f-99d3-9c0422e3d84d" (UID: "b602405c-162d-4b6f-99d3-9c0422e3d84d"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:59:04 crc kubenswrapper[4776]: I1125 10:59:04.509269 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b602405c-162d-4b6f-99d3-9c0422e3d84d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:04 crc kubenswrapper[4776]: I1125 10:59:04.509317 4776 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b602405c-162d-4b6f-99d3-9c0422e3d84d-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:04 crc kubenswrapper[4776]: I1125 10:59:04.509332 4776 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b602405c-162d-4b6f-99d3-9c0422e3d84d-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:04 crc kubenswrapper[4776]: I1125 10:59:04.509344 4776 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b602405c-162d-4b6f-99d3-9c0422e3d84d-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:04 crc kubenswrapper[4776]: I1125 10:59:04.509358 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zh4cw\" (UniqueName: \"kubernetes.io/projected/b602405c-162d-4b6f-99d3-9c0422e3d84d-kube-api-access-zh4cw\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:04 crc kubenswrapper[4776]: I1125 10:59:04.509371 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b602405c-162d-4b6f-99d3-9c0422e3d84d-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:04 crc kubenswrapper[4776]: I1125 10:59:04.509384 4776 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b602405c-162d-4b6f-99d3-9c0422e3d84d-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:05 crc kubenswrapper[4776]: I1125 10:59:05.336326 4776 generic.go:334] "Generic (PLEG): container finished" podID="26290360-835e-4f34-a1ad-1803bce72437" containerID="28081dcf1472877d0fb63c17e5eaf89e932e095527c9dbae4498cb43d792e427" exitCode=0 Nov 25 10:59:05 crc kubenswrapper[4776]: I1125 10:59:05.336646 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-dqc6q" Nov 25 10:59:05 crc kubenswrapper[4776]: I1125 10:59:05.343308 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67d94b9bc9-5cqlc" event={"ID":"26290360-835e-4f34-a1ad-1803bce72437","Type":"ContainerDied","Data":"28081dcf1472877d0fb63c17e5eaf89e932e095527c9dbae4498cb43d792e427"} Nov 25 10:59:05 crc kubenswrapper[4776]: I1125 10:59:05.426656 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-dqc6q"] Nov 25 10:59:05 crc kubenswrapper[4776]: I1125 10:59:05.452083 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-dqc6q"] Nov 25 10:59:05 crc kubenswrapper[4776]: I1125 10:59:05.675530 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b602405c-162d-4b6f-99d3-9c0422e3d84d" path="/var/lib/kubelet/pods/b602405c-162d-4b6f-99d3-9c0422e3d84d/volumes" Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.073486 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-65f4dfb8b6-bqpbj"] Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.075037 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.079881 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.079969 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-8mb9k" Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.080183 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.080323 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.083253 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-65f4dfb8b6-bqpbj"] Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.156658 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-etc-swift\") pod \"swift-proxy-65f4dfb8b6-bqpbj\" (UID: \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\") " pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.156726 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtzvw\" (UniqueName: \"kubernetes.io/projected/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-kube-api-access-mtzvw\") pod \"swift-proxy-65f4dfb8b6-bqpbj\" (UID: \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\") " pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.156816 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-run-httpd\") pod \"swift-proxy-65f4dfb8b6-bqpbj\" (UID: \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\") " pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.156878 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-combined-ca-bundle\") pod \"swift-proxy-65f4dfb8b6-bqpbj\" (UID: \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\") " pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.157022 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-log-httpd\") pod \"swift-proxy-65f4dfb8b6-bqpbj\" (UID: \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\") " pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.157200 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-config-data\") pod \"swift-proxy-65f4dfb8b6-bqpbj\" (UID: \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\") " pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.258622 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-log-httpd\") pod \"swift-proxy-65f4dfb8b6-bqpbj\" (UID: \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\") " pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.258701 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-config-data\") pod \"swift-proxy-65f4dfb8b6-bqpbj\" (UID: \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\") " pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.258742 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-etc-swift\") pod \"swift-proxy-65f4dfb8b6-bqpbj\" (UID: \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\") " pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.259119 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtzvw\" (UniqueName: \"kubernetes.io/projected/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-kube-api-access-mtzvw\") pod \"swift-proxy-65f4dfb8b6-bqpbj\" (UID: \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\") " pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.259245 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-run-httpd\") pod \"swift-proxy-65f4dfb8b6-bqpbj\" (UID: \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\") " pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.259341 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-combined-ca-bundle\") pod \"swift-proxy-65f4dfb8b6-bqpbj\" (UID: \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\") " pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.259680 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-run-httpd\") pod \"swift-proxy-65f4dfb8b6-bqpbj\" (UID: \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\") " pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.260008 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-log-httpd\") pod \"swift-proxy-65f4dfb8b6-bqpbj\" (UID: \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\") " pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.266946 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-combined-ca-bundle\") pod \"swift-proxy-65f4dfb8b6-bqpbj\" (UID: \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\") " pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.267631 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-config-data\") pod \"swift-proxy-65f4dfb8b6-bqpbj\" (UID: \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\") " pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.278589 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-etc-swift\") pod \"swift-proxy-65f4dfb8b6-bqpbj\" (UID: \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\") " pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.281200 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtzvw\" (UniqueName: \"kubernetes.io/projected/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-kube-api-access-mtzvw\") pod \"swift-proxy-65f4dfb8b6-bqpbj\" (UID: \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\") " pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.344897 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67d94b9bc9-5cqlc" event={"ID":"26290360-835e-4f34-a1ad-1803bce72437","Type":"ContainerStarted","Data":"548a8359c5df63cb48083eb603e31cff145deffb674aa59e73df33a69f5f0f06"} Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.345492 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67d94b9bc9-5cqlc" Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.364158 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67d94b9bc9-5cqlc" podStartSLOduration=3.364135781 podStartE2EDuration="3.364135781s" podCreationTimestamp="2025-11-25 10:59:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:59:06.360297214 +0000 UTC m=+5691.401356767" watchObservedRunningTime="2025-11-25 10:59:06.364135781 +0000 UTC m=+5691.405195334" Nov 25 10:59:06 crc kubenswrapper[4776]: I1125 10:59:06.396086 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.093464 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-65f4dfb8b6-bqpbj"] Nov 25 10:59:07 crc kubenswrapper[4776]: W1125 10:59:07.104769 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8919d1d_aa43_45f4_a1c3_37a9910e8a32.slice/crio-4d78e0674d16f3de9de4fd1dde9df11cf1e1ffac3583b01d6a983902e9cfa731 WatchSource:0}: Error finding container 4d78e0674d16f3de9de4fd1dde9df11cf1e1ffac3583b01d6a983902e9cfa731: Status 404 returned error can't find the container with id 4d78e0674d16f3de9de4fd1dde9df11cf1e1ffac3583b01d6a983902e9cfa731 Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.357115 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" event={"ID":"e8919d1d-aa43-45f4-a1c3-37a9910e8a32","Type":"ContainerStarted","Data":"34b982b5c2ce60541ba1de95d03cc6801c4015343223e772fa8d358b3a3b5394"} Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.357509 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" event={"ID":"e8919d1d-aa43-45f4-a1c3-37a9910e8a32","Type":"ContainerStarted","Data":"4d78e0674d16f3de9de4fd1dde9df11cf1e1ffac3583b01d6a983902e9cfa731"} Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.539362 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-5c6d9c7954-pvk2n"] Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.541149 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.543898 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.551616 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.552622 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5c6d9c7954-pvk2n"] Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.588102 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0dc0a5c5-9aa1-4190-985f-047a9406c24a-public-tls-certs\") pod \"swift-proxy-5c6d9c7954-pvk2n\" (UID: \"0dc0a5c5-9aa1-4190-985f-047a9406c24a\") " pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.588157 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0dc0a5c5-9aa1-4190-985f-047a9406c24a-etc-swift\") pod \"swift-proxy-5c6d9c7954-pvk2n\" (UID: \"0dc0a5c5-9aa1-4190-985f-047a9406c24a\") " pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.588190 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7jtp\" (UniqueName: \"kubernetes.io/projected/0dc0a5c5-9aa1-4190-985f-047a9406c24a-kube-api-access-d7jtp\") pod \"swift-proxy-5c6d9c7954-pvk2n\" (UID: \"0dc0a5c5-9aa1-4190-985f-047a9406c24a\") " pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.588223 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0dc0a5c5-9aa1-4190-985f-047a9406c24a-log-httpd\") pod \"swift-proxy-5c6d9c7954-pvk2n\" (UID: \"0dc0a5c5-9aa1-4190-985f-047a9406c24a\") " pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.588271 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dc0a5c5-9aa1-4190-985f-047a9406c24a-combined-ca-bundle\") pod \"swift-proxy-5c6d9c7954-pvk2n\" (UID: \"0dc0a5c5-9aa1-4190-985f-047a9406c24a\") " pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.588342 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dc0a5c5-9aa1-4190-985f-047a9406c24a-config-data\") pod \"swift-proxy-5c6d9c7954-pvk2n\" (UID: \"0dc0a5c5-9aa1-4190-985f-047a9406c24a\") " pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.588430 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0dc0a5c5-9aa1-4190-985f-047a9406c24a-internal-tls-certs\") pod \"swift-proxy-5c6d9c7954-pvk2n\" (UID: \"0dc0a5c5-9aa1-4190-985f-047a9406c24a\") " pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.588468 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0dc0a5c5-9aa1-4190-985f-047a9406c24a-run-httpd\") pod \"swift-proxy-5c6d9c7954-pvk2n\" (UID: \"0dc0a5c5-9aa1-4190-985f-047a9406c24a\") " pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.690176 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0dc0a5c5-9aa1-4190-985f-047a9406c24a-etc-swift\") pod \"swift-proxy-5c6d9c7954-pvk2n\" (UID: \"0dc0a5c5-9aa1-4190-985f-047a9406c24a\") " pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.690839 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7jtp\" (UniqueName: \"kubernetes.io/projected/0dc0a5c5-9aa1-4190-985f-047a9406c24a-kube-api-access-d7jtp\") pod \"swift-proxy-5c6d9c7954-pvk2n\" (UID: \"0dc0a5c5-9aa1-4190-985f-047a9406c24a\") " pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.690885 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0dc0a5c5-9aa1-4190-985f-047a9406c24a-log-httpd\") pod \"swift-proxy-5c6d9c7954-pvk2n\" (UID: \"0dc0a5c5-9aa1-4190-985f-047a9406c24a\") " pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.690906 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dc0a5c5-9aa1-4190-985f-047a9406c24a-combined-ca-bundle\") pod \"swift-proxy-5c6d9c7954-pvk2n\" (UID: \"0dc0a5c5-9aa1-4190-985f-047a9406c24a\") " pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.690940 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dc0a5c5-9aa1-4190-985f-047a9406c24a-config-data\") pod \"swift-proxy-5c6d9c7954-pvk2n\" (UID: \"0dc0a5c5-9aa1-4190-985f-047a9406c24a\") " pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.691093 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0dc0a5c5-9aa1-4190-985f-047a9406c24a-internal-tls-certs\") pod \"swift-proxy-5c6d9c7954-pvk2n\" (UID: \"0dc0a5c5-9aa1-4190-985f-047a9406c24a\") " pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.691126 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0dc0a5c5-9aa1-4190-985f-047a9406c24a-run-httpd\") pod \"swift-proxy-5c6d9c7954-pvk2n\" (UID: \"0dc0a5c5-9aa1-4190-985f-047a9406c24a\") " pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.691195 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0dc0a5c5-9aa1-4190-985f-047a9406c24a-public-tls-certs\") pod \"swift-proxy-5c6d9c7954-pvk2n\" (UID: \"0dc0a5c5-9aa1-4190-985f-047a9406c24a\") " pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.691500 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0dc0a5c5-9aa1-4190-985f-047a9406c24a-log-httpd\") pod \"swift-proxy-5c6d9c7954-pvk2n\" (UID: \"0dc0a5c5-9aa1-4190-985f-047a9406c24a\") " pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.692394 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0dc0a5c5-9aa1-4190-985f-047a9406c24a-run-httpd\") pod \"swift-proxy-5c6d9c7954-pvk2n\" (UID: \"0dc0a5c5-9aa1-4190-985f-047a9406c24a\") " pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.696676 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0dc0a5c5-9aa1-4190-985f-047a9406c24a-public-tls-certs\") pod \"swift-proxy-5c6d9c7954-pvk2n\" (UID: \"0dc0a5c5-9aa1-4190-985f-047a9406c24a\") " pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.697701 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dc0a5c5-9aa1-4190-985f-047a9406c24a-combined-ca-bundle\") pod \"swift-proxy-5c6d9c7954-pvk2n\" (UID: \"0dc0a5c5-9aa1-4190-985f-047a9406c24a\") " pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.700745 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dc0a5c5-9aa1-4190-985f-047a9406c24a-config-data\") pod \"swift-proxy-5c6d9c7954-pvk2n\" (UID: \"0dc0a5c5-9aa1-4190-985f-047a9406c24a\") " pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.703575 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/0dc0a5c5-9aa1-4190-985f-047a9406c24a-etc-swift\") pod \"swift-proxy-5c6d9c7954-pvk2n\" (UID: \"0dc0a5c5-9aa1-4190-985f-047a9406c24a\") " pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.709199 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0dc0a5c5-9aa1-4190-985f-047a9406c24a-internal-tls-certs\") pod \"swift-proxy-5c6d9c7954-pvk2n\" (UID: \"0dc0a5c5-9aa1-4190-985f-047a9406c24a\") " pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.709995 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7jtp\" (UniqueName: \"kubernetes.io/projected/0dc0a5c5-9aa1-4190-985f-047a9406c24a-kube-api-access-d7jtp\") pod \"swift-proxy-5c6d9c7954-pvk2n\" (UID: \"0dc0a5c5-9aa1-4190-985f-047a9406c24a\") " pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:07 crc kubenswrapper[4776]: I1125 10:59:07.857926 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:08 crc kubenswrapper[4776]: I1125 10:59:08.366166 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" event={"ID":"e8919d1d-aa43-45f4-a1c3-37a9910e8a32","Type":"ContainerStarted","Data":"df43f247d9591bacf4720ae864b30b0c3c66b9c2707d98c0aecd08dff0cb913e"} Nov 25 10:59:08 crc kubenswrapper[4776]: I1125 10:59:08.366663 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" Nov 25 10:59:08 crc kubenswrapper[4776]: I1125 10:59:08.366683 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" Nov 25 10:59:08 crc kubenswrapper[4776]: I1125 10:59:08.386585 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" podStartSLOduration=2.386565965 podStartE2EDuration="2.386565965s" podCreationTimestamp="2025-11-25 10:59:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:59:08.382550424 +0000 UTC m=+5693.423609977" watchObservedRunningTime="2025-11-25 10:59:08.386565965 +0000 UTC m=+5693.427625518" Nov 25 10:59:08 crc kubenswrapper[4776]: I1125 10:59:08.500061 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5c6d9c7954-pvk2n"] Nov 25 10:59:09 crc kubenswrapper[4776]: I1125 10:59:09.387560 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qqqvj"] Nov 25 10:59:09 crc kubenswrapper[4776]: I1125 10:59:09.393156 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5c6d9c7954-pvk2n" event={"ID":"0dc0a5c5-9aa1-4190-985f-047a9406c24a","Type":"ContainerStarted","Data":"2609a337b8d4052651ebe96896de957e99ad896c302a46044423f0f21a8fbec6"} Nov 25 10:59:09 crc kubenswrapper[4776]: I1125 10:59:09.393196 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5c6d9c7954-pvk2n" event={"ID":"0dc0a5c5-9aa1-4190-985f-047a9406c24a","Type":"ContainerStarted","Data":"6a36f9bfdb54facd66fc336d540d61e4eb27e6d4b16e5aacae41ec7042d60e5e"} Nov 25 10:59:09 crc kubenswrapper[4776]: I1125 10:59:09.393274 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qqqvj" Nov 25 10:59:09 crc kubenswrapper[4776]: I1125 10:59:09.406584 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qqqvj"] Nov 25 10:59:09 crc kubenswrapper[4776]: I1125 10:59:09.421978 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2jxc\" (UniqueName: \"kubernetes.io/projected/9f14e517-3052-4e3c-99c5-121293a8a542-kube-api-access-x2jxc\") pod \"redhat-marketplace-qqqvj\" (UID: \"9f14e517-3052-4e3c-99c5-121293a8a542\") " pod="openshift-marketplace/redhat-marketplace-qqqvj" Nov 25 10:59:09 crc kubenswrapper[4776]: I1125 10:59:09.422186 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f14e517-3052-4e3c-99c5-121293a8a542-catalog-content\") pod \"redhat-marketplace-qqqvj\" (UID: \"9f14e517-3052-4e3c-99c5-121293a8a542\") " pod="openshift-marketplace/redhat-marketplace-qqqvj" Nov 25 10:59:09 crc kubenswrapper[4776]: I1125 10:59:09.422265 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f14e517-3052-4e3c-99c5-121293a8a542-utilities\") pod \"redhat-marketplace-qqqvj\" (UID: \"9f14e517-3052-4e3c-99c5-121293a8a542\") " pod="openshift-marketplace/redhat-marketplace-qqqvj" Nov 25 10:59:09 crc kubenswrapper[4776]: I1125 10:59:09.524235 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f14e517-3052-4e3c-99c5-121293a8a542-catalog-content\") pod \"redhat-marketplace-qqqvj\" (UID: \"9f14e517-3052-4e3c-99c5-121293a8a542\") " pod="openshift-marketplace/redhat-marketplace-qqqvj" Nov 25 10:59:09 crc kubenswrapper[4776]: I1125 10:59:09.524340 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f14e517-3052-4e3c-99c5-121293a8a542-utilities\") pod \"redhat-marketplace-qqqvj\" (UID: \"9f14e517-3052-4e3c-99c5-121293a8a542\") " pod="openshift-marketplace/redhat-marketplace-qqqvj" Nov 25 10:59:09 crc kubenswrapper[4776]: I1125 10:59:09.524384 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2jxc\" (UniqueName: \"kubernetes.io/projected/9f14e517-3052-4e3c-99c5-121293a8a542-kube-api-access-x2jxc\") pod \"redhat-marketplace-qqqvj\" (UID: \"9f14e517-3052-4e3c-99c5-121293a8a542\") " pod="openshift-marketplace/redhat-marketplace-qqqvj" Nov 25 10:59:09 crc kubenswrapper[4776]: I1125 10:59:09.525274 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f14e517-3052-4e3c-99c5-121293a8a542-catalog-content\") pod \"redhat-marketplace-qqqvj\" (UID: \"9f14e517-3052-4e3c-99c5-121293a8a542\") " pod="openshift-marketplace/redhat-marketplace-qqqvj" Nov 25 10:59:09 crc kubenswrapper[4776]: I1125 10:59:09.525424 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f14e517-3052-4e3c-99c5-121293a8a542-utilities\") pod \"redhat-marketplace-qqqvj\" (UID: \"9f14e517-3052-4e3c-99c5-121293a8a542\") " pod="openshift-marketplace/redhat-marketplace-qqqvj" Nov 25 10:59:09 crc kubenswrapper[4776]: I1125 10:59:09.543580 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2jxc\" (UniqueName: \"kubernetes.io/projected/9f14e517-3052-4e3c-99c5-121293a8a542-kube-api-access-x2jxc\") pod \"redhat-marketplace-qqqvj\" (UID: \"9f14e517-3052-4e3c-99c5-121293a8a542\") " pod="openshift-marketplace/redhat-marketplace-qqqvj" Nov 25 10:59:09 crc kubenswrapper[4776]: I1125 10:59:09.720389 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qqqvj" Nov 25 10:59:10 crc kubenswrapper[4776]: I1125 10:59:10.223983 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qqqvj"] Nov 25 10:59:10 crc kubenswrapper[4776]: I1125 10:59:10.400329 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5c6d9c7954-pvk2n" event={"ID":"0dc0a5c5-9aa1-4190-985f-047a9406c24a","Type":"ContainerStarted","Data":"1c0c238d2821ce1f0174b508e0b8cfe5cd618dfd9e0c959e7f20125adc4f5d7e"} Nov 25 10:59:10 crc kubenswrapper[4776]: I1125 10:59:10.400451 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:10 crc kubenswrapper[4776]: I1125 10:59:10.401941 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qqqvj" event={"ID":"9f14e517-3052-4e3c-99c5-121293a8a542","Type":"ContainerStarted","Data":"5845e754bc368122b043f0bcd0b42db5d33cc6afab5d750cce7c6797a9fa6725"} Nov 25 10:59:10 crc kubenswrapper[4776]: I1125 10:59:10.427845 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-5c6d9c7954-pvk2n" podStartSLOduration=3.427823691 podStartE2EDuration="3.427823691s" podCreationTimestamp="2025-11-25 10:59:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:59:10.421697098 +0000 UTC m=+5695.462756651" watchObservedRunningTime="2025-11-25 10:59:10.427823691 +0000 UTC m=+5695.468883244" Nov 25 10:59:10 crc kubenswrapper[4776]: E1125 10:59:10.737841 4776 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f14e517_3052_4e3c_99c5_121293a8a542.slice/crio-conmon-6feb09434cdc3ae1fd8ba6d7f85241c4fc9285dd96afb717abb52822756e69f9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f14e517_3052_4e3c_99c5_121293a8a542.slice/crio-6feb09434cdc3ae1fd8ba6d7f85241c4fc9285dd96afb717abb52822756e69f9.scope\": RecentStats: unable to find data in memory cache]" Nov 25 10:59:11 crc kubenswrapper[4776]: I1125 10:59:11.410513 4776 generic.go:334] "Generic (PLEG): container finished" podID="9f14e517-3052-4e3c-99c5-121293a8a542" containerID="6feb09434cdc3ae1fd8ba6d7f85241c4fc9285dd96afb717abb52822756e69f9" exitCode=0 Nov 25 10:59:11 crc kubenswrapper[4776]: I1125 10:59:11.410639 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qqqvj" event={"ID":"9f14e517-3052-4e3c-99c5-121293a8a542","Type":"ContainerDied","Data":"6feb09434cdc3ae1fd8ba6d7f85241c4fc9285dd96afb717abb52822756e69f9"} Nov 25 10:59:11 crc kubenswrapper[4776]: I1125 10:59:11.411084 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:11 crc kubenswrapper[4776]: I1125 10:59:11.413405 4776 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 10:59:12 crc kubenswrapper[4776]: I1125 10:59:12.419378 4776 generic.go:334] "Generic (PLEG): container finished" podID="9f14e517-3052-4e3c-99c5-121293a8a542" containerID="47dd73561a5135188df8f2e28fccb30b8e2bb685ef792e0e27b5890741aefaae" exitCode=0 Nov 25 10:59:12 crc kubenswrapper[4776]: I1125 10:59:12.419456 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qqqvj" event={"ID":"9f14e517-3052-4e3c-99c5-121293a8a542","Type":"ContainerDied","Data":"47dd73561a5135188df8f2e28fccb30b8e2bb685ef792e0e27b5890741aefaae"} Nov 25 10:59:13 crc kubenswrapper[4776]: I1125 10:59:13.816301 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67d94b9bc9-5cqlc" Nov 25 10:59:13 crc kubenswrapper[4776]: I1125 10:59:13.888302 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d5c556d87-vf246"] Nov 25 10:59:14 crc kubenswrapper[4776]: I1125 10:59:14.437827 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d5c556d87-vf246" podUID="7fdc4742-4288-4756-8dfd-3f3f0d2174e2" containerName="dnsmasq-dns" containerID="cri-o://f46bb4d4a6a370a5680b4d71418cbdc0fa0aed30d3492279f265d8dfb0aa84f1" gracePeriod=10 Nov 25 10:59:14 crc kubenswrapper[4776]: I1125 10:59:14.438852 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qqqvj" event={"ID":"9f14e517-3052-4e3c-99c5-121293a8a542","Type":"ContainerStarted","Data":"cd0888d7369c3a96c99c268c08dc855dfdfec3b7555aa9aa833e56f7454edf13"} Nov 25 10:59:14 crc kubenswrapper[4776]: I1125 10:59:14.464509 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qqqvj" podStartSLOduration=3.54006555 podStartE2EDuration="5.464487615s" podCreationTimestamp="2025-11-25 10:59:09 +0000 UTC" firstStartedPulling="2025-11-25 10:59:11.413145764 +0000 UTC m=+5696.454205327" lastFinishedPulling="2025-11-25 10:59:13.337567839 +0000 UTC m=+5698.378627392" observedRunningTime="2025-11-25 10:59:14.458538776 +0000 UTC m=+5699.499598339" watchObservedRunningTime="2025-11-25 10:59:14.464487615 +0000 UTC m=+5699.505547168" Nov 25 10:59:14 crc kubenswrapper[4776]: I1125 10:59:14.865700 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d5c556d87-vf246" Nov 25 10:59:14 crc kubenswrapper[4776]: I1125 10:59:14.932837 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-ovsdbserver-nb\") pod \"7fdc4742-4288-4756-8dfd-3f3f0d2174e2\" (UID: \"7fdc4742-4288-4756-8dfd-3f3f0d2174e2\") " Nov 25 10:59:14 crc kubenswrapper[4776]: I1125 10:59:14.932917 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-dns-svc\") pod \"7fdc4742-4288-4756-8dfd-3f3f0d2174e2\" (UID: \"7fdc4742-4288-4756-8dfd-3f3f0d2174e2\") " Nov 25 10:59:14 crc kubenswrapper[4776]: I1125 10:59:14.933020 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-ovsdbserver-sb\") pod \"7fdc4742-4288-4756-8dfd-3f3f0d2174e2\" (UID: \"7fdc4742-4288-4756-8dfd-3f3f0d2174e2\") " Nov 25 10:59:14 crc kubenswrapper[4776]: I1125 10:59:14.933056 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6zsc\" (UniqueName: \"kubernetes.io/projected/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-kube-api-access-l6zsc\") pod \"7fdc4742-4288-4756-8dfd-3f3f0d2174e2\" (UID: \"7fdc4742-4288-4756-8dfd-3f3f0d2174e2\") " Nov 25 10:59:14 crc kubenswrapper[4776]: I1125 10:59:14.933103 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-config\") pod \"7fdc4742-4288-4756-8dfd-3f3f0d2174e2\" (UID: \"7fdc4742-4288-4756-8dfd-3f3f0d2174e2\") " Nov 25 10:59:14 crc kubenswrapper[4776]: I1125 10:59:14.938927 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-kube-api-access-l6zsc" (OuterVolumeSpecName: "kube-api-access-l6zsc") pod "7fdc4742-4288-4756-8dfd-3f3f0d2174e2" (UID: "7fdc4742-4288-4756-8dfd-3f3f0d2174e2"). InnerVolumeSpecName "kube-api-access-l6zsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:59:14 crc kubenswrapper[4776]: I1125 10:59:14.977416 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7fdc4742-4288-4756-8dfd-3f3f0d2174e2" (UID: "7fdc4742-4288-4756-8dfd-3f3f0d2174e2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:14 crc kubenswrapper[4776]: I1125 10:59:14.978825 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7fdc4742-4288-4756-8dfd-3f3f0d2174e2" (UID: "7fdc4742-4288-4756-8dfd-3f3f0d2174e2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:14 crc kubenswrapper[4776]: I1125 10:59:14.980011 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7fdc4742-4288-4756-8dfd-3f3f0d2174e2" (UID: "7fdc4742-4288-4756-8dfd-3f3f0d2174e2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:14 crc kubenswrapper[4776]: I1125 10:59:14.983903 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-config" (OuterVolumeSpecName: "config") pod "7fdc4742-4288-4756-8dfd-3f3f0d2174e2" (UID: "7fdc4742-4288-4756-8dfd-3f3f0d2174e2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:15 crc kubenswrapper[4776]: I1125 10:59:15.035003 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:15 crc kubenswrapper[4776]: I1125 10:59:15.035040 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:15 crc kubenswrapper[4776]: I1125 10:59:15.035055 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:15 crc kubenswrapper[4776]: I1125 10:59:15.035088 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:15 crc kubenswrapper[4776]: I1125 10:59:15.035102 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6zsc\" (UniqueName: \"kubernetes.io/projected/7fdc4742-4288-4756-8dfd-3f3f0d2174e2-kube-api-access-l6zsc\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:15 crc kubenswrapper[4776]: I1125 10:59:15.446897 4776 generic.go:334] "Generic (PLEG): container finished" podID="7fdc4742-4288-4756-8dfd-3f3f0d2174e2" containerID="f46bb4d4a6a370a5680b4d71418cbdc0fa0aed30d3492279f265d8dfb0aa84f1" exitCode=0 Nov 25 10:59:15 crc kubenswrapper[4776]: I1125 10:59:15.446941 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d5c556d87-vf246" event={"ID":"7fdc4742-4288-4756-8dfd-3f3f0d2174e2","Type":"ContainerDied","Data":"f46bb4d4a6a370a5680b4d71418cbdc0fa0aed30d3492279f265d8dfb0aa84f1"} Nov 25 10:59:15 crc kubenswrapper[4776]: I1125 10:59:15.447009 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d5c556d87-vf246" Nov 25 10:59:15 crc kubenswrapper[4776]: I1125 10:59:15.447239 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d5c556d87-vf246" event={"ID":"7fdc4742-4288-4756-8dfd-3f3f0d2174e2","Type":"ContainerDied","Data":"aafebdf2e101ad2a8a6a5a09e9273cb8e9da432f29abc8692d0d0b5cf699d1a6"} Nov 25 10:59:15 crc kubenswrapper[4776]: I1125 10:59:15.447273 4776 scope.go:117] "RemoveContainer" containerID="f46bb4d4a6a370a5680b4d71418cbdc0fa0aed30d3492279f265d8dfb0aa84f1" Nov 25 10:59:15 crc kubenswrapper[4776]: I1125 10:59:15.475968 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d5c556d87-vf246"] Nov 25 10:59:15 crc kubenswrapper[4776]: I1125 10:59:15.476163 4776 scope.go:117] "RemoveContainer" containerID="0fec0a1c6a9f05a8bee620a8b81d7e77e7f5245e6ec480c3bb1babf947c2091c" Nov 25 10:59:15 crc kubenswrapper[4776]: I1125 10:59:15.484165 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d5c556d87-vf246"] Nov 25 10:59:15 crc kubenswrapper[4776]: I1125 10:59:15.506944 4776 scope.go:117] "RemoveContainer" containerID="f46bb4d4a6a370a5680b4d71418cbdc0fa0aed30d3492279f265d8dfb0aa84f1" Nov 25 10:59:15 crc kubenswrapper[4776]: E1125 10:59:15.507421 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f46bb4d4a6a370a5680b4d71418cbdc0fa0aed30d3492279f265d8dfb0aa84f1\": container with ID starting with f46bb4d4a6a370a5680b4d71418cbdc0fa0aed30d3492279f265d8dfb0aa84f1 not found: ID does not exist" containerID="f46bb4d4a6a370a5680b4d71418cbdc0fa0aed30d3492279f265d8dfb0aa84f1" Nov 25 10:59:15 crc kubenswrapper[4776]: I1125 10:59:15.507461 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f46bb4d4a6a370a5680b4d71418cbdc0fa0aed30d3492279f265d8dfb0aa84f1"} err="failed to get container status \"f46bb4d4a6a370a5680b4d71418cbdc0fa0aed30d3492279f265d8dfb0aa84f1\": rpc error: code = NotFound desc = could not find container \"f46bb4d4a6a370a5680b4d71418cbdc0fa0aed30d3492279f265d8dfb0aa84f1\": container with ID starting with f46bb4d4a6a370a5680b4d71418cbdc0fa0aed30d3492279f265d8dfb0aa84f1 not found: ID does not exist" Nov 25 10:59:15 crc kubenswrapper[4776]: I1125 10:59:15.507486 4776 scope.go:117] "RemoveContainer" containerID="0fec0a1c6a9f05a8bee620a8b81d7e77e7f5245e6ec480c3bb1babf947c2091c" Nov 25 10:59:15 crc kubenswrapper[4776]: E1125 10:59:15.507826 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fec0a1c6a9f05a8bee620a8b81d7e77e7f5245e6ec480c3bb1babf947c2091c\": container with ID starting with 0fec0a1c6a9f05a8bee620a8b81d7e77e7f5245e6ec480c3bb1babf947c2091c not found: ID does not exist" containerID="0fec0a1c6a9f05a8bee620a8b81d7e77e7f5245e6ec480c3bb1babf947c2091c" Nov 25 10:59:15 crc kubenswrapper[4776]: I1125 10:59:15.507853 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fec0a1c6a9f05a8bee620a8b81d7e77e7f5245e6ec480c3bb1babf947c2091c"} err="failed to get container status \"0fec0a1c6a9f05a8bee620a8b81d7e77e7f5245e6ec480c3bb1babf947c2091c\": rpc error: code = NotFound desc = could not find container \"0fec0a1c6a9f05a8bee620a8b81d7e77e7f5245e6ec480c3bb1babf947c2091c\": container with ID starting with 0fec0a1c6a9f05a8bee620a8b81d7e77e7f5245e6ec480c3bb1babf947c2091c not found: ID does not exist" Nov 25 10:59:15 crc kubenswrapper[4776]: I1125 10:59:15.674533 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fdc4742-4288-4756-8dfd-3f3f0d2174e2" path="/var/lib/kubelet/pods/7fdc4742-4288-4756-8dfd-3f3f0d2174e2/volumes" Nov 25 10:59:16 crc kubenswrapper[4776]: I1125 10:59:16.399631 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" Nov 25 10:59:16 crc kubenswrapper[4776]: I1125 10:59:16.399704 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" Nov 25 10:59:17 crc kubenswrapper[4776]: I1125 10:59:17.864855 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:17 crc kubenswrapper[4776]: I1125 10:59:17.865236 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5c6d9c7954-pvk2n" Nov 25 10:59:17 crc kubenswrapper[4776]: I1125 10:59:17.966870 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-65f4dfb8b6-bqpbj"] Nov 25 10:59:17 crc kubenswrapper[4776]: I1125 10:59:17.967124 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" podUID="e8919d1d-aa43-45f4-a1c3-37a9910e8a32" containerName="proxy-httpd" containerID="cri-o://34b982b5c2ce60541ba1de95d03cc6801c4015343223e772fa8d358b3a3b5394" gracePeriod=30 Nov 25 10:59:17 crc kubenswrapper[4776]: I1125 10:59:17.967248 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" podUID="e8919d1d-aa43-45f4-a1c3-37a9910e8a32" containerName="proxy-server" containerID="cri-o://df43f247d9591bacf4720ae864b30b0c3c66b9c2707d98c0aecd08dff0cb913e" gracePeriod=30 Nov 25 10:59:18 crc kubenswrapper[4776]: I1125 10:59:18.490472 4776 generic.go:334] "Generic (PLEG): container finished" podID="e8919d1d-aa43-45f4-a1c3-37a9910e8a32" containerID="34b982b5c2ce60541ba1de95d03cc6801c4015343223e772fa8d358b3a3b5394" exitCode=0 Nov 25 10:59:18 crc kubenswrapper[4776]: I1125 10:59:18.490580 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" event={"ID":"e8919d1d-aa43-45f4-a1c3-37a9910e8a32","Type":"ContainerDied","Data":"34b982b5c2ce60541ba1de95d03cc6801c4015343223e772fa8d358b3a3b5394"} Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.242748 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.319629 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-log-httpd\") pod \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\" (UID: \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\") " Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.319715 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtzvw\" (UniqueName: \"kubernetes.io/projected/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-kube-api-access-mtzvw\") pod \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\" (UID: \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\") " Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.319771 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-run-httpd\") pod \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\" (UID: \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\") " Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.319823 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-config-data\") pod \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\" (UID: \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\") " Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.320267 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e8919d1d-aa43-45f4-a1c3-37a9910e8a32" (UID: "e8919d1d-aa43-45f4-a1c3-37a9910e8a32"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.320367 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e8919d1d-aa43-45f4-a1c3-37a9910e8a32" (UID: "e8919d1d-aa43-45f4-a1c3-37a9910e8a32"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.320535 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-etc-swift\") pod \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\" (UID: \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\") " Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.321096 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-combined-ca-bundle\") pod \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\" (UID: \"e8919d1d-aa43-45f4-a1c3-37a9910e8a32\") " Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.322042 4776 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.322230 4776 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.334599 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "e8919d1d-aa43-45f4-a1c3-37a9910e8a32" (UID: "e8919d1d-aa43-45f4-a1c3-37a9910e8a32"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.350291 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-kube-api-access-mtzvw" (OuterVolumeSpecName: "kube-api-access-mtzvw") pod "e8919d1d-aa43-45f4-a1c3-37a9910e8a32" (UID: "e8919d1d-aa43-45f4-a1c3-37a9910e8a32"). InnerVolumeSpecName "kube-api-access-mtzvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.373362 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-config-data" (OuterVolumeSpecName: "config-data") pod "e8919d1d-aa43-45f4-a1c3-37a9910e8a32" (UID: "e8919d1d-aa43-45f4-a1c3-37a9910e8a32"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.381665 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e8919d1d-aa43-45f4-a1c3-37a9910e8a32" (UID: "e8919d1d-aa43-45f4-a1c3-37a9910e8a32"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.424074 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtzvw\" (UniqueName: \"kubernetes.io/projected/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-kube-api-access-mtzvw\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.424106 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.424115 4776 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.424123 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8919d1d-aa43-45f4-a1c3-37a9910e8a32-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.502778 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.502810 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" event={"ID":"e8919d1d-aa43-45f4-a1c3-37a9910e8a32","Type":"ContainerDied","Data":"df43f247d9591bacf4720ae864b30b0c3c66b9c2707d98c0aecd08dff0cb913e"} Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.502855 4776 scope.go:117] "RemoveContainer" containerID="df43f247d9591bacf4720ae864b30b0c3c66b9c2707d98c0aecd08dff0cb913e" Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.502782 4776 generic.go:334] "Generic (PLEG): container finished" podID="e8919d1d-aa43-45f4-a1c3-37a9910e8a32" containerID="df43f247d9591bacf4720ae864b30b0c3c66b9c2707d98c0aecd08dff0cb913e" exitCode=0 Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.502972 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-65f4dfb8b6-bqpbj" event={"ID":"e8919d1d-aa43-45f4-a1c3-37a9910e8a32","Type":"ContainerDied","Data":"4d78e0674d16f3de9de4fd1dde9df11cf1e1ffac3583b01d6a983902e9cfa731"} Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.544342 4776 scope.go:117] "RemoveContainer" containerID="34b982b5c2ce60541ba1de95d03cc6801c4015343223e772fa8d358b3a3b5394" Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.545744 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-65f4dfb8b6-bqpbj"] Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.556637 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-65f4dfb8b6-bqpbj"] Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.572157 4776 scope.go:117] "RemoveContainer" containerID="df43f247d9591bacf4720ae864b30b0c3c66b9c2707d98c0aecd08dff0cb913e" Nov 25 10:59:19 crc kubenswrapper[4776]: E1125 10:59:19.572818 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df43f247d9591bacf4720ae864b30b0c3c66b9c2707d98c0aecd08dff0cb913e\": container with ID starting with df43f247d9591bacf4720ae864b30b0c3c66b9c2707d98c0aecd08dff0cb913e not found: ID does not exist" containerID="df43f247d9591bacf4720ae864b30b0c3c66b9c2707d98c0aecd08dff0cb913e" Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.573089 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df43f247d9591bacf4720ae864b30b0c3c66b9c2707d98c0aecd08dff0cb913e"} err="failed to get container status \"df43f247d9591bacf4720ae864b30b0c3c66b9c2707d98c0aecd08dff0cb913e\": rpc error: code = NotFound desc = could not find container \"df43f247d9591bacf4720ae864b30b0c3c66b9c2707d98c0aecd08dff0cb913e\": container with ID starting with df43f247d9591bacf4720ae864b30b0c3c66b9c2707d98c0aecd08dff0cb913e not found: ID does not exist" Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.573255 4776 scope.go:117] "RemoveContainer" containerID="34b982b5c2ce60541ba1de95d03cc6801c4015343223e772fa8d358b3a3b5394" Nov 25 10:59:19 crc kubenswrapper[4776]: E1125 10:59:19.573805 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34b982b5c2ce60541ba1de95d03cc6801c4015343223e772fa8d358b3a3b5394\": container with ID starting with 34b982b5c2ce60541ba1de95d03cc6801c4015343223e772fa8d358b3a3b5394 not found: ID does not exist" containerID="34b982b5c2ce60541ba1de95d03cc6801c4015343223e772fa8d358b3a3b5394" Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.573839 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34b982b5c2ce60541ba1de95d03cc6801c4015343223e772fa8d358b3a3b5394"} err="failed to get container status \"34b982b5c2ce60541ba1de95d03cc6801c4015343223e772fa8d358b3a3b5394\": rpc error: code = NotFound desc = could not find container \"34b982b5c2ce60541ba1de95d03cc6801c4015343223e772fa8d358b3a3b5394\": container with ID starting with 34b982b5c2ce60541ba1de95d03cc6801c4015343223e772fa8d358b3a3b5394 not found: ID does not exist" Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.673078 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8919d1d-aa43-45f4-a1c3-37a9910e8a32" path="/var/lib/kubelet/pods/e8919d1d-aa43-45f4-a1c3-37a9910e8a32/volumes" Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.721738 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qqqvj" Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.722393 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qqqvj" Nov 25 10:59:19 crc kubenswrapper[4776]: I1125 10:59:19.773807 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qqqvj" Nov 25 10:59:20 crc kubenswrapper[4776]: I1125 10:59:20.561311 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qqqvj" Nov 25 10:59:20 crc kubenswrapper[4776]: I1125 10:59:20.608525 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qqqvj"] Nov 25 10:59:22 crc kubenswrapper[4776]: I1125 10:59:22.526696 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qqqvj" podUID="9f14e517-3052-4e3c-99c5-121293a8a542" containerName="registry-server" containerID="cri-o://cd0888d7369c3a96c99c268c08dc855dfdfec3b7555aa9aa833e56f7454edf13" gracePeriod=2 Nov 25 10:59:23 crc kubenswrapper[4776]: I1125 10:59:23.006899 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qqqvj" Nov 25 10:59:23 crc kubenswrapper[4776]: I1125 10:59:23.095550 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f14e517-3052-4e3c-99c5-121293a8a542-catalog-content\") pod \"9f14e517-3052-4e3c-99c5-121293a8a542\" (UID: \"9f14e517-3052-4e3c-99c5-121293a8a542\") " Nov 25 10:59:23 crc kubenswrapper[4776]: I1125 10:59:23.095611 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2jxc\" (UniqueName: \"kubernetes.io/projected/9f14e517-3052-4e3c-99c5-121293a8a542-kube-api-access-x2jxc\") pod \"9f14e517-3052-4e3c-99c5-121293a8a542\" (UID: \"9f14e517-3052-4e3c-99c5-121293a8a542\") " Nov 25 10:59:23 crc kubenswrapper[4776]: I1125 10:59:23.095721 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f14e517-3052-4e3c-99c5-121293a8a542-utilities\") pod \"9f14e517-3052-4e3c-99c5-121293a8a542\" (UID: \"9f14e517-3052-4e3c-99c5-121293a8a542\") " Nov 25 10:59:23 crc kubenswrapper[4776]: I1125 10:59:23.097165 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f14e517-3052-4e3c-99c5-121293a8a542-utilities" (OuterVolumeSpecName: "utilities") pod "9f14e517-3052-4e3c-99c5-121293a8a542" (UID: "9f14e517-3052-4e3c-99c5-121293a8a542"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:59:23 crc kubenswrapper[4776]: I1125 10:59:23.103990 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f14e517-3052-4e3c-99c5-121293a8a542-kube-api-access-x2jxc" (OuterVolumeSpecName: "kube-api-access-x2jxc") pod "9f14e517-3052-4e3c-99c5-121293a8a542" (UID: "9f14e517-3052-4e3c-99c5-121293a8a542"). InnerVolumeSpecName "kube-api-access-x2jxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:59:23 crc kubenswrapper[4776]: I1125 10:59:23.123192 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f14e517-3052-4e3c-99c5-121293a8a542-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9f14e517-3052-4e3c-99c5-121293a8a542" (UID: "9f14e517-3052-4e3c-99c5-121293a8a542"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:59:23 crc kubenswrapper[4776]: I1125 10:59:23.198949 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f14e517-3052-4e3c-99c5-121293a8a542-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:23 crc kubenswrapper[4776]: I1125 10:59:23.199357 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f14e517-3052-4e3c-99c5-121293a8a542-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:23 crc kubenswrapper[4776]: I1125 10:59:23.199462 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2jxc\" (UniqueName: \"kubernetes.io/projected/9f14e517-3052-4e3c-99c5-121293a8a542-kube-api-access-x2jxc\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:23 crc kubenswrapper[4776]: I1125 10:59:23.536487 4776 generic.go:334] "Generic (PLEG): container finished" podID="9f14e517-3052-4e3c-99c5-121293a8a542" containerID="cd0888d7369c3a96c99c268c08dc855dfdfec3b7555aa9aa833e56f7454edf13" exitCode=0 Nov 25 10:59:23 crc kubenswrapper[4776]: I1125 10:59:23.536544 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qqqvj" Nov 25 10:59:23 crc kubenswrapper[4776]: I1125 10:59:23.536544 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qqqvj" event={"ID":"9f14e517-3052-4e3c-99c5-121293a8a542","Type":"ContainerDied","Data":"cd0888d7369c3a96c99c268c08dc855dfdfec3b7555aa9aa833e56f7454edf13"} Nov 25 10:59:23 crc kubenswrapper[4776]: I1125 10:59:23.536616 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qqqvj" event={"ID":"9f14e517-3052-4e3c-99c5-121293a8a542","Type":"ContainerDied","Data":"5845e754bc368122b043f0bcd0b42db5d33cc6afab5d750cce7c6797a9fa6725"} Nov 25 10:59:23 crc kubenswrapper[4776]: I1125 10:59:23.536659 4776 scope.go:117] "RemoveContainer" containerID="cd0888d7369c3a96c99c268c08dc855dfdfec3b7555aa9aa833e56f7454edf13" Nov 25 10:59:23 crc kubenswrapper[4776]: I1125 10:59:23.565572 4776 scope.go:117] "RemoveContainer" containerID="47dd73561a5135188df8f2e28fccb30b8e2bb685ef792e0e27b5890741aefaae" Nov 25 10:59:23 crc kubenswrapper[4776]: I1125 10:59:23.580567 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qqqvj"] Nov 25 10:59:23 crc kubenswrapper[4776]: I1125 10:59:23.591840 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qqqvj"] Nov 25 10:59:23 crc kubenswrapper[4776]: I1125 10:59:23.593132 4776 scope.go:117] "RemoveContainer" containerID="6feb09434cdc3ae1fd8ba6d7f85241c4fc9285dd96afb717abb52822756e69f9" Nov 25 10:59:23 crc kubenswrapper[4776]: I1125 10:59:23.641538 4776 scope.go:117] "RemoveContainer" containerID="cd0888d7369c3a96c99c268c08dc855dfdfec3b7555aa9aa833e56f7454edf13" Nov 25 10:59:23 crc kubenswrapper[4776]: E1125 10:59:23.642023 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd0888d7369c3a96c99c268c08dc855dfdfec3b7555aa9aa833e56f7454edf13\": container with ID starting with cd0888d7369c3a96c99c268c08dc855dfdfec3b7555aa9aa833e56f7454edf13 not found: ID does not exist" containerID="cd0888d7369c3a96c99c268c08dc855dfdfec3b7555aa9aa833e56f7454edf13" Nov 25 10:59:23 crc kubenswrapper[4776]: I1125 10:59:23.642286 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd0888d7369c3a96c99c268c08dc855dfdfec3b7555aa9aa833e56f7454edf13"} err="failed to get container status \"cd0888d7369c3a96c99c268c08dc855dfdfec3b7555aa9aa833e56f7454edf13\": rpc error: code = NotFound desc = could not find container \"cd0888d7369c3a96c99c268c08dc855dfdfec3b7555aa9aa833e56f7454edf13\": container with ID starting with cd0888d7369c3a96c99c268c08dc855dfdfec3b7555aa9aa833e56f7454edf13 not found: ID does not exist" Nov 25 10:59:23 crc kubenswrapper[4776]: I1125 10:59:23.642438 4776 scope.go:117] "RemoveContainer" containerID="47dd73561a5135188df8f2e28fccb30b8e2bb685ef792e0e27b5890741aefaae" Nov 25 10:59:23 crc kubenswrapper[4776]: E1125 10:59:23.642980 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47dd73561a5135188df8f2e28fccb30b8e2bb685ef792e0e27b5890741aefaae\": container with ID starting with 47dd73561a5135188df8f2e28fccb30b8e2bb685ef792e0e27b5890741aefaae not found: ID does not exist" containerID="47dd73561a5135188df8f2e28fccb30b8e2bb685ef792e0e27b5890741aefaae" Nov 25 10:59:23 crc kubenswrapper[4776]: I1125 10:59:23.643139 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47dd73561a5135188df8f2e28fccb30b8e2bb685ef792e0e27b5890741aefaae"} err="failed to get container status \"47dd73561a5135188df8f2e28fccb30b8e2bb685ef792e0e27b5890741aefaae\": rpc error: code = NotFound desc = could not find container \"47dd73561a5135188df8f2e28fccb30b8e2bb685ef792e0e27b5890741aefaae\": container with ID starting with 47dd73561a5135188df8f2e28fccb30b8e2bb685ef792e0e27b5890741aefaae not found: ID does not exist" Nov 25 10:59:23 crc kubenswrapper[4776]: I1125 10:59:23.643275 4776 scope.go:117] "RemoveContainer" containerID="6feb09434cdc3ae1fd8ba6d7f85241c4fc9285dd96afb717abb52822756e69f9" Nov 25 10:59:23 crc kubenswrapper[4776]: E1125 10:59:23.643608 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6feb09434cdc3ae1fd8ba6d7f85241c4fc9285dd96afb717abb52822756e69f9\": container with ID starting with 6feb09434cdc3ae1fd8ba6d7f85241c4fc9285dd96afb717abb52822756e69f9 not found: ID does not exist" containerID="6feb09434cdc3ae1fd8ba6d7f85241c4fc9285dd96afb717abb52822756e69f9" Nov 25 10:59:23 crc kubenswrapper[4776]: I1125 10:59:23.643766 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6feb09434cdc3ae1fd8ba6d7f85241c4fc9285dd96afb717abb52822756e69f9"} err="failed to get container status \"6feb09434cdc3ae1fd8ba6d7f85241c4fc9285dd96afb717abb52822756e69f9\": rpc error: code = NotFound desc = could not find container \"6feb09434cdc3ae1fd8ba6d7f85241c4fc9285dd96afb717abb52822756e69f9\": container with ID starting with 6feb09434cdc3ae1fd8ba6d7f85241c4fc9285dd96afb717abb52822756e69f9 not found: ID does not exist" Nov 25 10:59:23 crc kubenswrapper[4776]: I1125 10:59:23.690800 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f14e517-3052-4e3c-99c5-121293a8a542" path="/var/lib/kubelet/pods/9f14e517-3052-4e3c-99c5-121293a8a542/volumes" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.178766 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-99m2h"] Nov 25 10:59:24 crc kubenswrapper[4776]: E1125 10:59:24.179718 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8919d1d-aa43-45f4-a1c3-37a9910e8a32" containerName="proxy-httpd" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.179802 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8919d1d-aa43-45f4-a1c3-37a9910e8a32" containerName="proxy-httpd" Nov 25 10:59:24 crc kubenswrapper[4776]: E1125 10:59:24.179872 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fdc4742-4288-4756-8dfd-3f3f0d2174e2" containerName="init" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.179924 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fdc4742-4288-4756-8dfd-3f3f0d2174e2" containerName="init" Nov 25 10:59:24 crc kubenswrapper[4776]: E1125 10:59:24.179973 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f14e517-3052-4e3c-99c5-121293a8a542" containerName="registry-server" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.180043 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f14e517-3052-4e3c-99c5-121293a8a542" containerName="registry-server" Nov 25 10:59:24 crc kubenswrapper[4776]: E1125 10:59:24.180128 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f14e517-3052-4e3c-99c5-121293a8a542" containerName="extract-content" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.180186 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f14e517-3052-4e3c-99c5-121293a8a542" containerName="extract-content" Nov 25 10:59:24 crc kubenswrapper[4776]: E1125 10:59:24.180251 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f14e517-3052-4e3c-99c5-121293a8a542" containerName="extract-utilities" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.180309 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f14e517-3052-4e3c-99c5-121293a8a542" containerName="extract-utilities" Nov 25 10:59:24 crc kubenswrapper[4776]: E1125 10:59:24.180369 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fdc4742-4288-4756-8dfd-3f3f0d2174e2" containerName="dnsmasq-dns" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.180419 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fdc4742-4288-4756-8dfd-3f3f0d2174e2" containerName="dnsmasq-dns" Nov 25 10:59:24 crc kubenswrapper[4776]: E1125 10:59:24.180483 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8919d1d-aa43-45f4-a1c3-37a9910e8a32" containerName="proxy-server" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.180538 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8919d1d-aa43-45f4-a1c3-37a9910e8a32" containerName="proxy-server" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.180754 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f14e517-3052-4e3c-99c5-121293a8a542" containerName="registry-server" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.180827 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8919d1d-aa43-45f4-a1c3-37a9910e8a32" containerName="proxy-httpd" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.180888 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fdc4742-4288-4756-8dfd-3f3f0d2174e2" containerName="dnsmasq-dns" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.180952 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8919d1d-aa43-45f4-a1c3-37a9910e8a32" containerName="proxy-server" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.181594 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-99m2h" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.189117 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-99m2h"] Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.195899 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-ea9d-account-create-gtwm6"] Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.197109 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ea9d-account-create-gtwm6" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.232006 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-th6lr\" (UniqueName: \"kubernetes.io/projected/54887439-da3f-418c-85bd-3754d009e466-kube-api-access-th6lr\") pod \"cinder-db-create-99m2h\" (UID: \"54887439-da3f-418c-85bd-3754d009e466\") " pod="openstack/cinder-db-create-99m2h" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.232150 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54887439-da3f-418c-85bd-3754d009e466-operator-scripts\") pod \"cinder-db-create-99m2h\" (UID: \"54887439-da3f-418c-85bd-3754d009e466\") " pod="openstack/cinder-db-create-99m2h" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.232563 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.270943 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ea9d-account-create-gtwm6"] Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.334121 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55d65b36-5e84-4842-8184-2bf6ed987f45-operator-scripts\") pod \"cinder-ea9d-account-create-gtwm6\" (UID: \"55d65b36-5e84-4842-8184-2bf6ed987f45\") " pod="openstack/cinder-ea9d-account-create-gtwm6" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.334211 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-th6lr\" (UniqueName: \"kubernetes.io/projected/54887439-da3f-418c-85bd-3754d009e466-kube-api-access-th6lr\") pod \"cinder-db-create-99m2h\" (UID: \"54887439-da3f-418c-85bd-3754d009e466\") " pod="openstack/cinder-db-create-99m2h" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.334262 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsr7c\" (UniqueName: \"kubernetes.io/projected/55d65b36-5e84-4842-8184-2bf6ed987f45-kube-api-access-dsr7c\") pod \"cinder-ea9d-account-create-gtwm6\" (UID: \"55d65b36-5e84-4842-8184-2bf6ed987f45\") " pod="openstack/cinder-ea9d-account-create-gtwm6" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.334323 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54887439-da3f-418c-85bd-3754d009e466-operator-scripts\") pod \"cinder-db-create-99m2h\" (UID: \"54887439-da3f-418c-85bd-3754d009e466\") " pod="openstack/cinder-db-create-99m2h" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.335295 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54887439-da3f-418c-85bd-3754d009e466-operator-scripts\") pod \"cinder-db-create-99m2h\" (UID: \"54887439-da3f-418c-85bd-3754d009e466\") " pod="openstack/cinder-db-create-99m2h" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.366267 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-th6lr\" (UniqueName: \"kubernetes.io/projected/54887439-da3f-418c-85bd-3754d009e466-kube-api-access-th6lr\") pod \"cinder-db-create-99m2h\" (UID: \"54887439-da3f-418c-85bd-3754d009e466\") " pod="openstack/cinder-db-create-99m2h" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.435595 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55d65b36-5e84-4842-8184-2bf6ed987f45-operator-scripts\") pod \"cinder-ea9d-account-create-gtwm6\" (UID: \"55d65b36-5e84-4842-8184-2bf6ed987f45\") " pod="openstack/cinder-ea9d-account-create-gtwm6" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.435693 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsr7c\" (UniqueName: \"kubernetes.io/projected/55d65b36-5e84-4842-8184-2bf6ed987f45-kube-api-access-dsr7c\") pod \"cinder-ea9d-account-create-gtwm6\" (UID: \"55d65b36-5e84-4842-8184-2bf6ed987f45\") " pod="openstack/cinder-ea9d-account-create-gtwm6" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.436640 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55d65b36-5e84-4842-8184-2bf6ed987f45-operator-scripts\") pod \"cinder-ea9d-account-create-gtwm6\" (UID: \"55d65b36-5e84-4842-8184-2bf6ed987f45\") " pod="openstack/cinder-ea9d-account-create-gtwm6" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.452688 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsr7c\" (UniqueName: \"kubernetes.io/projected/55d65b36-5e84-4842-8184-2bf6ed987f45-kube-api-access-dsr7c\") pod \"cinder-ea9d-account-create-gtwm6\" (UID: \"55d65b36-5e84-4842-8184-2bf6ed987f45\") " pod="openstack/cinder-ea9d-account-create-gtwm6" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.557843 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-99m2h" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.571984 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ea9d-account-create-gtwm6" Nov 25 10:59:24 crc kubenswrapper[4776]: I1125 10:59:24.999558 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-99m2h"] Nov 25 10:59:25 crc kubenswrapper[4776]: W1125 10:59:25.003152 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54887439_da3f_418c_85bd_3754d009e466.slice/crio-2627f6561900c1a5142b433139921a9faf11e0053305fd8ac3deca187f6e886a WatchSource:0}: Error finding container 2627f6561900c1a5142b433139921a9faf11e0053305fd8ac3deca187f6e886a: Status 404 returned error can't find the container with id 2627f6561900c1a5142b433139921a9faf11e0053305fd8ac3deca187f6e886a Nov 25 10:59:25 crc kubenswrapper[4776]: I1125 10:59:25.110449 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ea9d-account-create-gtwm6"] Nov 25 10:59:25 crc kubenswrapper[4776]: I1125 10:59:25.561512 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ea9d-account-create-gtwm6" event={"ID":"55d65b36-5e84-4842-8184-2bf6ed987f45","Type":"ContainerStarted","Data":"4dade33a1326ebc318b0c703187ab1716a56503c32445a3856d0ad68bf7c7564"} Nov 25 10:59:25 crc kubenswrapper[4776]: I1125 10:59:25.561558 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ea9d-account-create-gtwm6" event={"ID":"55d65b36-5e84-4842-8184-2bf6ed987f45","Type":"ContainerStarted","Data":"61dbef1423da6ba0b9fb88562b7a439ce3d356c3d6e8c2324e3e539154bdfeaa"} Nov 25 10:59:25 crc kubenswrapper[4776]: I1125 10:59:25.563738 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-99m2h" event={"ID":"54887439-da3f-418c-85bd-3754d009e466","Type":"ContainerStarted","Data":"ceadafb6f7dd12bf36e87137b5b7f7a26ddf7fac48ce172f88cbba9e25611f51"} Nov 25 10:59:25 crc kubenswrapper[4776]: I1125 10:59:25.563771 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-99m2h" event={"ID":"54887439-da3f-418c-85bd-3754d009e466","Type":"ContainerStarted","Data":"2627f6561900c1a5142b433139921a9faf11e0053305fd8ac3deca187f6e886a"} Nov 25 10:59:25 crc kubenswrapper[4776]: I1125 10:59:25.577853 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-ea9d-account-create-gtwm6" podStartSLOduration=1.577837027 podStartE2EDuration="1.577837027s" podCreationTimestamp="2025-11-25 10:59:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:59:25.576483064 +0000 UTC m=+5710.617542617" watchObservedRunningTime="2025-11-25 10:59:25.577837027 +0000 UTC m=+5710.618896580" Nov 25 10:59:25 crc kubenswrapper[4776]: I1125 10:59:25.595092 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-99m2h" podStartSLOduration=1.5950501400000001 podStartE2EDuration="1.59505014s" podCreationTimestamp="2025-11-25 10:59:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:59:25.591425229 +0000 UTC m=+5710.632484802" watchObservedRunningTime="2025-11-25 10:59:25.59505014 +0000 UTC m=+5710.636109693" Nov 25 10:59:26 crc kubenswrapper[4776]: I1125 10:59:26.572738 4776 generic.go:334] "Generic (PLEG): container finished" podID="55d65b36-5e84-4842-8184-2bf6ed987f45" containerID="4dade33a1326ebc318b0c703187ab1716a56503c32445a3856d0ad68bf7c7564" exitCode=0 Nov 25 10:59:26 crc kubenswrapper[4776]: I1125 10:59:26.572806 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ea9d-account-create-gtwm6" event={"ID":"55d65b36-5e84-4842-8184-2bf6ed987f45","Type":"ContainerDied","Data":"4dade33a1326ebc318b0c703187ab1716a56503c32445a3856d0ad68bf7c7564"} Nov 25 10:59:26 crc kubenswrapper[4776]: I1125 10:59:26.575683 4776 generic.go:334] "Generic (PLEG): container finished" podID="54887439-da3f-418c-85bd-3754d009e466" containerID="ceadafb6f7dd12bf36e87137b5b7f7a26ddf7fac48ce172f88cbba9e25611f51" exitCode=0 Nov 25 10:59:26 crc kubenswrapper[4776]: I1125 10:59:26.575777 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-99m2h" event={"ID":"54887439-da3f-418c-85bd-3754d009e466","Type":"ContainerDied","Data":"ceadafb6f7dd12bf36e87137b5b7f7a26ddf7fac48ce172f88cbba9e25611f51"} Nov 25 10:59:28 crc kubenswrapper[4776]: I1125 10:59:28.042455 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-99m2h" Nov 25 10:59:28 crc kubenswrapper[4776]: I1125 10:59:28.051354 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ea9d-account-create-gtwm6" Nov 25 10:59:28 crc kubenswrapper[4776]: I1125 10:59:28.120651 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-th6lr\" (UniqueName: \"kubernetes.io/projected/54887439-da3f-418c-85bd-3754d009e466-kube-api-access-th6lr\") pod \"54887439-da3f-418c-85bd-3754d009e466\" (UID: \"54887439-da3f-418c-85bd-3754d009e466\") " Nov 25 10:59:28 crc kubenswrapper[4776]: I1125 10:59:28.120760 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54887439-da3f-418c-85bd-3754d009e466-operator-scripts\") pod \"54887439-da3f-418c-85bd-3754d009e466\" (UID: \"54887439-da3f-418c-85bd-3754d009e466\") " Nov 25 10:59:28 crc kubenswrapper[4776]: I1125 10:59:28.120962 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55d65b36-5e84-4842-8184-2bf6ed987f45-operator-scripts\") pod \"55d65b36-5e84-4842-8184-2bf6ed987f45\" (UID: \"55d65b36-5e84-4842-8184-2bf6ed987f45\") " Nov 25 10:59:28 crc kubenswrapper[4776]: I1125 10:59:28.120989 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsr7c\" (UniqueName: \"kubernetes.io/projected/55d65b36-5e84-4842-8184-2bf6ed987f45-kube-api-access-dsr7c\") pod \"55d65b36-5e84-4842-8184-2bf6ed987f45\" (UID: \"55d65b36-5e84-4842-8184-2bf6ed987f45\") " Nov 25 10:59:28 crc kubenswrapper[4776]: I1125 10:59:28.121629 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54887439-da3f-418c-85bd-3754d009e466-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "54887439-da3f-418c-85bd-3754d009e466" (UID: "54887439-da3f-418c-85bd-3754d009e466"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:28 crc kubenswrapper[4776]: I1125 10:59:28.121892 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55d65b36-5e84-4842-8184-2bf6ed987f45-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "55d65b36-5e84-4842-8184-2bf6ed987f45" (UID: "55d65b36-5e84-4842-8184-2bf6ed987f45"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:28 crc kubenswrapper[4776]: I1125 10:59:28.129822 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54887439-da3f-418c-85bd-3754d009e466-kube-api-access-th6lr" (OuterVolumeSpecName: "kube-api-access-th6lr") pod "54887439-da3f-418c-85bd-3754d009e466" (UID: "54887439-da3f-418c-85bd-3754d009e466"). InnerVolumeSpecName "kube-api-access-th6lr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:59:28 crc kubenswrapper[4776]: I1125 10:59:28.130097 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55d65b36-5e84-4842-8184-2bf6ed987f45-kube-api-access-dsr7c" (OuterVolumeSpecName: "kube-api-access-dsr7c") pod "55d65b36-5e84-4842-8184-2bf6ed987f45" (UID: "55d65b36-5e84-4842-8184-2bf6ed987f45"). InnerVolumeSpecName "kube-api-access-dsr7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:59:28 crc kubenswrapper[4776]: I1125 10:59:28.223366 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-th6lr\" (UniqueName: \"kubernetes.io/projected/54887439-da3f-418c-85bd-3754d009e466-kube-api-access-th6lr\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:28 crc kubenswrapper[4776]: I1125 10:59:28.223417 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54887439-da3f-418c-85bd-3754d009e466-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:28 crc kubenswrapper[4776]: I1125 10:59:28.223429 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55d65b36-5e84-4842-8184-2bf6ed987f45-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:28 crc kubenswrapper[4776]: I1125 10:59:28.223441 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsr7c\" (UniqueName: \"kubernetes.io/projected/55d65b36-5e84-4842-8184-2bf6ed987f45-kube-api-access-dsr7c\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:28 crc kubenswrapper[4776]: I1125 10:59:28.596282 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ea9d-account-create-gtwm6" event={"ID":"55d65b36-5e84-4842-8184-2bf6ed987f45","Type":"ContainerDied","Data":"61dbef1423da6ba0b9fb88562b7a439ce3d356c3d6e8c2324e3e539154bdfeaa"} Nov 25 10:59:28 crc kubenswrapper[4776]: I1125 10:59:28.596323 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61dbef1423da6ba0b9fb88562b7a439ce3d356c3d6e8c2324e3e539154bdfeaa" Nov 25 10:59:28 crc kubenswrapper[4776]: I1125 10:59:28.596375 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ea9d-account-create-gtwm6" Nov 25 10:59:28 crc kubenswrapper[4776]: I1125 10:59:28.597868 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-99m2h" event={"ID":"54887439-da3f-418c-85bd-3754d009e466","Type":"ContainerDied","Data":"2627f6561900c1a5142b433139921a9faf11e0053305fd8ac3deca187f6e886a"} Nov 25 10:59:28 crc kubenswrapper[4776]: I1125 10:59:28.597927 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-99m2h" Nov 25 10:59:28 crc kubenswrapper[4776]: I1125 10:59:28.597938 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2627f6561900c1a5142b433139921a9faf11e0053305fd8ac3deca187f6e886a" Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.350108 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-4f94t"] Nov 25 10:59:29 crc kubenswrapper[4776]: E1125 10:59:29.351030 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55d65b36-5e84-4842-8184-2bf6ed987f45" containerName="mariadb-account-create" Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.351053 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="55d65b36-5e84-4842-8184-2bf6ed987f45" containerName="mariadb-account-create" Nov 25 10:59:29 crc kubenswrapper[4776]: E1125 10:59:29.351130 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54887439-da3f-418c-85bd-3754d009e466" containerName="mariadb-database-create" Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.351137 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="54887439-da3f-418c-85bd-3754d009e466" containerName="mariadb-database-create" Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.351550 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="54887439-da3f-418c-85bd-3754d009e466" containerName="mariadb-database-create" Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.351570 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="55d65b36-5e84-4842-8184-2bf6ed987f45" containerName="mariadb-account-create" Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.352435 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-4f94t" Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.355389 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-cqpk5" Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.355448 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.370026 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-4f94t"] Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.370767 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.446986 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90693935-44c2-4e50-a8ad-5e2967ac582f-config-data\") pod \"cinder-db-sync-4f94t\" (UID: \"90693935-44c2-4e50-a8ad-5e2967ac582f\") " pod="openstack/cinder-db-sync-4f94t" Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.447041 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/90693935-44c2-4e50-a8ad-5e2967ac582f-etc-machine-id\") pod \"cinder-db-sync-4f94t\" (UID: \"90693935-44c2-4e50-a8ad-5e2967ac582f\") " pod="openstack/cinder-db-sync-4f94t" Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.447082 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90693935-44c2-4e50-a8ad-5e2967ac582f-scripts\") pod \"cinder-db-sync-4f94t\" (UID: \"90693935-44c2-4e50-a8ad-5e2967ac582f\") " pod="openstack/cinder-db-sync-4f94t" Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.447127 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wjtl\" (UniqueName: \"kubernetes.io/projected/90693935-44c2-4e50-a8ad-5e2967ac582f-kube-api-access-2wjtl\") pod \"cinder-db-sync-4f94t\" (UID: \"90693935-44c2-4e50-a8ad-5e2967ac582f\") " pod="openstack/cinder-db-sync-4f94t" Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.447545 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90693935-44c2-4e50-a8ad-5e2967ac582f-combined-ca-bundle\") pod \"cinder-db-sync-4f94t\" (UID: \"90693935-44c2-4e50-a8ad-5e2967ac582f\") " pod="openstack/cinder-db-sync-4f94t" Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.448053 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/90693935-44c2-4e50-a8ad-5e2967ac582f-db-sync-config-data\") pod \"cinder-db-sync-4f94t\" (UID: \"90693935-44c2-4e50-a8ad-5e2967ac582f\") " pod="openstack/cinder-db-sync-4f94t" Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.549678 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90693935-44c2-4e50-a8ad-5e2967ac582f-config-data\") pod \"cinder-db-sync-4f94t\" (UID: \"90693935-44c2-4e50-a8ad-5e2967ac582f\") " pod="openstack/cinder-db-sync-4f94t" Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.549759 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/90693935-44c2-4e50-a8ad-5e2967ac582f-etc-machine-id\") pod \"cinder-db-sync-4f94t\" (UID: \"90693935-44c2-4e50-a8ad-5e2967ac582f\") " pod="openstack/cinder-db-sync-4f94t" Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.549784 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90693935-44c2-4e50-a8ad-5e2967ac582f-scripts\") pod \"cinder-db-sync-4f94t\" (UID: \"90693935-44c2-4e50-a8ad-5e2967ac582f\") " pod="openstack/cinder-db-sync-4f94t" Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.549817 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wjtl\" (UniqueName: \"kubernetes.io/projected/90693935-44c2-4e50-a8ad-5e2967ac582f-kube-api-access-2wjtl\") pod \"cinder-db-sync-4f94t\" (UID: \"90693935-44c2-4e50-a8ad-5e2967ac582f\") " pod="openstack/cinder-db-sync-4f94t" Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.549825 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/90693935-44c2-4e50-a8ad-5e2967ac582f-etc-machine-id\") pod \"cinder-db-sync-4f94t\" (UID: \"90693935-44c2-4e50-a8ad-5e2967ac582f\") " pod="openstack/cinder-db-sync-4f94t" Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.549932 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90693935-44c2-4e50-a8ad-5e2967ac582f-combined-ca-bundle\") pod \"cinder-db-sync-4f94t\" (UID: \"90693935-44c2-4e50-a8ad-5e2967ac582f\") " pod="openstack/cinder-db-sync-4f94t" Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.550008 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/90693935-44c2-4e50-a8ad-5e2967ac582f-db-sync-config-data\") pod \"cinder-db-sync-4f94t\" (UID: \"90693935-44c2-4e50-a8ad-5e2967ac582f\") " pod="openstack/cinder-db-sync-4f94t" Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.556283 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90693935-44c2-4e50-a8ad-5e2967ac582f-scripts\") pod \"cinder-db-sync-4f94t\" (UID: \"90693935-44c2-4e50-a8ad-5e2967ac582f\") " pod="openstack/cinder-db-sync-4f94t" Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.558501 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90693935-44c2-4e50-a8ad-5e2967ac582f-config-data\") pod \"cinder-db-sync-4f94t\" (UID: \"90693935-44c2-4e50-a8ad-5e2967ac582f\") " pod="openstack/cinder-db-sync-4f94t" Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.559467 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/90693935-44c2-4e50-a8ad-5e2967ac582f-db-sync-config-data\") pod \"cinder-db-sync-4f94t\" (UID: \"90693935-44c2-4e50-a8ad-5e2967ac582f\") " pod="openstack/cinder-db-sync-4f94t" Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.560629 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90693935-44c2-4e50-a8ad-5e2967ac582f-combined-ca-bundle\") pod \"cinder-db-sync-4f94t\" (UID: \"90693935-44c2-4e50-a8ad-5e2967ac582f\") " pod="openstack/cinder-db-sync-4f94t" Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.567876 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wjtl\" (UniqueName: \"kubernetes.io/projected/90693935-44c2-4e50-a8ad-5e2967ac582f-kube-api-access-2wjtl\") pod \"cinder-db-sync-4f94t\" (UID: \"90693935-44c2-4e50-a8ad-5e2967ac582f\") " pod="openstack/cinder-db-sync-4f94t" Nov 25 10:59:29 crc kubenswrapper[4776]: I1125 10:59:29.680022 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-4f94t" Nov 25 10:59:30 crc kubenswrapper[4776]: I1125 10:59:30.184346 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-4f94t"] Nov 25 10:59:30 crc kubenswrapper[4776]: W1125 10:59:30.189878 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90693935_44c2_4e50_a8ad_5e2967ac582f.slice/crio-a50d48760965ddd55050ec311f5c98725704e49d5f8afae9905feaa90bc310e3 WatchSource:0}: Error finding container a50d48760965ddd55050ec311f5c98725704e49d5f8afae9905feaa90bc310e3: Status 404 returned error can't find the container with id a50d48760965ddd55050ec311f5c98725704e49d5f8afae9905feaa90bc310e3 Nov 25 10:59:30 crc kubenswrapper[4776]: I1125 10:59:30.654233 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-4f94t" event={"ID":"90693935-44c2-4e50-a8ad-5e2967ac582f","Type":"ContainerStarted","Data":"a50d48760965ddd55050ec311f5c98725704e49d5f8afae9905feaa90bc310e3"} Nov 25 10:59:31 crc kubenswrapper[4776]: I1125 10:59:31.670652 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-4f94t" event={"ID":"90693935-44c2-4e50-a8ad-5e2967ac582f","Type":"ContainerStarted","Data":"2b63f488b275331fa7a54b18739ebdbff1487a47277829a841d45990dd323cb6"} Nov 25 10:59:31 crc kubenswrapper[4776]: I1125 10:59:31.685104 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-4f94t" podStartSLOduration=2.685036163 podStartE2EDuration="2.685036163s" podCreationTimestamp="2025-11-25 10:59:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:59:31.680685683 +0000 UTC m=+5716.721745246" watchObservedRunningTime="2025-11-25 10:59:31.685036163 +0000 UTC m=+5716.726095716" Nov 25 10:59:32 crc kubenswrapper[4776]: I1125 10:59:32.392770 4776 scope.go:117] "RemoveContainer" containerID="ef0b5866a0822f88197714a524be43881cd9ff6f5151bd4079365aff874aa509" Nov 25 10:59:37 crc kubenswrapper[4776]: I1125 10:59:37.728806 4776 generic.go:334] "Generic (PLEG): container finished" podID="90693935-44c2-4e50-a8ad-5e2967ac582f" containerID="2b63f488b275331fa7a54b18739ebdbff1487a47277829a841d45990dd323cb6" exitCode=0 Nov 25 10:59:37 crc kubenswrapper[4776]: I1125 10:59:37.728894 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-4f94t" event={"ID":"90693935-44c2-4e50-a8ad-5e2967ac582f","Type":"ContainerDied","Data":"2b63f488b275331fa7a54b18739ebdbff1487a47277829a841d45990dd323cb6"} Nov 25 10:59:39 crc kubenswrapper[4776]: I1125 10:59:39.034099 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-4f94t" Nov 25 10:59:39 crc kubenswrapper[4776]: I1125 10:59:39.181809 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/90693935-44c2-4e50-a8ad-5e2967ac582f-db-sync-config-data\") pod \"90693935-44c2-4e50-a8ad-5e2967ac582f\" (UID: \"90693935-44c2-4e50-a8ad-5e2967ac582f\") " Nov 25 10:59:39 crc kubenswrapper[4776]: I1125 10:59:39.181910 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90693935-44c2-4e50-a8ad-5e2967ac582f-config-data\") pod \"90693935-44c2-4e50-a8ad-5e2967ac582f\" (UID: \"90693935-44c2-4e50-a8ad-5e2967ac582f\") " Nov 25 10:59:39 crc kubenswrapper[4776]: I1125 10:59:39.182113 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90693935-44c2-4e50-a8ad-5e2967ac582f-scripts\") pod \"90693935-44c2-4e50-a8ad-5e2967ac582f\" (UID: \"90693935-44c2-4e50-a8ad-5e2967ac582f\") " Nov 25 10:59:39 crc kubenswrapper[4776]: I1125 10:59:39.182160 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90693935-44c2-4e50-a8ad-5e2967ac582f-combined-ca-bundle\") pod \"90693935-44c2-4e50-a8ad-5e2967ac582f\" (UID: \"90693935-44c2-4e50-a8ad-5e2967ac582f\") " Nov 25 10:59:39 crc kubenswrapper[4776]: I1125 10:59:39.182210 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/90693935-44c2-4e50-a8ad-5e2967ac582f-etc-machine-id\") pod \"90693935-44c2-4e50-a8ad-5e2967ac582f\" (UID: \"90693935-44c2-4e50-a8ad-5e2967ac582f\") " Nov 25 10:59:39 crc kubenswrapper[4776]: I1125 10:59:39.182294 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wjtl\" (UniqueName: \"kubernetes.io/projected/90693935-44c2-4e50-a8ad-5e2967ac582f-kube-api-access-2wjtl\") pod \"90693935-44c2-4e50-a8ad-5e2967ac582f\" (UID: \"90693935-44c2-4e50-a8ad-5e2967ac582f\") " Nov 25 10:59:39 crc kubenswrapper[4776]: I1125 10:59:39.182866 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/90693935-44c2-4e50-a8ad-5e2967ac582f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "90693935-44c2-4e50-a8ad-5e2967ac582f" (UID: "90693935-44c2-4e50-a8ad-5e2967ac582f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:59:39 crc kubenswrapper[4776]: I1125 10:59:39.189653 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90693935-44c2-4e50-a8ad-5e2967ac582f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "90693935-44c2-4e50-a8ad-5e2967ac582f" (UID: "90693935-44c2-4e50-a8ad-5e2967ac582f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:59:39 crc kubenswrapper[4776]: I1125 10:59:39.213417 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90693935-44c2-4e50-a8ad-5e2967ac582f-kube-api-access-2wjtl" (OuterVolumeSpecName: "kube-api-access-2wjtl") pod "90693935-44c2-4e50-a8ad-5e2967ac582f" (UID: "90693935-44c2-4e50-a8ad-5e2967ac582f"). InnerVolumeSpecName "kube-api-access-2wjtl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:59:39 crc kubenswrapper[4776]: I1125 10:59:39.213484 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90693935-44c2-4e50-a8ad-5e2967ac582f-scripts" (OuterVolumeSpecName: "scripts") pod "90693935-44c2-4e50-a8ad-5e2967ac582f" (UID: "90693935-44c2-4e50-a8ad-5e2967ac582f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:59:39 crc kubenswrapper[4776]: I1125 10:59:39.263243 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90693935-44c2-4e50-a8ad-5e2967ac582f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "90693935-44c2-4e50-a8ad-5e2967ac582f" (UID: "90693935-44c2-4e50-a8ad-5e2967ac582f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:59:39 crc kubenswrapper[4776]: I1125 10:59:39.291396 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90693935-44c2-4e50-a8ad-5e2967ac582f-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:39 crc kubenswrapper[4776]: I1125 10:59:39.291458 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90693935-44c2-4e50-a8ad-5e2967ac582f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:39 crc kubenswrapper[4776]: I1125 10:59:39.291472 4776 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/90693935-44c2-4e50-a8ad-5e2967ac582f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:39 crc kubenswrapper[4776]: I1125 10:59:39.291486 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wjtl\" (UniqueName: \"kubernetes.io/projected/90693935-44c2-4e50-a8ad-5e2967ac582f-kube-api-access-2wjtl\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:39 crc kubenswrapper[4776]: I1125 10:59:39.291496 4776 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/90693935-44c2-4e50-a8ad-5e2967ac582f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:39 crc kubenswrapper[4776]: I1125 10:59:39.323444 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90693935-44c2-4e50-a8ad-5e2967ac582f-config-data" (OuterVolumeSpecName: "config-data") pod "90693935-44c2-4e50-a8ad-5e2967ac582f" (UID: "90693935-44c2-4e50-a8ad-5e2967ac582f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:59:39 crc kubenswrapper[4776]: I1125 10:59:39.393885 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90693935-44c2-4e50-a8ad-5e2967ac582f-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:39 crc kubenswrapper[4776]: I1125 10:59:39.747158 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-4f94t" event={"ID":"90693935-44c2-4e50-a8ad-5e2967ac582f","Type":"ContainerDied","Data":"a50d48760965ddd55050ec311f5c98725704e49d5f8afae9905feaa90bc310e3"} Nov 25 10:59:39 crc kubenswrapper[4776]: I1125 10:59:39.747701 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a50d48760965ddd55050ec311f5c98725704e49d5f8afae9905feaa90bc310e3" Nov 25 10:59:39 crc kubenswrapper[4776]: I1125 10:59:39.747312 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-4f94t" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.142405 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78994c4b75-5c25c"] Nov 25 10:59:40 crc kubenswrapper[4776]: E1125 10:59:40.142783 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90693935-44c2-4e50-a8ad-5e2967ac582f" containerName="cinder-db-sync" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.142801 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="90693935-44c2-4e50-a8ad-5e2967ac582f" containerName="cinder-db-sync" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.142999 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="90693935-44c2-4e50-a8ad-5e2967ac582f" containerName="cinder-db-sync" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.143973 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78994c4b75-5c25c" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.169963 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78994c4b75-5c25c"] Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.208955 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-ovsdbserver-nb\") pod \"dnsmasq-dns-78994c4b75-5c25c\" (UID: \"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9\") " pod="openstack/dnsmasq-dns-78994c4b75-5c25c" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.209054 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh2fv\" (UniqueName: \"kubernetes.io/projected/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-kube-api-access-dh2fv\") pod \"dnsmasq-dns-78994c4b75-5c25c\" (UID: \"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9\") " pod="openstack/dnsmasq-dns-78994c4b75-5c25c" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.209132 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-ovsdbserver-sb\") pod \"dnsmasq-dns-78994c4b75-5c25c\" (UID: \"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9\") " pod="openstack/dnsmasq-dns-78994c4b75-5c25c" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.209179 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-config\") pod \"dnsmasq-dns-78994c4b75-5c25c\" (UID: \"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9\") " pod="openstack/dnsmasq-dns-78994c4b75-5c25c" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.209234 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-dns-svc\") pod \"dnsmasq-dns-78994c4b75-5c25c\" (UID: \"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9\") " pod="openstack/dnsmasq-dns-78994c4b75-5c25c" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.247448 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.249298 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.252451 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-cqpk5" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.252821 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.253020 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.253036 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.260307 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.310802 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-scripts\") pod \"cinder-api-0\" (UID: \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\") " pod="openstack/cinder-api-0" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.310870 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh2fv\" (UniqueName: \"kubernetes.io/projected/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-kube-api-access-dh2fv\") pod \"dnsmasq-dns-78994c4b75-5c25c\" (UID: \"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9\") " pod="openstack/dnsmasq-dns-78994c4b75-5c25c" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.310913 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-config-data\") pod \"cinder-api-0\" (UID: \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\") " pod="openstack/cinder-api-0" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.310951 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-config-data-custom\") pod \"cinder-api-0\" (UID: \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\") " pod="openstack/cinder-api-0" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.310981 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\") " pod="openstack/cinder-api-0" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.311011 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-ovsdbserver-sb\") pod \"dnsmasq-dns-78994c4b75-5c25c\" (UID: \"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9\") " pod="openstack/dnsmasq-dns-78994c4b75-5c25c" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.311050 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\") " pod="openstack/cinder-api-0" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.311098 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-config\") pod \"dnsmasq-dns-78994c4b75-5c25c\" (UID: \"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9\") " pod="openstack/dnsmasq-dns-78994c4b75-5c25c" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.311159 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qp8k\" (UniqueName: \"kubernetes.io/projected/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-kube-api-access-4qp8k\") pod \"cinder-api-0\" (UID: \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\") " pod="openstack/cinder-api-0" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.311186 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-dns-svc\") pod \"dnsmasq-dns-78994c4b75-5c25c\" (UID: \"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9\") " pod="openstack/dnsmasq-dns-78994c4b75-5c25c" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.311230 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-logs\") pod \"cinder-api-0\" (UID: \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\") " pod="openstack/cinder-api-0" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.311256 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-ovsdbserver-nb\") pod \"dnsmasq-dns-78994c4b75-5c25c\" (UID: \"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9\") " pod="openstack/dnsmasq-dns-78994c4b75-5c25c" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.312367 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-config\") pod \"dnsmasq-dns-78994c4b75-5c25c\" (UID: \"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9\") " pod="openstack/dnsmasq-dns-78994c4b75-5c25c" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.312381 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-ovsdbserver-sb\") pod \"dnsmasq-dns-78994c4b75-5c25c\" (UID: \"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9\") " pod="openstack/dnsmasq-dns-78994c4b75-5c25c" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.312888 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-dns-svc\") pod \"dnsmasq-dns-78994c4b75-5c25c\" (UID: \"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9\") " pod="openstack/dnsmasq-dns-78994c4b75-5c25c" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.313393 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-ovsdbserver-nb\") pod \"dnsmasq-dns-78994c4b75-5c25c\" (UID: \"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9\") " pod="openstack/dnsmasq-dns-78994c4b75-5c25c" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.350740 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh2fv\" (UniqueName: \"kubernetes.io/projected/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-kube-api-access-dh2fv\") pod \"dnsmasq-dns-78994c4b75-5c25c\" (UID: \"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9\") " pod="openstack/dnsmasq-dns-78994c4b75-5c25c" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.412139 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-scripts\") pod \"cinder-api-0\" (UID: \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\") " pod="openstack/cinder-api-0" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.412214 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-config-data\") pod \"cinder-api-0\" (UID: \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\") " pod="openstack/cinder-api-0" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.412239 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-config-data-custom\") pod \"cinder-api-0\" (UID: \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\") " pod="openstack/cinder-api-0" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.412260 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\") " pod="openstack/cinder-api-0" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.412290 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\") " pod="openstack/cinder-api-0" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.412329 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qp8k\" (UniqueName: \"kubernetes.io/projected/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-kube-api-access-4qp8k\") pod \"cinder-api-0\" (UID: \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\") " pod="openstack/cinder-api-0" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.412362 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-logs\") pod \"cinder-api-0\" (UID: \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\") " pod="openstack/cinder-api-0" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.412713 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-logs\") pod \"cinder-api-0\" (UID: \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\") " pod="openstack/cinder-api-0" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.412801 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\") " pod="openstack/cinder-api-0" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.416348 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\") " pod="openstack/cinder-api-0" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.417443 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-config-data-custom\") pod \"cinder-api-0\" (UID: \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\") " pod="openstack/cinder-api-0" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.418187 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-config-data\") pod \"cinder-api-0\" (UID: \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\") " pod="openstack/cinder-api-0" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.430849 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-scripts\") pod \"cinder-api-0\" (UID: \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\") " pod="openstack/cinder-api-0" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.435147 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qp8k\" (UniqueName: \"kubernetes.io/projected/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-kube-api-access-4qp8k\") pod \"cinder-api-0\" (UID: \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\") " pod="openstack/cinder-api-0" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.461818 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78994c4b75-5c25c" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.573934 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 10:59:40 crc kubenswrapper[4776]: I1125 10:59:40.966981 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78994c4b75-5c25c"] Nov 25 10:59:41 crc kubenswrapper[4776]: I1125 10:59:41.150240 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:59:41 crc kubenswrapper[4776]: W1125 10:59:41.159211 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c8c4d0a_6cf7_47a9_b5a3_b23249e6a7ca.slice/crio-1086e70e3e4a66c5737c3f2a87b38155aac24bd04b91d3b08d6e5a6bf3648272 WatchSource:0}: Error finding container 1086e70e3e4a66c5737c3f2a87b38155aac24bd04b91d3b08d6e5a6bf3648272: Status 404 returned error can't find the container with id 1086e70e3e4a66c5737c3f2a87b38155aac24bd04b91d3b08d6e5a6bf3648272 Nov 25 10:59:41 crc kubenswrapper[4776]: E1125 10:59:41.525315 4776 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb812d47_f7bf_4a44_a2f0_abc9c9f0e5f9.slice/crio-3e3d519217585fa073450cb159b471ba8a4c166e7697200f22002f3e6b5d9501.scope\": RecentStats: unable to find data in memory cache]" Nov 25 10:59:41 crc kubenswrapper[4776]: I1125 10:59:41.785352 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca","Type":"ContainerStarted","Data":"1086e70e3e4a66c5737c3f2a87b38155aac24bd04b91d3b08d6e5a6bf3648272"} Nov 25 10:59:41 crc kubenswrapper[4776]: I1125 10:59:41.796007 4776 generic.go:334] "Generic (PLEG): container finished" podID="eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9" containerID="3e3d519217585fa073450cb159b471ba8a4c166e7697200f22002f3e6b5d9501" exitCode=0 Nov 25 10:59:41 crc kubenswrapper[4776]: I1125 10:59:41.796058 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78994c4b75-5c25c" event={"ID":"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9","Type":"ContainerDied","Data":"3e3d519217585fa073450cb159b471ba8a4c166e7697200f22002f3e6b5d9501"} Nov 25 10:59:41 crc kubenswrapper[4776]: I1125 10:59:41.796103 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78994c4b75-5c25c" event={"ID":"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9","Type":"ContainerStarted","Data":"59dd63dcf3b1f5909dfdc3086bf39b611a044e3330957ce93cd583d3662fecbf"} Nov 25 10:59:42 crc kubenswrapper[4776]: I1125 10:59:42.619924 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:59:42 crc kubenswrapper[4776]: I1125 10:59:42.806813 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78994c4b75-5c25c" event={"ID":"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9","Type":"ContainerStarted","Data":"98e8c48ee2c38d5323691911b3fb543dd2e31fcffacde20b6b7b79bf4a2692fa"} Nov 25 10:59:42 crc kubenswrapper[4776]: I1125 10:59:42.808260 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78994c4b75-5c25c" Nov 25 10:59:42 crc kubenswrapper[4776]: I1125 10:59:42.810464 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca","Type":"ContainerStarted","Data":"a4a0d9d6d1aa0bb83e43a7bf2eb64b6b5513c31468fe85abf2be14dfcf17b29f"} Nov 25 10:59:42 crc kubenswrapper[4776]: I1125 10:59:42.810492 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca","Type":"ContainerStarted","Data":"50dd17ac9f44532381fede2234a64841f0c4a84c565658a123734ba0a508be25"} Nov 25 10:59:42 crc kubenswrapper[4776]: I1125 10:59:42.810606 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca" containerName="cinder-api-log" containerID="cri-o://50dd17ac9f44532381fede2234a64841f0c4a84c565658a123734ba0a508be25" gracePeriod=30 Nov 25 10:59:42 crc kubenswrapper[4776]: I1125 10:59:42.810827 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 25 10:59:42 crc kubenswrapper[4776]: I1125 10:59:42.810866 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca" containerName="cinder-api" containerID="cri-o://a4a0d9d6d1aa0bb83e43a7bf2eb64b6b5513c31468fe85abf2be14dfcf17b29f" gracePeriod=30 Nov 25 10:59:42 crc kubenswrapper[4776]: I1125 10:59:42.831984 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78994c4b75-5c25c" podStartSLOduration=2.831962717 podStartE2EDuration="2.831962717s" podCreationTimestamp="2025-11-25 10:59:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:59:42.8304795 +0000 UTC m=+5727.871539053" watchObservedRunningTime="2025-11-25 10:59:42.831962717 +0000 UTC m=+5727.873022280" Nov 25 10:59:42 crc kubenswrapper[4776]: I1125 10:59:42.850295 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.850273557 podStartE2EDuration="2.850273557s" podCreationTimestamp="2025-11-25 10:59:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:59:42.845435015 +0000 UTC m=+5727.886494568" watchObservedRunningTime="2025-11-25 10:59:42.850273557 +0000 UTC m=+5727.891333120" Nov 25 10:59:43 crc kubenswrapper[4776]: I1125 10:59:43.820548 4776 generic.go:334] "Generic (PLEG): container finished" podID="0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca" containerID="50dd17ac9f44532381fede2234a64841f0c4a84c565658a123734ba0a508be25" exitCode=143 Nov 25 10:59:43 crc kubenswrapper[4776]: I1125 10:59:43.820822 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca","Type":"ContainerDied","Data":"50dd17ac9f44532381fede2234a64841f0c4a84c565658a123734ba0a508be25"} Nov 25 10:59:47 crc kubenswrapper[4776]: I1125 10:59:47.819126 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:59:47 crc kubenswrapper[4776]: I1125 10:59:47.819585 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:59:50 crc kubenswrapper[4776]: I1125 10:59:50.464384 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-78994c4b75-5c25c" Nov 25 10:59:50 crc kubenswrapper[4776]: I1125 10:59:50.558314 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67d94b9bc9-5cqlc"] Nov 25 10:59:50 crc kubenswrapper[4776]: I1125 10:59:50.558623 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67d94b9bc9-5cqlc" podUID="26290360-835e-4f34-a1ad-1803bce72437" containerName="dnsmasq-dns" containerID="cri-o://548a8359c5df63cb48083eb603e31cff145deffb674aa59e73df33a69f5f0f06" gracePeriod=10 Nov 25 10:59:50 crc kubenswrapper[4776]: I1125 10:59:50.891296 4776 generic.go:334] "Generic (PLEG): container finished" podID="26290360-835e-4f34-a1ad-1803bce72437" containerID="548a8359c5df63cb48083eb603e31cff145deffb674aa59e73df33a69f5f0f06" exitCode=0 Nov 25 10:59:50 crc kubenswrapper[4776]: I1125 10:59:50.891366 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67d94b9bc9-5cqlc" event={"ID":"26290360-835e-4f34-a1ad-1803bce72437","Type":"ContainerDied","Data":"548a8359c5df63cb48083eb603e31cff145deffb674aa59e73df33a69f5f0f06"} Nov 25 10:59:51 crc kubenswrapper[4776]: I1125 10:59:51.701140 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67d94b9bc9-5cqlc" Nov 25 10:59:51 crc kubenswrapper[4776]: I1125 10:59:51.844354 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26290360-835e-4f34-a1ad-1803bce72437-config\") pod \"26290360-835e-4f34-a1ad-1803bce72437\" (UID: \"26290360-835e-4f34-a1ad-1803bce72437\") " Nov 25 10:59:51 crc kubenswrapper[4776]: I1125 10:59:51.844764 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26290360-835e-4f34-a1ad-1803bce72437-ovsdbserver-nb\") pod \"26290360-835e-4f34-a1ad-1803bce72437\" (UID: \"26290360-835e-4f34-a1ad-1803bce72437\") " Nov 25 10:59:51 crc kubenswrapper[4776]: I1125 10:59:51.844794 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26290360-835e-4f34-a1ad-1803bce72437-dns-svc\") pod \"26290360-835e-4f34-a1ad-1803bce72437\" (UID: \"26290360-835e-4f34-a1ad-1803bce72437\") " Nov 25 10:59:51 crc kubenswrapper[4776]: I1125 10:59:51.844915 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klfk4\" (UniqueName: \"kubernetes.io/projected/26290360-835e-4f34-a1ad-1803bce72437-kube-api-access-klfk4\") pod \"26290360-835e-4f34-a1ad-1803bce72437\" (UID: \"26290360-835e-4f34-a1ad-1803bce72437\") " Nov 25 10:59:51 crc kubenswrapper[4776]: I1125 10:59:51.845102 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26290360-835e-4f34-a1ad-1803bce72437-ovsdbserver-sb\") pod \"26290360-835e-4f34-a1ad-1803bce72437\" (UID: \"26290360-835e-4f34-a1ad-1803bce72437\") " Nov 25 10:59:51 crc kubenswrapper[4776]: I1125 10:59:51.850846 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26290360-835e-4f34-a1ad-1803bce72437-kube-api-access-klfk4" (OuterVolumeSpecName: "kube-api-access-klfk4") pod "26290360-835e-4f34-a1ad-1803bce72437" (UID: "26290360-835e-4f34-a1ad-1803bce72437"). InnerVolumeSpecName "kube-api-access-klfk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:59:51 crc kubenswrapper[4776]: I1125 10:59:51.899635 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26290360-835e-4f34-a1ad-1803bce72437-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "26290360-835e-4f34-a1ad-1803bce72437" (UID: "26290360-835e-4f34-a1ad-1803bce72437"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:51 crc kubenswrapper[4776]: I1125 10:59:51.900893 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26290360-835e-4f34-a1ad-1803bce72437-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "26290360-835e-4f34-a1ad-1803bce72437" (UID: "26290360-835e-4f34-a1ad-1803bce72437"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:51 crc kubenswrapper[4776]: I1125 10:59:51.907402 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26290360-835e-4f34-a1ad-1803bce72437-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "26290360-835e-4f34-a1ad-1803bce72437" (UID: "26290360-835e-4f34-a1ad-1803bce72437"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:51 crc kubenswrapper[4776]: I1125 10:59:51.911795 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67d94b9bc9-5cqlc" event={"ID":"26290360-835e-4f34-a1ad-1803bce72437","Type":"ContainerDied","Data":"69a4797ee8efbc57a0adc79a051c65ba0cbab45a0d1738400eab53719af11e80"} Nov 25 10:59:51 crc kubenswrapper[4776]: I1125 10:59:51.911934 4776 scope.go:117] "RemoveContainer" containerID="548a8359c5df63cb48083eb603e31cff145deffb674aa59e73df33a69f5f0f06" Nov 25 10:59:51 crc kubenswrapper[4776]: I1125 10:59:51.912243 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67d94b9bc9-5cqlc" Nov 25 10:59:51 crc kubenswrapper[4776]: I1125 10:59:51.923509 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26290360-835e-4f34-a1ad-1803bce72437-config" (OuterVolumeSpecName: "config") pod "26290360-835e-4f34-a1ad-1803bce72437" (UID: "26290360-835e-4f34-a1ad-1803bce72437"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:59:51 crc kubenswrapper[4776]: I1125 10:59:51.947692 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26290360-835e-4f34-a1ad-1803bce72437-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:51 crc kubenswrapper[4776]: I1125 10:59:51.947732 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/26290360-835e-4f34-a1ad-1803bce72437-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:51 crc kubenswrapper[4776]: I1125 10:59:51.947748 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26290360-835e-4f34-a1ad-1803bce72437-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:51 crc kubenswrapper[4776]: I1125 10:59:51.947761 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klfk4\" (UniqueName: \"kubernetes.io/projected/26290360-835e-4f34-a1ad-1803bce72437-kube-api-access-klfk4\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:51 crc kubenswrapper[4776]: I1125 10:59:51.947773 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/26290360-835e-4f34-a1ad-1803bce72437-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:59:52 crc kubenswrapper[4776]: I1125 10:59:52.016424 4776 scope.go:117] "RemoveContainer" containerID="28081dcf1472877d0fb63c17e5eaf89e932e095527c9dbae4498cb43d792e427" Nov 25 10:59:52 crc kubenswrapper[4776]: I1125 10:59:52.248820 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67d94b9bc9-5cqlc"] Nov 25 10:59:52 crc kubenswrapper[4776]: I1125 10:59:52.261351 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67d94b9bc9-5cqlc"] Nov 25 10:59:52 crc kubenswrapper[4776]: I1125 10:59:52.755023 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 25 10:59:53 crc kubenswrapper[4776]: I1125 10:59:53.673314 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26290360-835e-4f34-a1ad-1803bce72437" path="/var/lib/kubelet/pods/26290360-835e-4f34-a1ad-1803bce72437/volumes" Nov 25 11:00:00 crc kubenswrapper[4776]: I1125 11:00:00.134722 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401140-xc252"] Nov 25 11:00:00 crc kubenswrapper[4776]: E1125 11:00:00.135918 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26290360-835e-4f34-a1ad-1803bce72437" containerName="dnsmasq-dns" Nov 25 11:00:00 crc kubenswrapper[4776]: I1125 11:00:00.135942 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="26290360-835e-4f34-a1ad-1803bce72437" containerName="dnsmasq-dns" Nov 25 11:00:00 crc kubenswrapper[4776]: E1125 11:00:00.136205 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26290360-835e-4f34-a1ad-1803bce72437" containerName="init" Nov 25 11:00:00 crc kubenswrapper[4776]: I1125 11:00:00.136221 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="26290360-835e-4f34-a1ad-1803bce72437" containerName="init" Nov 25 11:00:00 crc kubenswrapper[4776]: I1125 11:00:00.136560 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="26290360-835e-4f34-a1ad-1803bce72437" containerName="dnsmasq-dns" Nov 25 11:00:00 crc kubenswrapper[4776]: I1125 11:00:00.137558 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-xc252" Nov 25 11:00:00 crc kubenswrapper[4776]: I1125 11:00:00.139943 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 11:00:00 crc kubenswrapper[4776]: I1125 11:00:00.140052 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 11:00:00 crc kubenswrapper[4776]: I1125 11:00:00.145014 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401140-xc252"] Nov 25 11:00:00 crc kubenswrapper[4776]: I1125 11:00:00.299598 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51d095bb-1d12-4939-aeb6-104e88e0163f-config-volume\") pod \"collect-profiles-29401140-xc252\" (UID: \"51d095bb-1d12-4939-aeb6-104e88e0163f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-xc252" Nov 25 11:00:00 crc kubenswrapper[4776]: I1125 11:00:00.299778 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rr7nf\" (UniqueName: \"kubernetes.io/projected/51d095bb-1d12-4939-aeb6-104e88e0163f-kube-api-access-rr7nf\") pod \"collect-profiles-29401140-xc252\" (UID: \"51d095bb-1d12-4939-aeb6-104e88e0163f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-xc252" Nov 25 11:00:00 crc kubenswrapper[4776]: I1125 11:00:00.299815 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/51d095bb-1d12-4939-aeb6-104e88e0163f-secret-volume\") pod \"collect-profiles-29401140-xc252\" (UID: \"51d095bb-1d12-4939-aeb6-104e88e0163f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-xc252" Nov 25 11:00:00 crc kubenswrapper[4776]: I1125 11:00:00.401641 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rr7nf\" (UniqueName: \"kubernetes.io/projected/51d095bb-1d12-4939-aeb6-104e88e0163f-kube-api-access-rr7nf\") pod \"collect-profiles-29401140-xc252\" (UID: \"51d095bb-1d12-4939-aeb6-104e88e0163f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-xc252" Nov 25 11:00:00 crc kubenswrapper[4776]: I1125 11:00:00.401782 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/51d095bb-1d12-4939-aeb6-104e88e0163f-secret-volume\") pod \"collect-profiles-29401140-xc252\" (UID: \"51d095bb-1d12-4939-aeb6-104e88e0163f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-xc252" Nov 25 11:00:00 crc kubenswrapper[4776]: I1125 11:00:00.402099 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51d095bb-1d12-4939-aeb6-104e88e0163f-config-volume\") pod \"collect-profiles-29401140-xc252\" (UID: \"51d095bb-1d12-4939-aeb6-104e88e0163f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-xc252" Nov 25 11:00:00 crc kubenswrapper[4776]: I1125 11:00:00.403400 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51d095bb-1d12-4939-aeb6-104e88e0163f-config-volume\") pod \"collect-profiles-29401140-xc252\" (UID: \"51d095bb-1d12-4939-aeb6-104e88e0163f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-xc252" Nov 25 11:00:00 crc kubenswrapper[4776]: I1125 11:00:00.412914 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/51d095bb-1d12-4939-aeb6-104e88e0163f-secret-volume\") pod \"collect-profiles-29401140-xc252\" (UID: \"51d095bb-1d12-4939-aeb6-104e88e0163f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-xc252" Nov 25 11:00:00 crc kubenswrapper[4776]: I1125 11:00:00.443122 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rr7nf\" (UniqueName: \"kubernetes.io/projected/51d095bb-1d12-4939-aeb6-104e88e0163f-kube-api-access-rr7nf\") pod \"collect-profiles-29401140-xc252\" (UID: \"51d095bb-1d12-4939-aeb6-104e88e0163f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-xc252" Nov 25 11:00:00 crc kubenswrapper[4776]: I1125 11:00:00.460919 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-xc252" Nov 25 11:00:00 crc kubenswrapper[4776]: W1125 11:00:00.918520 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod51d095bb_1d12_4939_aeb6_104e88e0163f.slice/crio-a0f3242a61ca3c31125975c1e578ba276acfa6b52f565af1a5b3363ff43ee0f5 WatchSource:0}: Error finding container a0f3242a61ca3c31125975c1e578ba276acfa6b52f565af1a5b3363ff43ee0f5: Status 404 returned error can't find the container with id a0f3242a61ca3c31125975c1e578ba276acfa6b52f565af1a5b3363ff43ee0f5 Nov 25 11:00:00 crc kubenswrapper[4776]: I1125 11:00:00.932619 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401140-xc252"] Nov 25 11:00:01 crc kubenswrapper[4776]: I1125 11:00:01.021016 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-xc252" event={"ID":"51d095bb-1d12-4939-aeb6-104e88e0163f","Type":"ContainerStarted","Data":"a0f3242a61ca3c31125975c1e578ba276acfa6b52f565af1a5b3363ff43ee0f5"} Nov 25 11:00:02 crc kubenswrapper[4776]: I1125 11:00:02.033587 4776 generic.go:334] "Generic (PLEG): container finished" podID="51d095bb-1d12-4939-aeb6-104e88e0163f" containerID="ba23c2ddb4c24f894d2222b9501f72d438e8d6a63f91eba6b3c8c1c692bd9bdf" exitCode=0 Nov 25 11:00:02 crc kubenswrapper[4776]: I1125 11:00:02.033631 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-xc252" event={"ID":"51d095bb-1d12-4939-aeb6-104e88e0163f","Type":"ContainerDied","Data":"ba23c2ddb4c24f894d2222b9501f72d438e8d6a63f91eba6b3c8c1c692bd9bdf"} Nov 25 11:00:03 crc kubenswrapper[4776]: I1125 11:00:03.420914 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-xc252" Nov 25 11:00:03 crc kubenswrapper[4776]: I1125 11:00:03.570955 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51d095bb-1d12-4939-aeb6-104e88e0163f-config-volume\") pod \"51d095bb-1d12-4939-aeb6-104e88e0163f\" (UID: \"51d095bb-1d12-4939-aeb6-104e88e0163f\") " Nov 25 11:00:03 crc kubenswrapper[4776]: I1125 11:00:03.571242 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rr7nf\" (UniqueName: \"kubernetes.io/projected/51d095bb-1d12-4939-aeb6-104e88e0163f-kube-api-access-rr7nf\") pod \"51d095bb-1d12-4939-aeb6-104e88e0163f\" (UID: \"51d095bb-1d12-4939-aeb6-104e88e0163f\") " Nov 25 11:00:03 crc kubenswrapper[4776]: I1125 11:00:03.571311 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/51d095bb-1d12-4939-aeb6-104e88e0163f-secret-volume\") pod \"51d095bb-1d12-4939-aeb6-104e88e0163f\" (UID: \"51d095bb-1d12-4939-aeb6-104e88e0163f\") " Nov 25 11:00:03 crc kubenswrapper[4776]: I1125 11:00:03.571708 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51d095bb-1d12-4939-aeb6-104e88e0163f-config-volume" (OuterVolumeSpecName: "config-volume") pod "51d095bb-1d12-4939-aeb6-104e88e0163f" (UID: "51d095bb-1d12-4939-aeb6-104e88e0163f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:00:03 crc kubenswrapper[4776]: I1125 11:00:03.572108 4776 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51d095bb-1d12-4939-aeb6-104e88e0163f-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:03 crc kubenswrapper[4776]: I1125 11:00:03.576587 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51d095bb-1d12-4939-aeb6-104e88e0163f-kube-api-access-rr7nf" (OuterVolumeSpecName: "kube-api-access-rr7nf") pod "51d095bb-1d12-4939-aeb6-104e88e0163f" (UID: "51d095bb-1d12-4939-aeb6-104e88e0163f"). InnerVolumeSpecName "kube-api-access-rr7nf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:00:03 crc kubenswrapper[4776]: I1125 11:00:03.579838 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51d095bb-1d12-4939-aeb6-104e88e0163f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "51d095bb-1d12-4939-aeb6-104e88e0163f" (UID: "51d095bb-1d12-4939-aeb6-104e88e0163f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:00:03 crc kubenswrapper[4776]: I1125 11:00:03.673701 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rr7nf\" (UniqueName: \"kubernetes.io/projected/51d095bb-1d12-4939-aeb6-104e88e0163f-kube-api-access-rr7nf\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:03 crc kubenswrapper[4776]: I1125 11:00:03.674029 4776 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/51d095bb-1d12-4939-aeb6-104e88e0163f-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:04 crc kubenswrapper[4776]: I1125 11:00:04.052897 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-xc252" event={"ID":"51d095bb-1d12-4939-aeb6-104e88e0163f","Type":"ContainerDied","Data":"a0f3242a61ca3c31125975c1e578ba276acfa6b52f565af1a5b3363ff43ee0f5"} Nov 25 11:00:04 crc kubenswrapper[4776]: I1125 11:00:04.052937 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0f3242a61ca3c31125975c1e578ba276acfa6b52f565af1a5b3363ff43ee0f5" Nov 25 11:00:04 crc kubenswrapper[4776]: I1125 11:00:04.066440 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401140-xc252" Nov 25 11:00:04 crc kubenswrapper[4776]: I1125 11:00:04.487302 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401095-rnclm"] Nov 25 11:00:04 crc kubenswrapper[4776]: I1125 11:00:04.494236 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401095-rnclm"] Nov 25 11:00:05 crc kubenswrapper[4776]: I1125 11:00:05.682660 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="625731b7-d813-461a-b2b3-8b7e64622ff4" path="/var/lib/kubelet/pods/625731b7-d813-461a-b2b3-8b7e64622ff4/volumes" Nov 25 11:00:13 crc kubenswrapper[4776]: I1125 11:00:13.920219 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.069606 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-etc-machine-id\") pod \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\" (UID: \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\") " Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.069712 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca" (UID: "0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.070391 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-logs\") pod \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\" (UID: \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\") " Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.070619 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-config-data-custom\") pod \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\" (UID: \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\") " Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.070839 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-scripts\") pod \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\" (UID: \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\") " Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.070986 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-logs" (OuterVolumeSpecName: "logs") pod "0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca" (UID: "0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.071174 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-combined-ca-bundle\") pod \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\" (UID: \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\") " Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.071249 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-config-data\") pod \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\" (UID: \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\") " Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.071287 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qp8k\" (UniqueName: \"kubernetes.io/projected/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-kube-api-access-4qp8k\") pod \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\" (UID: \"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca\") " Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.072319 4776 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.072340 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-logs\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.081648 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-scripts" (OuterVolumeSpecName: "scripts") pod "0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca" (UID: "0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.082021 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca" (UID: "0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.086687 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-kube-api-access-4qp8k" (OuterVolumeSpecName: "kube-api-access-4qp8k") pod "0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca" (UID: "0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca"). InnerVolumeSpecName "kube-api-access-4qp8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.101775 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca" (UID: "0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.141810 4776 generic.go:334] "Generic (PLEG): container finished" podID="0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca" containerID="a4a0d9d6d1aa0bb83e43a7bf2eb64b6b5513c31468fe85abf2be14dfcf17b29f" exitCode=137 Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.141853 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca","Type":"ContainerDied","Data":"a4a0d9d6d1aa0bb83e43a7bf2eb64b6b5513c31468fe85abf2be14dfcf17b29f"} Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.141879 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca","Type":"ContainerDied","Data":"1086e70e3e4a66c5737c3f2a87b38155aac24bd04b91d3b08d6e5a6bf3648272"} Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.141897 4776 scope.go:117] "RemoveContainer" containerID="a4a0d9d6d1aa0bb83e43a7bf2eb64b6b5513c31468fe85abf2be14dfcf17b29f" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.142033 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.145984 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-config-data" (OuterVolumeSpecName: "config-data") pod "0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca" (UID: "0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.179033 4776 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.179093 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.179107 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.179118 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.179131 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qp8k\" (UniqueName: \"kubernetes.io/projected/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca-kube-api-access-4qp8k\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.253966 4776 scope.go:117] "RemoveContainer" containerID="50dd17ac9f44532381fede2234a64841f0c4a84c565658a123734ba0a508be25" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.270384 4776 scope.go:117] "RemoveContainer" containerID="a4a0d9d6d1aa0bb83e43a7bf2eb64b6b5513c31468fe85abf2be14dfcf17b29f" Nov 25 11:00:14 crc kubenswrapper[4776]: E1125 11:00:14.270808 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4a0d9d6d1aa0bb83e43a7bf2eb64b6b5513c31468fe85abf2be14dfcf17b29f\": container with ID starting with a4a0d9d6d1aa0bb83e43a7bf2eb64b6b5513c31468fe85abf2be14dfcf17b29f not found: ID does not exist" containerID="a4a0d9d6d1aa0bb83e43a7bf2eb64b6b5513c31468fe85abf2be14dfcf17b29f" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.270839 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4a0d9d6d1aa0bb83e43a7bf2eb64b6b5513c31468fe85abf2be14dfcf17b29f"} err="failed to get container status \"a4a0d9d6d1aa0bb83e43a7bf2eb64b6b5513c31468fe85abf2be14dfcf17b29f\": rpc error: code = NotFound desc = could not find container \"a4a0d9d6d1aa0bb83e43a7bf2eb64b6b5513c31468fe85abf2be14dfcf17b29f\": container with ID starting with a4a0d9d6d1aa0bb83e43a7bf2eb64b6b5513c31468fe85abf2be14dfcf17b29f not found: ID does not exist" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.270862 4776 scope.go:117] "RemoveContainer" containerID="50dd17ac9f44532381fede2234a64841f0c4a84c565658a123734ba0a508be25" Nov 25 11:00:14 crc kubenswrapper[4776]: E1125 11:00:14.271272 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50dd17ac9f44532381fede2234a64841f0c4a84c565658a123734ba0a508be25\": container with ID starting with 50dd17ac9f44532381fede2234a64841f0c4a84c565658a123734ba0a508be25 not found: ID does not exist" containerID="50dd17ac9f44532381fede2234a64841f0c4a84c565658a123734ba0a508be25" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.271294 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50dd17ac9f44532381fede2234a64841f0c4a84c565658a123734ba0a508be25"} err="failed to get container status \"50dd17ac9f44532381fede2234a64841f0c4a84c565658a123734ba0a508be25\": rpc error: code = NotFound desc = could not find container \"50dd17ac9f44532381fede2234a64841f0c4a84c565658a123734ba0a508be25\": container with ID starting with 50dd17ac9f44532381fede2234a64841f0c4a84c565658a123734ba0a508be25 not found: ID does not exist" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.470933 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.480789 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.498692 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 25 11:00:14 crc kubenswrapper[4776]: E1125 11:00:14.499321 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca" containerName="cinder-api-log" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.499407 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca" containerName="cinder-api-log" Nov 25 11:00:14 crc kubenswrapper[4776]: E1125 11:00:14.499486 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51d095bb-1d12-4939-aeb6-104e88e0163f" containerName="collect-profiles" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.499540 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="51d095bb-1d12-4939-aeb6-104e88e0163f" containerName="collect-profiles" Nov 25 11:00:14 crc kubenswrapper[4776]: E1125 11:00:14.499618 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca" containerName="cinder-api" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.499677 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca" containerName="cinder-api" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.499877 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca" containerName="cinder-api-log" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.499945 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca" containerName="cinder-api" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.500006 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="51d095bb-1d12-4939-aeb6-104e88e0163f" containerName="collect-profiles" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.501137 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.504399 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.504485 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.504579 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-cqpk5" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.504604 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.504611 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.504624 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.513287 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.685891 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-config-data\") pod \"cinder-api-0\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.686128 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-scripts\") pod \"cinder-api-0\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.686306 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddbe9150-55f7-4f9f-98db-44fefa0ef987-logs\") pod \"cinder-api-0\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.686627 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.686704 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-public-tls-certs\") pod \"cinder-api-0\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.686798 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ddbe9150-55f7-4f9f-98db-44fefa0ef987-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.686884 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zg4vc\" (UniqueName: \"kubernetes.io/projected/ddbe9150-55f7-4f9f-98db-44fefa0ef987-kube-api-access-zg4vc\") pod \"cinder-api-0\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.686919 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-config-data-custom\") pod \"cinder-api-0\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.686956 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.788631 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-scripts\") pod \"cinder-api-0\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.788703 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddbe9150-55f7-4f9f-98db-44fefa0ef987-logs\") pod \"cinder-api-0\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.788771 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.788799 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-public-tls-certs\") pod \"cinder-api-0\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.788822 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ddbe9150-55f7-4f9f-98db-44fefa0ef987-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.788859 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zg4vc\" (UniqueName: \"kubernetes.io/projected/ddbe9150-55f7-4f9f-98db-44fefa0ef987-kube-api-access-zg4vc\") pod \"cinder-api-0\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.788886 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-config-data-custom\") pod \"cinder-api-0\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.788920 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.788970 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-config-data\") pod \"cinder-api-0\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.788972 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ddbe9150-55f7-4f9f-98db-44fefa0ef987-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.789476 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddbe9150-55f7-4f9f-98db-44fefa0ef987-logs\") pod \"cinder-api-0\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.793920 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-config-data-custom\") pod \"cinder-api-0\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.794992 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.795174 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-public-tls-certs\") pod \"cinder-api-0\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.796889 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-config-data\") pod \"cinder-api-0\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.798747 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.798873 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-scripts\") pod \"cinder-api-0\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.813640 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zg4vc\" (UniqueName: \"kubernetes.io/projected/ddbe9150-55f7-4f9f-98db-44fefa0ef987-kube-api-access-zg4vc\") pod \"cinder-api-0\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " pod="openstack/cinder-api-0" Nov 25 11:00:14 crc kubenswrapper[4776]: I1125 11:00:14.819542 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 11:00:15 crc kubenswrapper[4776]: I1125 11:00:15.269006 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 11:00:15 crc kubenswrapper[4776]: I1125 11:00:15.675017 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca" path="/var/lib/kubelet/pods/0c8c4d0a-6cf7-47a9-b5a3-b23249e6a7ca/volumes" Nov 25 11:00:16 crc kubenswrapper[4776]: I1125 11:00:16.157925 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ddbe9150-55f7-4f9f-98db-44fefa0ef987","Type":"ContainerStarted","Data":"4fd9eee5f071bd57285b29964248fcfee9919631ed5111a17492329a3b130027"} Nov 25 11:00:17 crc kubenswrapper[4776]: I1125 11:00:17.168296 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ddbe9150-55f7-4f9f-98db-44fefa0ef987","Type":"ContainerStarted","Data":"77966b22f32bf6be4d314b4131ad426c71fbae0ecf8a625bf1e6dfe8fdf40447"} Nov 25 11:00:17 crc kubenswrapper[4776]: I1125 11:00:17.168975 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ddbe9150-55f7-4f9f-98db-44fefa0ef987","Type":"ContainerStarted","Data":"b9e7ac47c1cd882f94aa42ce1600780b4ae95276ee5166e9e1e003ef48e99972"} Nov 25 11:00:17 crc kubenswrapper[4776]: I1125 11:00:17.169001 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 25 11:00:17 crc kubenswrapper[4776]: I1125 11:00:17.195539 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.195508994 podStartE2EDuration="3.195508994s" podCreationTimestamp="2025-11-25 11:00:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:00:17.188301373 +0000 UTC m=+5762.229360986" watchObservedRunningTime="2025-11-25 11:00:17.195508994 +0000 UTC m=+5762.236568557" Nov 25 11:00:17 crc kubenswrapper[4776]: I1125 11:00:17.819186 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:00:17 crc kubenswrapper[4776]: I1125 11:00:17.819308 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:00:26 crc kubenswrapper[4776]: I1125 11:00:26.640179 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 25 11:00:32 crc kubenswrapper[4776]: I1125 11:00:32.496781 4776 scope.go:117] "RemoveContainer" containerID="fa22650a109f2356b9687b9bec7a6d93988d1cac25c975f4f83548d0996703a4" Nov 25 11:00:32 crc kubenswrapper[4776]: I1125 11:00:32.555130 4776 scope.go:117] "RemoveContainer" containerID="c189a2b86eca91bc5be33fbc2bafdcdda468548845cf5f1546dd2249836d2c54" Nov 25 11:00:32 crc kubenswrapper[4776]: I1125 11:00:32.591291 4776 scope.go:117] "RemoveContainer" containerID="ba422d27a9918f591c7ba0af0fe8f866b7b0b1bb352015072379ae794b2b2e4a" Nov 25 11:00:46 crc kubenswrapper[4776]: I1125 11:00:46.912881 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 11:00:46 crc kubenswrapper[4776]: I1125 11:00:46.916724 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 11:00:46 crc kubenswrapper[4776]: I1125 11:00:46.920329 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 25 11:00:46 crc kubenswrapper[4776]: I1125 11:00:46.922585 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 11:00:47 crc kubenswrapper[4776]: I1125 11:00:47.078497 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43532459-8c82-4389-abed-0e6757ee89a0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"43532459-8c82-4389-abed-0e6757ee89a0\") " pod="openstack/cinder-scheduler-0" Nov 25 11:00:47 crc kubenswrapper[4776]: I1125 11:00:47.078551 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/43532459-8c82-4389-abed-0e6757ee89a0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"43532459-8c82-4389-abed-0e6757ee89a0\") " pod="openstack/cinder-scheduler-0" Nov 25 11:00:47 crc kubenswrapper[4776]: I1125 11:00:47.078583 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43532459-8c82-4389-abed-0e6757ee89a0-config-data\") pod \"cinder-scheduler-0\" (UID: \"43532459-8c82-4389-abed-0e6757ee89a0\") " pod="openstack/cinder-scheduler-0" Nov 25 11:00:47 crc kubenswrapper[4776]: I1125 11:00:47.078721 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xd8gr\" (UniqueName: \"kubernetes.io/projected/43532459-8c82-4389-abed-0e6757ee89a0-kube-api-access-xd8gr\") pod \"cinder-scheduler-0\" (UID: \"43532459-8c82-4389-abed-0e6757ee89a0\") " pod="openstack/cinder-scheduler-0" Nov 25 11:00:47 crc kubenswrapper[4776]: I1125 11:00:47.078917 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43532459-8c82-4389-abed-0e6757ee89a0-scripts\") pod \"cinder-scheduler-0\" (UID: \"43532459-8c82-4389-abed-0e6757ee89a0\") " pod="openstack/cinder-scheduler-0" Nov 25 11:00:47 crc kubenswrapper[4776]: I1125 11:00:47.078954 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/43532459-8c82-4389-abed-0e6757ee89a0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"43532459-8c82-4389-abed-0e6757ee89a0\") " pod="openstack/cinder-scheduler-0" Nov 25 11:00:47 crc kubenswrapper[4776]: I1125 11:00:47.180883 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43532459-8c82-4389-abed-0e6757ee89a0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"43532459-8c82-4389-abed-0e6757ee89a0\") " pod="openstack/cinder-scheduler-0" Nov 25 11:00:47 crc kubenswrapper[4776]: I1125 11:00:47.180941 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/43532459-8c82-4389-abed-0e6757ee89a0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"43532459-8c82-4389-abed-0e6757ee89a0\") " pod="openstack/cinder-scheduler-0" Nov 25 11:00:47 crc kubenswrapper[4776]: I1125 11:00:47.180984 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43532459-8c82-4389-abed-0e6757ee89a0-config-data\") pod \"cinder-scheduler-0\" (UID: \"43532459-8c82-4389-abed-0e6757ee89a0\") " pod="openstack/cinder-scheduler-0" Nov 25 11:00:47 crc kubenswrapper[4776]: I1125 11:00:47.181021 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xd8gr\" (UniqueName: \"kubernetes.io/projected/43532459-8c82-4389-abed-0e6757ee89a0-kube-api-access-xd8gr\") pod \"cinder-scheduler-0\" (UID: \"43532459-8c82-4389-abed-0e6757ee89a0\") " pod="openstack/cinder-scheduler-0" Nov 25 11:00:47 crc kubenswrapper[4776]: I1125 11:00:47.181103 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43532459-8c82-4389-abed-0e6757ee89a0-scripts\") pod \"cinder-scheduler-0\" (UID: \"43532459-8c82-4389-abed-0e6757ee89a0\") " pod="openstack/cinder-scheduler-0" Nov 25 11:00:47 crc kubenswrapper[4776]: I1125 11:00:47.181130 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/43532459-8c82-4389-abed-0e6757ee89a0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"43532459-8c82-4389-abed-0e6757ee89a0\") " pod="openstack/cinder-scheduler-0" Nov 25 11:00:47 crc kubenswrapper[4776]: I1125 11:00:47.181290 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/43532459-8c82-4389-abed-0e6757ee89a0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"43532459-8c82-4389-abed-0e6757ee89a0\") " pod="openstack/cinder-scheduler-0" Nov 25 11:00:47 crc kubenswrapper[4776]: I1125 11:00:47.186483 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43532459-8c82-4389-abed-0e6757ee89a0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"43532459-8c82-4389-abed-0e6757ee89a0\") " pod="openstack/cinder-scheduler-0" Nov 25 11:00:47 crc kubenswrapper[4776]: I1125 11:00:47.186524 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43532459-8c82-4389-abed-0e6757ee89a0-scripts\") pod \"cinder-scheduler-0\" (UID: \"43532459-8c82-4389-abed-0e6757ee89a0\") " pod="openstack/cinder-scheduler-0" Nov 25 11:00:47 crc kubenswrapper[4776]: I1125 11:00:47.186535 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/43532459-8c82-4389-abed-0e6757ee89a0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"43532459-8c82-4389-abed-0e6757ee89a0\") " pod="openstack/cinder-scheduler-0" Nov 25 11:00:47 crc kubenswrapper[4776]: I1125 11:00:47.194060 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43532459-8c82-4389-abed-0e6757ee89a0-config-data\") pod \"cinder-scheduler-0\" (UID: \"43532459-8c82-4389-abed-0e6757ee89a0\") " pod="openstack/cinder-scheduler-0" Nov 25 11:00:47 crc kubenswrapper[4776]: I1125 11:00:47.199773 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xd8gr\" (UniqueName: \"kubernetes.io/projected/43532459-8c82-4389-abed-0e6757ee89a0-kube-api-access-xd8gr\") pod \"cinder-scheduler-0\" (UID: \"43532459-8c82-4389-abed-0e6757ee89a0\") " pod="openstack/cinder-scheduler-0" Nov 25 11:00:47 crc kubenswrapper[4776]: I1125 11:00:47.256968 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 11:00:47 crc kubenswrapper[4776]: I1125 11:00:47.704232 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 11:00:47 crc kubenswrapper[4776]: I1125 11:00:47.820030 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:00:47 crc kubenswrapper[4776]: I1125 11:00:47.820116 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:00:47 crc kubenswrapper[4776]: I1125 11:00:47.820166 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 11:00:47 crc kubenswrapper[4776]: I1125 11:00:47.820861 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8fb1377358b27002d683fed0c035aa47a3bd9a7c30a5ee14f1b728159b2160e7"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 11:00:47 crc kubenswrapper[4776]: I1125 11:00:47.820931 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://8fb1377358b27002d683fed0c035aa47a3bd9a7c30a5ee14f1b728159b2160e7" gracePeriod=600 Nov 25 11:00:48 crc kubenswrapper[4776]: I1125 11:00:48.207939 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 25 11:00:48 crc kubenswrapper[4776]: I1125 11:00:48.208487 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="ddbe9150-55f7-4f9f-98db-44fefa0ef987" containerName="cinder-api-log" containerID="cri-o://b9e7ac47c1cd882f94aa42ce1600780b4ae95276ee5166e9e1e003ef48e99972" gracePeriod=30 Nov 25 11:00:48 crc kubenswrapper[4776]: I1125 11:00:48.208894 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="ddbe9150-55f7-4f9f-98db-44fefa0ef987" containerName="cinder-api" containerID="cri-o://77966b22f32bf6be4d314b4131ad426c71fbae0ecf8a625bf1e6dfe8fdf40447" gracePeriod=30 Nov 25 11:00:48 crc kubenswrapper[4776]: I1125 11:00:48.444887 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"43532459-8c82-4389-abed-0e6757ee89a0","Type":"ContainerStarted","Data":"2e2b3b36d030ad665cfb318780ce74cbee7637a03cabb385c9575c04099d53cb"} Nov 25 11:00:48 crc kubenswrapper[4776]: I1125 11:00:48.445228 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"43532459-8c82-4389-abed-0e6757ee89a0","Type":"ContainerStarted","Data":"1acffdd214840be74e16a52b90ebcc7c916f44d7985ed1afe29590b0b8ac9415"} Nov 25 11:00:48 crc kubenswrapper[4776]: I1125 11:00:48.447768 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="8fb1377358b27002d683fed0c035aa47a3bd9a7c30a5ee14f1b728159b2160e7" exitCode=0 Nov 25 11:00:48 crc kubenswrapper[4776]: I1125 11:00:48.447825 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"8fb1377358b27002d683fed0c035aa47a3bd9a7c30a5ee14f1b728159b2160e7"} Nov 25 11:00:48 crc kubenswrapper[4776]: I1125 11:00:48.447856 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698"} Nov 25 11:00:48 crc kubenswrapper[4776]: I1125 11:00:48.447873 4776 scope.go:117] "RemoveContainer" containerID="f828df88450a4b0fc3635a5f3c22d5195948f2d6dbda85072ec7e9dbcef83c40" Nov 25 11:00:48 crc kubenswrapper[4776]: I1125 11:00:48.460244 4776 generic.go:334] "Generic (PLEG): container finished" podID="ddbe9150-55f7-4f9f-98db-44fefa0ef987" containerID="b9e7ac47c1cd882f94aa42ce1600780b4ae95276ee5166e9e1e003ef48e99972" exitCode=143 Nov 25 11:00:48 crc kubenswrapper[4776]: I1125 11:00:48.460290 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ddbe9150-55f7-4f9f-98db-44fefa0ef987","Type":"ContainerDied","Data":"b9e7ac47c1cd882f94aa42ce1600780b4ae95276ee5166e9e1e003ef48e99972"} Nov 25 11:00:49 crc kubenswrapper[4776]: I1125 11:00:49.472905 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"43532459-8c82-4389-abed-0e6757ee89a0","Type":"ContainerStarted","Data":"261a09bb9832c31e0eaeca206b32016f57c59e89546c1f1b25c3bbacc0e1bd92"} Nov 25 11:00:49 crc kubenswrapper[4776]: I1125 11:00:49.502156 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.502138042 podStartE2EDuration="3.502138042s" podCreationTimestamp="2025-11-25 11:00:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:00:49.495928346 +0000 UTC m=+5794.536987919" watchObservedRunningTime="2025-11-25 11:00:49.502138042 +0000 UTC m=+5794.543197595" Nov 25 11:00:51 crc kubenswrapper[4776]: I1125 11:00:51.363185 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="ddbe9150-55f7-4f9f-98db-44fefa0ef987" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.1.57:8776/healthcheck\": read tcp 10.217.0.2:53678->10.217.1.57:8776: read: connection reset by peer" Nov 25 11:00:51 crc kubenswrapper[4776]: I1125 11:00:51.492795 4776 generic.go:334] "Generic (PLEG): container finished" podID="ddbe9150-55f7-4f9f-98db-44fefa0ef987" containerID="77966b22f32bf6be4d314b4131ad426c71fbae0ecf8a625bf1e6dfe8fdf40447" exitCode=0 Nov 25 11:00:51 crc kubenswrapper[4776]: I1125 11:00:51.492953 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ddbe9150-55f7-4f9f-98db-44fefa0ef987","Type":"ContainerDied","Data":"77966b22f32bf6be4d314b4131ad426c71fbae0ecf8a625bf1e6dfe8fdf40447"} Nov 25 11:00:51 crc kubenswrapper[4776]: I1125 11:00:51.854885 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 11:00:51 crc kubenswrapper[4776]: I1125 11:00:51.978902 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ddbe9150-55f7-4f9f-98db-44fefa0ef987-etc-machine-id\") pod \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " Nov 25 11:00:51 crc kubenswrapper[4776]: I1125 11:00:51.979281 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-public-tls-certs\") pod \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " Nov 25 11:00:51 crc kubenswrapper[4776]: I1125 11:00:51.979228 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ddbe9150-55f7-4f9f-98db-44fefa0ef987-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ddbe9150-55f7-4f9f-98db-44fefa0ef987" (UID: "ddbe9150-55f7-4f9f-98db-44fefa0ef987"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 11:00:51 crc kubenswrapper[4776]: I1125 11:00:51.979356 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zg4vc\" (UniqueName: \"kubernetes.io/projected/ddbe9150-55f7-4f9f-98db-44fefa0ef987-kube-api-access-zg4vc\") pod \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " Nov 25 11:00:51 crc kubenswrapper[4776]: I1125 11:00:51.980153 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-config-data-custom\") pod \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " Nov 25 11:00:51 crc kubenswrapper[4776]: I1125 11:00:51.980180 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddbe9150-55f7-4f9f-98db-44fefa0ef987-logs\") pod \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " Nov 25 11:00:51 crc kubenswrapper[4776]: I1125 11:00:51.980209 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-config-data\") pod \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " Nov 25 11:00:51 crc kubenswrapper[4776]: I1125 11:00:51.980233 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-scripts\") pod \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " Nov 25 11:00:51 crc kubenswrapper[4776]: I1125 11:00:51.980272 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-internal-tls-certs\") pod \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " Nov 25 11:00:51 crc kubenswrapper[4776]: I1125 11:00:51.980299 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-combined-ca-bundle\") pod \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\" (UID: \"ddbe9150-55f7-4f9f-98db-44fefa0ef987\") " Nov 25 11:00:51 crc kubenswrapper[4776]: I1125 11:00:51.981314 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ddbe9150-55f7-4f9f-98db-44fefa0ef987-logs" (OuterVolumeSpecName: "logs") pod "ddbe9150-55f7-4f9f-98db-44fefa0ef987" (UID: "ddbe9150-55f7-4f9f-98db-44fefa0ef987"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.002291 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-scripts" (OuterVolumeSpecName: "scripts") pod "ddbe9150-55f7-4f9f-98db-44fefa0ef987" (UID: "ddbe9150-55f7-4f9f-98db-44fefa0ef987"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.004556 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddbe9150-55f7-4f9f-98db-44fefa0ef987-kube-api-access-zg4vc" (OuterVolumeSpecName: "kube-api-access-zg4vc") pod "ddbe9150-55f7-4f9f-98db-44fefa0ef987" (UID: "ddbe9150-55f7-4f9f-98db-44fefa0ef987"). InnerVolumeSpecName "kube-api-access-zg4vc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.018290 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ddbe9150-55f7-4f9f-98db-44fefa0ef987" (UID: "ddbe9150-55f7-4f9f-98db-44fefa0ef987"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.043284 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ddbe9150-55f7-4f9f-98db-44fefa0ef987" (UID: "ddbe9150-55f7-4f9f-98db-44fefa0ef987"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.088397 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ddbe9150-55f7-4f9f-98db-44fefa0ef987" (UID: "ddbe9150-55f7-4f9f-98db-44fefa0ef987"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.088411 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.088486 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.088502 4776 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ddbe9150-55f7-4f9f-98db-44fefa0ef987-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.088515 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zg4vc\" (UniqueName: \"kubernetes.io/projected/ddbe9150-55f7-4f9f-98db-44fefa0ef987-kube-api-access-zg4vc\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.088529 4776 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.088540 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ddbe9150-55f7-4f9f-98db-44fefa0ef987-logs\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.111656 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ddbe9150-55f7-4f9f-98db-44fefa0ef987" (UID: "ddbe9150-55f7-4f9f-98db-44fefa0ef987"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.114337 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-config-data" (OuterVolumeSpecName: "config-data") pod "ddbe9150-55f7-4f9f-98db-44fefa0ef987" (UID: "ddbe9150-55f7-4f9f-98db-44fefa0ef987"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.190655 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.190692 4776 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.190704 4776 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddbe9150-55f7-4f9f-98db-44fefa0ef987-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.257977 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.503667 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ddbe9150-55f7-4f9f-98db-44fefa0ef987","Type":"ContainerDied","Data":"4fd9eee5f071bd57285b29964248fcfee9919631ed5111a17492329a3b130027"} Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.504006 4776 scope.go:117] "RemoveContainer" containerID="77966b22f32bf6be4d314b4131ad426c71fbae0ecf8a625bf1e6dfe8fdf40447" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.503900 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.533727 4776 scope.go:117] "RemoveContainer" containerID="b9e7ac47c1cd882f94aa42ce1600780b4ae95276ee5166e9e1e003ef48e99972" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.545196 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.567823 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.574480 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 25 11:00:52 crc kubenswrapper[4776]: E1125 11:00:52.576225 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddbe9150-55f7-4f9f-98db-44fefa0ef987" containerName="cinder-api" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.576264 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddbe9150-55f7-4f9f-98db-44fefa0ef987" containerName="cinder-api" Nov 25 11:00:52 crc kubenswrapper[4776]: E1125 11:00:52.576285 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddbe9150-55f7-4f9f-98db-44fefa0ef987" containerName="cinder-api-log" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.576292 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddbe9150-55f7-4f9f-98db-44fefa0ef987" containerName="cinder-api-log" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.576523 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddbe9150-55f7-4f9f-98db-44fefa0ef987" containerName="cinder-api-log" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.576549 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddbe9150-55f7-4f9f-98db-44fefa0ef987" containerName="cinder-api" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.577618 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.579549 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.580784 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.580842 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.584763 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.700770 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d076c50-21b1-4e6d-b345-a8c46bde8e8f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6d076c50-21b1-4e6d-b345-a8c46bde8e8f\") " pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.700837 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vs8fg\" (UniqueName: \"kubernetes.io/projected/6d076c50-21b1-4e6d-b345-a8c46bde8e8f-kube-api-access-vs8fg\") pod \"cinder-api-0\" (UID: \"6d076c50-21b1-4e6d-b345-a8c46bde8e8f\") " pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.700910 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d076c50-21b1-4e6d-b345-a8c46bde8e8f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"6d076c50-21b1-4e6d-b345-a8c46bde8e8f\") " pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.700994 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d076c50-21b1-4e6d-b345-a8c46bde8e8f-config-data\") pod \"cinder-api-0\" (UID: \"6d076c50-21b1-4e6d-b345-a8c46bde8e8f\") " pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.701238 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6d076c50-21b1-4e6d-b345-a8c46bde8e8f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6d076c50-21b1-4e6d-b345-a8c46bde8e8f\") " pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.701350 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d076c50-21b1-4e6d-b345-a8c46bde8e8f-scripts\") pod \"cinder-api-0\" (UID: \"6d076c50-21b1-4e6d-b345-a8c46bde8e8f\") " pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.701429 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d076c50-21b1-4e6d-b345-a8c46bde8e8f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"6d076c50-21b1-4e6d-b345-a8c46bde8e8f\") " pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.701462 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6d076c50-21b1-4e6d-b345-a8c46bde8e8f-config-data-custom\") pod \"cinder-api-0\" (UID: \"6d076c50-21b1-4e6d-b345-a8c46bde8e8f\") " pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.701490 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d076c50-21b1-4e6d-b345-a8c46bde8e8f-logs\") pod \"cinder-api-0\" (UID: \"6d076c50-21b1-4e6d-b345-a8c46bde8e8f\") " pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.803056 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6d076c50-21b1-4e6d-b345-a8c46bde8e8f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6d076c50-21b1-4e6d-b345-a8c46bde8e8f\") " pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.803168 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d076c50-21b1-4e6d-b345-a8c46bde8e8f-scripts\") pod \"cinder-api-0\" (UID: \"6d076c50-21b1-4e6d-b345-a8c46bde8e8f\") " pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.803191 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6d076c50-21b1-4e6d-b345-a8c46bde8e8f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6d076c50-21b1-4e6d-b345-a8c46bde8e8f\") " pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.803275 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d076c50-21b1-4e6d-b345-a8c46bde8e8f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"6d076c50-21b1-4e6d-b345-a8c46bde8e8f\") " pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.803300 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6d076c50-21b1-4e6d-b345-a8c46bde8e8f-config-data-custom\") pod \"cinder-api-0\" (UID: \"6d076c50-21b1-4e6d-b345-a8c46bde8e8f\") " pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.803321 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d076c50-21b1-4e6d-b345-a8c46bde8e8f-logs\") pod \"cinder-api-0\" (UID: \"6d076c50-21b1-4e6d-b345-a8c46bde8e8f\") " pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.803400 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d076c50-21b1-4e6d-b345-a8c46bde8e8f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6d076c50-21b1-4e6d-b345-a8c46bde8e8f\") " pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.803422 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vs8fg\" (UniqueName: \"kubernetes.io/projected/6d076c50-21b1-4e6d-b345-a8c46bde8e8f-kube-api-access-vs8fg\") pod \"cinder-api-0\" (UID: \"6d076c50-21b1-4e6d-b345-a8c46bde8e8f\") " pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.803455 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d076c50-21b1-4e6d-b345-a8c46bde8e8f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"6d076c50-21b1-4e6d-b345-a8c46bde8e8f\") " pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.803485 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d076c50-21b1-4e6d-b345-a8c46bde8e8f-config-data\") pod \"cinder-api-0\" (UID: \"6d076c50-21b1-4e6d-b345-a8c46bde8e8f\") " pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.804301 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d076c50-21b1-4e6d-b345-a8c46bde8e8f-logs\") pod \"cinder-api-0\" (UID: \"6d076c50-21b1-4e6d-b345-a8c46bde8e8f\") " pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.808673 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d076c50-21b1-4e6d-b345-a8c46bde8e8f-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"6d076c50-21b1-4e6d-b345-a8c46bde8e8f\") " pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.808701 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d076c50-21b1-4e6d-b345-a8c46bde8e8f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6d076c50-21b1-4e6d-b345-a8c46bde8e8f\") " pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.808715 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d076c50-21b1-4e6d-b345-a8c46bde8e8f-public-tls-certs\") pod \"cinder-api-0\" (UID: \"6d076c50-21b1-4e6d-b345-a8c46bde8e8f\") " pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.810081 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6d076c50-21b1-4e6d-b345-a8c46bde8e8f-config-data-custom\") pod \"cinder-api-0\" (UID: \"6d076c50-21b1-4e6d-b345-a8c46bde8e8f\") " pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.810289 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d076c50-21b1-4e6d-b345-a8c46bde8e8f-config-data\") pod \"cinder-api-0\" (UID: \"6d076c50-21b1-4e6d-b345-a8c46bde8e8f\") " pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.810692 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d076c50-21b1-4e6d-b345-a8c46bde8e8f-scripts\") pod \"cinder-api-0\" (UID: \"6d076c50-21b1-4e6d-b345-a8c46bde8e8f\") " pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.824220 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vs8fg\" (UniqueName: \"kubernetes.io/projected/6d076c50-21b1-4e6d-b345-a8c46bde8e8f-kube-api-access-vs8fg\") pod \"cinder-api-0\" (UID: \"6d076c50-21b1-4e6d-b345-a8c46bde8e8f\") " pod="openstack/cinder-api-0" Nov 25 11:00:52 crc kubenswrapper[4776]: I1125 11:00:52.902777 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 11:00:53 crc kubenswrapper[4776]: I1125 11:00:53.375060 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 11:00:53 crc kubenswrapper[4776]: I1125 11:00:53.516578 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6d076c50-21b1-4e6d-b345-a8c46bde8e8f","Type":"ContainerStarted","Data":"89da2e10065de0f655a8c7fbb03a9f2963cb3e1d7d82d4af750eaeded8d71636"} Nov 25 11:00:53 crc kubenswrapper[4776]: I1125 11:00:53.675675 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddbe9150-55f7-4f9f-98db-44fefa0ef987" path="/var/lib/kubelet/pods/ddbe9150-55f7-4f9f-98db-44fefa0ef987/volumes" Nov 25 11:00:54 crc kubenswrapper[4776]: I1125 11:00:54.532740 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6d076c50-21b1-4e6d-b345-a8c46bde8e8f","Type":"ContainerStarted","Data":"6b2168d14ec294c45c71dd5df676bfb659dcf66adc9e0a723458d32914b2d28c"} Nov 25 11:00:54 crc kubenswrapper[4776]: I1125 11:00:54.533111 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6d076c50-21b1-4e6d-b345-a8c46bde8e8f","Type":"ContainerStarted","Data":"fa81a97e5247f7ca66a54f418c95b1a92735160d0277268d54a9badb75403057"} Nov 25 11:00:54 crc kubenswrapper[4776]: I1125 11:00:54.533377 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 25 11:00:54 crc kubenswrapper[4776]: I1125 11:00:54.555112 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.555072683 podStartE2EDuration="2.555072683s" podCreationTimestamp="2025-11-25 11:00:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:00:54.549229816 +0000 UTC m=+5799.590289369" watchObservedRunningTime="2025-11-25 11:00:54.555072683 +0000 UTC m=+5799.596132236" Nov 25 11:00:57 crc kubenswrapper[4776]: I1125 11:00:57.464899 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 25 11:00:57 crc kubenswrapper[4776]: I1125 11:00:57.518504 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 11:00:57 crc kubenswrapper[4776]: I1125 11:00:57.558345 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="43532459-8c82-4389-abed-0e6757ee89a0" containerName="cinder-scheduler" containerID="cri-o://2e2b3b36d030ad665cfb318780ce74cbee7637a03cabb385c9575c04099d53cb" gracePeriod=30 Nov 25 11:00:57 crc kubenswrapper[4776]: I1125 11:00:57.558903 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="43532459-8c82-4389-abed-0e6757ee89a0" containerName="probe" containerID="cri-o://261a09bb9832c31e0eaeca206b32016f57c59e89546c1f1b25c3bbacc0e1bd92" gracePeriod=30 Nov 25 11:00:58 crc kubenswrapper[4776]: I1125 11:00:58.569898 4776 generic.go:334] "Generic (PLEG): container finished" podID="43532459-8c82-4389-abed-0e6757ee89a0" containerID="261a09bb9832c31e0eaeca206b32016f57c59e89546c1f1b25c3bbacc0e1bd92" exitCode=0 Nov 25 11:00:58 crc kubenswrapper[4776]: I1125 11:00:58.569965 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"43532459-8c82-4389-abed-0e6757ee89a0","Type":"ContainerDied","Data":"261a09bb9832c31e0eaeca206b32016f57c59e89546c1f1b25c3bbacc0e1bd92"} Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.431711 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.533043 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xd8gr\" (UniqueName: \"kubernetes.io/projected/43532459-8c82-4389-abed-0e6757ee89a0-kube-api-access-xd8gr\") pod \"43532459-8c82-4389-abed-0e6757ee89a0\" (UID: \"43532459-8c82-4389-abed-0e6757ee89a0\") " Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.533471 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43532459-8c82-4389-abed-0e6757ee89a0-scripts\") pod \"43532459-8c82-4389-abed-0e6757ee89a0\" (UID: \"43532459-8c82-4389-abed-0e6757ee89a0\") " Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.533536 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43532459-8c82-4389-abed-0e6757ee89a0-config-data\") pod \"43532459-8c82-4389-abed-0e6757ee89a0\" (UID: \"43532459-8c82-4389-abed-0e6757ee89a0\") " Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.533584 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43532459-8c82-4389-abed-0e6757ee89a0-combined-ca-bundle\") pod \"43532459-8c82-4389-abed-0e6757ee89a0\" (UID: \"43532459-8c82-4389-abed-0e6757ee89a0\") " Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.533654 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/43532459-8c82-4389-abed-0e6757ee89a0-config-data-custom\") pod \"43532459-8c82-4389-abed-0e6757ee89a0\" (UID: \"43532459-8c82-4389-abed-0e6757ee89a0\") " Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.533791 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/43532459-8c82-4389-abed-0e6757ee89a0-etc-machine-id\") pod \"43532459-8c82-4389-abed-0e6757ee89a0\" (UID: \"43532459-8c82-4389-abed-0e6757ee89a0\") " Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.533923 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/43532459-8c82-4389-abed-0e6757ee89a0-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "43532459-8c82-4389-abed-0e6757ee89a0" (UID: "43532459-8c82-4389-abed-0e6757ee89a0"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.534420 4776 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/43532459-8c82-4389-abed-0e6757ee89a0-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.540508 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43532459-8c82-4389-abed-0e6757ee89a0-scripts" (OuterVolumeSpecName: "scripts") pod "43532459-8c82-4389-abed-0e6757ee89a0" (UID: "43532459-8c82-4389-abed-0e6757ee89a0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.551236 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43532459-8c82-4389-abed-0e6757ee89a0-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "43532459-8c82-4389-abed-0e6757ee89a0" (UID: "43532459-8c82-4389-abed-0e6757ee89a0"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.551299 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43532459-8c82-4389-abed-0e6757ee89a0-kube-api-access-xd8gr" (OuterVolumeSpecName: "kube-api-access-xd8gr") pod "43532459-8c82-4389-abed-0e6757ee89a0" (UID: "43532459-8c82-4389-abed-0e6757ee89a0"). InnerVolumeSpecName "kube-api-access-xd8gr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.597710 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43532459-8c82-4389-abed-0e6757ee89a0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "43532459-8c82-4389-abed-0e6757ee89a0" (UID: "43532459-8c82-4389-abed-0e6757ee89a0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.606118 4776 generic.go:334] "Generic (PLEG): container finished" podID="43532459-8c82-4389-abed-0e6757ee89a0" containerID="2e2b3b36d030ad665cfb318780ce74cbee7637a03cabb385c9575c04099d53cb" exitCode=0 Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.606172 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"43532459-8c82-4389-abed-0e6757ee89a0","Type":"ContainerDied","Data":"2e2b3b36d030ad665cfb318780ce74cbee7637a03cabb385c9575c04099d53cb"} Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.606202 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"43532459-8c82-4389-abed-0e6757ee89a0","Type":"ContainerDied","Data":"1acffdd214840be74e16a52b90ebcc7c916f44d7985ed1afe29590b0b8ac9415"} Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.606222 4776 scope.go:117] "RemoveContainer" containerID="261a09bb9832c31e0eaeca206b32016f57c59e89546c1f1b25c3bbacc0e1bd92" Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.606365 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.633239 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43532459-8c82-4389-abed-0e6757ee89a0-config-data" (OuterVolumeSpecName: "config-data") pod "43532459-8c82-4389-abed-0e6757ee89a0" (UID: "43532459-8c82-4389-abed-0e6757ee89a0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.636304 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xd8gr\" (UniqueName: \"kubernetes.io/projected/43532459-8c82-4389-abed-0e6757ee89a0-kube-api-access-xd8gr\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.636331 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/43532459-8c82-4389-abed-0e6757ee89a0-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.636339 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43532459-8c82-4389-abed-0e6757ee89a0-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.636348 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43532459-8c82-4389-abed-0e6757ee89a0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.636356 4776 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/43532459-8c82-4389-abed-0e6757ee89a0-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.700766 4776 scope.go:117] "RemoveContainer" containerID="2e2b3b36d030ad665cfb318780ce74cbee7637a03cabb385c9575c04099d53cb" Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.726674 4776 scope.go:117] "RemoveContainer" containerID="261a09bb9832c31e0eaeca206b32016f57c59e89546c1f1b25c3bbacc0e1bd92" Nov 25 11:00:59 crc kubenswrapper[4776]: E1125 11:00:59.727304 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"261a09bb9832c31e0eaeca206b32016f57c59e89546c1f1b25c3bbacc0e1bd92\": container with ID starting with 261a09bb9832c31e0eaeca206b32016f57c59e89546c1f1b25c3bbacc0e1bd92 not found: ID does not exist" containerID="261a09bb9832c31e0eaeca206b32016f57c59e89546c1f1b25c3bbacc0e1bd92" Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.727423 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"261a09bb9832c31e0eaeca206b32016f57c59e89546c1f1b25c3bbacc0e1bd92"} err="failed to get container status \"261a09bb9832c31e0eaeca206b32016f57c59e89546c1f1b25c3bbacc0e1bd92\": rpc error: code = NotFound desc = could not find container \"261a09bb9832c31e0eaeca206b32016f57c59e89546c1f1b25c3bbacc0e1bd92\": container with ID starting with 261a09bb9832c31e0eaeca206b32016f57c59e89546c1f1b25c3bbacc0e1bd92 not found: ID does not exist" Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.727449 4776 scope.go:117] "RemoveContainer" containerID="2e2b3b36d030ad665cfb318780ce74cbee7637a03cabb385c9575c04099d53cb" Nov 25 11:00:59 crc kubenswrapper[4776]: E1125 11:00:59.727804 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e2b3b36d030ad665cfb318780ce74cbee7637a03cabb385c9575c04099d53cb\": container with ID starting with 2e2b3b36d030ad665cfb318780ce74cbee7637a03cabb385c9575c04099d53cb not found: ID does not exist" containerID="2e2b3b36d030ad665cfb318780ce74cbee7637a03cabb385c9575c04099d53cb" Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.727844 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e2b3b36d030ad665cfb318780ce74cbee7637a03cabb385c9575c04099d53cb"} err="failed to get container status \"2e2b3b36d030ad665cfb318780ce74cbee7637a03cabb385c9575c04099d53cb\": rpc error: code = NotFound desc = could not find container \"2e2b3b36d030ad665cfb318780ce74cbee7637a03cabb385c9575c04099d53cb\": container with ID starting with 2e2b3b36d030ad665cfb318780ce74cbee7637a03cabb385c9575c04099d53cb not found: ID does not exist" Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.932870 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.940276 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.957266 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 11:00:59 crc kubenswrapper[4776]: E1125 11:00:59.957778 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43532459-8c82-4389-abed-0e6757ee89a0" containerName="probe" Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.957800 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="43532459-8c82-4389-abed-0e6757ee89a0" containerName="probe" Nov 25 11:00:59 crc kubenswrapper[4776]: E1125 11:00:59.957820 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43532459-8c82-4389-abed-0e6757ee89a0" containerName="cinder-scheduler" Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.957829 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="43532459-8c82-4389-abed-0e6757ee89a0" containerName="cinder-scheduler" Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.958608 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="43532459-8c82-4389-abed-0e6757ee89a0" containerName="cinder-scheduler" Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.958645 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="43532459-8c82-4389-abed-0e6757ee89a0" containerName="probe" Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.959712 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.963244 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 25 11:00:59 crc kubenswrapper[4776]: I1125 11:00:59.976469 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.043009 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18d53345-e467-4604-8446-450375280e28-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"18d53345-e467-4604-8446-450375280e28\") " pod="openstack/cinder-scheduler-0" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.043095 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-947fj\" (UniqueName: \"kubernetes.io/projected/18d53345-e467-4604-8446-450375280e28-kube-api-access-947fj\") pod \"cinder-scheduler-0\" (UID: \"18d53345-e467-4604-8446-450375280e28\") " pod="openstack/cinder-scheduler-0" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.043156 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/18d53345-e467-4604-8446-450375280e28-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"18d53345-e467-4604-8446-450375280e28\") " pod="openstack/cinder-scheduler-0" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.043206 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18d53345-e467-4604-8446-450375280e28-config-data\") pod \"cinder-scheduler-0\" (UID: \"18d53345-e467-4604-8446-450375280e28\") " pod="openstack/cinder-scheduler-0" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.043235 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/18d53345-e467-4604-8446-450375280e28-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"18d53345-e467-4604-8446-450375280e28\") " pod="openstack/cinder-scheduler-0" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.043326 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18d53345-e467-4604-8446-450375280e28-scripts\") pod \"cinder-scheduler-0\" (UID: \"18d53345-e467-4604-8446-450375280e28\") " pod="openstack/cinder-scheduler-0" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.142363 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29401141-9cqm2"] Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.143836 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29401141-9cqm2" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.144539 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/18d53345-e467-4604-8446-450375280e28-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"18d53345-e467-4604-8446-450375280e28\") " pod="openstack/cinder-scheduler-0" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.144586 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18d53345-e467-4604-8446-450375280e28-scripts\") pod \"cinder-scheduler-0\" (UID: \"18d53345-e467-4604-8446-450375280e28\") " pod="openstack/cinder-scheduler-0" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.144655 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18d53345-e467-4604-8446-450375280e28-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"18d53345-e467-4604-8446-450375280e28\") " pod="openstack/cinder-scheduler-0" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.144688 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-947fj\" (UniqueName: \"kubernetes.io/projected/18d53345-e467-4604-8446-450375280e28-kube-api-access-947fj\") pod \"cinder-scheduler-0\" (UID: \"18d53345-e467-4604-8446-450375280e28\") " pod="openstack/cinder-scheduler-0" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.144731 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/18d53345-e467-4604-8446-450375280e28-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"18d53345-e467-4604-8446-450375280e28\") " pod="openstack/cinder-scheduler-0" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.144754 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18d53345-e467-4604-8446-450375280e28-config-data\") pod \"cinder-scheduler-0\" (UID: \"18d53345-e467-4604-8446-450375280e28\") " pod="openstack/cinder-scheduler-0" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.144655 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/18d53345-e467-4604-8446-450375280e28-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"18d53345-e467-4604-8446-450375280e28\") " pod="openstack/cinder-scheduler-0" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.149329 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/18d53345-e467-4604-8446-450375280e28-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"18d53345-e467-4604-8446-450375280e28\") " pod="openstack/cinder-scheduler-0" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.154567 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18d53345-e467-4604-8446-450375280e28-scripts\") pod \"cinder-scheduler-0\" (UID: \"18d53345-e467-4604-8446-450375280e28\") " pod="openstack/cinder-scheduler-0" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.155569 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18d53345-e467-4604-8446-450375280e28-config-data\") pod \"cinder-scheduler-0\" (UID: \"18d53345-e467-4604-8446-450375280e28\") " pod="openstack/cinder-scheduler-0" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.157489 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29401141-9cqm2"] Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.160839 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18d53345-e467-4604-8446-450375280e28-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"18d53345-e467-4604-8446-450375280e28\") " pod="openstack/cinder-scheduler-0" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.171147 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-947fj\" (UniqueName: \"kubernetes.io/projected/18d53345-e467-4604-8446-450375280e28-kube-api-access-947fj\") pod \"cinder-scheduler-0\" (UID: \"18d53345-e467-4604-8446-450375280e28\") " pod="openstack/cinder-scheduler-0" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.246771 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ad73ef14-feaa-40da-98f8-4ac5687bf72b-fernet-keys\") pod \"keystone-cron-29401141-9cqm2\" (UID: \"ad73ef14-feaa-40da-98f8-4ac5687bf72b\") " pod="openstack/keystone-cron-29401141-9cqm2" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.246862 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad73ef14-feaa-40da-98f8-4ac5687bf72b-combined-ca-bundle\") pod \"keystone-cron-29401141-9cqm2\" (UID: \"ad73ef14-feaa-40da-98f8-4ac5687bf72b\") " pod="openstack/keystone-cron-29401141-9cqm2" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.246986 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9cf7\" (UniqueName: \"kubernetes.io/projected/ad73ef14-feaa-40da-98f8-4ac5687bf72b-kube-api-access-c9cf7\") pod \"keystone-cron-29401141-9cqm2\" (UID: \"ad73ef14-feaa-40da-98f8-4ac5687bf72b\") " pod="openstack/keystone-cron-29401141-9cqm2" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.247024 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad73ef14-feaa-40da-98f8-4ac5687bf72b-config-data\") pod \"keystone-cron-29401141-9cqm2\" (UID: \"ad73ef14-feaa-40da-98f8-4ac5687bf72b\") " pod="openstack/keystone-cron-29401141-9cqm2" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.283459 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.348417 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad73ef14-feaa-40da-98f8-4ac5687bf72b-combined-ca-bundle\") pod \"keystone-cron-29401141-9cqm2\" (UID: \"ad73ef14-feaa-40da-98f8-4ac5687bf72b\") " pod="openstack/keystone-cron-29401141-9cqm2" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.350148 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9cf7\" (UniqueName: \"kubernetes.io/projected/ad73ef14-feaa-40da-98f8-4ac5687bf72b-kube-api-access-c9cf7\") pod \"keystone-cron-29401141-9cqm2\" (UID: \"ad73ef14-feaa-40da-98f8-4ac5687bf72b\") " pod="openstack/keystone-cron-29401141-9cqm2" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.350216 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad73ef14-feaa-40da-98f8-4ac5687bf72b-config-data\") pod \"keystone-cron-29401141-9cqm2\" (UID: \"ad73ef14-feaa-40da-98f8-4ac5687bf72b\") " pod="openstack/keystone-cron-29401141-9cqm2" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.350275 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ad73ef14-feaa-40da-98f8-4ac5687bf72b-fernet-keys\") pod \"keystone-cron-29401141-9cqm2\" (UID: \"ad73ef14-feaa-40da-98f8-4ac5687bf72b\") " pod="openstack/keystone-cron-29401141-9cqm2" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.353626 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad73ef14-feaa-40da-98f8-4ac5687bf72b-combined-ca-bundle\") pod \"keystone-cron-29401141-9cqm2\" (UID: \"ad73ef14-feaa-40da-98f8-4ac5687bf72b\") " pod="openstack/keystone-cron-29401141-9cqm2" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.356220 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad73ef14-feaa-40da-98f8-4ac5687bf72b-config-data\") pod \"keystone-cron-29401141-9cqm2\" (UID: \"ad73ef14-feaa-40da-98f8-4ac5687bf72b\") " pod="openstack/keystone-cron-29401141-9cqm2" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.356640 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ad73ef14-feaa-40da-98f8-4ac5687bf72b-fernet-keys\") pod \"keystone-cron-29401141-9cqm2\" (UID: \"ad73ef14-feaa-40da-98f8-4ac5687bf72b\") " pod="openstack/keystone-cron-29401141-9cqm2" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.373499 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9cf7\" (UniqueName: \"kubernetes.io/projected/ad73ef14-feaa-40da-98f8-4ac5687bf72b-kube-api-access-c9cf7\") pod \"keystone-cron-29401141-9cqm2\" (UID: \"ad73ef14-feaa-40da-98f8-4ac5687bf72b\") " pod="openstack/keystone-cron-29401141-9cqm2" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.502754 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29401141-9cqm2" Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.549109 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.628407 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"18d53345-e467-4604-8446-450375280e28","Type":"ContainerStarted","Data":"7cd2e2d292a8c96354e5a332930e328546c5fd0910065f1af932abf1dd8b22df"} Nov 25 11:01:00 crc kubenswrapper[4776]: I1125 11:01:00.961022 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29401141-9cqm2"] Nov 25 11:01:01 crc kubenswrapper[4776]: I1125 11:01:01.647930 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29401141-9cqm2" event={"ID":"ad73ef14-feaa-40da-98f8-4ac5687bf72b","Type":"ContainerStarted","Data":"3cb451b5cfa856f6e9e0407cca0f0642232e5ec88fa577cf1385abe0a8b3568e"} Nov 25 11:01:01 crc kubenswrapper[4776]: I1125 11:01:01.648415 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29401141-9cqm2" event={"ID":"ad73ef14-feaa-40da-98f8-4ac5687bf72b","Type":"ContainerStarted","Data":"72fb0bd677117ae8442abe5e4cafbdd996b68796dd8d8f7661c1758675732100"} Nov 25 11:01:01 crc kubenswrapper[4776]: I1125 11:01:01.657651 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"18d53345-e467-4604-8446-450375280e28","Type":"ContainerStarted","Data":"1a2d82ec72ac54b429a7de45d2dae9c740c17edc18185ffd60db12feebc90cd5"} Nov 25 11:01:01 crc kubenswrapper[4776]: I1125 11:01:01.669156 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29401141-9cqm2" podStartSLOduration=1.669134502 podStartE2EDuration="1.669134502s" podCreationTimestamp="2025-11-25 11:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:01:01.663644214 +0000 UTC m=+5806.704703817" watchObservedRunningTime="2025-11-25 11:01:01.669134502 +0000 UTC m=+5806.710194075" Nov 25 11:01:01 crc kubenswrapper[4776]: I1125 11:01:01.677350 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43532459-8c82-4389-abed-0e6757ee89a0" path="/var/lib/kubelet/pods/43532459-8c82-4389-abed-0e6757ee89a0/volumes" Nov 25 11:01:02 crc kubenswrapper[4776]: I1125 11:01:02.668947 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"18d53345-e467-4604-8446-450375280e28","Type":"ContainerStarted","Data":"48e24d783bdd42b19c10fa0430c034e9857c68b9711a817e2564ab39310920be"} Nov 25 11:01:02 crc kubenswrapper[4776]: I1125 11:01:02.705483 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.705460341 podStartE2EDuration="3.705460341s" podCreationTimestamp="2025-11-25 11:00:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:01:02.695349837 +0000 UTC m=+5807.736409430" watchObservedRunningTime="2025-11-25 11:01:02.705460341 +0000 UTC m=+5807.746519914" Nov 25 11:01:03 crc kubenswrapper[4776]: I1125 11:01:03.707955 4776 generic.go:334] "Generic (PLEG): container finished" podID="ad73ef14-feaa-40da-98f8-4ac5687bf72b" containerID="3cb451b5cfa856f6e9e0407cca0f0642232e5ec88fa577cf1385abe0a8b3568e" exitCode=0 Nov 25 11:01:03 crc kubenswrapper[4776]: I1125 11:01:03.708049 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29401141-9cqm2" event={"ID":"ad73ef14-feaa-40da-98f8-4ac5687bf72b","Type":"ContainerDied","Data":"3cb451b5cfa856f6e9e0407cca0f0642232e5ec88fa577cf1385abe0a8b3568e"} Nov 25 11:01:04 crc kubenswrapper[4776]: I1125 11:01:04.852663 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 25 11:01:05 crc kubenswrapper[4776]: I1125 11:01:05.074536 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29401141-9cqm2" Nov 25 11:01:05 crc kubenswrapper[4776]: I1125 11:01:05.143212 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad73ef14-feaa-40da-98f8-4ac5687bf72b-combined-ca-bundle\") pod \"ad73ef14-feaa-40da-98f8-4ac5687bf72b\" (UID: \"ad73ef14-feaa-40da-98f8-4ac5687bf72b\") " Nov 25 11:01:05 crc kubenswrapper[4776]: I1125 11:01:05.143288 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9cf7\" (UniqueName: \"kubernetes.io/projected/ad73ef14-feaa-40da-98f8-4ac5687bf72b-kube-api-access-c9cf7\") pod \"ad73ef14-feaa-40da-98f8-4ac5687bf72b\" (UID: \"ad73ef14-feaa-40da-98f8-4ac5687bf72b\") " Nov 25 11:01:05 crc kubenswrapper[4776]: I1125 11:01:05.143347 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad73ef14-feaa-40da-98f8-4ac5687bf72b-config-data\") pod \"ad73ef14-feaa-40da-98f8-4ac5687bf72b\" (UID: \"ad73ef14-feaa-40da-98f8-4ac5687bf72b\") " Nov 25 11:01:05 crc kubenswrapper[4776]: I1125 11:01:05.143379 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ad73ef14-feaa-40da-98f8-4ac5687bf72b-fernet-keys\") pod \"ad73ef14-feaa-40da-98f8-4ac5687bf72b\" (UID: \"ad73ef14-feaa-40da-98f8-4ac5687bf72b\") " Nov 25 11:01:05 crc kubenswrapper[4776]: I1125 11:01:05.151043 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad73ef14-feaa-40da-98f8-4ac5687bf72b-kube-api-access-c9cf7" (OuterVolumeSpecName: "kube-api-access-c9cf7") pod "ad73ef14-feaa-40da-98f8-4ac5687bf72b" (UID: "ad73ef14-feaa-40da-98f8-4ac5687bf72b"). InnerVolumeSpecName "kube-api-access-c9cf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:01:05 crc kubenswrapper[4776]: I1125 11:01:05.151513 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad73ef14-feaa-40da-98f8-4ac5687bf72b-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ad73ef14-feaa-40da-98f8-4ac5687bf72b" (UID: "ad73ef14-feaa-40da-98f8-4ac5687bf72b"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:01:05 crc kubenswrapper[4776]: I1125 11:01:05.194228 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad73ef14-feaa-40da-98f8-4ac5687bf72b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ad73ef14-feaa-40da-98f8-4ac5687bf72b" (UID: "ad73ef14-feaa-40da-98f8-4ac5687bf72b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:01:05 crc kubenswrapper[4776]: I1125 11:01:05.195303 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad73ef14-feaa-40da-98f8-4ac5687bf72b-config-data" (OuterVolumeSpecName: "config-data") pod "ad73ef14-feaa-40da-98f8-4ac5687bf72b" (UID: "ad73ef14-feaa-40da-98f8-4ac5687bf72b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:01:05 crc kubenswrapper[4776]: I1125 11:01:05.247615 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad73ef14-feaa-40da-98f8-4ac5687bf72b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:05 crc kubenswrapper[4776]: I1125 11:01:05.247691 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9cf7\" (UniqueName: \"kubernetes.io/projected/ad73ef14-feaa-40da-98f8-4ac5687bf72b-kube-api-access-c9cf7\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:05 crc kubenswrapper[4776]: I1125 11:01:05.247708 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad73ef14-feaa-40da-98f8-4ac5687bf72b-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:05 crc kubenswrapper[4776]: I1125 11:01:05.247723 4776 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ad73ef14-feaa-40da-98f8-4ac5687bf72b-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:05 crc kubenswrapper[4776]: I1125 11:01:05.284639 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 25 11:01:05 crc kubenswrapper[4776]: I1125 11:01:05.727883 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29401141-9cqm2" event={"ID":"ad73ef14-feaa-40da-98f8-4ac5687bf72b","Type":"ContainerDied","Data":"72fb0bd677117ae8442abe5e4cafbdd996b68796dd8d8f7661c1758675732100"} Nov 25 11:01:05 crc kubenswrapper[4776]: I1125 11:01:05.727920 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72fb0bd677117ae8442abe5e4cafbdd996b68796dd8d8f7661c1758675732100" Nov 25 11:01:05 crc kubenswrapper[4776]: I1125 11:01:05.727946 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29401141-9cqm2" Nov 25 11:01:10 crc kubenswrapper[4776]: I1125 11:01:10.499596 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 25 11:01:13 crc kubenswrapper[4776]: I1125 11:01:13.219036 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-pnmqh"] Nov 25 11:01:13 crc kubenswrapper[4776]: E1125 11:01:13.219848 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad73ef14-feaa-40da-98f8-4ac5687bf72b" containerName="keystone-cron" Nov 25 11:01:13 crc kubenswrapper[4776]: I1125 11:01:13.219868 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad73ef14-feaa-40da-98f8-4ac5687bf72b" containerName="keystone-cron" Nov 25 11:01:13 crc kubenswrapper[4776]: I1125 11:01:13.220124 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad73ef14-feaa-40da-98f8-4ac5687bf72b" containerName="keystone-cron" Nov 25 11:01:13 crc kubenswrapper[4776]: I1125 11:01:13.220751 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pnmqh" Nov 25 11:01:13 crc kubenswrapper[4776]: I1125 11:01:13.232808 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-pnmqh"] Nov 25 11:01:13 crc kubenswrapper[4776]: I1125 11:01:13.303561 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5t45x\" (UniqueName: \"kubernetes.io/projected/e5e277db-7467-4a4f-9715-6854e235cac1-kube-api-access-5t45x\") pod \"glance-db-create-pnmqh\" (UID: \"e5e277db-7467-4a4f-9715-6854e235cac1\") " pod="openstack/glance-db-create-pnmqh" Nov 25 11:01:13 crc kubenswrapper[4776]: I1125 11:01:13.303729 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5e277db-7467-4a4f-9715-6854e235cac1-operator-scripts\") pod \"glance-db-create-pnmqh\" (UID: \"e5e277db-7467-4a4f-9715-6854e235cac1\") " pod="openstack/glance-db-create-pnmqh" Nov 25 11:01:13 crc kubenswrapper[4776]: I1125 11:01:13.323021 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-f6ca-account-create-qsbv2"] Nov 25 11:01:13 crc kubenswrapper[4776]: I1125 11:01:13.324561 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f6ca-account-create-qsbv2" Nov 25 11:01:13 crc kubenswrapper[4776]: I1125 11:01:13.326954 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 25 11:01:13 crc kubenswrapper[4776]: I1125 11:01:13.337155 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-f6ca-account-create-qsbv2"] Nov 25 11:01:13 crc kubenswrapper[4776]: I1125 11:01:13.405476 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6984d\" (UniqueName: \"kubernetes.io/projected/e356929a-2335-4a51-b410-39521ac9f4d1-kube-api-access-6984d\") pod \"glance-f6ca-account-create-qsbv2\" (UID: \"e356929a-2335-4a51-b410-39521ac9f4d1\") " pod="openstack/glance-f6ca-account-create-qsbv2" Nov 25 11:01:13 crc kubenswrapper[4776]: I1125 11:01:13.405592 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5t45x\" (UniqueName: \"kubernetes.io/projected/e5e277db-7467-4a4f-9715-6854e235cac1-kube-api-access-5t45x\") pod \"glance-db-create-pnmqh\" (UID: \"e5e277db-7467-4a4f-9715-6854e235cac1\") " pod="openstack/glance-db-create-pnmqh" Nov 25 11:01:13 crc kubenswrapper[4776]: I1125 11:01:13.405636 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e356929a-2335-4a51-b410-39521ac9f4d1-operator-scripts\") pod \"glance-f6ca-account-create-qsbv2\" (UID: \"e356929a-2335-4a51-b410-39521ac9f4d1\") " pod="openstack/glance-f6ca-account-create-qsbv2" Nov 25 11:01:13 crc kubenswrapper[4776]: I1125 11:01:13.405709 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5e277db-7467-4a4f-9715-6854e235cac1-operator-scripts\") pod \"glance-db-create-pnmqh\" (UID: \"e5e277db-7467-4a4f-9715-6854e235cac1\") " pod="openstack/glance-db-create-pnmqh" Nov 25 11:01:13 crc kubenswrapper[4776]: I1125 11:01:13.406529 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5e277db-7467-4a4f-9715-6854e235cac1-operator-scripts\") pod \"glance-db-create-pnmqh\" (UID: \"e5e277db-7467-4a4f-9715-6854e235cac1\") " pod="openstack/glance-db-create-pnmqh" Nov 25 11:01:13 crc kubenswrapper[4776]: I1125 11:01:13.423782 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5t45x\" (UniqueName: \"kubernetes.io/projected/e5e277db-7467-4a4f-9715-6854e235cac1-kube-api-access-5t45x\") pod \"glance-db-create-pnmqh\" (UID: \"e5e277db-7467-4a4f-9715-6854e235cac1\") " pod="openstack/glance-db-create-pnmqh" Nov 25 11:01:13 crc kubenswrapper[4776]: I1125 11:01:13.507034 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6984d\" (UniqueName: \"kubernetes.io/projected/e356929a-2335-4a51-b410-39521ac9f4d1-kube-api-access-6984d\") pod \"glance-f6ca-account-create-qsbv2\" (UID: \"e356929a-2335-4a51-b410-39521ac9f4d1\") " pod="openstack/glance-f6ca-account-create-qsbv2" Nov 25 11:01:13 crc kubenswrapper[4776]: I1125 11:01:13.507139 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e356929a-2335-4a51-b410-39521ac9f4d1-operator-scripts\") pod \"glance-f6ca-account-create-qsbv2\" (UID: \"e356929a-2335-4a51-b410-39521ac9f4d1\") " pod="openstack/glance-f6ca-account-create-qsbv2" Nov 25 11:01:13 crc kubenswrapper[4776]: I1125 11:01:13.507958 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e356929a-2335-4a51-b410-39521ac9f4d1-operator-scripts\") pod \"glance-f6ca-account-create-qsbv2\" (UID: \"e356929a-2335-4a51-b410-39521ac9f4d1\") " pod="openstack/glance-f6ca-account-create-qsbv2" Nov 25 11:01:13 crc kubenswrapper[4776]: I1125 11:01:13.527802 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6984d\" (UniqueName: \"kubernetes.io/projected/e356929a-2335-4a51-b410-39521ac9f4d1-kube-api-access-6984d\") pod \"glance-f6ca-account-create-qsbv2\" (UID: \"e356929a-2335-4a51-b410-39521ac9f4d1\") " pod="openstack/glance-f6ca-account-create-qsbv2" Nov 25 11:01:13 crc kubenswrapper[4776]: I1125 11:01:13.539808 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pnmqh" Nov 25 11:01:13 crc kubenswrapper[4776]: I1125 11:01:13.644383 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f6ca-account-create-qsbv2" Nov 25 11:01:13 crc kubenswrapper[4776]: I1125 11:01:13.987593 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-pnmqh"] Nov 25 11:01:13 crc kubenswrapper[4776]: W1125 11:01:13.992202 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5e277db_7467_4a4f_9715_6854e235cac1.slice/crio-c39721c970d7bc187bb563690da66396dd47518288feda72501e4d72d05dafdd WatchSource:0}: Error finding container c39721c970d7bc187bb563690da66396dd47518288feda72501e4d72d05dafdd: Status 404 returned error can't find the container with id c39721c970d7bc187bb563690da66396dd47518288feda72501e4d72d05dafdd Nov 25 11:01:14 crc kubenswrapper[4776]: I1125 11:01:14.098182 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-f6ca-account-create-qsbv2"] Nov 25 11:01:14 crc kubenswrapper[4776]: W1125 11:01:14.103952 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode356929a_2335_4a51_b410_39521ac9f4d1.slice/crio-646868b00886d08b7858da96faef67bcc17725c93275a02b03ce001793828e44 WatchSource:0}: Error finding container 646868b00886d08b7858da96faef67bcc17725c93275a02b03ce001793828e44: Status 404 returned error can't find the container with id 646868b00886d08b7858da96faef67bcc17725c93275a02b03ce001793828e44 Nov 25 11:01:14 crc kubenswrapper[4776]: I1125 11:01:14.814050 4776 generic.go:334] "Generic (PLEG): container finished" podID="e356929a-2335-4a51-b410-39521ac9f4d1" containerID="fd5d2241a662f5b682bc3191a53df6f5849ca0002fef8d45ee81d122a9b632d3" exitCode=0 Nov 25 11:01:14 crc kubenswrapper[4776]: I1125 11:01:14.814169 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-f6ca-account-create-qsbv2" event={"ID":"e356929a-2335-4a51-b410-39521ac9f4d1","Type":"ContainerDied","Data":"fd5d2241a662f5b682bc3191a53df6f5849ca0002fef8d45ee81d122a9b632d3"} Nov 25 11:01:14 crc kubenswrapper[4776]: I1125 11:01:14.814217 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-f6ca-account-create-qsbv2" event={"ID":"e356929a-2335-4a51-b410-39521ac9f4d1","Type":"ContainerStarted","Data":"646868b00886d08b7858da96faef67bcc17725c93275a02b03ce001793828e44"} Nov 25 11:01:14 crc kubenswrapper[4776]: I1125 11:01:14.816163 4776 generic.go:334] "Generic (PLEG): container finished" podID="e5e277db-7467-4a4f-9715-6854e235cac1" containerID="9cdb2b9569c3e16ed00198fe74fc270e2aedadd09e725000eb60c8f716c2afcc" exitCode=0 Nov 25 11:01:14 crc kubenswrapper[4776]: I1125 11:01:14.816201 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-pnmqh" event={"ID":"e5e277db-7467-4a4f-9715-6854e235cac1","Type":"ContainerDied","Data":"9cdb2b9569c3e16ed00198fe74fc270e2aedadd09e725000eb60c8f716c2afcc"} Nov 25 11:01:14 crc kubenswrapper[4776]: I1125 11:01:14.816221 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-pnmqh" event={"ID":"e5e277db-7467-4a4f-9715-6854e235cac1","Type":"ContainerStarted","Data":"c39721c970d7bc187bb563690da66396dd47518288feda72501e4d72d05dafdd"} Nov 25 11:01:16 crc kubenswrapper[4776]: I1125 11:01:16.169752 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pnmqh" Nov 25 11:01:16 crc kubenswrapper[4776]: I1125 11:01:16.175993 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f6ca-account-create-qsbv2" Nov 25 11:01:16 crc kubenswrapper[4776]: I1125 11:01:16.261470 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6984d\" (UniqueName: \"kubernetes.io/projected/e356929a-2335-4a51-b410-39521ac9f4d1-kube-api-access-6984d\") pod \"e356929a-2335-4a51-b410-39521ac9f4d1\" (UID: \"e356929a-2335-4a51-b410-39521ac9f4d1\") " Nov 25 11:01:16 crc kubenswrapper[4776]: I1125 11:01:16.261766 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5t45x\" (UniqueName: \"kubernetes.io/projected/e5e277db-7467-4a4f-9715-6854e235cac1-kube-api-access-5t45x\") pod \"e5e277db-7467-4a4f-9715-6854e235cac1\" (UID: \"e5e277db-7467-4a4f-9715-6854e235cac1\") " Nov 25 11:01:16 crc kubenswrapper[4776]: I1125 11:01:16.261983 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e356929a-2335-4a51-b410-39521ac9f4d1-operator-scripts\") pod \"e356929a-2335-4a51-b410-39521ac9f4d1\" (UID: \"e356929a-2335-4a51-b410-39521ac9f4d1\") " Nov 25 11:01:16 crc kubenswrapper[4776]: I1125 11:01:16.262141 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5e277db-7467-4a4f-9715-6854e235cac1-operator-scripts\") pod \"e5e277db-7467-4a4f-9715-6854e235cac1\" (UID: \"e5e277db-7467-4a4f-9715-6854e235cac1\") " Nov 25 11:01:16 crc kubenswrapper[4776]: I1125 11:01:16.262774 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e356929a-2335-4a51-b410-39521ac9f4d1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e356929a-2335-4a51-b410-39521ac9f4d1" (UID: "e356929a-2335-4a51-b410-39521ac9f4d1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:01:16 crc kubenswrapper[4776]: I1125 11:01:16.262799 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5e277db-7467-4a4f-9715-6854e235cac1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e5e277db-7467-4a4f-9715-6854e235cac1" (UID: "e5e277db-7467-4a4f-9715-6854e235cac1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:01:16 crc kubenswrapper[4776]: I1125 11:01:16.263039 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e356929a-2335-4a51-b410-39521ac9f4d1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:16 crc kubenswrapper[4776]: I1125 11:01:16.263133 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5e277db-7467-4a4f-9715-6854e235cac1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:16 crc kubenswrapper[4776]: I1125 11:01:16.266624 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e356929a-2335-4a51-b410-39521ac9f4d1-kube-api-access-6984d" (OuterVolumeSpecName: "kube-api-access-6984d") pod "e356929a-2335-4a51-b410-39521ac9f4d1" (UID: "e356929a-2335-4a51-b410-39521ac9f4d1"). InnerVolumeSpecName "kube-api-access-6984d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:01:16 crc kubenswrapper[4776]: I1125 11:01:16.266704 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5e277db-7467-4a4f-9715-6854e235cac1-kube-api-access-5t45x" (OuterVolumeSpecName: "kube-api-access-5t45x") pod "e5e277db-7467-4a4f-9715-6854e235cac1" (UID: "e5e277db-7467-4a4f-9715-6854e235cac1"). InnerVolumeSpecName "kube-api-access-5t45x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:01:16 crc kubenswrapper[4776]: I1125 11:01:16.365197 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6984d\" (UniqueName: \"kubernetes.io/projected/e356929a-2335-4a51-b410-39521ac9f4d1-kube-api-access-6984d\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:16 crc kubenswrapper[4776]: I1125 11:01:16.365484 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5t45x\" (UniqueName: \"kubernetes.io/projected/e5e277db-7467-4a4f-9715-6854e235cac1-kube-api-access-5t45x\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:16 crc kubenswrapper[4776]: I1125 11:01:16.838579 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-pnmqh" event={"ID":"e5e277db-7467-4a4f-9715-6854e235cac1","Type":"ContainerDied","Data":"c39721c970d7bc187bb563690da66396dd47518288feda72501e4d72d05dafdd"} Nov 25 11:01:16 crc kubenswrapper[4776]: I1125 11:01:16.838617 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c39721c970d7bc187bb563690da66396dd47518288feda72501e4d72d05dafdd" Nov 25 11:01:16 crc kubenswrapper[4776]: I1125 11:01:16.838621 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pnmqh" Nov 25 11:01:16 crc kubenswrapper[4776]: I1125 11:01:16.840313 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-f6ca-account-create-qsbv2" event={"ID":"e356929a-2335-4a51-b410-39521ac9f4d1","Type":"ContainerDied","Data":"646868b00886d08b7858da96faef67bcc17725c93275a02b03ce001793828e44"} Nov 25 11:01:16 crc kubenswrapper[4776]: I1125 11:01:16.840336 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="646868b00886d08b7858da96faef67bcc17725c93275a02b03ce001793828e44" Nov 25 11:01:16 crc kubenswrapper[4776]: I1125 11:01:16.840369 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f6ca-account-create-qsbv2" Nov 25 11:01:18 crc kubenswrapper[4776]: I1125 11:01:18.482640 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-kz9jg"] Nov 25 11:01:18 crc kubenswrapper[4776]: E1125 11:01:18.483410 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5e277db-7467-4a4f-9715-6854e235cac1" containerName="mariadb-database-create" Nov 25 11:01:18 crc kubenswrapper[4776]: I1125 11:01:18.483426 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5e277db-7467-4a4f-9715-6854e235cac1" containerName="mariadb-database-create" Nov 25 11:01:18 crc kubenswrapper[4776]: E1125 11:01:18.483466 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e356929a-2335-4a51-b410-39521ac9f4d1" containerName="mariadb-account-create" Nov 25 11:01:18 crc kubenswrapper[4776]: I1125 11:01:18.483475 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="e356929a-2335-4a51-b410-39521ac9f4d1" containerName="mariadb-account-create" Nov 25 11:01:18 crc kubenswrapper[4776]: I1125 11:01:18.483678 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5e277db-7467-4a4f-9715-6854e235cac1" containerName="mariadb-database-create" Nov 25 11:01:18 crc kubenswrapper[4776]: I1125 11:01:18.483694 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="e356929a-2335-4a51-b410-39521ac9f4d1" containerName="mariadb-account-create" Nov 25 11:01:18 crc kubenswrapper[4776]: I1125 11:01:18.484443 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-kz9jg" Nov 25 11:01:18 crc kubenswrapper[4776]: I1125 11:01:18.486777 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 25 11:01:18 crc kubenswrapper[4776]: I1125 11:01:18.487047 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-xlswr" Nov 25 11:01:18 crc kubenswrapper[4776]: I1125 11:01:18.496428 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-kz9jg"] Nov 25 11:01:18 crc kubenswrapper[4776]: I1125 11:01:18.605597 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72g99\" (UniqueName: \"kubernetes.io/projected/fcb0fc87-e3b2-475c-a203-aeda2c4e21da-kube-api-access-72g99\") pod \"glance-db-sync-kz9jg\" (UID: \"fcb0fc87-e3b2-475c-a203-aeda2c4e21da\") " pod="openstack/glance-db-sync-kz9jg" Nov 25 11:01:18 crc kubenswrapper[4776]: I1125 11:01:18.605920 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcb0fc87-e3b2-475c-a203-aeda2c4e21da-combined-ca-bundle\") pod \"glance-db-sync-kz9jg\" (UID: \"fcb0fc87-e3b2-475c-a203-aeda2c4e21da\") " pod="openstack/glance-db-sync-kz9jg" Nov 25 11:01:18 crc kubenswrapper[4776]: I1125 11:01:18.606048 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fcb0fc87-e3b2-475c-a203-aeda2c4e21da-db-sync-config-data\") pod \"glance-db-sync-kz9jg\" (UID: \"fcb0fc87-e3b2-475c-a203-aeda2c4e21da\") " pod="openstack/glance-db-sync-kz9jg" Nov 25 11:01:18 crc kubenswrapper[4776]: I1125 11:01:18.606116 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcb0fc87-e3b2-475c-a203-aeda2c4e21da-config-data\") pod \"glance-db-sync-kz9jg\" (UID: \"fcb0fc87-e3b2-475c-a203-aeda2c4e21da\") " pod="openstack/glance-db-sync-kz9jg" Nov 25 11:01:18 crc kubenswrapper[4776]: I1125 11:01:18.707677 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72g99\" (UniqueName: \"kubernetes.io/projected/fcb0fc87-e3b2-475c-a203-aeda2c4e21da-kube-api-access-72g99\") pod \"glance-db-sync-kz9jg\" (UID: \"fcb0fc87-e3b2-475c-a203-aeda2c4e21da\") " pod="openstack/glance-db-sync-kz9jg" Nov 25 11:01:18 crc kubenswrapper[4776]: I1125 11:01:18.707800 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcb0fc87-e3b2-475c-a203-aeda2c4e21da-combined-ca-bundle\") pod \"glance-db-sync-kz9jg\" (UID: \"fcb0fc87-e3b2-475c-a203-aeda2c4e21da\") " pod="openstack/glance-db-sync-kz9jg" Nov 25 11:01:18 crc kubenswrapper[4776]: I1125 11:01:18.707846 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fcb0fc87-e3b2-475c-a203-aeda2c4e21da-db-sync-config-data\") pod \"glance-db-sync-kz9jg\" (UID: \"fcb0fc87-e3b2-475c-a203-aeda2c4e21da\") " pod="openstack/glance-db-sync-kz9jg" Nov 25 11:01:18 crc kubenswrapper[4776]: I1125 11:01:18.707868 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcb0fc87-e3b2-475c-a203-aeda2c4e21da-config-data\") pod \"glance-db-sync-kz9jg\" (UID: \"fcb0fc87-e3b2-475c-a203-aeda2c4e21da\") " pod="openstack/glance-db-sync-kz9jg" Nov 25 11:01:18 crc kubenswrapper[4776]: I1125 11:01:18.714346 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcb0fc87-e3b2-475c-a203-aeda2c4e21da-combined-ca-bundle\") pod \"glance-db-sync-kz9jg\" (UID: \"fcb0fc87-e3b2-475c-a203-aeda2c4e21da\") " pod="openstack/glance-db-sync-kz9jg" Nov 25 11:01:18 crc kubenswrapper[4776]: I1125 11:01:18.714469 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcb0fc87-e3b2-475c-a203-aeda2c4e21da-config-data\") pod \"glance-db-sync-kz9jg\" (UID: \"fcb0fc87-e3b2-475c-a203-aeda2c4e21da\") " pod="openstack/glance-db-sync-kz9jg" Nov 25 11:01:18 crc kubenswrapper[4776]: I1125 11:01:18.714664 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fcb0fc87-e3b2-475c-a203-aeda2c4e21da-db-sync-config-data\") pod \"glance-db-sync-kz9jg\" (UID: \"fcb0fc87-e3b2-475c-a203-aeda2c4e21da\") " pod="openstack/glance-db-sync-kz9jg" Nov 25 11:01:18 crc kubenswrapper[4776]: I1125 11:01:18.724398 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72g99\" (UniqueName: \"kubernetes.io/projected/fcb0fc87-e3b2-475c-a203-aeda2c4e21da-kube-api-access-72g99\") pod \"glance-db-sync-kz9jg\" (UID: \"fcb0fc87-e3b2-475c-a203-aeda2c4e21da\") " pod="openstack/glance-db-sync-kz9jg" Nov 25 11:01:18 crc kubenswrapper[4776]: I1125 11:01:18.802821 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-kz9jg" Nov 25 11:01:19 crc kubenswrapper[4776]: I1125 11:01:19.414442 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-kz9jg"] Nov 25 11:01:19 crc kubenswrapper[4776]: W1125 11:01:19.415696 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfcb0fc87_e3b2_475c_a203_aeda2c4e21da.slice/crio-0571b2f115dd7b836b187b29895fdd9a99c32b941976dc8bff565204b90c46b3 WatchSource:0}: Error finding container 0571b2f115dd7b836b187b29895fdd9a99c32b941976dc8bff565204b90c46b3: Status 404 returned error can't find the container with id 0571b2f115dd7b836b187b29895fdd9a99c32b941976dc8bff565204b90c46b3 Nov 25 11:01:19 crc kubenswrapper[4776]: I1125 11:01:19.866774 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-kz9jg" event={"ID":"fcb0fc87-e3b2-475c-a203-aeda2c4e21da","Type":"ContainerStarted","Data":"0571b2f115dd7b836b187b29895fdd9a99c32b941976dc8bff565204b90c46b3"} Nov 25 11:01:20 crc kubenswrapper[4776]: I1125 11:01:20.878217 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-kz9jg" event={"ID":"fcb0fc87-e3b2-475c-a203-aeda2c4e21da","Type":"ContainerStarted","Data":"ac78b904326fa8bea78f7c60190b901320665430e39ce84ba4d6a49f0cbc8473"} Nov 25 11:01:20 crc kubenswrapper[4776]: I1125 11:01:20.905267 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-kz9jg" podStartSLOduration=2.905243274 podStartE2EDuration="2.905243274s" podCreationTimestamp="2025-11-25 11:01:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:01:20.894036802 +0000 UTC m=+5825.935096355" watchObservedRunningTime="2025-11-25 11:01:20.905243274 +0000 UTC m=+5825.946302827" Nov 25 11:01:23 crc kubenswrapper[4776]: I1125 11:01:23.904659 4776 generic.go:334] "Generic (PLEG): container finished" podID="fcb0fc87-e3b2-475c-a203-aeda2c4e21da" containerID="ac78b904326fa8bea78f7c60190b901320665430e39ce84ba4d6a49f0cbc8473" exitCode=0 Nov 25 11:01:23 crc kubenswrapper[4776]: I1125 11:01:23.904759 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-kz9jg" event={"ID":"fcb0fc87-e3b2-475c-a203-aeda2c4e21da","Type":"ContainerDied","Data":"ac78b904326fa8bea78f7c60190b901320665430e39ce84ba4d6a49f0cbc8473"} Nov 25 11:01:25 crc kubenswrapper[4776]: I1125 11:01:25.287280 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-kz9jg" Nov 25 11:01:25 crc kubenswrapper[4776]: I1125 11:01:25.425372 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72g99\" (UniqueName: \"kubernetes.io/projected/fcb0fc87-e3b2-475c-a203-aeda2c4e21da-kube-api-access-72g99\") pod \"fcb0fc87-e3b2-475c-a203-aeda2c4e21da\" (UID: \"fcb0fc87-e3b2-475c-a203-aeda2c4e21da\") " Nov 25 11:01:25 crc kubenswrapper[4776]: I1125 11:01:25.425493 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcb0fc87-e3b2-475c-a203-aeda2c4e21da-combined-ca-bundle\") pod \"fcb0fc87-e3b2-475c-a203-aeda2c4e21da\" (UID: \"fcb0fc87-e3b2-475c-a203-aeda2c4e21da\") " Nov 25 11:01:25 crc kubenswrapper[4776]: I1125 11:01:25.425591 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcb0fc87-e3b2-475c-a203-aeda2c4e21da-config-data\") pod \"fcb0fc87-e3b2-475c-a203-aeda2c4e21da\" (UID: \"fcb0fc87-e3b2-475c-a203-aeda2c4e21da\") " Nov 25 11:01:25 crc kubenswrapper[4776]: I1125 11:01:25.425738 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fcb0fc87-e3b2-475c-a203-aeda2c4e21da-db-sync-config-data\") pod \"fcb0fc87-e3b2-475c-a203-aeda2c4e21da\" (UID: \"fcb0fc87-e3b2-475c-a203-aeda2c4e21da\") " Nov 25 11:01:25 crc kubenswrapper[4776]: I1125 11:01:25.430466 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcb0fc87-e3b2-475c-a203-aeda2c4e21da-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "fcb0fc87-e3b2-475c-a203-aeda2c4e21da" (UID: "fcb0fc87-e3b2-475c-a203-aeda2c4e21da"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:01:25 crc kubenswrapper[4776]: I1125 11:01:25.430588 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcb0fc87-e3b2-475c-a203-aeda2c4e21da-kube-api-access-72g99" (OuterVolumeSpecName: "kube-api-access-72g99") pod "fcb0fc87-e3b2-475c-a203-aeda2c4e21da" (UID: "fcb0fc87-e3b2-475c-a203-aeda2c4e21da"). InnerVolumeSpecName "kube-api-access-72g99". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:01:25 crc kubenswrapper[4776]: I1125 11:01:25.454971 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcb0fc87-e3b2-475c-a203-aeda2c4e21da-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fcb0fc87-e3b2-475c-a203-aeda2c4e21da" (UID: "fcb0fc87-e3b2-475c-a203-aeda2c4e21da"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:01:25 crc kubenswrapper[4776]: I1125 11:01:25.468553 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcb0fc87-e3b2-475c-a203-aeda2c4e21da-config-data" (OuterVolumeSpecName: "config-data") pod "fcb0fc87-e3b2-475c-a203-aeda2c4e21da" (UID: "fcb0fc87-e3b2-475c-a203-aeda2c4e21da"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:01:25 crc kubenswrapper[4776]: I1125 11:01:25.527458 4776 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fcb0fc87-e3b2-475c-a203-aeda2c4e21da-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:25 crc kubenswrapper[4776]: I1125 11:01:25.527816 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72g99\" (UniqueName: \"kubernetes.io/projected/fcb0fc87-e3b2-475c-a203-aeda2c4e21da-kube-api-access-72g99\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:25 crc kubenswrapper[4776]: I1125 11:01:25.527831 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcb0fc87-e3b2-475c-a203-aeda2c4e21da-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:25 crc kubenswrapper[4776]: I1125 11:01:25.527843 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcb0fc87-e3b2-475c-a203-aeda2c4e21da-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:25 crc kubenswrapper[4776]: I1125 11:01:25.922534 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-kz9jg" event={"ID":"fcb0fc87-e3b2-475c-a203-aeda2c4e21da","Type":"ContainerDied","Data":"0571b2f115dd7b836b187b29895fdd9a99c32b941976dc8bff565204b90c46b3"} Nov 25 11:01:25 crc kubenswrapper[4776]: I1125 11:01:25.922580 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0571b2f115dd7b836b187b29895fdd9a99c32b941976dc8bff565204b90c46b3" Nov 25 11:01:25 crc kubenswrapper[4776]: I1125 11:01:25.922648 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-kz9jg" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.194465 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 11:01:26 crc kubenswrapper[4776]: E1125 11:01:26.195142 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcb0fc87-e3b2-475c-a203-aeda2c4e21da" containerName="glance-db-sync" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.195161 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcb0fc87-e3b2-475c-a203-aeda2c4e21da" containerName="glance-db-sync" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.195317 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcb0fc87-e3b2-475c-a203-aeda2c4e21da" containerName="glance-db-sync" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.196564 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.200116 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-xlswr" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.200365 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.200661 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.211991 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.251108 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.251189 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.251259 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-scripts\") pod \"glance-default-external-api-0\" (UID: \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.251299 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-logs\") pod \"glance-default-external-api-0\" (UID: \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.251452 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9p85\" (UniqueName: \"kubernetes.io/projected/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-kube-api-access-v9p85\") pod \"glance-default-external-api-0\" (UID: \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.251563 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-config-data\") pod \"glance-default-external-api-0\" (UID: \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.311410 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d8cfd89b5-6pvwz"] Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.312821 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d8cfd89b5-6pvwz" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.336878 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d8cfd89b5-6pvwz"] Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.352714 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-scripts\") pod \"glance-default-external-api-0\" (UID: \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.352793 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-logs\") pod \"glance-default-external-api-0\" (UID: \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.352859 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9p85\" (UniqueName: \"kubernetes.io/projected/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-kube-api-access-v9p85\") pod \"glance-default-external-api-0\" (UID: \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.352925 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-config-data\") pod \"glance-default-external-api-0\" (UID: \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.352947 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.352998 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.357207 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-logs\") pod \"glance-default-external-api-0\" (UID: \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.357728 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.361214 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.367518 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-scripts\") pod \"glance-default-external-api-0\" (UID: \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.370244 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-config-data\") pod \"glance-default-external-api-0\" (UID: \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.390307 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9p85\" (UniqueName: \"kubernetes.io/projected/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-kube-api-access-v9p85\") pod \"glance-default-external-api-0\" (UID: \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.454160 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mgqp\" (UniqueName: \"kubernetes.io/projected/147b6859-7722-443e-b424-f8a61aee320d-kube-api-access-5mgqp\") pod \"dnsmasq-dns-6d8cfd89b5-6pvwz\" (UID: \"147b6859-7722-443e-b424-f8a61aee320d\") " pod="openstack/dnsmasq-dns-6d8cfd89b5-6pvwz" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.454422 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/147b6859-7722-443e-b424-f8a61aee320d-ovsdbserver-sb\") pod \"dnsmasq-dns-6d8cfd89b5-6pvwz\" (UID: \"147b6859-7722-443e-b424-f8a61aee320d\") " pod="openstack/dnsmasq-dns-6d8cfd89b5-6pvwz" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.454584 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/147b6859-7722-443e-b424-f8a61aee320d-dns-svc\") pod \"dnsmasq-dns-6d8cfd89b5-6pvwz\" (UID: \"147b6859-7722-443e-b424-f8a61aee320d\") " pod="openstack/dnsmasq-dns-6d8cfd89b5-6pvwz" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.454661 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/147b6859-7722-443e-b424-f8a61aee320d-config\") pod \"dnsmasq-dns-6d8cfd89b5-6pvwz\" (UID: \"147b6859-7722-443e-b424-f8a61aee320d\") " pod="openstack/dnsmasq-dns-6d8cfd89b5-6pvwz" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.454772 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/147b6859-7722-443e-b424-f8a61aee320d-ovsdbserver-nb\") pod \"dnsmasq-dns-6d8cfd89b5-6pvwz\" (UID: \"147b6859-7722-443e-b424-f8a61aee320d\") " pod="openstack/dnsmasq-dns-6d8cfd89b5-6pvwz" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.480125 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.486973 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.494467 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.498050 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.536025 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.557154 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/147b6859-7722-443e-b424-f8a61aee320d-ovsdbserver-sb\") pod \"dnsmasq-dns-6d8cfd89b5-6pvwz\" (UID: \"147b6859-7722-443e-b424-f8a61aee320d\") " pod="openstack/dnsmasq-dns-6d8cfd89b5-6pvwz" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.557265 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/147b6859-7722-443e-b424-f8a61aee320d-dns-svc\") pod \"dnsmasq-dns-6d8cfd89b5-6pvwz\" (UID: \"147b6859-7722-443e-b424-f8a61aee320d\") " pod="openstack/dnsmasq-dns-6d8cfd89b5-6pvwz" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.557291 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/147b6859-7722-443e-b424-f8a61aee320d-config\") pod \"dnsmasq-dns-6d8cfd89b5-6pvwz\" (UID: \"147b6859-7722-443e-b424-f8a61aee320d\") " pod="openstack/dnsmasq-dns-6d8cfd89b5-6pvwz" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.557353 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/147b6859-7722-443e-b424-f8a61aee320d-ovsdbserver-nb\") pod \"dnsmasq-dns-6d8cfd89b5-6pvwz\" (UID: \"147b6859-7722-443e-b424-f8a61aee320d\") " pod="openstack/dnsmasq-dns-6d8cfd89b5-6pvwz" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.557384 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mgqp\" (UniqueName: \"kubernetes.io/projected/147b6859-7722-443e-b424-f8a61aee320d-kube-api-access-5mgqp\") pod \"dnsmasq-dns-6d8cfd89b5-6pvwz\" (UID: \"147b6859-7722-443e-b424-f8a61aee320d\") " pod="openstack/dnsmasq-dns-6d8cfd89b5-6pvwz" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.558588 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/147b6859-7722-443e-b424-f8a61aee320d-ovsdbserver-sb\") pod \"dnsmasq-dns-6d8cfd89b5-6pvwz\" (UID: \"147b6859-7722-443e-b424-f8a61aee320d\") " pod="openstack/dnsmasq-dns-6d8cfd89b5-6pvwz" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.559736 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/147b6859-7722-443e-b424-f8a61aee320d-ovsdbserver-nb\") pod \"dnsmasq-dns-6d8cfd89b5-6pvwz\" (UID: \"147b6859-7722-443e-b424-f8a61aee320d\") " pod="openstack/dnsmasq-dns-6d8cfd89b5-6pvwz" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.559820 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/147b6859-7722-443e-b424-f8a61aee320d-config\") pod \"dnsmasq-dns-6d8cfd89b5-6pvwz\" (UID: \"147b6859-7722-443e-b424-f8a61aee320d\") " pod="openstack/dnsmasq-dns-6d8cfd89b5-6pvwz" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.560005 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/147b6859-7722-443e-b424-f8a61aee320d-dns-svc\") pod \"dnsmasq-dns-6d8cfd89b5-6pvwz\" (UID: \"147b6859-7722-443e-b424-f8a61aee320d\") " pod="openstack/dnsmasq-dns-6d8cfd89b5-6pvwz" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.576627 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mgqp\" (UniqueName: \"kubernetes.io/projected/147b6859-7722-443e-b424-f8a61aee320d-kube-api-access-5mgqp\") pod \"dnsmasq-dns-6d8cfd89b5-6pvwz\" (UID: \"147b6859-7722-443e-b424-f8a61aee320d\") " pod="openstack/dnsmasq-dns-6d8cfd89b5-6pvwz" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.644756 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d8cfd89b5-6pvwz" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.658782 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.658832 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-logs\") pod \"glance-default-internal-api-0\" (UID: \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.658873 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.658931 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwwkv\" (UniqueName: \"kubernetes.io/projected/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-kube-api-access-wwwkv\") pod \"glance-default-internal-api-0\" (UID: \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.658961 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.659046 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.760507 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.760673 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.760707 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.760727 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-logs\") pod \"glance-default-internal-api-0\" (UID: \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.760773 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.760839 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwwkv\" (UniqueName: \"kubernetes.io/projected/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-kube-api-access-wwwkv\") pod \"glance-default-internal-api-0\" (UID: \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.761279 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.761770 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-logs\") pod \"glance-default-internal-api-0\" (UID: \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.767142 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.767793 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.769391 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.779617 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwwkv\" (UniqueName: \"kubernetes.io/projected/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-kube-api-access-wwwkv\") pod \"glance-default-internal-api-0\" (UID: \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:26 crc kubenswrapper[4776]: I1125 11:01:26.818540 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 11:01:27 crc kubenswrapper[4776]: I1125 11:01:27.152302 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 11:01:27 crc kubenswrapper[4776]: W1125 11:01:27.152733 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda78be2da_2747_4e3f_8cbd_7ee6bf8e8dbc.slice/crio-9a16bf03a1316a32ba0247158a664630fa0830db8cdd5bca91cde21b179ceafe WatchSource:0}: Error finding container 9a16bf03a1316a32ba0247158a664630fa0830db8cdd5bca91cde21b179ceafe: Status 404 returned error can't find the container with id 9a16bf03a1316a32ba0247158a664630fa0830db8cdd5bca91cde21b179ceafe Nov 25 11:01:27 crc kubenswrapper[4776]: I1125 11:01:27.254500 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d8cfd89b5-6pvwz"] Nov 25 11:01:27 crc kubenswrapper[4776]: W1125 11:01:27.260726 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod147b6859_7722_443e_b424_f8a61aee320d.slice/crio-0cee2b1f82b705f156e44c3e33fc3a4aabb4bbe0f1919ccce8cff31dcbda07d9 WatchSource:0}: Error finding container 0cee2b1f82b705f156e44c3e33fc3a4aabb4bbe0f1919ccce8cff31dcbda07d9: Status 404 returned error can't find the container with id 0cee2b1f82b705f156e44c3e33fc3a4aabb4bbe0f1919ccce8cff31dcbda07d9 Nov 25 11:01:27 crc kubenswrapper[4776]: I1125 11:01:27.422137 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 11:01:27 crc kubenswrapper[4776]: I1125 11:01:27.504950 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 11:01:27 crc kubenswrapper[4776]: W1125 11:01:27.544306 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod345d87f1_ea72_4bcb_b5d8_7d5921f0aa5b.slice/crio-6508a80db6832d76f32c0ddf45ae8eec73cc624ecda06cb3740a63f3ddb9ee85 WatchSource:0}: Error finding container 6508a80db6832d76f32c0ddf45ae8eec73cc624ecda06cb3740a63f3ddb9ee85: Status 404 returned error can't find the container with id 6508a80db6832d76f32c0ddf45ae8eec73cc624ecda06cb3740a63f3ddb9ee85 Nov 25 11:01:27 crc kubenswrapper[4776]: I1125 11:01:27.953975 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc","Type":"ContainerStarted","Data":"62f99b47441ff17760618b5b803fae0efd2aebefed3d736a761f18426dd1d2e3"} Nov 25 11:01:27 crc kubenswrapper[4776]: I1125 11:01:27.954365 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc","Type":"ContainerStarted","Data":"9a16bf03a1316a32ba0247158a664630fa0830db8cdd5bca91cde21b179ceafe"} Nov 25 11:01:27 crc kubenswrapper[4776]: I1125 11:01:27.956674 4776 generic.go:334] "Generic (PLEG): container finished" podID="147b6859-7722-443e-b424-f8a61aee320d" containerID="d483500b8db3124e16840bcd7d8ccc12156fa86f6634181067004cbea2149103" exitCode=0 Nov 25 11:01:27 crc kubenswrapper[4776]: I1125 11:01:27.956719 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d8cfd89b5-6pvwz" event={"ID":"147b6859-7722-443e-b424-f8a61aee320d","Type":"ContainerDied","Data":"d483500b8db3124e16840bcd7d8ccc12156fa86f6634181067004cbea2149103"} Nov 25 11:01:27 crc kubenswrapper[4776]: I1125 11:01:27.956735 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d8cfd89b5-6pvwz" event={"ID":"147b6859-7722-443e-b424-f8a61aee320d","Type":"ContainerStarted","Data":"0cee2b1f82b705f156e44c3e33fc3a4aabb4bbe0f1919ccce8cff31dcbda07d9"} Nov 25 11:01:27 crc kubenswrapper[4776]: I1125 11:01:27.982807 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b","Type":"ContainerStarted","Data":"6508a80db6832d76f32c0ddf45ae8eec73cc624ecda06cb3740a63f3ddb9ee85"} Nov 25 11:01:28 crc kubenswrapper[4776]: I1125 11:01:28.696942 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 11:01:28 crc kubenswrapper[4776]: I1125 11:01:28.999419 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b","Type":"ContainerStarted","Data":"b719f697fb6d47a1467b1bfb291bf90f7f2ac9e737dba0b1c3f2f5fd4adafca4"} Nov 25 11:01:28 crc kubenswrapper[4776]: I1125 11:01:28.999795 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b","Type":"ContainerStarted","Data":"a72854ecf751815c10fecd2592725bb62e38c7e07a6f3bd8c0f7e17102cf8fd5"} Nov 25 11:01:28 crc kubenswrapper[4776]: I1125 11:01:28.999603 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b" containerName="glance-httpd" containerID="cri-o://b719f697fb6d47a1467b1bfb291bf90f7f2ac9e737dba0b1c3f2f5fd4adafca4" gracePeriod=30 Nov 25 11:01:28 crc kubenswrapper[4776]: I1125 11:01:28.999532 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b" containerName="glance-log" containerID="cri-o://a72854ecf751815c10fecd2592725bb62e38c7e07a6f3bd8c0f7e17102cf8fd5" gracePeriod=30 Nov 25 11:01:29 crc kubenswrapper[4776]: I1125 11:01:29.007272 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc","Type":"ContainerStarted","Data":"f5fd94e6048195b37b6c9f09840f447cf43c985657bb9cee783b389511135404"} Nov 25 11:01:29 crc kubenswrapper[4776]: I1125 11:01:29.007358 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc" containerName="glance-log" containerID="cri-o://62f99b47441ff17760618b5b803fae0efd2aebefed3d736a761f18426dd1d2e3" gracePeriod=30 Nov 25 11:01:29 crc kubenswrapper[4776]: I1125 11:01:29.007374 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc" containerName="glance-httpd" containerID="cri-o://f5fd94e6048195b37b6c9f09840f447cf43c985657bb9cee783b389511135404" gracePeriod=30 Nov 25 11:01:29 crc kubenswrapper[4776]: I1125 11:01:29.014484 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d8cfd89b5-6pvwz" event={"ID":"147b6859-7722-443e-b424-f8a61aee320d","Type":"ContainerStarted","Data":"8d2953a60f0e316d160745c681ba01c995f50a349f09228314de9609280fded3"} Nov 25 11:01:29 crc kubenswrapper[4776]: I1125 11:01:29.015330 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d8cfd89b5-6pvwz" Nov 25 11:01:29 crc kubenswrapper[4776]: I1125 11:01:29.050663 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.050639336 podStartE2EDuration="3.050639336s" podCreationTimestamp="2025-11-25 11:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:01:29.020600012 +0000 UTC m=+5834.061659575" watchObservedRunningTime="2025-11-25 11:01:29.050639336 +0000 UTC m=+5834.091698889" Nov 25 11:01:29 crc kubenswrapper[4776]: I1125 11:01:29.052922 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.052911393 podStartE2EDuration="3.052911393s" podCreationTimestamp="2025-11-25 11:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:01:29.043900597 +0000 UTC m=+5834.084960150" watchObservedRunningTime="2025-11-25 11:01:29.052911393 +0000 UTC m=+5834.093970956" Nov 25 11:01:29 crc kubenswrapper[4776]: I1125 11:01:29.062525 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d8cfd89b5-6pvwz" podStartSLOduration=3.062508694 podStartE2EDuration="3.062508694s" podCreationTimestamp="2025-11-25 11:01:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:01:29.061905399 +0000 UTC m=+5834.102964952" watchObservedRunningTime="2025-11-25 11:01:29.062508694 +0000 UTC m=+5834.103568247" Nov 25 11:01:29 crc kubenswrapper[4776]: I1125 11:01:29.705248 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 11:01:29 crc kubenswrapper[4776]: I1125 11:01:29.827796 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-httpd-run\") pod \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\" (UID: \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\") " Nov 25 11:01:29 crc kubenswrapper[4776]: I1125 11:01:29.827874 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-combined-ca-bundle\") pod \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\" (UID: \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\") " Nov 25 11:01:29 crc kubenswrapper[4776]: I1125 11:01:29.828001 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-config-data\") pod \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\" (UID: \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\") " Nov 25 11:01:29 crc kubenswrapper[4776]: I1125 11:01:29.828038 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-logs\") pod \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\" (UID: \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\") " Nov 25 11:01:29 crc kubenswrapper[4776]: I1125 11:01:29.828082 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9p85\" (UniqueName: \"kubernetes.io/projected/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-kube-api-access-v9p85\") pod \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\" (UID: \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\") " Nov 25 11:01:29 crc kubenswrapper[4776]: I1125 11:01:29.828185 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-scripts\") pod \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\" (UID: \"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc\") " Nov 25 11:01:29 crc kubenswrapper[4776]: I1125 11:01:29.828955 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc" (UID: "a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:01:29 crc kubenswrapper[4776]: I1125 11:01:29.829210 4776 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:29 crc kubenswrapper[4776]: I1125 11:01:29.830517 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-logs" (OuterVolumeSpecName: "logs") pod "a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc" (UID: "a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:01:29 crc kubenswrapper[4776]: I1125 11:01:29.835752 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-scripts" (OuterVolumeSpecName: "scripts") pod "a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc" (UID: "a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:01:29 crc kubenswrapper[4776]: I1125 11:01:29.844129 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-kube-api-access-v9p85" (OuterVolumeSpecName: "kube-api-access-v9p85") pod "a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc" (UID: "a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc"). InnerVolumeSpecName "kube-api-access-v9p85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:01:29 crc kubenswrapper[4776]: I1125 11:01:29.867161 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc" (UID: "a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:01:29 crc kubenswrapper[4776]: I1125 11:01:29.890435 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-config-data" (OuterVolumeSpecName: "config-data") pod "a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc" (UID: "a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:01:29 crc kubenswrapper[4776]: I1125 11:01:29.931334 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9p85\" (UniqueName: \"kubernetes.io/projected/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-kube-api-access-v9p85\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:29 crc kubenswrapper[4776]: I1125 11:01:29.931371 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:29 crc kubenswrapper[4776]: I1125 11:01:29.931381 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:29 crc kubenswrapper[4776]: I1125 11:01:29.931390 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:29 crc kubenswrapper[4776]: I1125 11:01:29.931400 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc-logs\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:29 crc kubenswrapper[4776]: I1125 11:01:29.938427 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.027806 4776 generic.go:334] "Generic (PLEG): container finished" podID="345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b" containerID="b719f697fb6d47a1467b1bfb291bf90f7f2ac9e737dba0b1c3f2f5fd4adafca4" exitCode=0 Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.027852 4776 generic.go:334] "Generic (PLEG): container finished" podID="345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b" containerID="a72854ecf751815c10fecd2592725bb62e38c7e07a6f3bd8c0f7e17102cf8fd5" exitCode=143 Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.027903 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b","Type":"ContainerDied","Data":"b719f697fb6d47a1467b1bfb291bf90f7f2ac9e737dba0b1c3f2f5fd4adafca4"} Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.028010 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b","Type":"ContainerDied","Data":"a72854ecf751815c10fecd2592725bb62e38c7e07a6f3bd8c0f7e17102cf8fd5"} Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.028052 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b","Type":"ContainerDied","Data":"6508a80db6832d76f32c0ddf45ae8eec73cc624ecda06cb3740a63f3ddb9ee85"} Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.028125 4776 scope.go:117] "RemoveContainer" containerID="b719f697fb6d47a1467b1bfb291bf90f7f2ac9e737dba0b1c3f2f5fd4adafca4" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.028448 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.032036 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-logs\") pod \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\" (UID: \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\") " Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.032155 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwwkv\" (UniqueName: \"kubernetes.io/projected/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-kube-api-access-wwwkv\") pod \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\" (UID: \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\") " Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.032213 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-httpd-run\") pod \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\" (UID: \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\") " Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.032389 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-scripts\") pod \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\" (UID: \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\") " Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.032472 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-config-data\") pod \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\" (UID: \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\") " Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.032570 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-combined-ca-bundle\") pod \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\" (UID: \"345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b\") " Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.032824 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b" (UID: "345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.033399 4776 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.033802 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-logs" (OuterVolumeSpecName: "logs") pod "345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b" (UID: "345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.036952 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-kube-api-access-wwwkv" (OuterVolumeSpecName: "kube-api-access-wwwkv") pod "345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b" (UID: "345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b"). InnerVolumeSpecName "kube-api-access-wwwkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.037447 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-scripts" (OuterVolumeSpecName: "scripts") pod "345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b" (UID: "345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.040270 4776 generic.go:334] "Generic (PLEG): container finished" podID="a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc" containerID="f5fd94e6048195b37b6c9f09840f447cf43c985657bb9cee783b389511135404" exitCode=0 Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.040313 4776 generic.go:334] "Generic (PLEG): container finished" podID="a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc" containerID="62f99b47441ff17760618b5b803fae0efd2aebefed3d736a761f18426dd1d2e3" exitCode=143 Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.040360 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.040376 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc","Type":"ContainerDied","Data":"f5fd94e6048195b37b6c9f09840f447cf43c985657bb9cee783b389511135404"} Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.040414 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc","Type":"ContainerDied","Data":"62f99b47441ff17760618b5b803fae0efd2aebefed3d736a761f18426dd1d2e3"} Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.040427 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc","Type":"ContainerDied","Data":"9a16bf03a1316a32ba0247158a664630fa0830db8cdd5bca91cde21b179ceafe"} Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.069578 4776 scope.go:117] "RemoveContainer" containerID="a72854ecf751815c10fecd2592725bb62e38c7e07a6f3bd8c0f7e17102cf8fd5" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.085815 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b" (UID: "345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.090081 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.107593 4776 scope.go:117] "RemoveContainer" containerID="b719f697fb6d47a1467b1bfb291bf90f7f2ac9e737dba0b1c3f2f5fd4adafca4" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.110230 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 11:01:30 crc kubenswrapper[4776]: E1125 11:01:30.110284 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b719f697fb6d47a1467b1bfb291bf90f7f2ac9e737dba0b1c3f2f5fd4adafca4\": container with ID starting with b719f697fb6d47a1467b1bfb291bf90f7f2ac9e737dba0b1c3f2f5fd4adafca4 not found: ID does not exist" containerID="b719f697fb6d47a1467b1bfb291bf90f7f2ac9e737dba0b1c3f2f5fd4adafca4" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.110328 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b719f697fb6d47a1467b1bfb291bf90f7f2ac9e737dba0b1c3f2f5fd4adafca4"} err="failed to get container status \"b719f697fb6d47a1467b1bfb291bf90f7f2ac9e737dba0b1c3f2f5fd4adafca4\": rpc error: code = NotFound desc = could not find container \"b719f697fb6d47a1467b1bfb291bf90f7f2ac9e737dba0b1c3f2f5fd4adafca4\": container with ID starting with b719f697fb6d47a1467b1bfb291bf90f7f2ac9e737dba0b1c3f2f5fd4adafca4 not found: ID does not exist" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.110360 4776 scope.go:117] "RemoveContainer" containerID="a72854ecf751815c10fecd2592725bb62e38c7e07a6f3bd8c0f7e17102cf8fd5" Nov 25 11:01:30 crc kubenswrapper[4776]: E1125 11:01:30.110775 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a72854ecf751815c10fecd2592725bb62e38c7e07a6f3bd8c0f7e17102cf8fd5\": container with ID starting with a72854ecf751815c10fecd2592725bb62e38c7e07a6f3bd8c0f7e17102cf8fd5 not found: ID does not exist" containerID="a72854ecf751815c10fecd2592725bb62e38c7e07a6f3bd8c0f7e17102cf8fd5" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.110804 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a72854ecf751815c10fecd2592725bb62e38c7e07a6f3bd8c0f7e17102cf8fd5"} err="failed to get container status \"a72854ecf751815c10fecd2592725bb62e38c7e07a6f3bd8c0f7e17102cf8fd5\": rpc error: code = NotFound desc = could not find container \"a72854ecf751815c10fecd2592725bb62e38c7e07a6f3bd8c0f7e17102cf8fd5\": container with ID starting with a72854ecf751815c10fecd2592725bb62e38c7e07a6f3bd8c0f7e17102cf8fd5 not found: ID does not exist" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.110821 4776 scope.go:117] "RemoveContainer" containerID="b719f697fb6d47a1467b1bfb291bf90f7f2ac9e737dba0b1c3f2f5fd4adafca4" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.111099 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b719f697fb6d47a1467b1bfb291bf90f7f2ac9e737dba0b1c3f2f5fd4adafca4"} err="failed to get container status \"b719f697fb6d47a1467b1bfb291bf90f7f2ac9e737dba0b1c3f2f5fd4adafca4\": rpc error: code = NotFound desc = could not find container \"b719f697fb6d47a1467b1bfb291bf90f7f2ac9e737dba0b1c3f2f5fd4adafca4\": container with ID starting with b719f697fb6d47a1467b1bfb291bf90f7f2ac9e737dba0b1c3f2f5fd4adafca4 not found: ID does not exist" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.111127 4776 scope.go:117] "RemoveContainer" containerID="a72854ecf751815c10fecd2592725bb62e38c7e07a6f3bd8c0f7e17102cf8fd5" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.111477 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a72854ecf751815c10fecd2592725bb62e38c7e07a6f3bd8c0f7e17102cf8fd5"} err="failed to get container status \"a72854ecf751815c10fecd2592725bb62e38c7e07a6f3bd8c0f7e17102cf8fd5\": rpc error: code = NotFound desc = could not find container \"a72854ecf751815c10fecd2592725bb62e38c7e07a6f3bd8c0f7e17102cf8fd5\": container with ID starting with a72854ecf751815c10fecd2592725bb62e38c7e07a6f3bd8c0f7e17102cf8fd5 not found: ID does not exist" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.111503 4776 scope.go:117] "RemoveContainer" containerID="f5fd94e6048195b37b6c9f09840f447cf43c985657bb9cee783b389511135404" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.120233 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 11:01:30 crc kubenswrapper[4776]: E1125 11:01:30.124051 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc" containerName="glance-log" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.124110 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc" containerName="glance-log" Nov 25 11:01:30 crc kubenswrapper[4776]: E1125 11:01:30.124134 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc" containerName="glance-httpd" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.124143 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc" containerName="glance-httpd" Nov 25 11:01:30 crc kubenswrapper[4776]: E1125 11:01:30.124158 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b" containerName="glance-httpd" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.124165 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b" containerName="glance-httpd" Nov 25 11:01:30 crc kubenswrapper[4776]: E1125 11:01:30.124190 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b" containerName="glance-log" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.124198 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b" containerName="glance-log" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.124403 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b" containerName="glance-httpd" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.124421 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b" containerName="glance-log" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.124439 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc" containerName="glance-log" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.124453 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc" containerName="glance-httpd" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.124881 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-config-data" (OuterVolumeSpecName: "config-data") pod "345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b" (UID: "345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.125603 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.128769 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.128960 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.135906 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-logs\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.135952 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwwkv\" (UniqueName: \"kubernetes.io/projected/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-kube-api-access-wwwkv\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.135971 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.135985 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.135998 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.141340 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.158425 4776 scope.go:117] "RemoveContainer" containerID="62f99b47441ff17760618b5b803fae0efd2aebefed3d736a761f18426dd1d2e3" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.184862 4776 scope.go:117] "RemoveContainer" containerID="f5fd94e6048195b37b6c9f09840f447cf43c985657bb9cee783b389511135404" Nov 25 11:01:30 crc kubenswrapper[4776]: E1125 11:01:30.185351 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5fd94e6048195b37b6c9f09840f447cf43c985657bb9cee783b389511135404\": container with ID starting with f5fd94e6048195b37b6c9f09840f447cf43c985657bb9cee783b389511135404 not found: ID does not exist" containerID="f5fd94e6048195b37b6c9f09840f447cf43c985657bb9cee783b389511135404" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.185394 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5fd94e6048195b37b6c9f09840f447cf43c985657bb9cee783b389511135404"} err="failed to get container status \"f5fd94e6048195b37b6c9f09840f447cf43c985657bb9cee783b389511135404\": rpc error: code = NotFound desc = could not find container \"f5fd94e6048195b37b6c9f09840f447cf43c985657bb9cee783b389511135404\": container with ID starting with f5fd94e6048195b37b6c9f09840f447cf43c985657bb9cee783b389511135404 not found: ID does not exist" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.185438 4776 scope.go:117] "RemoveContainer" containerID="62f99b47441ff17760618b5b803fae0efd2aebefed3d736a761f18426dd1d2e3" Nov 25 11:01:30 crc kubenswrapper[4776]: E1125 11:01:30.186385 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62f99b47441ff17760618b5b803fae0efd2aebefed3d736a761f18426dd1d2e3\": container with ID starting with 62f99b47441ff17760618b5b803fae0efd2aebefed3d736a761f18426dd1d2e3 not found: ID does not exist" containerID="62f99b47441ff17760618b5b803fae0efd2aebefed3d736a761f18426dd1d2e3" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.186448 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62f99b47441ff17760618b5b803fae0efd2aebefed3d736a761f18426dd1d2e3"} err="failed to get container status \"62f99b47441ff17760618b5b803fae0efd2aebefed3d736a761f18426dd1d2e3\": rpc error: code = NotFound desc = could not find container \"62f99b47441ff17760618b5b803fae0efd2aebefed3d736a761f18426dd1d2e3\": container with ID starting with 62f99b47441ff17760618b5b803fae0efd2aebefed3d736a761f18426dd1d2e3 not found: ID does not exist" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.186481 4776 scope.go:117] "RemoveContainer" containerID="f5fd94e6048195b37b6c9f09840f447cf43c985657bb9cee783b389511135404" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.186854 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5fd94e6048195b37b6c9f09840f447cf43c985657bb9cee783b389511135404"} err="failed to get container status \"f5fd94e6048195b37b6c9f09840f447cf43c985657bb9cee783b389511135404\": rpc error: code = NotFound desc = could not find container \"f5fd94e6048195b37b6c9f09840f447cf43c985657bb9cee783b389511135404\": container with ID starting with f5fd94e6048195b37b6c9f09840f447cf43c985657bb9cee783b389511135404 not found: ID does not exist" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.186885 4776 scope.go:117] "RemoveContainer" containerID="62f99b47441ff17760618b5b803fae0efd2aebefed3d736a761f18426dd1d2e3" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.187153 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62f99b47441ff17760618b5b803fae0efd2aebefed3d736a761f18426dd1d2e3"} err="failed to get container status \"62f99b47441ff17760618b5b803fae0efd2aebefed3d736a761f18426dd1d2e3\": rpc error: code = NotFound desc = could not find container \"62f99b47441ff17760618b5b803fae0efd2aebefed3d736a761f18426dd1d2e3\": container with ID starting with 62f99b47441ff17760618b5b803fae0efd2aebefed3d736a761f18426dd1d2e3 not found: ID does not exist" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.237515 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbpzr\" (UniqueName: \"kubernetes.io/projected/afe1ae15-e85d-4d25-81ef-445a574d8222-kube-api-access-kbpzr\") pod \"glance-default-external-api-0\" (UID: \"afe1ae15-e85d-4d25-81ef-445a574d8222\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.237654 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afe1ae15-e85d-4d25-81ef-445a574d8222-logs\") pod \"glance-default-external-api-0\" (UID: \"afe1ae15-e85d-4d25-81ef-445a574d8222\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.237680 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/afe1ae15-e85d-4d25-81ef-445a574d8222-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"afe1ae15-e85d-4d25-81ef-445a574d8222\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.237705 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afe1ae15-e85d-4d25-81ef-445a574d8222-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"afe1ae15-e85d-4d25-81ef-445a574d8222\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.237732 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afe1ae15-e85d-4d25-81ef-445a574d8222-scripts\") pod \"glance-default-external-api-0\" (UID: \"afe1ae15-e85d-4d25-81ef-445a574d8222\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.237846 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afe1ae15-e85d-4d25-81ef-445a574d8222-config-data\") pod \"glance-default-external-api-0\" (UID: \"afe1ae15-e85d-4d25-81ef-445a574d8222\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.237897 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/afe1ae15-e85d-4d25-81ef-445a574d8222-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"afe1ae15-e85d-4d25-81ef-445a574d8222\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.339553 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afe1ae15-e85d-4d25-81ef-445a574d8222-logs\") pod \"glance-default-external-api-0\" (UID: \"afe1ae15-e85d-4d25-81ef-445a574d8222\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.339601 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/afe1ae15-e85d-4d25-81ef-445a574d8222-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"afe1ae15-e85d-4d25-81ef-445a574d8222\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.339639 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afe1ae15-e85d-4d25-81ef-445a574d8222-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"afe1ae15-e85d-4d25-81ef-445a574d8222\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.339666 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afe1ae15-e85d-4d25-81ef-445a574d8222-scripts\") pod \"glance-default-external-api-0\" (UID: \"afe1ae15-e85d-4d25-81ef-445a574d8222\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.339762 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afe1ae15-e85d-4d25-81ef-445a574d8222-config-data\") pod \"glance-default-external-api-0\" (UID: \"afe1ae15-e85d-4d25-81ef-445a574d8222\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.339838 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/afe1ae15-e85d-4d25-81ef-445a574d8222-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"afe1ae15-e85d-4d25-81ef-445a574d8222\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.339895 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbpzr\" (UniqueName: \"kubernetes.io/projected/afe1ae15-e85d-4d25-81ef-445a574d8222-kube-api-access-kbpzr\") pod \"glance-default-external-api-0\" (UID: \"afe1ae15-e85d-4d25-81ef-445a574d8222\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.340143 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afe1ae15-e85d-4d25-81ef-445a574d8222-logs\") pod \"glance-default-external-api-0\" (UID: \"afe1ae15-e85d-4d25-81ef-445a574d8222\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.340514 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/afe1ae15-e85d-4d25-81ef-445a574d8222-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"afe1ae15-e85d-4d25-81ef-445a574d8222\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.344463 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afe1ae15-e85d-4d25-81ef-445a574d8222-scripts\") pod \"glance-default-external-api-0\" (UID: \"afe1ae15-e85d-4d25-81ef-445a574d8222\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.345593 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afe1ae15-e85d-4d25-81ef-445a574d8222-config-data\") pod \"glance-default-external-api-0\" (UID: \"afe1ae15-e85d-4d25-81ef-445a574d8222\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.346191 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/afe1ae15-e85d-4d25-81ef-445a574d8222-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"afe1ae15-e85d-4d25-81ef-445a574d8222\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.346318 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afe1ae15-e85d-4d25-81ef-445a574d8222-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"afe1ae15-e85d-4d25-81ef-445a574d8222\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.370330 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbpzr\" (UniqueName: \"kubernetes.io/projected/afe1ae15-e85d-4d25-81ef-445a574d8222-kube-api-access-kbpzr\") pod \"glance-default-external-api-0\" (UID: \"afe1ae15-e85d-4d25-81ef-445a574d8222\") " pod="openstack/glance-default-external-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.383396 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.396008 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.409245 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.411412 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.414560 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.420255 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.426573 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.443918 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.543384 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.543489 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-logs\") pod \"glance-default-internal-api-0\" (UID: \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.543575 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cvhn\" (UniqueName: \"kubernetes.io/projected/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-kube-api-access-6cvhn\") pod \"glance-default-internal-api-0\" (UID: \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.543723 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.543851 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.543887 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.543986 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.645313 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.645720 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.645794 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-logs\") pod \"glance-default-internal-api-0\" (UID: \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.645846 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cvhn\" (UniqueName: \"kubernetes.io/projected/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-kube-api-access-6cvhn\") pod \"glance-default-internal-api-0\" (UID: \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.645927 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.645975 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.646005 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.647835 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-logs\") pod \"glance-default-internal-api-0\" (UID: \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.649341 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.663790 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.664573 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.666515 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.666804 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.666898 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cvhn\" (UniqueName: \"kubernetes.io/projected/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-kube-api-access-6cvhn\") pod \"glance-default-internal-api-0\" (UID: \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:01:30 crc kubenswrapper[4776]: I1125 11:01:30.736256 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 11:01:31 crc kubenswrapper[4776]: I1125 11:01:31.073204 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 11:01:31 crc kubenswrapper[4776]: W1125 11:01:31.076842 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafe1ae15_e85d_4d25_81ef_445a574d8222.slice/crio-5fe6069ad61b78b91b5e7db1494a23f0e4776a679e6abbd5081e8bf56edefc26 WatchSource:0}: Error finding container 5fe6069ad61b78b91b5e7db1494a23f0e4776a679e6abbd5081e8bf56edefc26: Status 404 returned error can't find the container with id 5fe6069ad61b78b91b5e7db1494a23f0e4776a679e6abbd5081e8bf56edefc26 Nov 25 11:01:31 crc kubenswrapper[4776]: W1125 11:01:31.264217 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bdf5cf7_3d1f_43b3_8d5a_b7b5253a502a.slice/crio-92594241f6fc33b4adbb62148ccdf2e313b680db092c248427a6a0eb82e8c4ea WatchSource:0}: Error finding container 92594241f6fc33b4adbb62148ccdf2e313b680db092c248427a6a0eb82e8c4ea: Status 404 returned error can't find the container with id 92594241f6fc33b4adbb62148ccdf2e313b680db092c248427a6a0eb82e8c4ea Nov 25 11:01:31 crc kubenswrapper[4776]: I1125 11:01:31.269866 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 11:01:31 crc kubenswrapper[4776]: I1125 11:01:31.674029 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b" path="/var/lib/kubelet/pods/345d87f1-ea72-4bcb-b5d8-7d5921f0aa5b/volumes" Nov 25 11:01:31 crc kubenswrapper[4776]: I1125 11:01:31.675318 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc" path="/var/lib/kubelet/pods/a78be2da-2747-4e3f-8cbd-7ee6bf8e8dbc/volumes" Nov 25 11:01:32 crc kubenswrapper[4776]: I1125 11:01:32.062680 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"afe1ae15-e85d-4d25-81ef-445a574d8222","Type":"ContainerStarted","Data":"4b90d93d9326d7485e4ab371e06f94e76c0b64f3ce6adcbb172886a6fa5348d8"} Nov 25 11:01:32 crc kubenswrapper[4776]: I1125 11:01:32.062742 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"afe1ae15-e85d-4d25-81ef-445a574d8222","Type":"ContainerStarted","Data":"5fe6069ad61b78b91b5e7db1494a23f0e4776a679e6abbd5081e8bf56edefc26"} Nov 25 11:01:32 crc kubenswrapper[4776]: I1125 11:01:32.065528 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a","Type":"ContainerStarted","Data":"5d69cfffdca6866fa1479b16f66e7d4638a5e2afa489a7050cef817c49652ad8"} Nov 25 11:01:32 crc kubenswrapper[4776]: I1125 11:01:32.065552 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a","Type":"ContainerStarted","Data":"92594241f6fc33b4adbb62148ccdf2e313b680db092c248427a6a0eb82e8c4ea"} Nov 25 11:01:32 crc kubenswrapper[4776]: I1125 11:01:32.748978 4776 scope.go:117] "RemoveContainer" containerID="1bdc2fc4eeba1050b8a93b69e842c63b8967de131f1dfc0bf3dd2b51d11017eb" Nov 25 11:01:33 crc kubenswrapper[4776]: I1125 11:01:33.075505 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a","Type":"ContainerStarted","Data":"d67251a37a2930b6bc7b71c9e7bce4cd60e86e67780a1487e294e57b5a93ca21"} Nov 25 11:01:33 crc kubenswrapper[4776]: I1125 11:01:33.078276 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"afe1ae15-e85d-4d25-81ef-445a574d8222","Type":"ContainerStarted","Data":"614824f19f67266106422717756035ef58d54d132e6357b434ced6ff82605737"} Nov 25 11:01:33 crc kubenswrapper[4776]: I1125 11:01:33.100375 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.10035271 podStartE2EDuration="3.10035271s" podCreationTimestamp="2025-11-25 11:01:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:01:33.098681649 +0000 UTC m=+5838.139741192" watchObservedRunningTime="2025-11-25 11:01:33.10035271 +0000 UTC m=+5838.141412263" Nov 25 11:01:33 crc kubenswrapper[4776]: I1125 11:01:33.141739 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.141721589 podStartE2EDuration="3.141721589s" podCreationTimestamp="2025-11-25 11:01:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:01:33.134868367 +0000 UTC m=+5838.175927920" watchObservedRunningTime="2025-11-25 11:01:33.141721589 +0000 UTC m=+5838.182781142" Nov 25 11:01:36 crc kubenswrapper[4776]: I1125 11:01:36.646437 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d8cfd89b5-6pvwz" Nov 25 11:01:36 crc kubenswrapper[4776]: I1125 11:01:36.718602 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78994c4b75-5c25c"] Nov 25 11:01:36 crc kubenswrapper[4776]: I1125 11:01:36.718869 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-78994c4b75-5c25c" podUID="eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9" containerName="dnsmasq-dns" containerID="cri-o://98e8c48ee2c38d5323691911b3fb543dd2e31fcffacde20b6b7b79bf4a2692fa" gracePeriod=10 Nov 25 11:01:37 crc kubenswrapper[4776]: I1125 11:01:37.115939 4776 generic.go:334] "Generic (PLEG): container finished" podID="eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9" containerID="98e8c48ee2c38d5323691911b3fb543dd2e31fcffacde20b6b7b79bf4a2692fa" exitCode=0 Nov 25 11:01:37 crc kubenswrapper[4776]: I1125 11:01:37.116191 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78994c4b75-5c25c" event={"ID":"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9","Type":"ContainerDied","Data":"98e8c48ee2c38d5323691911b3fb543dd2e31fcffacde20b6b7b79bf4a2692fa"} Nov 25 11:01:37 crc kubenswrapper[4776]: I1125 11:01:37.229606 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78994c4b75-5c25c" Nov 25 11:01:37 crc kubenswrapper[4776]: I1125 11:01:37.373082 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-dns-svc\") pod \"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9\" (UID: \"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9\") " Nov 25 11:01:37 crc kubenswrapper[4776]: I1125 11:01:37.373206 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-ovsdbserver-nb\") pod \"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9\" (UID: \"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9\") " Nov 25 11:01:37 crc kubenswrapper[4776]: I1125 11:01:37.373237 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-config\") pod \"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9\" (UID: \"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9\") " Nov 25 11:01:37 crc kubenswrapper[4776]: I1125 11:01:37.373291 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-ovsdbserver-sb\") pod \"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9\" (UID: \"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9\") " Nov 25 11:01:37 crc kubenswrapper[4776]: I1125 11:01:37.373334 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dh2fv\" (UniqueName: \"kubernetes.io/projected/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-kube-api-access-dh2fv\") pod \"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9\" (UID: \"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9\") " Nov 25 11:01:37 crc kubenswrapper[4776]: I1125 11:01:37.390372 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-kube-api-access-dh2fv" (OuterVolumeSpecName: "kube-api-access-dh2fv") pod "eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9" (UID: "eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9"). InnerVolumeSpecName "kube-api-access-dh2fv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:01:37 crc kubenswrapper[4776]: I1125 11:01:37.421974 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-config" (OuterVolumeSpecName: "config") pod "eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9" (UID: "eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:01:37 crc kubenswrapper[4776]: I1125 11:01:37.422351 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9" (UID: "eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:01:37 crc kubenswrapper[4776]: I1125 11:01:37.423009 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9" (UID: "eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:01:37 crc kubenswrapper[4776]: I1125 11:01:37.426226 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9" (UID: "eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:01:37 crc kubenswrapper[4776]: I1125 11:01:37.474813 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:37 crc kubenswrapper[4776]: I1125 11:01:37.474989 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:37 crc kubenswrapper[4776]: I1125 11:01:37.475043 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-config\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:37 crc kubenswrapper[4776]: I1125 11:01:37.475149 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:37 crc kubenswrapper[4776]: I1125 11:01:37.475260 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dh2fv\" (UniqueName: \"kubernetes.io/projected/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9-kube-api-access-dh2fv\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:38 crc kubenswrapper[4776]: I1125 11:01:38.127465 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78994c4b75-5c25c" event={"ID":"eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9","Type":"ContainerDied","Data":"59dd63dcf3b1f5909dfdc3086bf39b611a044e3330957ce93cd583d3662fecbf"} Nov 25 11:01:38 crc kubenswrapper[4776]: I1125 11:01:38.127739 4776 scope.go:117] "RemoveContainer" containerID="98e8c48ee2c38d5323691911b3fb543dd2e31fcffacde20b6b7b79bf4a2692fa" Nov 25 11:01:38 crc kubenswrapper[4776]: I1125 11:01:38.127552 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78994c4b75-5c25c" Nov 25 11:01:38 crc kubenswrapper[4776]: I1125 11:01:38.147864 4776 scope.go:117] "RemoveContainer" containerID="3e3d519217585fa073450cb159b471ba8a4c166e7697200f22002f3e6b5d9501" Nov 25 11:01:38 crc kubenswrapper[4776]: I1125 11:01:38.150735 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78994c4b75-5c25c"] Nov 25 11:01:38 crc kubenswrapper[4776]: I1125 11:01:38.157124 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78994c4b75-5c25c"] Nov 25 11:01:39 crc kubenswrapper[4776]: I1125 11:01:39.673872 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9" path="/var/lib/kubelet/pods/eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9/volumes" Nov 25 11:01:40 crc kubenswrapper[4776]: I1125 11:01:40.445090 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 25 11:01:40 crc kubenswrapper[4776]: I1125 11:01:40.445401 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 25 11:01:40 crc kubenswrapper[4776]: I1125 11:01:40.473518 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 25 11:01:40 crc kubenswrapper[4776]: I1125 11:01:40.482827 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 25 11:01:40 crc kubenswrapper[4776]: I1125 11:01:40.737180 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 25 11:01:40 crc kubenswrapper[4776]: I1125 11:01:40.737276 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 25 11:01:40 crc kubenswrapper[4776]: I1125 11:01:40.769048 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 25 11:01:40 crc kubenswrapper[4776]: I1125 11:01:40.793718 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 25 11:01:41 crc kubenswrapper[4776]: I1125 11:01:41.152301 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 25 11:01:41 crc kubenswrapper[4776]: I1125 11:01:41.152516 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 25 11:01:41 crc kubenswrapper[4776]: I1125 11:01:41.152675 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 25 11:01:41 crc kubenswrapper[4776]: I1125 11:01:41.153213 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 25 11:01:43 crc kubenswrapper[4776]: I1125 11:01:43.039352 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 25 11:01:43 crc kubenswrapper[4776]: I1125 11:01:43.165845 4776 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 11:01:43 crc kubenswrapper[4776]: I1125 11:01:43.165875 4776 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 11:01:43 crc kubenswrapper[4776]: I1125 11:01:43.165952 4776 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 11:01:43 crc kubenswrapper[4776]: I1125 11:01:43.196155 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 25 11:01:43 crc kubenswrapper[4776]: I1125 11:01:43.257976 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 25 11:01:43 crc kubenswrapper[4776]: I1125 11:01:43.266710 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 25 11:01:51 crc kubenswrapper[4776]: I1125 11:01:51.225855 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-78bsr"] Nov 25 11:01:51 crc kubenswrapper[4776]: E1125 11:01:51.228934 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9" containerName="dnsmasq-dns" Nov 25 11:01:51 crc kubenswrapper[4776]: I1125 11:01:51.229056 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9" containerName="dnsmasq-dns" Nov 25 11:01:51 crc kubenswrapper[4776]: E1125 11:01:51.229167 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9" containerName="init" Nov 25 11:01:51 crc kubenswrapper[4776]: I1125 11:01:51.229263 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9" containerName="init" Nov 25 11:01:51 crc kubenswrapper[4776]: I1125 11:01:51.229584 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb812d47-f7bf-4a44-a2f0-abc9c9f0e5f9" containerName="dnsmasq-dns" Nov 25 11:01:51 crc kubenswrapper[4776]: I1125 11:01:51.230968 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-78bsr" Nov 25 11:01:51 crc kubenswrapper[4776]: I1125 11:01:51.237783 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-78bsr"] Nov 25 11:01:51 crc kubenswrapper[4776]: I1125 11:01:51.319958 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-4595-account-create-rrgbl"] Nov 25 11:01:51 crc kubenswrapper[4776]: I1125 11:01:51.321537 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4595-account-create-rrgbl" Nov 25 11:01:51 crc kubenswrapper[4776]: I1125 11:01:51.323265 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 25 11:01:51 crc kubenswrapper[4776]: I1125 11:01:51.327420 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ppjz\" (UniqueName: \"kubernetes.io/projected/59ac55fd-b5a0-47f9-aae2-9d1384594315-kube-api-access-8ppjz\") pod \"placement-db-create-78bsr\" (UID: \"59ac55fd-b5a0-47f9-aae2-9d1384594315\") " pod="openstack/placement-db-create-78bsr" Nov 25 11:01:51 crc kubenswrapper[4776]: I1125 11:01:51.327478 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59ac55fd-b5a0-47f9-aae2-9d1384594315-operator-scripts\") pod \"placement-db-create-78bsr\" (UID: \"59ac55fd-b5a0-47f9-aae2-9d1384594315\") " pod="openstack/placement-db-create-78bsr" Nov 25 11:01:51 crc kubenswrapper[4776]: I1125 11:01:51.328758 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-4595-account-create-rrgbl"] Nov 25 11:01:51 crc kubenswrapper[4776]: I1125 11:01:51.428788 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/865436be-d605-4159-a85e-38a7403703e9-operator-scripts\") pod \"placement-4595-account-create-rrgbl\" (UID: \"865436be-d605-4159-a85e-38a7403703e9\") " pod="openstack/placement-4595-account-create-rrgbl" Nov 25 11:01:51 crc kubenswrapper[4776]: I1125 11:01:51.428884 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ppjz\" (UniqueName: \"kubernetes.io/projected/59ac55fd-b5a0-47f9-aae2-9d1384594315-kube-api-access-8ppjz\") pod \"placement-db-create-78bsr\" (UID: \"59ac55fd-b5a0-47f9-aae2-9d1384594315\") " pod="openstack/placement-db-create-78bsr" Nov 25 11:01:51 crc kubenswrapper[4776]: I1125 11:01:51.428927 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59ac55fd-b5a0-47f9-aae2-9d1384594315-operator-scripts\") pod \"placement-db-create-78bsr\" (UID: \"59ac55fd-b5a0-47f9-aae2-9d1384594315\") " pod="openstack/placement-db-create-78bsr" Nov 25 11:01:51 crc kubenswrapper[4776]: I1125 11:01:51.428953 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59qdh\" (UniqueName: \"kubernetes.io/projected/865436be-d605-4159-a85e-38a7403703e9-kube-api-access-59qdh\") pod \"placement-4595-account-create-rrgbl\" (UID: \"865436be-d605-4159-a85e-38a7403703e9\") " pod="openstack/placement-4595-account-create-rrgbl" Nov 25 11:01:51 crc kubenswrapper[4776]: I1125 11:01:51.429739 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59ac55fd-b5a0-47f9-aae2-9d1384594315-operator-scripts\") pod \"placement-db-create-78bsr\" (UID: \"59ac55fd-b5a0-47f9-aae2-9d1384594315\") " pod="openstack/placement-db-create-78bsr" Nov 25 11:01:51 crc kubenswrapper[4776]: I1125 11:01:51.448053 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ppjz\" (UniqueName: \"kubernetes.io/projected/59ac55fd-b5a0-47f9-aae2-9d1384594315-kube-api-access-8ppjz\") pod \"placement-db-create-78bsr\" (UID: \"59ac55fd-b5a0-47f9-aae2-9d1384594315\") " pod="openstack/placement-db-create-78bsr" Nov 25 11:01:51 crc kubenswrapper[4776]: I1125 11:01:51.530832 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59qdh\" (UniqueName: \"kubernetes.io/projected/865436be-d605-4159-a85e-38a7403703e9-kube-api-access-59qdh\") pod \"placement-4595-account-create-rrgbl\" (UID: \"865436be-d605-4159-a85e-38a7403703e9\") " pod="openstack/placement-4595-account-create-rrgbl" Nov 25 11:01:51 crc kubenswrapper[4776]: I1125 11:01:51.530992 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/865436be-d605-4159-a85e-38a7403703e9-operator-scripts\") pod \"placement-4595-account-create-rrgbl\" (UID: \"865436be-d605-4159-a85e-38a7403703e9\") " pod="openstack/placement-4595-account-create-rrgbl" Nov 25 11:01:51 crc kubenswrapper[4776]: I1125 11:01:51.531858 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/865436be-d605-4159-a85e-38a7403703e9-operator-scripts\") pod \"placement-4595-account-create-rrgbl\" (UID: \"865436be-d605-4159-a85e-38a7403703e9\") " pod="openstack/placement-4595-account-create-rrgbl" Nov 25 11:01:51 crc kubenswrapper[4776]: I1125 11:01:51.549336 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59qdh\" (UniqueName: \"kubernetes.io/projected/865436be-d605-4159-a85e-38a7403703e9-kube-api-access-59qdh\") pod \"placement-4595-account-create-rrgbl\" (UID: \"865436be-d605-4159-a85e-38a7403703e9\") " pod="openstack/placement-4595-account-create-rrgbl" Nov 25 11:01:51 crc kubenswrapper[4776]: I1125 11:01:51.551130 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-78bsr" Nov 25 11:01:51 crc kubenswrapper[4776]: I1125 11:01:51.640452 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4595-account-create-rrgbl" Nov 25 11:01:52 crc kubenswrapper[4776]: W1125 11:01:52.072965 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59ac55fd_b5a0_47f9_aae2_9d1384594315.slice/crio-1da1c85cfe2ea105fd6fc81f4bab1f9dcec7008cd7b806e4dfb0e342d52cdab9 WatchSource:0}: Error finding container 1da1c85cfe2ea105fd6fc81f4bab1f9dcec7008cd7b806e4dfb0e342d52cdab9: Status 404 returned error can't find the container with id 1da1c85cfe2ea105fd6fc81f4bab1f9dcec7008cd7b806e4dfb0e342d52cdab9 Nov 25 11:01:52 crc kubenswrapper[4776]: I1125 11:01:52.074577 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-78bsr"] Nov 25 11:01:52 crc kubenswrapper[4776]: I1125 11:01:52.194455 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-4595-account-create-rrgbl"] Nov 25 11:01:52 crc kubenswrapper[4776]: W1125 11:01:52.194647 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod865436be_d605_4159_a85e_38a7403703e9.slice/crio-035e59a58a5fff6c3a06afd2bf549f684c95ce64688a22e7280e3f76a3f45f1f WatchSource:0}: Error finding container 035e59a58a5fff6c3a06afd2bf549f684c95ce64688a22e7280e3f76a3f45f1f: Status 404 returned error can't find the container with id 035e59a58a5fff6c3a06afd2bf549f684c95ce64688a22e7280e3f76a3f45f1f Nov 25 11:01:52 crc kubenswrapper[4776]: I1125 11:01:52.261106 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4595-account-create-rrgbl" event={"ID":"865436be-d605-4159-a85e-38a7403703e9","Type":"ContainerStarted","Data":"035e59a58a5fff6c3a06afd2bf549f684c95ce64688a22e7280e3f76a3f45f1f"} Nov 25 11:01:52 crc kubenswrapper[4776]: I1125 11:01:52.267559 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-78bsr" event={"ID":"59ac55fd-b5a0-47f9-aae2-9d1384594315","Type":"ContainerStarted","Data":"3e5e58c668fec192a0a49e5e7e1abd071ab2d9e0be25bb4a9f499a2eafd26898"} Nov 25 11:01:52 crc kubenswrapper[4776]: I1125 11:01:52.267622 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-78bsr" event={"ID":"59ac55fd-b5a0-47f9-aae2-9d1384594315","Type":"ContainerStarted","Data":"1da1c85cfe2ea105fd6fc81f4bab1f9dcec7008cd7b806e4dfb0e342d52cdab9"} Nov 25 11:01:52 crc kubenswrapper[4776]: I1125 11:01:52.283658 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-78bsr" podStartSLOduration=1.283633897 podStartE2EDuration="1.283633897s" podCreationTimestamp="2025-11-25 11:01:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:01:52.283017001 +0000 UTC m=+5857.324076554" watchObservedRunningTime="2025-11-25 11:01:52.283633897 +0000 UTC m=+5857.324693450" Nov 25 11:01:53 crc kubenswrapper[4776]: I1125 11:01:53.278358 4776 generic.go:334] "Generic (PLEG): container finished" podID="59ac55fd-b5a0-47f9-aae2-9d1384594315" containerID="3e5e58c668fec192a0a49e5e7e1abd071ab2d9e0be25bb4a9f499a2eafd26898" exitCode=0 Nov 25 11:01:53 crc kubenswrapper[4776]: I1125 11:01:53.278752 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-78bsr" event={"ID":"59ac55fd-b5a0-47f9-aae2-9d1384594315","Type":"ContainerDied","Data":"3e5e58c668fec192a0a49e5e7e1abd071ab2d9e0be25bb4a9f499a2eafd26898"} Nov 25 11:01:53 crc kubenswrapper[4776]: I1125 11:01:53.281106 4776 generic.go:334] "Generic (PLEG): container finished" podID="865436be-d605-4159-a85e-38a7403703e9" containerID="e036f950b949a43764acf5a11e58886283eb69f531f34baf6e18e175c1ab7588" exitCode=0 Nov 25 11:01:53 crc kubenswrapper[4776]: I1125 11:01:53.281162 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4595-account-create-rrgbl" event={"ID":"865436be-d605-4159-a85e-38a7403703e9","Type":"ContainerDied","Data":"e036f950b949a43764acf5a11e58886283eb69f531f34baf6e18e175c1ab7588"} Nov 25 11:01:54 crc kubenswrapper[4776]: I1125 11:01:54.714520 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-78bsr" Nov 25 11:01:54 crc kubenswrapper[4776]: I1125 11:01:54.719768 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4595-account-create-rrgbl" Nov 25 11:01:54 crc kubenswrapper[4776]: I1125 11:01:54.793867 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ppjz\" (UniqueName: \"kubernetes.io/projected/59ac55fd-b5a0-47f9-aae2-9d1384594315-kube-api-access-8ppjz\") pod \"59ac55fd-b5a0-47f9-aae2-9d1384594315\" (UID: \"59ac55fd-b5a0-47f9-aae2-9d1384594315\") " Nov 25 11:01:54 crc kubenswrapper[4776]: I1125 11:01:54.794044 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59ac55fd-b5a0-47f9-aae2-9d1384594315-operator-scripts\") pod \"59ac55fd-b5a0-47f9-aae2-9d1384594315\" (UID: \"59ac55fd-b5a0-47f9-aae2-9d1384594315\") " Nov 25 11:01:54 crc kubenswrapper[4776]: I1125 11:01:54.794165 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59qdh\" (UniqueName: \"kubernetes.io/projected/865436be-d605-4159-a85e-38a7403703e9-kube-api-access-59qdh\") pod \"865436be-d605-4159-a85e-38a7403703e9\" (UID: \"865436be-d605-4159-a85e-38a7403703e9\") " Nov 25 11:01:54 crc kubenswrapper[4776]: I1125 11:01:54.794308 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/865436be-d605-4159-a85e-38a7403703e9-operator-scripts\") pod \"865436be-d605-4159-a85e-38a7403703e9\" (UID: \"865436be-d605-4159-a85e-38a7403703e9\") " Nov 25 11:01:54 crc kubenswrapper[4776]: I1125 11:01:54.794959 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/865436be-d605-4159-a85e-38a7403703e9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "865436be-d605-4159-a85e-38a7403703e9" (UID: "865436be-d605-4159-a85e-38a7403703e9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:01:54 crc kubenswrapper[4776]: I1125 11:01:54.795290 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59ac55fd-b5a0-47f9-aae2-9d1384594315-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "59ac55fd-b5a0-47f9-aae2-9d1384594315" (UID: "59ac55fd-b5a0-47f9-aae2-9d1384594315"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:01:54 crc kubenswrapper[4776]: I1125 11:01:54.799747 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59ac55fd-b5a0-47f9-aae2-9d1384594315-kube-api-access-8ppjz" (OuterVolumeSpecName: "kube-api-access-8ppjz") pod "59ac55fd-b5a0-47f9-aae2-9d1384594315" (UID: "59ac55fd-b5a0-47f9-aae2-9d1384594315"). InnerVolumeSpecName "kube-api-access-8ppjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:01:54 crc kubenswrapper[4776]: I1125 11:01:54.801714 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/865436be-d605-4159-a85e-38a7403703e9-kube-api-access-59qdh" (OuterVolumeSpecName: "kube-api-access-59qdh") pod "865436be-d605-4159-a85e-38a7403703e9" (UID: "865436be-d605-4159-a85e-38a7403703e9"). InnerVolumeSpecName "kube-api-access-59qdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:01:54 crc kubenswrapper[4776]: I1125 11:01:54.896177 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59qdh\" (UniqueName: \"kubernetes.io/projected/865436be-d605-4159-a85e-38a7403703e9-kube-api-access-59qdh\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:54 crc kubenswrapper[4776]: I1125 11:01:54.896228 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/865436be-d605-4159-a85e-38a7403703e9-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:54 crc kubenswrapper[4776]: I1125 11:01:54.896243 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ppjz\" (UniqueName: \"kubernetes.io/projected/59ac55fd-b5a0-47f9-aae2-9d1384594315-kube-api-access-8ppjz\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:54 crc kubenswrapper[4776]: I1125 11:01:54.896256 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/59ac55fd-b5a0-47f9-aae2-9d1384594315-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:01:55 crc kubenswrapper[4776]: I1125 11:01:55.301289 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4595-account-create-rrgbl" event={"ID":"865436be-d605-4159-a85e-38a7403703e9","Type":"ContainerDied","Data":"035e59a58a5fff6c3a06afd2bf549f684c95ce64688a22e7280e3f76a3f45f1f"} Nov 25 11:01:55 crc kubenswrapper[4776]: I1125 11:01:55.301647 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="035e59a58a5fff6c3a06afd2bf549f684c95ce64688a22e7280e3f76a3f45f1f" Nov 25 11:01:55 crc kubenswrapper[4776]: I1125 11:01:55.301314 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4595-account-create-rrgbl" Nov 25 11:01:55 crc kubenswrapper[4776]: I1125 11:01:55.303010 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-78bsr" event={"ID":"59ac55fd-b5a0-47f9-aae2-9d1384594315","Type":"ContainerDied","Data":"1da1c85cfe2ea105fd6fc81f4bab1f9dcec7008cd7b806e4dfb0e342d52cdab9"} Nov 25 11:01:55 crc kubenswrapper[4776]: I1125 11:01:55.303055 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-78bsr" Nov 25 11:01:55 crc kubenswrapper[4776]: I1125 11:01:55.303058 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1da1c85cfe2ea105fd6fc81f4bab1f9dcec7008cd7b806e4dfb0e342d52cdab9" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.538704 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-674c55699c-vp7jm"] Nov 25 11:01:56 crc kubenswrapper[4776]: E1125 11:01:56.542076 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59ac55fd-b5a0-47f9-aae2-9d1384594315" containerName="mariadb-database-create" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.542104 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="59ac55fd-b5a0-47f9-aae2-9d1384594315" containerName="mariadb-database-create" Nov 25 11:01:56 crc kubenswrapper[4776]: E1125 11:01:56.542117 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="865436be-d605-4159-a85e-38a7403703e9" containerName="mariadb-account-create" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.542123 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="865436be-d605-4159-a85e-38a7403703e9" containerName="mariadb-account-create" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.542318 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="865436be-d605-4159-a85e-38a7403703e9" containerName="mariadb-account-create" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.542338 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="59ac55fd-b5a0-47f9-aae2-9d1384594315" containerName="mariadb-database-create" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.547039 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-674c55699c-vp7jm" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.573713 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-674c55699c-vp7jm"] Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.605855 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-gdfkv"] Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.607375 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gdfkv" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.609960 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.610013 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.610374 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-72h98" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.624518 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfxx7\" (UniqueName: \"kubernetes.io/projected/6bdfe28a-e199-4082-9f5a-3d8b38131d49-kube-api-access-zfxx7\") pod \"dnsmasq-dns-674c55699c-vp7jm\" (UID: \"6bdfe28a-e199-4082-9f5a-3d8b38131d49\") " pod="openstack/dnsmasq-dns-674c55699c-vp7jm" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.624578 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bdfe28a-e199-4082-9f5a-3d8b38131d49-config\") pod \"dnsmasq-dns-674c55699c-vp7jm\" (UID: \"6bdfe28a-e199-4082-9f5a-3d8b38131d49\") " pod="openstack/dnsmasq-dns-674c55699c-vp7jm" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.624654 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6bdfe28a-e199-4082-9f5a-3d8b38131d49-ovsdbserver-sb\") pod \"dnsmasq-dns-674c55699c-vp7jm\" (UID: \"6bdfe28a-e199-4082-9f5a-3d8b38131d49\") " pod="openstack/dnsmasq-dns-674c55699c-vp7jm" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.624741 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6bdfe28a-e199-4082-9f5a-3d8b38131d49-ovsdbserver-nb\") pod \"dnsmasq-dns-674c55699c-vp7jm\" (UID: \"6bdfe28a-e199-4082-9f5a-3d8b38131d49\") " pod="openstack/dnsmasq-dns-674c55699c-vp7jm" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.624817 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6bdfe28a-e199-4082-9f5a-3d8b38131d49-dns-svc\") pod \"dnsmasq-dns-674c55699c-vp7jm\" (UID: \"6bdfe28a-e199-4082-9f5a-3d8b38131d49\") " pod="openstack/dnsmasq-dns-674c55699c-vp7jm" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.629572 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-gdfkv"] Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.726598 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6bdfe28a-e199-4082-9f5a-3d8b38131d49-dns-svc\") pod \"dnsmasq-dns-674c55699c-vp7jm\" (UID: \"6bdfe28a-e199-4082-9f5a-3d8b38131d49\") " pod="openstack/dnsmasq-dns-674c55699c-vp7jm" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.726742 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfxx7\" (UniqueName: \"kubernetes.io/projected/6bdfe28a-e199-4082-9f5a-3d8b38131d49-kube-api-access-zfxx7\") pod \"dnsmasq-dns-674c55699c-vp7jm\" (UID: \"6bdfe28a-e199-4082-9f5a-3d8b38131d49\") " pod="openstack/dnsmasq-dns-674c55699c-vp7jm" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.726793 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bdfe28a-e199-4082-9f5a-3d8b38131d49-config\") pod \"dnsmasq-dns-674c55699c-vp7jm\" (UID: \"6bdfe28a-e199-4082-9f5a-3d8b38131d49\") " pod="openstack/dnsmasq-dns-674c55699c-vp7jm" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.726849 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5b03bdd-3ede-487b-8403-073190b9f68b-logs\") pod \"placement-db-sync-gdfkv\" (UID: \"b5b03bdd-3ede-487b-8403-073190b9f68b\") " pod="openstack/placement-db-sync-gdfkv" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.726891 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltqcf\" (UniqueName: \"kubernetes.io/projected/b5b03bdd-3ede-487b-8403-073190b9f68b-kube-api-access-ltqcf\") pod \"placement-db-sync-gdfkv\" (UID: \"b5b03bdd-3ede-487b-8403-073190b9f68b\") " pod="openstack/placement-db-sync-gdfkv" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.726916 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6bdfe28a-e199-4082-9f5a-3d8b38131d49-ovsdbserver-sb\") pod \"dnsmasq-dns-674c55699c-vp7jm\" (UID: \"6bdfe28a-e199-4082-9f5a-3d8b38131d49\") " pod="openstack/dnsmasq-dns-674c55699c-vp7jm" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.726951 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6bdfe28a-e199-4082-9f5a-3d8b38131d49-ovsdbserver-nb\") pod \"dnsmasq-dns-674c55699c-vp7jm\" (UID: \"6bdfe28a-e199-4082-9f5a-3d8b38131d49\") " pod="openstack/dnsmasq-dns-674c55699c-vp7jm" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.726977 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5b03bdd-3ede-487b-8403-073190b9f68b-config-data\") pod \"placement-db-sync-gdfkv\" (UID: \"b5b03bdd-3ede-487b-8403-073190b9f68b\") " pod="openstack/placement-db-sync-gdfkv" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.727018 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5b03bdd-3ede-487b-8403-073190b9f68b-scripts\") pod \"placement-db-sync-gdfkv\" (UID: \"b5b03bdd-3ede-487b-8403-073190b9f68b\") " pod="openstack/placement-db-sync-gdfkv" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.727039 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5b03bdd-3ede-487b-8403-073190b9f68b-combined-ca-bundle\") pod \"placement-db-sync-gdfkv\" (UID: \"b5b03bdd-3ede-487b-8403-073190b9f68b\") " pod="openstack/placement-db-sync-gdfkv" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.728644 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6bdfe28a-e199-4082-9f5a-3d8b38131d49-dns-svc\") pod \"dnsmasq-dns-674c55699c-vp7jm\" (UID: \"6bdfe28a-e199-4082-9f5a-3d8b38131d49\") " pod="openstack/dnsmasq-dns-674c55699c-vp7jm" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.728689 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6bdfe28a-e199-4082-9f5a-3d8b38131d49-ovsdbserver-sb\") pod \"dnsmasq-dns-674c55699c-vp7jm\" (UID: \"6bdfe28a-e199-4082-9f5a-3d8b38131d49\") " pod="openstack/dnsmasq-dns-674c55699c-vp7jm" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.728756 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bdfe28a-e199-4082-9f5a-3d8b38131d49-config\") pod \"dnsmasq-dns-674c55699c-vp7jm\" (UID: \"6bdfe28a-e199-4082-9f5a-3d8b38131d49\") " pod="openstack/dnsmasq-dns-674c55699c-vp7jm" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.728886 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6bdfe28a-e199-4082-9f5a-3d8b38131d49-ovsdbserver-nb\") pod \"dnsmasq-dns-674c55699c-vp7jm\" (UID: \"6bdfe28a-e199-4082-9f5a-3d8b38131d49\") " pod="openstack/dnsmasq-dns-674c55699c-vp7jm" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.747538 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfxx7\" (UniqueName: \"kubernetes.io/projected/6bdfe28a-e199-4082-9f5a-3d8b38131d49-kube-api-access-zfxx7\") pod \"dnsmasq-dns-674c55699c-vp7jm\" (UID: \"6bdfe28a-e199-4082-9f5a-3d8b38131d49\") " pod="openstack/dnsmasq-dns-674c55699c-vp7jm" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.828310 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5b03bdd-3ede-487b-8403-073190b9f68b-logs\") pod \"placement-db-sync-gdfkv\" (UID: \"b5b03bdd-3ede-487b-8403-073190b9f68b\") " pod="openstack/placement-db-sync-gdfkv" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.828377 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltqcf\" (UniqueName: \"kubernetes.io/projected/b5b03bdd-3ede-487b-8403-073190b9f68b-kube-api-access-ltqcf\") pod \"placement-db-sync-gdfkv\" (UID: \"b5b03bdd-3ede-487b-8403-073190b9f68b\") " pod="openstack/placement-db-sync-gdfkv" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.828463 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5b03bdd-3ede-487b-8403-073190b9f68b-config-data\") pod \"placement-db-sync-gdfkv\" (UID: \"b5b03bdd-3ede-487b-8403-073190b9f68b\") " pod="openstack/placement-db-sync-gdfkv" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.828512 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5b03bdd-3ede-487b-8403-073190b9f68b-scripts\") pod \"placement-db-sync-gdfkv\" (UID: \"b5b03bdd-3ede-487b-8403-073190b9f68b\") " pod="openstack/placement-db-sync-gdfkv" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.828536 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5b03bdd-3ede-487b-8403-073190b9f68b-combined-ca-bundle\") pod \"placement-db-sync-gdfkv\" (UID: \"b5b03bdd-3ede-487b-8403-073190b9f68b\") " pod="openstack/placement-db-sync-gdfkv" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.829433 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5b03bdd-3ede-487b-8403-073190b9f68b-logs\") pod \"placement-db-sync-gdfkv\" (UID: \"b5b03bdd-3ede-487b-8403-073190b9f68b\") " pod="openstack/placement-db-sync-gdfkv" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.832778 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5b03bdd-3ede-487b-8403-073190b9f68b-combined-ca-bundle\") pod \"placement-db-sync-gdfkv\" (UID: \"b5b03bdd-3ede-487b-8403-073190b9f68b\") " pod="openstack/placement-db-sync-gdfkv" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.832977 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5b03bdd-3ede-487b-8403-073190b9f68b-scripts\") pod \"placement-db-sync-gdfkv\" (UID: \"b5b03bdd-3ede-487b-8403-073190b9f68b\") " pod="openstack/placement-db-sync-gdfkv" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.833557 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5b03bdd-3ede-487b-8403-073190b9f68b-config-data\") pod \"placement-db-sync-gdfkv\" (UID: \"b5b03bdd-3ede-487b-8403-073190b9f68b\") " pod="openstack/placement-db-sync-gdfkv" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.846975 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltqcf\" (UniqueName: \"kubernetes.io/projected/b5b03bdd-3ede-487b-8403-073190b9f68b-kube-api-access-ltqcf\") pod \"placement-db-sync-gdfkv\" (UID: \"b5b03bdd-3ede-487b-8403-073190b9f68b\") " pod="openstack/placement-db-sync-gdfkv" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.876755 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-674c55699c-vp7jm" Nov 25 11:01:56 crc kubenswrapper[4776]: I1125 11:01:56.923874 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gdfkv" Nov 25 11:01:57 crc kubenswrapper[4776]: I1125 11:01:57.418730 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-674c55699c-vp7jm"] Nov 25 11:01:57 crc kubenswrapper[4776]: I1125 11:01:57.469499 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-gdfkv"] Nov 25 11:01:57 crc kubenswrapper[4776]: W1125 11:01:57.489859 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5b03bdd_3ede_487b_8403_073190b9f68b.slice/crio-e946556b39feca755a4eac2338ab08430f6e192cad39e455ba99053968756104 WatchSource:0}: Error finding container e946556b39feca755a4eac2338ab08430f6e192cad39e455ba99053968756104: Status 404 returned error can't find the container with id e946556b39feca755a4eac2338ab08430f6e192cad39e455ba99053968756104 Nov 25 11:01:58 crc kubenswrapper[4776]: I1125 11:01:58.329427 4776 generic.go:334] "Generic (PLEG): container finished" podID="6bdfe28a-e199-4082-9f5a-3d8b38131d49" containerID="571a5c8b23e91c84b58876c8c205e770d328659cd087426a20c82f701aec677c" exitCode=0 Nov 25 11:01:58 crc kubenswrapper[4776]: I1125 11:01:58.329504 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-674c55699c-vp7jm" event={"ID":"6bdfe28a-e199-4082-9f5a-3d8b38131d49","Type":"ContainerDied","Data":"571a5c8b23e91c84b58876c8c205e770d328659cd087426a20c82f701aec677c"} Nov 25 11:01:58 crc kubenswrapper[4776]: I1125 11:01:58.330114 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-674c55699c-vp7jm" event={"ID":"6bdfe28a-e199-4082-9f5a-3d8b38131d49","Type":"ContainerStarted","Data":"0614e906961a923b101c6dd5c1435b3b8db6826744ab370b43722e5181168de5"} Nov 25 11:01:58 crc kubenswrapper[4776]: I1125 11:01:58.333947 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gdfkv" event={"ID":"b5b03bdd-3ede-487b-8403-073190b9f68b","Type":"ContainerStarted","Data":"084262977fe869689b12fd1e2ff0368c44c32915ea0383fc5dd5c733933bcf6c"} Nov 25 11:01:58 crc kubenswrapper[4776]: I1125 11:01:58.334161 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gdfkv" event={"ID":"b5b03bdd-3ede-487b-8403-073190b9f68b","Type":"ContainerStarted","Data":"e946556b39feca755a4eac2338ab08430f6e192cad39e455ba99053968756104"} Nov 25 11:01:58 crc kubenswrapper[4776]: I1125 11:01:58.377685 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-gdfkv" podStartSLOduration=2.377660347 podStartE2EDuration="2.377660347s" podCreationTimestamp="2025-11-25 11:01:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:01:58.370129908 +0000 UTC m=+5863.411189451" watchObservedRunningTime="2025-11-25 11:01:58.377660347 +0000 UTC m=+5863.418719900" Nov 25 11:01:59 crc kubenswrapper[4776]: I1125 11:01:59.888303 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-674c55699c-vp7jm" event={"ID":"6bdfe28a-e199-4082-9f5a-3d8b38131d49","Type":"ContainerStarted","Data":"b3a7dd9ebc25e0f457dc46848f9e46d2f30504b2d54a45e2705416e9ef580e93"} Nov 25 11:02:00 crc kubenswrapper[4776]: I1125 11:02:00.896599 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-674c55699c-vp7jm" Nov 25 11:02:00 crc kubenswrapper[4776]: I1125 11:02:00.921003 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-674c55699c-vp7jm" podStartSLOduration=4.9209823660000005 podStartE2EDuration="4.920982366s" podCreationTimestamp="2025-11-25 11:01:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:02:00.914199895 +0000 UTC m=+5865.955259468" watchObservedRunningTime="2025-11-25 11:02:00.920982366 +0000 UTC m=+5865.962041919" Nov 25 11:02:03 crc kubenswrapper[4776]: I1125 11:02:03.933490 4776 generic.go:334] "Generic (PLEG): container finished" podID="b5b03bdd-3ede-487b-8403-073190b9f68b" containerID="084262977fe869689b12fd1e2ff0368c44c32915ea0383fc5dd5c733933bcf6c" exitCode=0 Nov 25 11:02:03 crc kubenswrapper[4776]: I1125 11:02:03.933621 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gdfkv" event={"ID":"b5b03bdd-3ede-487b-8403-073190b9f68b","Type":"ContainerDied","Data":"084262977fe869689b12fd1e2ff0368c44c32915ea0383fc5dd5c733933bcf6c"} Nov 25 11:02:05 crc kubenswrapper[4776]: I1125 11:02:05.262721 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gdfkv" Nov 25 11:02:05 crc kubenswrapper[4776]: I1125 11:02:05.396835 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltqcf\" (UniqueName: \"kubernetes.io/projected/b5b03bdd-3ede-487b-8403-073190b9f68b-kube-api-access-ltqcf\") pod \"b5b03bdd-3ede-487b-8403-073190b9f68b\" (UID: \"b5b03bdd-3ede-487b-8403-073190b9f68b\") " Nov 25 11:02:05 crc kubenswrapper[4776]: I1125 11:02:05.396949 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5b03bdd-3ede-487b-8403-073190b9f68b-scripts\") pod \"b5b03bdd-3ede-487b-8403-073190b9f68b\" (UID: \"b5b03bdd-3ede-487b-8403-073190b9f68b\") " Nov 25 11:02:05 crc kubenswrapper[4776]: I1125 11:02:05.397025 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5b03bdd-3ede-487b-8403-073190b9f68b-config-data\") pod \"b5b03bdd-3ede-487b-8403-073190b9f68b\" (UID: \"b5b03bdd-3ede-487b-8403-073190b9f68b\") " Nov 25 11:02:05 crc kubenswrapper[4776]: I1125 11:02:05.397106 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5b03bdd-3ede-487b-8403-073190b9f68b-logs\") pod \"b5b03bdd-3ede-487b-8403-073190b9f68b\" (UID: \"b5b03bdd-3ede-487b-8403-073190b9f68b\") " Nov 25 11:02:05 crc kubenswrapper[4776]: I1125 11:02:05.397189 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5b03bdd-3ede-487b-8403-073190b9f68b-combined-ca-bundle\") pod \"b5b03bdd-3ede-487b-8403-073190b9f68b\" (UID: \"b5b03bdd-3ede-487b-8403-073190b9f68b\") " Nov 25 11:02:05 crc kubenswrapper[4776]: I1125 11:02:05.397604 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5b03bdd-3ede-487b-8403-073190b9f68b-logs" (OuterVolumeSpecName: "logs") pod "b5b03bdd-3ede-487b-8403-073190b9f68b" (UID: "b5b03bdd-3ede-487b-8403-073190b9f68b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:02:05 crc kubenswrapper[4776]: I1125 11:02:05.402373 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5b03bdd-3ede-487b-8403-073190b9f68b-scripts" (OuterVolumeSpecName: "scripts") pod "b5b03bdd-3ede-487b-8403-073190b9f68b" (UID: "b5b03bdd-3ede-487b-8403-073190b9f68b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:02:05 crc kubenswrapper[4776]: I1125 11:02:05.402502 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5b03bdd-3ede-487b-8403-073190b9f68b-kube-api-access-ltqcf" (OuterVolumeSpecName: "kube-api-access-ltqcf") pod "b5b03bdd-3ede-487b-8403-073190b9f68b" (UID: "b5b03bdd-3ede-487b-8403-073190b9f68b"). InnerVolumeSpecName "kube-api-access-ltqcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:02:05 crc kubenswrapper[4776]: I1125 11:02:05.421197 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5b03bdd-3ede-487b-8403-073190b9f68b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b5b03bdd-3ede-487b-8403-073190b9f68b" (UID: "b5b03bdd-3ede-487b-8403-073190b9f68b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:02:05 crc kubenswrapper[4776]: I1125 11:02:05.422208 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5b03bdd-3ede-487b-8403-073190b9f68b-config-data" (OuterVolumeSpecName: "config-data") pod "b5b03bdd-3ede-487b-8403-073190b9f68b" (UID: "b5b03bdd-3ede-487b-8403-073190b9f68b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:02:05 crc kubenswrapper[4776]: I1125 11:02:05.499255 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5b03bdd-3ede-487b-8403-073190b9f68b-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:02:05 crc kubenswrapper[4776]: I1125 11:02:05.499286 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5b03bdd-3ede-487b-8403-073190b9f68b-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:02:05 crc kubenswrapper[4776]: I1125 11:02:05.499297 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5b03bdd-3ede-487b-8403-073190b9f68b-logs\") on node \"crc\" DevicePath \"\"" Nov 25 11:02:05 crc kubenswrapper[4776]: I1125 11:02:05.499307 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5b03bdd-3ede-487b-8403-073190b9f68b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:02:05 crc kubenswrapper[4776]: I1125 11:02:05.499318 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltqcf\" (UniqueName: \"kubernetes.io/projected/b5b03bdd-3ede-487b-8403-073190b9f68b-kube-api-access-ltqcf\") on node \"crc\" DevicePath \"\"" Nov 25 11:02:05 crc kubenswrapper[4776]: I1125 11:02:05.950782 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-gdfkv" event={"ID":"b5b03bdd-3ede-487b-8403-073190b9f68b","Type":"ContainerDied","Data":"e946556b39feca755a4eac2338ab08430f6e192cad39e455ba99053968756104"} Nov 25 11:02:05 crc kubenswrapper[4776]: I1125 11:02:05.950822 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e946556b39feca755a4eac2338ab08430f6e192cad39e455ba99053968756104" Nov 25 11:02:05 crc kubenswrapper[4776]: I1125 11:02:05.950839 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-gdfkv" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.036322 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5488dbdb98-xmdn2"] Nov 25 11:02:06 crc kubenswrapper[4776]: E1125 11:02:06.036936 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5b03bdd-3ede-487b-8403-073190b9f68b" containerName="placement-db-sync" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.036970 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5b03bdd-3ede-487b-8403-073190b9f68b" containerName="placement-db-sync" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.037339 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5b03bdd-3ede-487b-8403-073190b9f68b" containerName="placement-db-sync" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.038798 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5488dbdb98-xmdn2" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.054019 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5488dbdb98-xmdn2"] Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.054998 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.056415 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.056537 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-72h98" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.056685 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.057605 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.210711 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5720535b-1e84-4ca1-8a09-67ad129337f1-internal-tls-certs\") pod \"placement-5488dbdb98-xmdn2\" (UID: \"5720535b-1e84-4ca1-8a09-67ad129337f1\") " pod="openstack/placement-5488dbdb98-xmdn2" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.210789 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkslg\" (UniqueName: \"kubernetes.io/projected/5720535b-1e84-4ca1-8a09-67ad129337f1-kube-api-access-zkslg\") pod \"placement-5488dbdb98-xmdn2\" (UID: \"5720535b-1e84-4ca1-8a09-67ad129337f1\") " pod="openstack/placement-5488dbdb98-xmdn2" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.210849 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5720535b-1e84-4ca1-8a09-67ad129337f1-scripts\") pod \"placement-5488dbdb98-xmdn2\" (UID: \"5720535b-1e84-4ca1-8a09-67ad129337f1\") " pod="openstack/placement-5488dbdb98-xmdn2" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.210888 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5720535b-1e84-4ca1-8a09-67ad129337f1-logs\") pod \"placement-5488dbdb98-xmdn2\" (UID: \"5720535b-1e84-4ca1-8a09-67ad129337f1\") " pod="openstack/placement-5488dbdb98-xmdn2" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.210942 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5720535b-1e84-4ca1-8a09-67ad129337f1-config-data\") pod \"placement-5488dbdb98-xmdn2\" (UID: \"5720535b-1e84-4ca1-8a09-67ad129337f1\") " pod="openstack/placement-5488dbdb98-xmdn2" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.210968 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5720535b-1e84-4ca1-8a09-67ad129337f1-combined-ca-bundle\") pod \"placement-5488dbdb98-xmdn2\" (UID: \"5720535b-1e84-4ca1-8a09-67ad129337f1\") " pod="openstack/placement-5488dbdb98-xmdn2" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.211000 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5720535b-1e84-4ca1-8a09-67ad129337f1-public-tls-certs\") pod \"placement-5488dbdb98-xmdn2\" (UID: \"5720535b-1e84-4ca1-8a09-67ad129337f1\") " pod="openstack/placement-5488dbdb98-xmdn2" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.312197 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5720535b-1e84-4ca1-8a09-67ad129337f1-scripts\") pod \"placement-5488dbdb98-xmdn2\" (UID: \"5720535b-1e84-4ca1-8a09-67ad129337f1\") " pod="openstack/placement-5488dbdb98-xmdn2" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.312260 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5720535b-1e84-4ca1-8a09-67ad129337f1-logs\") pod \"placement-5488dbdb98-xmdn2\" (UID: \"5720535b-1e84-4ca1-8a09-67ad129337f1\") " pod="openstack/placement-5488dbdb98-xmdn2" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.312321 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5720535b-1e84-4ca1-8a09-67ad129337f1-config-data\") pod \"placement-5488dbdb98-xmdn2\" (UID: \"5720535b-1e84-4ca1-8a09-67ad129337f1\") " pod="openstack/placement-5488dbdb98-xmdn2" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.312348 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5720535b-1e84-4ca1-8a09-67ad129337f1-combined-ca-bundle\") pod \"placement-5488dbdb98-xmdn2\" (UID: \"5720535b-1e84-4ca1-8a09-67ad129337f1\") " pod="openstack/placement-5488dbdb98-xmdn2" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.312383 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5720535b-1e84-4ca1-8a09-67ad129337f1-public-tls-certs\") pod \"placement-5488dbdb98-xmdn2\" (UID: \"5720535b-1e84-4ca1-8a09-67ad129337f1\") " pod="openstack/placement-5488dbdb98-xmdn2" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.312490 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5720535b-1e84-4ca1-8a09-67ad129337f1-internal-tls-certs\") pod \"placement-5488dbdb98-xmdn2\" (UID: \"5720535b-1e84-4ca1-8a09-67ad129337f1\") " pod="openstack/placement-5488dbdb98-xmdn2" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.312521 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkslg\" (UniqueName: \"kubernetes.io/projected/5720535b-1e84-4ca1-8a09-67ad129337f1-kube-api-access-zkslg\") pod \"placement-5488dbdb98-xmdn2\" (UID: \"5720535b-1e84-4ca1-8a09-67ad129337f1\") " pod="openstack/placement-5488dbdb98-xmdn2" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.314797 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5720535b-1e84-4ca1-8a09-67ad129337f1-logs\") pod \"placement-5488dbdb98-xmdn2\" (UID: \"5720535b-1e84-4ca1-8a09-67ad129337f1\") " pod="openstack/placement-5488dbdb98-xmdn2" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.317249 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5720535b-1e84-4ca1-8a09-67ad129337f1-internal-tls-certs\") pod \"placement-5488dbdb98-xmdn2\" (UID: \"5720535b-1e84-4ca1-8a09-67ad129337f1\") " pod="openstack/placement-5488dbdb98-xmdn2" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.317605 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5720535b-1e84-4ca1-8a09-67ad129337f1-combined-ca-bundle\") pod \"placement-5488dbdb98-xmdn2\" (UID: \"5720535b-1e84-4ca1-8a09-67ad129337f1\") " pod="openstack/placement-5488dbdb98-xmdn2" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.318661 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5720535b-1e84-4ca1-8a09-67ad129337f1-config-data\") pod \"placement-5488dbdb98-xmdn2\" (UID: \"5720535b-1e84-4ca1-8a09-67ad129337f1\") " pod="openstack/placement-5488dbdb98-xmdn2" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.320733 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5720535b-1e84-4ca1-8a09-67ad129337f1-public-tls-certs\") pod \"placement-5488dbdb98-xmdn2\" (UID: \"5720535b-1e84-4ca1-8a09-67ad129337f1\") " pod="openstack/placement-5488dbdb98-xmdn2" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.329525 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5720535b-1e84-4ca1-8a09-67ad129337f1-scripts\") pod \"placement-5488dbdb98-xmdn2\" (UID: \"5720535b-1e84-4ca1-8a09-67ad129337f1\") " pod="openstack/placement-5488dbdb98-xmdn2" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.332754 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkslg\" (UniqueName: \"kubernetes.io/projected/5720535b-1e84-4ca1-8a09-67ad129337f1-kube-api-access-zkslg\") pod \"placement-5488dbdb98-xmdn2\" (UID: \"5720535b-1e84-4ca1-8a09-67ad129337f1\") " pod="openstack/placement-5488dbdb98-xmdn2" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.363436 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5488dbdb98-xmdn2" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.523709 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pck8n"] Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.528875 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pck8n" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.535259 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pck8n"] Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.618906 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/324736fc-a692-40df-ba02-46ff4240e970-utilities\") pod \"certified-operators-pck8n\" (UID: \"324736fc-a692-40df-ba02-46ff4240e970\") " pod="openshift-marketplace/certified-operators-pck8n" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.619267 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/324736fc-a692-40df-ba02-46ff4240e970-catalog-content\") pod \"certified-operators-pck8n\" (UID: \"324736fc-a692-40df-ba02-46ff4240e970\") " pod="openshift-marketplace/certified-operators-pck8n" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.619362 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72v9f\" (UniqueName: \"kubernetes.io/projected/324736fc-a692-40df-ba02-46ff4240e970-kube-api-access-72v9f\") pod \"certified-operators-pck8n\" (UID: \"324736fc-a692-40df-ba02-46ff4240e970\") " pod="openshift-marketplace/certified-operators-pck8n" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.720692 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/324736fc-a692-40df-ba02-46ff4240e970-catalog-content\") pod \"certified-operators-pck8n\" (UID: \"324736fc-a692-40df-ba02-46ff4240e970\") " pod="openshift-marketplace/certified-operators-pck8n" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.721277 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/324736fc-a692-40df-ba02-46ff4240e970-catalog-content\") pod \"certified-operators-pck8n\" (UID: \"324736fc-a692-40df-ba02-46ff4240e970\") " pod="openshift-marketplace/certified-operators-pck8n" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.721408 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72v9f\" (UniqueName: \"kubernetes.io/projected/324736fc-a692-40df-ba02-46ff4240e970-kube-api-access-72v9f\") pod \"certified-operators-pck8n\" (UID: \"324736fc-a692-40df-ba02-46ff4240e970\") " pod="openshift-marketplace/certified-operators-pck8n" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.722293 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/324736fc-a692-40df-ba02-46ff4240e970-utilities\") pod \"certified-operators-pck8n\" (UID: \"324736fc-a692-40df-ba02-46ff4240e970\") " pod="openshift-marketplace/certified-operators-pck8n" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.722761 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/324736fc-a692-40df-ba02-46ff4240e970-utilities\") pod \"certified-operators-pck8n\" (UID: \"324736fc-a692-40df-ba02-46ff4240e970\") " pod="openshift-marketplace/certified-operators-pck8n" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.744754 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72v9f\" (UniqueName: \"kubernetes.io/projected/324736fc-a692-40df-ba02-46ff4240e970-kube-api-access-72v9f\") pod \"certified-operators-pck8n\" (UID: \"324736fc-a692-40df-ba02-46ff4240e970\") " pod="openshift-marketplace/certified-operators-pck8n" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.809751 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5488dbdb98-xmdn2"] Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.856645 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pck8n" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.878046 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-674c55699c-vp7jm" Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.968863 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d8cfd89b5-6pvwz"] Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.969144 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d8cfd89b5-6pvwz" podUID="147b6859-7722-443e-b424-f8a61aee320d" containerName="dnsmasq-dns" containerID="cri-o://8d2953a60f0e316d160745c681ba01c995f50a349f09228314de9609280fded3" gracePeriod=10 Nov 25 11:02:06 crc kubenswrapper[4776]: I1125 11:02:06.996007 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5488dbdb98-xmdn2" event={"ID":"5720535b-1e84-4ca1-8a09-67ad129337f1","Type":"ContainerStarted","Data":"7908ae6d49a3840be6cc5a0f2901784dd669a62982bea4502d407d4cd2324f5d"} Nov 25 11:02:07 crc kubenswrapper[4776]: I1125 11:02:07.470883 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pck8n"] Nov 25 11:02:08 crc kubenswrapper[4776]: I1125 11:02:08.016698 4776 generic.go:334] "Generic (PLEG): container finished" podID="324736fc-a692-40df-ba02-46ff4240e970" containerID="3ead85cec1b564a0baf1855cf472132f528aa3815696d7abeaf0725780c14d38" exitCode=0 Nov 25 11:02:08 crc kubenswrapper[4776]: I1125 11:02:08.017091 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pck8n" event={"ID":"324736fc-a692-40df-ba02-46ff4240e970","Type":"ContainerDied","Data":"3ead85cec1b564a0baf1855cf472132f528aa3815696d7abeaf0725780c14d38"} Nov 25 11:02:08 crc kubenswrapper[4776]: I1125 11:02:08.017153 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pck8n" event={"ID":"324736fc-a692-40df-ba02-46ff4240e970","Type":"ContainerStarted","Data":"f9399793f8000882deb464a021bd5664627e05b723ef1d764ce6c8bf9135b498"} Nov 25 11:02:08 crc kubenswrapper[4776]: I1125 11:02:08.023777 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5488dbdb98-xmdn2" event={"ID":"5720535b-1e84-4ca1-8a09-67ad129337f1","Type":"ContainerStarted","Data":"509d6fb47fe2f80aa4f96ace97c3b7e8e911f1e9717213d8fb6f73a5f942f571"} Nov 25 11:02:08 crc kubenswrapper[4776]: I1125 11:02:08.023818 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5488dbdb98-xmdn2" event={"ID":"5720535b-1e84-4ca1-8a09-67ad129337f1","Type":"ContainerStarted","Data":"3817589200b730c4edca543b636609e380bd8dd21c884162082e57e4c9f98e97"} Nov 25 11:02:08 crc kubenswrapper[4776]: I1125 11:02:08.024887 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5488dbdb98-xmdn2" Nov 25 11:02:08 crc kubenswrapper[4776]: I1125 11:02:08.024907 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5488dbdb98-xmdn2" Nov 25 11:02:08 crc kubenswrapper[4776]: I1125 11:02:08.026849 4776 generic.go:334] "Generic (PLEG): container finished" podID="147b6859-7722-443e-b424-f8a61aee320d" containerID="8d2953a60f0e316d160745c681ba01c995f50a349f09228314de9609280fded3" exitCode=0 Nov 25 11:02:08 crc kubenswrapper[4776]: I1125 11:02:08.026885 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d8cfd89b5-6pvwz" event={"ID":"147b6859-7722-443e-b424-f8a61aee320d","Type":"ContainerDied","Data":"8d2953a60f0e316d160745c681ba01c995f50a349f09228314de9609280fded3"} Nov 25 11:02:08 crc kubenswrapper[4776]: I1125 11:02:08.061739 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-5488dbdb98-xmdn2" podStartSLOduration=2.061724235 podStartE2EDuration="2.061724235s" podCreationTimestamp="2025-11-25 11:02:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:02:08.050335019 +0000 UTC m=+5873.091394572" watchObservedRunningTime="2025-11-25 11:02:08.061724235 +0000 UTC m=+5873.102783788" Nov 25 11:02:08 crc kubenswrapper[4776]: I1125 11:02:08.074385 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d8cfd89b5-6pvwz" Nov 25 11:02:08 crc kubenswrapper[4776]: I1125 11:02:08.259686 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mgqp\" (UniqueName: \"kubernetes.io/projected/147b6859-7722-443e-b424-f8a61aee320d-kube-api-access-5mgqp\") pod \"147b6859-7722-443e-b424-f8a61aee320d\" (UID: \"147b6859-7722-443e-b424-f8a61aee320d\") " Nov 25 11:02:08 crc kubenswrapper[4776]: I1125 11:02:08.259884 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/147b6859-7722-443e-b424-f8a61aee320d-config\") pod \"147b6859-7722-443e-b424-f8a61aee320d\" (UID: \"147b6859-7722-443e-b424-f8a61aee320d\") " Nov 25 11:02:08 crc kubenswrapper[4776]: I1125 11:02:08.259916 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/147b6859-7722-443e-b424-f8a61aee320d-ovsdbserver-nb\") pod \"147b6859-7722-443e-b424-f8a61aee320d\" (UID: \"147b6859-7722-443e-b424-f8a61aee320d\") " Nov 25 11:02:08 crc kubenswrapper[4776]: I1125 11:02:08.260014 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/147b6859-7722-443e-b424-f8a61aee320d-dns-svc\") pod \"147b6859-7722-443e-b424-f8a61aee320d\" (UID: \"147b6859-7722-443e-b424-f8a61aee320d\") " Nov 25 11:02:08 crc kubenswrapper[4776]: I1125 11:02:08.260052 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/147b6859-7722-443e-b424-f8a61aee320d-ovsdbserver-sb\") pod \"147b6859-7722-443e-b424-f8a61aee320d\" (UID: \"147b6859-7722-443e-b424-f8a61aee320d\") " Nov 25 11:02:08 crc kubenswrapper[4776]: I1125 11:02:08.265300 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/147b6859-7722-443e-b424-f8a61aee320d-kube-api-access-5mgqp" (OuterVolumeSpecName: "kube-api-access-5mgqp") pod "147b6859-7722-443e-b424-f8a61aee320d" (UID: "147b6859-7722-443e-b424-f8a61aee320d"). InnerVolumeSpecName "kube-api-access-5mgqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:02:08 crc kubenswrapper[4776]: I1125 11:02:08.314529 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/147b6859-7722-443e-b424-f8a61aee320d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "147b6859-7722-443e-b424-f8a61aee320d" (UID: "147b6859-7722-443e-b424-f8a61aee320d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:02:08 crc kubenswrapper[4776]: I1125 11:02:08.318622 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/147b6859-7722-443e-b424-f8a61aee320d-config" (OuterVolumeSpecName: "config") pod "147b6859-7722-443e-b424-f8a61aee320d" (UID: "147b6859-7722-443e-b424-f8a61aee320d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:02:08 crc kubenswrapper[4776]: I1125 11:02:08.327568 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/147b6859-7722-443e-b424-f8a61aee320d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "147b6859-7722-443e-b424-f8a61aee320d" (UID: "147b6859-7722-443e-b424-f8a61aee320d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:02:08 crc kubenswrapper[4776]: I1125 11:02:08.335878 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/147b6859-7722-443e-b424-f8a61aee320d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "147b6859-7722-443e-b424-f8a61aee320d" (UID: "147b6859-7722-443e-b424-f8a61aee320d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:02:08 crc kubenswrapper[4776]: I1125 11:02:08.362240 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/147b6859-7722-443e-b424-f8a61aee320d-config\") on node \"crc\" DevicePath \"\"" Nov 25 11:02:08 crc kubenswrapper[4776]: I1125 11:02:08.362277 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/147b6859-7722-443e-b424-f8a61aee320d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 11:02:08 crc kubenswrapper[4776]: I1125 11:02:08.362292 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/147b6859-7722-443e-b424-f8a61aee320d-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 11:02:08 crc kubenswrapper[4776]: I1125 11:02:08.362302 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/147b6859-7722-443e-b424-f8a61aee320d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 11:02:08 crc kubenswrapper[4776]: I1125 11:02:08.362314 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mgqp\" (UniqueName: \"kubernetes.io/projected/147b6859-7722-443e-b424-f8a61aee320d-kube-api-access-5mgqp\") on node \"crc\" DevicePath \"\"" Nov 25 11:02:09 crc kubenswrapper[4776]: I1125 11:02:09.038913 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d8cfd89b5-6pvwz" Nov 25 11:02:09 crc kubenswrapper[4776]: I1125 11:02:09.038930 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d8cfd89b5-6pvwz" event={"ID":"147b6859-7722-443e-b424-f8a61aee320d","Type":"ContainerDied","Data":"0cee2b1f82b705f156e44c3e33fc3a4aabb4bbe0f1919ccce8cff31dcbda07d9"} Nov 25 11:02:09 crc kubenswrapper[4776]: I1125 11:02:09.039395 4776 scope.go:117] "RemoveContainer" containerID="8d2953a60f0e316d160745c681ba01c995f50a349f09228314de9609280fded3" Nov 25 11:02:09 crc kubenswrapper[4776]: I1125 11:02:09.081297 4776 scope.go:117] "RemoveContainer" containerID="d483500b8db3124e16840bcd7d8ccc12156fa86f6634181067004cbea2149103" Nov 25 11:02:09 crc kubenswrapper[4776]: I1125 11:02:09.083451 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d8cfd89b5-6pvwz"] Nov 25 11:02:09 crc kubenswrapper[4776]: I1125 11:02:09.097891 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d8cfd89b5-6pvwz"] Nov 25 11:02:09 crc kubenswrapper[4776]: I1125 11:02:09.687179 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="147b6859-7722-443e-b424-f8a61aee320d" path="/var/lib/kubelet/pods/147b6859-7722-443e-b424-f8a61aee320d/volumes" Nov 25 11:02:10 crc kubenswrapper[4776]: I1125 11:02:10.055324 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pck8n" event={"ID":"324736fc-a692-40df-ba02-46ff4240e970","Type":"ContainerStarted","Data":"9fbe2fbef0227728290f39423df201f6670c53396a05a431366a978187b92edf"} Nov 25 11:02:11 crc kubenswrapper[4776]: I1125 11:02:11.073694 4776 generic.go:334] "Generic (PLEG): container finished" podID="324736fc-a692-40df-ba02-46ff4240e970" containerID="9fbe2fbef0227728290f39423df201f6670c53396a05a431366a978187b92edf" exitCode=0 Nov 25 11:02:11 crc kubenswrapper[4776]: I1125 11:02:11.073770 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pck8n" event={"ID":"324736fc-a692-40df-ba02-46ff4240e970","Type":"ContainerDied","Data":"9fbe2fbef0227728290f39423df201f6670c53396a05a431366a978187b92edf"} Nov 25 11:02:13 crc kubenswrapper[4776]: I1125 11:02:13.099524 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pck8n" event={"ID":"324736fc-a692-40df-ba02-46ff4240e970","Type":"ContainerStarted","Data":"3bc2e4acc542ba30c628b2623e83d7c0bb9a0f2d88dc635a6922d694691bb39b"} Nov 25 11:02:13 crc kubenswrapper[4776]: I1125 11:02:13.122150 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pck8n" podStartSLOduration=3.261472203 podStartE2EDuration="7.122135914s" podCreationTimestamp="2025-11-25 11:02:06 +0000 UTC" firstStartedPulling="2025-11-25 11:02:08.020951581 +0000 UTC m=+5873.062011134" lastFinishedPulling="2025-11-25 11:02:11.881615262 +0000 UTC m=+5876.922674845" observedRunningTime="2025-11-25 11:02:13.119062027 +0000 UTC m=+5878.160121580" watchObservedRunningTime="2025-11-25 11:02:13.122135914 +0000 UTC m=+5878.163195467" Nov 25 11:02:16 crc kubenswrapper[4776]: I1125 11:02:16.857551 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pck8n" Nov 25 11:02:16 crc kubenswrapper[4776]: I1125 11:02:16.858399 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pck8n" Nov 25 11:02:16 crc kubenswrapper[4776]: I1125 11:02:16.908635 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pck8n" Nov 25 11:02:17 crc kubenswrapper[4776]: I1125 11:02:17.196290 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pck8n" Nov 25 11:02:17 crc kubenswrapper[4776]: I1125 11:02:17.901887 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pck8n"] Nov 25 11:02:19 crc kubenswrapper[4776]: I1125 11:02:19.159977 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pck8n" podUID="324736fc-a692-40df-ba02-46ff4240e970" containerName="registry-server" containerID="cri-o://3bc2e4acc542ba30c628b2623e83d7c0bb9a0f2d88dc635a6922d694691bb39b" gracePeriod=2 Nov 25 11:02:19 crc kubenswrapper[4776]: I1125 11:02:19.631322 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pck8n" Nov 25 11:02:19 crc kubenswrapper[4776]: I1125 11:02:19.806829 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/324736fc-a692-40df-ba02-46ff4240e970-catalog-content\") pod \"324736fc-a692-40df-ba02-46ff4240e970\" (UID: \"324736fc-a692-40df-ba02-46ff4240e970\") " Nov 25 11:02:19 crc kubenswrapper[4776]: I1125 11:02:19.806948 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/324736fc-a692-40df-ba02-46ff4240e970-utilities\") pod \"324736fc-a692-40df-ba02-46ff4240e970\" (UID: \"324736fc-a692-40df-ba02-46ff4240e970\") " Nov 25 11:02:19 crc kubenswrapper[4776]: I1125 11:02:19.806974 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72v9f\" (UniqueName: \"kubernetes.io/projected/324736fc-a692-40df-ba02-46ff4240e970-kube-api-access-72v9f\") pod \"324736fc-a692-40df-ba02-46ff4240e970\" (UID: \"324736fc-a692-40df-ba02-46ff4240e970\") " Nov 25 11:02:19 crc kubenswrapper[4776]: I1125 11:02:19.808032 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/324736fc-a692-40df-ba02-46ff4240e970-utilities" (OuterVolumeSpecName: "utilities") pod "324736fc-a692-40df-ba02-46ff4240e970" (UID: "324736fc-a692-40df-ba02-46ff4240e970"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:02:19 crc kubenswrapper[4776]: I1125 11:02:19.812806 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/324736fc-a692-40df-ba02-46ff4240e970-kube-api-access-72v9f" (OuterVolumeSpecName: "kube-api-access-72v9f") pod "324736fc-a692-40df-ba02-46ff4240e970" (UID: "324736fc-a692-40df-ba02-46ff4240e970"). InnerVolumeSpecName "kube-api-access-72v9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:02:19 crc kubenswrapper[4776]: I1125 11:02:19.864548 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/324736fc-a692-40df-ba02-46ff4240e970-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "324736fc-a692-40df-ba02-46ff4240e970" (UID: "324736fc-a692-40df-ba02-46ff4240e970"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:02:19 crc kubenswrapper[4776]: I1125 11:02:19.909759 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/324736fc-a692-40df-ba02-46ff4240e970-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:02:19 crc kubenswrapper[4776]: I1125 11:02:19.909810 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/324736fc-a692-40df-ba02-46ff4240e970-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:02:19 crc kubenswrapper[4776]: I1125 11:02:19.909830 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72v9f\" (UniqueName: \"kubernetes.io/projected/324736fc-a692-40df-ba02-46ff4240e970-kube-api-access-72v9f\") on node \"crc\" DevicePath \"\"" Nov 25 11:02:20 crc kubenswrapper[4776]: I1125 11:02:20.171401 4776 generic.go:334] "Generic (PLEG): container finished" podID="324736fc-a692-40df-ba02-46ff4240e970" containerID="3bc2e4acc542ba30c628b2623e83d7c0bb9a0f2d88dc635a6922d694691bb39b" exitCode=0 Nov 25 11:02:20 crc kubenswrapper[4776]: I1125 11:02:20.171447 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pck8n" event={"ID":"324736fc-a692-40df-ba02-46ff4240e970","Type":"ContainerDied","Data":"3bc2e4acc542ba30c628b2623e83d7c0bb9a0f2d88dc635a6922d694691bb39b"} Nov 25 11:02:20 crc kubenswrapper[4776]: I1125 11:02:20.171483 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pck8n" Nov 25 11:02:20 crc kubenswrapper[4776]: I1125 11:02:20.171508 4776 scope.go:117] "RemoveContainer" containerID="3bc2e4acc542ba30c628b2623e83d7c0bb9a0f2d88dc635a6922d694691bb39b" Nov 25 11:02:20 crc kubenswrapper[4776]: I1125 11:02:20.171495 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pck8n" event={"ID":"324736fc-a692-40df-ba02-46ff4240e970","Type":"ContainerDied","Data":"f9399793f8000882deb464a021bd5664627e05b723ef1d764ce6c8bf9135b498"} Nov 25 11:02:20 crc kubenswrapper[4776]: I1125 11:02:20.203438 4776 scope.go:117] "RemoveContainer" containerID="9fbe2fbef0227728290f39423df201f6670c53396a05a431366a978187b92edf" Nov 25 11:02:20 crc kubenswrapper[4776]: I1125 11:02:20.221207 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pck8n"] Nov 25 11:02:20 crc kubenswrapper[4776]: I1125 11:02:20.228041 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pck8n"] Nov 25 11:02:20 crc kubenswrapper[4776]: I1125 11:02:20.234142 4776 scope.go:117] "RemoveContainer" containerID="3ead85cec1b564a0baf1855cf472132f528aa3815696d7abeaf0725780c14d38" Nov 25 11:02:20 crc kubenswrapper[4776]: I1125 11:02:20.284101 4776 scope.go:117] "RemoveContainer" containerID="3bc2e4acc542ba30c628b2623e83d7c0bb9a0f2d88dc635a6922d694691bb39b" Nov 25 11:02:20 crc kubenswrapper[4776]: E1125 11:02:20.284750 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bc2e4acc542ba30c628b2623e83d7c0bb9a0f2d88dc635a6922d694691bb39b\": container with ID starting with 3bc2e4acc542ba30c628b2623e83d7c0bb9a0f2d88dc635a6922d694691bb39b not found: ID does not exist" containerID="3bc2e4acc542ba30c628b2623e83d7c0bb9a0f2d88dc635a6922d694691bb39b" Nov 25 11:02:20 crc kubenswrapper[4776]: I1125 11:02:20.284839 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bc2e4acc542ba30c628b2623e83d7c0bb9a0f2d88dc635a6922d694691bb39b"} err="failed to get container status \"3bc2e4acc542ba30c628b2623e83d7c0bb9a0f2d88dc635a6922d694691bb39b\": rpc error: code = NotFound desc = could not find container \"3bc2e4acc542ba30c628b2623e83d7c0bb9a0f2d88dc635a6922d694691bb39b\": container with ID starting with 3bc2e4acc542ba30c628b2623e83d7c0bb9a0f2d88dc635a6922d694691bb39b not found: ID does not exist" Nov 25 11:02:20 crc kubenswrapper[4776]: I1125 11:02:20.284946 4776 scope.go:117] "RemoveContainer" containerID="9fbe2fbef0227728290f39423df201f6670c53396a05a431366a978187b92edf" Nov 25 11:02:20 crc kubenswrapper[4776]: E1125 11:02:20.285402 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fbe2fbef0227728290f39423df201f6670c53396a05a431366a978187b92edf\": container with ID starting with 9fbe2fbef0227728290f39423df201f6670c53396a05a431366a978187b92edf not found: ID does not exist" containerID="9fbe2fbef0227728290f39423df201f6670c53396a05a431366a978187b92edf" Nov 25 11:02:20 crc kubenswrapper[4776]: I1125 11:02:20.285492 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fbe2fbef0227728290f39423df201f6670c53396a05a431366a978187b92edf"} err="failed to get container status \"9fbe2fbef0227728290f39423df201f6670c53396a05a431366a978187b92edf\": rpc error: code = NotFound desc = could not find container \"9fbe2fbef0227728290f39423df201f6670c53396a05a431366a978187b92edf\": container with ID starting with 9fbe2fbef0227728290f39423df201f6670c53396a05a431366a978187b92edf not found: ID does not exist" Nov 25 11:02:20 crc kubenswrapper[4776]: I1125 11:02:20.285576 4776 scope.go:117] "RemoveContainer" containerID="3ead85cec1b564a0baf1855cf472132f528aa3815696d7abeaf0725780c14d38" Nov 25 11:02:20 crc kubenswrapper[4776]: E1125 11:02:20.285856 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ead85cec1b564a0baf1855cf472132f528aa3815696d7abeaf0725780c14d38\": container with ID starting with 3ead85cec1b564a0baf1855cf472132f528aa3815696d7abeaf0725780c14d38 not found: ID does not exist" containerID="3ead85cec1b564a0baf1855cf472132f528aa3815696d7abeaf0725780c14d38" Nov 25 11:02:20 crc kubenswrapper[4776]: I1125 11:02:20.285879 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ead85cec1b564a0baf1855cf472132f528aa3815696d7abeaf0725780c14d38"} err="failed to get container status \"3ead85cec1b564a0baf1855cf472132f528aa3815696d7abeaf0725780c14d38\": rpc error: code = NotFound desc = could not find container \"3ead85cec1b564a0baf1855cf472132f528aa3815696d7abeaf0725780c14d38\": container with ID starting with 3ead85cec1b564a0baf1855cf472132f528aa3815696d7abeaf0725780c14d38 not found: ID does not exist" Nov 25 11:02:21 crc kubenswrapper[4776]: I1125 11:02:21.676788 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="324736fc-a692-40df-ba02-46ff4240e970" path="/var/lib/kubelet/pods/324736fc-a692-40df-ba02-46ff4240e970/volumes" Nov 25 11:02:37 crc kubenswrapper[4776]: I1125 11:02:37.398619 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5488dbdb98-xmdn2" Nov 25 11:02:38 crc kubenswrapper[4776]: I1125 11:02:38.415905 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5488dbdb98-xmdn2" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.194713 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-8flmt"] Nov 25 11:02:59 crc kubenswrapper[4776]: E1125 11:02:59.195829 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="324736fc-a692-40df-ba02-46ff4240e970" containerName="extract-utilities" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.195849 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="324736fc-a692-40df-ba02-46ff4240e970" containerName="extract-utilities" Nov 25 11:02:59 crc kubenswrapper[4776]: E1125 11:02:59.195867 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="324736fc-a692-40df-ba02-46ff4240e970" containerName="registry-server" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.195876 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="324736fc-a692-40df-ba02-46ff4240e970" containerName="registry-server" Nov 25 11:02:59 crc kubenswrapper[4776]: E1125 11:02:59.195888 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="147b6859-7722-443e-b424-f8a61aee320d" containerName="dnsmasq-dns" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.195900 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="147b6859-7722-443e-b424-f8a61aee320d" containerName="dnsmasq-dns" Nov 25 11:02:59 crc kubenswrapper[4776]: E1125 11:02:59.195934 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="147b6859-7722-443e-b424-f8a61aee320d" containerName="init" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.195943 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="147b6859-7722-443e-b424-f8a61aee320d" containerName="init" Nov 25 11:02:59 crc kubenswrapper[4776]: E1125 11:02:59.195961 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="324736fc-a692-40df-ba02-46ff4240e970" containerName="extract-content" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.195968 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="324736fc-a692-40df-ba02-46ff4240e970" containerName="extract-content" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.196237 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="324736fc-a692-40df-ba02-46ff4240e970" containerName="registry-server" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.196262 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="147b6859-7722-443e-b424-f8a61aee320d" containerName="dnsmasq-dns" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.197104 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8flmt" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.213732 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-8flmt"] Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.394200 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-2x8x7"] Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.395158 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbk9k\" (UniqueName: \"kubernetes.io/projected/4988892b-e67d-4595-abb3-c097fa0808b5-kube-api-access-hbk9k\") pod \"nova-api-db-create-8flmt\" (UID: \"4988892b-e67d-4595-abb3-c097fa0808b5\") " pod="openstack/nova-api-db-create-8flmt" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.395250 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4988892b-e67d-4595-abb3-c097fa0808b5-operator-scripts\") pod \"nova-api-db-create-8flmt\" (UID: \"4988892b-e67d-4595-abb3-c097fa0808b5\") " pod="openstack/nova-api-db-create-8flmt" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.395736 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-2x8x7" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.404328 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-173a-account-create-2vr9z"] Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.406045 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-173a-account-create-2vr9z" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.411033 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.414522 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-2x8x7"] Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.424607 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-173a-account-create-2vr9z"] Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.494359 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-jj7jb"] Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.495599 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jj7jb" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.497942 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztp9m\" (UniqueName: \"kubernetes.io/projected/96dba744-9bb8-480c-8498-c8eedd4fdb88-kube-api-access-ztp9m\") pod \"nova-cell1-db-create-jj7jb\" (UID: \"96dba744-9bb8-480c-8498-c8eedd4fdb88\") " pod="openstack/nova-cell1-db-create-jj7jb" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.498026 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9ae59af-54ec-4b1d-8285-a248f0b5f50a-operator-scripts\") pod \"nova-api-173a-account-create-2vr9z\" (UID: \"a9ae59af-54ec-4b1d-8285-a248f0b5f50a\") " pod="openstack/nova-api-173a-account-create-2vr9z" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.498082 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6d7ec58-a5a8-48aa-920d-d7f81e603f76-operator-scripts\") pod \"nova-cell0-db-create-2x8x7\" (UID: \"d6d7ec58-a5a8-48aa-920d-d7f81e603f76\") " pod="openstack/nova-cell0-db-create-2x8x7" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.498154 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7s2xc\" (UniqueName: \"kubernetes.io/projected/a9ae59af-54ec-4b1d-8285-a248f0b5f50a-kube-api-access-7s2xc\") pod \"nova-api-173a-account-create-2vr9z\" (UID: \"a9ae59af-54ec-4b1d-8285-a248f0b5f50a\") " pod="openstack/nova-api-173a-account-create-2vr9z" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.498176 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pczmg\" (UniqueName: \"kubernetes.io/projected/d6d7ec58-a5a8-48aa-920d-d7f81e603f76-kube-api-access-pczmg\") pod \"nova-cell0-db-create-2x8x7\" (UID: \"d6d7ec58-a5a8-48aa-920d-d7f81e603f76\") " pod="openstack/nova-cell0-db-create-2x8x7" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.498250 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96dba744-9bb8-480c-8498-c8eedd4fdb88-operator-scripts\") pod \"nova-cell1-db-create-jj7jb\" (UID: \"96dba744-9bb8-480c-8498-c8eedd4fdb88\") " pod="openstack/nova-cell1-db-create-jj7jb" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.498372 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbk9k\" (UniqueName: \"kubernetes.io/projected/4988892b-e67d-4595-abb3-c097fa0808b5-kube-api-access-hbk9k\") pod \"nova-api-db-create-8flmt\" (UID: \"4988892b-e67d-4595-abb3-c097fa0808b5\") " pod="openstack/nova-api-db-create-8flmt" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.498680 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4988892b-e67d-4595-abb3-c097fa0808b5-operator-scripts\") pod \"nova-api-db-create-8flmt\" (UID: \"4988892b-e67d-4595-abb3-c097fa0808b5\") " pod="openstack/nova-api-db-create-8flmt" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.499920 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4988892b-e67d-4595-abb3-c097fa0808b5-operator-scripts\") pod \"nova-api-db-create-8flmt\" (UID: \"4988892b-e67d-4595-abb3-c097fa0808b5\") " pod="openstack/nova-api-db-create-8flmt" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.503409 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-jj7jb"] Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.520005 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbk9k\" (UniqueName: \"kubernetes.io/projected/4988892b-e67d-4595-abb3-c097fa0808b5-kube-api-access-hbk9k\") pod \"nova-api-db-create-8flmt\" (UID: \"4988892b-e67d-4595-abb3-c097fa0808b5\") " pod="openstack/nova-api-db-create-8flmt" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.520617 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8flmt" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.599675 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztp9m\" (UniqueName: \"kubernetes.io/projected/96dba744-9bb8-480c-8498-c8eedd4fdb88-kube-api-access-ztp9m\") pod \"nova-cell1-db-create-jj7jb\" (UID: \"96dba744-9bb8-480c-8498-c8eedd4fdb88\") " pod="openstack/nova-cell1-db-create-jj7jb" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.600503 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9ae59af-54ec-4b1d-8285-a248f0b5f50a-operator-scripts\") pod \"nova-api-173a-account-create-2vr9z\" (UID: \"a9ae59af-54ec-4b1d-8285-a248f0b5f50a\") " pod="openstack/nova-api-173a-account-create-2vr9z" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.600529 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6d7ec58-a5a8-48aa-920d-d7f81e603f76-operator-scripts\") pod \"nova-cell0-db-create-2x8x7\" (UID: \"d6d7ec58-a5a8-48aa-920d-d7f81e603f76\") " pod="openstack/nova-cell0-db-create-2x8x7" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.600561 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7s2xc\" (UniqueName: \"kubernetes.io/projected/a9ae59af-54ec-4b1d-8285-a248f0b5f50a-kube-api-access-7s2xc\") pod \"nova-api-173a-account-create-2vr9z\" (UID: \"a9ae59af-54ec-4b1d-8285-a248f0b5f50a\") " pod="openstack/nova-api-173a-account-create-2vr9z" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.600579 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pczmg\" (UniqueName: \"kubernetes.io/projected/d6d7ec58-a5a8-48aa-920d-d7f81e603f76-kube-api-access-pczmg\") pod \"nova-cell0-db-create-2x8x7\" (UID: \"d6d7ec58-a5a8-48aa-920d-d7f81e603f76\") " pod="openstack/nova-cell0-db-create-2x8x7" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.600612 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96dba744-9bb8-480c-8498-c8eedd4fdb88-operator-scripts\") pod \"nova-cell1-db-create-jj7jb\" (UID: \"96dba744-9bb8-480c-8498-c8eedd4fdb88\") " pod="openstack/nova-cell1-db-create-jj7jb" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.601351 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96dba744-9bb8-480c-8498-c8eedd4fdb88-operator-scripts\") pod \"nova-cell1-db-create-jj7jb\" (UID: \"96dba744-9bb8-480c-8498-c8eedd4fdb88\") " pod="openstack/nova-cell1-db-create-jj7jb" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.601841 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9ae59af-54ec-4b1d-8285-a248f0b5f50a-operator-scripts\") pod \"nova-api-173a-account-create-2vr9z\" (UID: \"a9ae59af-54ec-4b1d-8285-a248f0b5f50a\") " pod="openstack/nova-api-173a-account-create-2vr9z" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.602312 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6d7ec58-a5a8-48aa-920d-d7f81e603f76-operator-scripts\") pod \"nova-cell0-db-create-2x8x7\" (UID: \"d6d7ec58-a5a8-48aa-920d-d7f81e603f76\") " pod="openstack/nova-cell0-db-create-2x8x7" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.617123 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-e3f8-account-create-ht27m"] Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.618554 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e3f8-account-create-ht27m" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.627482 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.638855 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pczmg\" (UniqueName: \"kubernetes.io/projected/d6d7ec58-a5a8-48aa-920d-d7f81e603f76-kube-api-access-pczmg\") pod \"nova-cell0-db-create-2x8x7\" (UID: \"d6d7ec58-a5a8-48aa-920d-d7f81e603f76\") " pod="openstack/nova-cell0-db-create-2x8x7" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.638855 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztp9m\" (UniqueName: \"kubernetes.io/projected/96dba744-9bb8-480c-8498-c8eedd4fdb88-kube-api-access-ztp9m\") pod \"nova-cell1-db-create-jj7jb\" (UID: \"96dba744-9bb8-480c-8498-c8eedd4fdb88\") " pod="openstack/nova-cell1-db-create-jj7jb" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.639373 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7s2xc\" (UniqueName: \"kubernetes.io/projected/a9ae59af-54ec-4b1d-8285-a248f0b5f50a-kube-api-access-7s2xc\") pod \"nova-api-173a-account-create-2vr9z\" (UID: \"a9ae59af-54ec-4b1d-8285-a248f0b5f50a\") " pod="openstack/nova-api-173a-account-create-2vr9z" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.645867 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-e3f8-account-create-ht27m"] Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.701955 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/38ec4382-1ef1-4cab-af5b-26ffecd9354a-operator-scripts\") pod \"nova-cell0-e3f8-account-create-ht27m\" (UID: \"38ec4382-1ef1-4cab-af5b-26ffecd9354a\") " pod="openstack/nova-cell0-e3f8-account-create-ht27m" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.702232 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhlb8\" (UniqueName: \"kubernetes.io/projected/38ec4382-1ef1-4cab-af5b-26ffecd9354a-kube-api-access-qhlb8\") pod \"nova-cell0-e3f8-account-create-ht27m\" (UID: \"38ec4382-1ef1-4cab-af5b-26ffecd9354a\") " pod="openstack/nova-cell0-e3f8-account-create-ht27m" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.711480 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-2x8x7" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.733162 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-173a-account-create-2vr9z" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.800381 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-0d1a-account-create-mhvkn"] Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.801830 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-0d1a-account-create-mhvkn" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.804625 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bskx4\" (UniqueName: \"kubernetes.io/projected/2c7b6238-2ba3-4d7f-a7f2-8f74092ce260-kube-api-access-bskx4\") pod \"nova-cell1-0d1a-account-create-mhvkn\" (UID: \"2c7b6238-2ba3-4d7f-a7f2-8f74092ce260\") " pod="openstack/nova-cell1-0d1a-account-create-mhvkn" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.804719 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhlb8\" (UniqueName: \"kubernetes.io/projected/38ec4382-1ef1-4cab-af5b-26ffecd9354a-kube-api-access-qhlb8\") pod \"nova-cell0-e3f8-account-create-ht27m\" (UID: \"38ec4382-1ef1-4cab-af5b-26ffecd9354a\") " pod="openstack/nova-cell0-e3f8-account-create-ht27m" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.804760 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c7b6238-2ba3-4d7f-a7f2-8f74092ce260-operator-scripts\") pod \"nova-cell1-0d1a-account-create-mhvkn\" (UID: \"2c7b6238-2ba3-4d7f-a7f2-8f74092ce260\") " pod="openstack/nova-cell1-0d1a-account-create-mhvkn" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.804815 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/38ec4382-1ef1-4cab-af5b-26ffecd9354a-operator-scripts\") pod \"nova-cell0-e3f8-account-create-ht27m\" (UID: \"38ec4382-1ef1-4cab-af5b-26ffecd9354a\") " pod="openstack/nova-cell0-e3f8-account-create-ht27m" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.806209 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/38ec4382-1ef1-4cab-af5b-26ffecd9354a-operator-scripts\") pod \"nova-cell0-e3f8-account-create-ht27m\" (UID: \"38ec4382-1ef1-4cab-af5b-26ffecd9354a\") " pod="openstack/nova-cell0-e3f8-account-create-ht27m" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.807616 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.812521 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jj7jb" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.820084 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-0d1a-account-create-mhvkn"] Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.828522 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhlb8\" (UniqueName: \"kubernetes.io/projected/38ec4382-1ef1-4cab-af5b-26ffecd9354a-kube-api-access-qhlb8\") pod \"nova-cell0-e3f8-account-create-ht27m\" (UID: \"38ec4382-1ef1-4cab-af5b-26ffecd9354a\") " pod="openstack/nova-cell0-e3f8-account-create-ht27m" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.905959 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bskx4\" (UniqueName: \"kubernetes.io/projected/2c7b6238-2ba3-4d7f-a7f2-8f74092ce260-kube-api-access-bskx4\") pod \"nova-cell1-0d1a-account-create-mhvkn\" (UID: \"2c7b6238-2ba3-4d7f-a7f2-8f74092ce260\") " pod="openstack/nova-cell1-0d1a-account-create-mhvkn" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.906088 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c7b6238-2ba3-4d7f-a7f2-8f74092ce260-operator-scripts\") pod \"nova-cell1-0d1a-account-create-mhvkn\" (UID: \"2c7b6238-2ba3-4d7f-a7f2-8f74092ce260\") " pod="openstack/nova-cell1-0d1a-account-create-mhvkn" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.906817 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c7b6238-2ba3-4d7f-a7f2-8f74092ce260-operator-scripts\") pod \"nova-cell1-0d1a-account-create-mhvkn\" (UID: \"2c7b6238-2ba3-4d7f-a7f2-8f74092ce260\") " pod="openstack/nova-cell1-0d1a-account-create-mhvkn" Nov 25 11:02:59 crc kubenswrapper[4776]: I1125 11:02:59.926744 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bskx4\" (UniqueName: \"kubernetes.io/projected/2c7b6238-2ba3-4d7f-a7f2-8f74092ce260-kube-api-access-bskx4\") pod \"nova-cell1-0d1a-account-create-mhvkn\" (UID: \"2c7b6238-2ba3-4d7f-a7f2-8f74092ce260\") " pod="openstack/nova-cell1-0d1a-account-create-mhvkn" Nov 25 11:03:00 crc kubenswrapper[4776]: I1125 11:03:00.003377 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-8flmt"] Nov 25 11:03:00 crc kubenswrapper[4776]: I1125 11:03:00.038733 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e3f8-account-create-ht27m" Nov 25 11:03:00 crc kubenswrapper[4776]: I1125 11:03:00.125123 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-0d1a-account-create-mhvkn" Nov 25 11:03:00 crc kubenswrapper[4776]: I1125 11:03:00.310269 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-2x8x7"] Nov 25 11:03:00 crc kubenswrapper[4776]: I1125 11:03:00.357901 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-173a-account-create-2vr9z"] Nov 25 11:03:00 crc kubenswrapper[4776]: W1125 11:03:00.364862 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9ae59af_54ec_4b1d_8285_a248f0b5f50a.slice/crio-097edc68524d81ee6dc49c048cce2375296e1ae90a743d94eb0c0f60b8f28096 WatchSource:0}: Error finding container 097edc68524d81ee6dc49c048cce2375296e1ae90a743d94eb0c0f60b8f28096: Status 404 returned error can't find the container with id 097edc68524d81ee6dc49c048cce2375296e1ae90a743d94eb0c0f60b8f28096 Nov 25 11:03:00 crc kubenswrapper[4776]: I1125 11:03:00.556787 4776 generic.go:334] "Generic (PLEG): container finished" podID="4988892b-e67d-4595-abb3-c097fa0808b5" containerID="9f39dda292ef1dcaae6182d4beeb52ce0f0134af0a782ad5e3a3ee0df193fa6f" exitCode=0 Nov 25 11:03:00 crc kubenswrapper[4776]: I1125 11:03:00.556846 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8flmt" event={"ID":"4988892b-e67d-4595-abb3-c097fa0808b5","Type":"ContainerDied","Data":"9f39dda292ef1dcaae6182d4beeb52ce0f0134af0a782ad5e3a3ee0df193fa6f"} Nov 25 11:03:00 crc kubenswrapper[4776]: I1125 11:03:00.556870 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8flmt" event={"ID":"4988892b-e67d-4595-abb3-c097fa0808b5","Type":"ContainerStarted","Data":"44a9f401e95c02b65c173c4fe182f7e479cfbd1140bc9d8c6b454e16db52b239"} Nov 25 11:03:00 crc kubenswrapper[4776]: I1125 11:03:00.563468 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-2x8x7" event={"ID":"d6d7ec58-a5a8-48aa-920d-d7f81e603f76","Type":"ContainerStarted","Data":"2c2269cf05113666a597abe1a179abb3caf504221ea666cf04c88dd2225db787"} Nov 25 11:03:00 crc kubenswrapper[4776]: I1125 11:03:00.563502 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-2x8x7" event={"ID":"d6d7ec58-a5a8-48aa-920d-d7f81e603f76","Type":"ContainerStarted","Data":"10c18cc07e0636703bcc95f2d17e1ee2bd40a7be6af3967e1ba7870f08d2c635"} Nov 25 11:03:00 crc kubenswrapper[4776]: I1125 11:03:00.565976 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-173a-account-create-2vr9z" event={"ID":"a9ae59af-54ec-4b1d-8285-a248f0b5f50a","Type":"ContainerStarted","Data":"097edc68524d81ee6dc49c048cce2375296e1ae90a743d94eb0c0f60b8f28096"} Nov 25 11:03:00 crc kubenswrapper[4776]: I1125 11:03:00.576725 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-jj7jb"] Nov 25 11:03:00 crc kubenswrapper[4776]: I1125 11:03:00.590528 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-2x8x7" podStartSLOduration=1.5905109899999998 podStartE2EDuration="1.59051099s" podCreationTimestamp="2025-11-25 11:02:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:03:00.586438768 +0000 UTC m=+5925.627498321" watchObservedRunningTime="2025-11-25 11:03:00.59051099 +0000 UTC m=+5925.631570543" Nov 25 11:03:00 crc kubenswrapper[4776]: W1125 11:03:00.595186 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96dba744_9bb8_480c_8498_c8eedd4fdb88.slice/crio-a10c1536ee37010fdb3e97ecc30a20cbced35bb8da4484df19b8a979f1aad9fe WatchSource:0}: Error finding container a10c1536ee37010fdb3e97ecc30a20cbced35bb8da4484df19b8a979f1aad9fe: Status 404 returned error can't find the container with id a10c1536ee37010fdb3e97ecc30a20cbced35bb8da4484df19b8a979f1aad9fe Nov 25 11:03:00 crc kubenswrapper[4776]: I1125 11:03:00.726655 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-e3f8-account-create-ht27m"] Nov 25 11:03:00 crc kubenswrapper[4776]: I1125 11:03:00.734449 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-0d1a-account-create-mhvkn"] Nov 25 11:03:01 crc kubenswrapper[4776]: I1125 11:03:01.577949 4776 generic.go:334] "Generic (PLEG): container finished" podID="38ec4382-1ef1-4cab-af5b-26ffecd9354a" containerID="0dba43c5ebc402c61560b9689b802fb034db2eaaee31930a18c9f03c8b125ead" exitCode=0 Nov 25 11:03:01 crc kubenswrapper[4776]: I1125 11:03:01.578123 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-e3f8-account-create-ht27m" event={"ID":"38ec4382-1ef1-4cab-af5b-26ffecd9354a","Type":"ContainerDied","Data":"0dba43c5ebc402c61560b9689b802fb034db2eaaee31930a18c9f03c8b125ead"} Nov 25 11:03:01 crc kubenswrapper[4776]: I1125 11:03:01.578219 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-e3f8-account-create-ht27m" event={"ID":"38ec4382-1ef1-4cab-af5b-26ffecd9354a","Type":"ContainerStarted","Data":"779f9556c33f07a6d9b63bbcef9d7db9b107b679e0cbb49289d5394d0f6e1e9c"} Nov 25 11:03:01 crc kubenswrapper[4776]: I1125 11:03:01.582475 4776 generic.go:334] "Generic (PLEG): container finished" podID="96dba744-9bb8-480c-8498-c8eedd4fdb88" containerID="cc2712b0b16a7e0249e124ae15cf4c417ac55be025874fb5ac41712df54e1bfe" exitCode=0 Nov 25 11:03:01 crc kubenswrapper[4776]: I1125 11:03:01.582570 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-jj7jb" event={"ID":"96dba744-9bb8-480c-8498-c8eedd4fdb88","Type":"ContainerDied","Data":"cc2712b0b16a7e0249e124ae15cf4c417ac55be025874fb5ac41712df54e1bfe"} Nov 25 11:03:01 crc kubenswrapper[4776]: I1125 11:03:01.582621 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-jj7jb" event={"ID":"96dba744-9bb8-480c-8498-c8eedd4fdb88","Type":"ContainerStarted","Data":"a10c1536ee37010fdb3e97ecc30a20cbced35bb8da4484df19b8a979f1aad9fe"} Nov 25 11:03:01 crc kubenswrapper[4776]: I1125 11:03:01.584674 4776 generic.go:334] "Generic (PLEG): container finished" podID="2c7b6238-2ba3-4d7f-a7f2-8f74092ce260" containerID="373b463ab348319edd3b2608fb5610259960bbb7b5cc35a33261c75f8705420f" exitCode=0 Nov 25 11:03:01 crc kubenswrapper[4776]: I1125 11:03:01.584721 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-0d1a-account-create-mhvkn" event={"ID":"2c7b6238-2ba3-4d7f-a7f2-8f74092ce260","Type":"ContainerDied","Data":"373b463ab348319edd3b2608fb5610259960bbb7b5cc35a33261c75f8705420f"} Nov 25 11:03:01 crc kubenswrapper[4776]: I1125 11:03:01.584739 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-0d1a-account-create-mhvkn" event={"ID":"2c7b6238-2ba3-4d7f-a7f2-8f74092ce260","Type":"ContainerStarted","Data":"671de968679f0027a613b9a0b94cc14ac8ba915c5b8bf31621e7af9c187da49d"} Nov 25 11:03:01 crc kubenswrapper[4776]: I1125 11:03:01.586483 4776 generic.go:334] "Generic (PLEG): container finished" podID="a9ae59af-54ec-4b1d-8285-a248f0b5f50a" containerID="7496c12a89099167c778dc437fb6b9930c8f9c9172a07ecfbbd8b770ada7995b" exitCode=0 Nov 25 11:03:01 crc kubenswrapper[4776]: I1125 11:03:01.586531 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-173a-account-create-2vr9z" event={"ID":"a9ae59af-54ec-4b1d-8285-a248f0b5f50a","Type":"ContainerDied","Data":"7496c12a89099167c778dc437fb6b9930c8f9c9172a07ecfbbd8b770ada7995b"} Nov 25 11:03:01 crc kubenswrapper[4776]: I1125 11:03:01.588895 4776 generic.go:334] "Generic (PLEG): container finished" podID="d6d7ec58-a5a8-48aa-920d-d7f81e603f76" containerID="2c2269cf05113666a597abe1a179abb3caf504221ea666cf04c88dd2225db787" exitCode=0 Nov 25 11:03:01 crc kubenswrapper[4776]: I1125 11:03:01.589893 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-2x8x7" event={"ID":"d6d7ec58-a5a8-48aa-920d-d7f81e603f76","Type":"ContainerDied","Data":"2c2269cf05113666a597abe1a179abb3caf504221ea666cf04c88dd2225db787"} Nov 25 11:03:01 crc kubenswrapper[4776]: I1125 11:03:01.935231 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8flmt" Nov 25 11:03:01 crc kubenswrapper[4776]: I1125 11:03:01.955095 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4988892b-e67d-4595-abb3-c097fa0808b5-operator-scripts\") pod \"4988892b-e67d-4595-abb3-c097fa0808b5\" (UID: \"4988892b-e67d-4595-abb3-c097fa0808b5\") " Nov 25 11:03:01 crc kubenswrapper[4776]: I1125 11:03:01.955227 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbk9k\" (UniqueName: \"kubernetes.io/projected/4988892b-e67d-4595-abb3-c097fa0808b5-kube-api-access-hbk9k\") pod \"4988892b-e67d-4595-abb3-c097fa0808b5\" (UID: \"4988892b-e67d-4595-abb3-c097fa0808b5\") " Nov 25 11:03:01 crc kubenswrapper[4776]: I1125 11:03:01.958397 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4988892b-e67d-4595-abb3-c097fa0808b5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4988892b-e67d-4595-abb3-c097fa0808b5" (UID: "4988892b-e67d-4595-abb3-c097fa0808b5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:03:01 crc kubenswrapper[4776]: I1125 11:03:01.961851 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4988892b-e67d-4595-abb3-c097fa0808b5-kube-api-access-hbk9k" (OuterVolumeSpecName: "kube-api-access-hbk9k") pod "4988892b-e67d-4595-abb3-c097fa0808b5" (UID: "4988892b-e67d-4595-abb3-c097fa0808b5"). InnerVolumeSpecName "kube-api-access-hbk9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:03:02 crc kubenswrapper[4776]: I1125 11:03:02.059678 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4988892b-e67d-4595-abb3-c097fa0808b5-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:02 crc kubenswrapper[4776]: I1125 11:03:02.059752 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbk9k\" (UniqueName: \"kubernetes.io/projected/4988892b-e67d-4595-abb3-c097fa0808b5-kube-api-access-hbk9k\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:02 crc kubenswrapper[4776]: I1125 11:03:02.622930 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8flmt" event={"ID":"4988892b-e67d-4595-abb3-c097fa0808b5","Type":"ContainerDied","Data":"44a9f401e95c02b65c173c4fe182f7e479cfbd1140bc9d8c6b454e16db52b239"} Nov 25 11:03:02 crc kubenswrapper[4776]: I1125 11:03:02.623005 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44a9f401e95c02b65c173c4fe182f7e479cfbd1140bc9d8c6b454e16db52b239" Nov 25 11:03:02 crc kubenswrapper[4776]: I1125 11:03:02.623120 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8flmt" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.030687 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jj7jb" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.077764 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztp9m\" (UniqueName: \"kubernetes.io/projected/96dba744-9bb8-480c-8498-c8eedd4fdb88-kube-api-access-ztp9m\") pod \"96dba744-9bb8-480c-8498-c8eedd4fdb88\" (UID: \"96dba744-9bb8-480c-8498-c8eedd4fdb88\") " Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.077847 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96dba744-9bb8-480c-8498-c8eedd4fdb88-operator-scripts\") pod \"96dba744-9bb8-480c-8498-c8eedd4fdb88\" (UID: \"96dba744-9bb8-480c-8498-c8eedd4fdb88\") " Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.078680 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96dba744-9bb8-480c-8498-c8eedd4fdb88-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "96dba744-9bb8-480c-8498-c8eedd4fdb88" (UID: "96dba744-9bb8-480c-8498-c8eedd4fdb88"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.083422 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96dba744-9bb8-480c-8498-c8eedd4fdb88-kube-api-access-ztp9m" (OuterVolumeSpecName: "kube-api-access-ztp9m") pod "96dba744-9bb8-480c-8498-c8eedd4fdb88" (UID: "96dba744-9bb8-480c-8498-c8eedd4fdb88"). InnerVolumeSpecName "kube-api-access-ztp9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.151898 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-173a-account-create-2vr9z" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.158615 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-0d1a-account-create-mhvkn" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.170680 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-2x8x7" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.179520 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bskx4\" (UniqueName: \"kubernetes.io/projected/2c7b6238-2ba3-4d7f-a7f2-8f74092ce260-kube-api-access-bskx4\") pod \"2c7b6238-2ba3-4d7f-a7f2-8f74092ce260\" (UID: \"2c7b6238-2ba3-4d7f-a7f2-8f74092ce260\") " Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.179644 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c7b6238-2ba3-4d7f-a7f2-8f74092ce260-operator-scripts\") pod \"2c7b6238-2ba3-4d7f-a7f2-8f74092ce260\" (UID: \"2c7b6238-2ba3-4d7f-a7f2-8f74092ce260\") " Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.179735 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6d7ec58-a5a8-48aa-920d-d7f81e603f76-operator-scripts\") pod \"d6d7ec58-a5a8-48aa-920d-d7f81e603f76\" (UID: \"d6d7ec58-a5a8-48aa-920d-d7f81e603f76\") " Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.179769 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7s2xc\" (UniqueName: \"kubernetes.io/projected/a9ae59af-54ec-4b1d-8285-a248f0b5f50a-kube-api-access-7s2xc\") pod \"a9ae59af-54ec-4b1d-8285-a248f0b5f50a\" (UID: \"a9ae59af-54ec-4b1d-8285-a248f0b5f50a\") " Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.179827 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9ae59af-54ec-4b1d-8285-a248f0b5f50a-operator-scripts\") pod \"a9ae59af-54ec-4b1d-8285-a248f0b5f50a\" (UID: \"a9ae59af-54ec-4b1d-8285-a248f0b5f50a\") " Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.180197 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pczmg\" (UniqueName: \"kubernetes.io/projected/d6d7ec58-a5a8-48aa-920d-d7f81e603f76-kube-api-access-pczmg\") pod \"d6d7ec58-a5a8-48aa-920d-d7f81e603f76\" (UID: \"d6d7ec58-a5a8-48aa-920d-d7f81e603f76\") " Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.180842 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c7b6238-2ba3-4d7f-a7f2-8f74092ce260-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2c7b6238-2ba3-4d7f-a7f2-8f74092ce260" (UID: "2c7b6238-2ba3-4d7f-a7f2-8f74092ce260"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.181046 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e3f8-account-create-ht27m" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.181211 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztp9m\" (UniqueName: \"kubernetes.io/projected/96dba744-9bb8-480c-8498-c8eedd4fdb88-kube-api-access-ztp9m\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.181234 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/96dba744-9bb8-480c-8498-c8eedd4fdb88-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.181244 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c7b6238-2ba3-4d7f-a7f2-8f74092ce260-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.182109 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9ae59af-54ec-4b1d-8285-a248f0b5f50a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a9ae59af-54ec-4b1d-8285-a248f0b5f50a" (UID: "a9ae59af-54ec-4b1d-8285-a248f0b5f50a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.185460 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6d7ec58-a5a8-48aa-920d-d7f81e603f76-kube-api-access-pczmg" (OuterVolumeSpecName: "kube-api-access-pczmg") pod "d6d7ec58-a5a8-48aa-920d-d7f81e603f76" (UID: "d6d7ec58-a5a8-48aa-920d-d7f81e603f76"). InnerVolumeSpecName "kube-api-access-pczmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.186348 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6d7ec58-a5a8-48aa-920d-d7f81e603f76-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d6d7ec58-a5a8-48aa-920d-d7f81e603f76" (UID: "d6d7ec58-a5a8-48aa-920d-d7f81e603f76"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.187017 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9ae59af-54ec-4b1d-8285-a248f0b5f50a-kube-api-access-7s2xc" (OuterVolumeSpecName: "kube-api-access-7s2xc") pod "a9ae59af-54ec-4b1d-8285-a248f0b5f50a" (UID: "a9ae59af-54ec-4b1d-8285-a248f0b5f50a"). InnerVolumeSpecName "kube-api-access-7s2xc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.197268 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c7b6238-2ba3-4d7f-a7f2-8f74092ce260-kube-api-access-bskx4" (OuterVolumeSpecName: "kube-api-access-bskx4") pod "2c7b6238-2ba3-4d7f-a7f2-8f74092ce260" (UID: "2c7b6238-2ba3-4d7f-a7f2-8f74092ce260"). InnerVolumeSpecName "kube-api-access-bskx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.282839 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/38ec4382-1ef1-4cab-af5b-26ffecd9354a-operator-scripts\") pod \"38ec4382-1ef1-4cab-af5b-26ffecd9354a\" (UID: \"38ec4382-1ef1-4cab-af5b-26ffecd9354a\") " Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.283053 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhlb8\" (UniqueName: \"kubernetes.io/projected/38ec4382-1ef1-4cab-af5b-26ffecd9354a-kube-api-access-qhlb8\") pod \"38ec4382-1ef1-4cab-af5b-26ffecd9354a\" (UID: \"38ec4382-1ef1-4cab-af5b-26ffecd9354a\") " Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.283537 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38ec4382-1ef1-4cab-af5b-26ffecd9354a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "38ec4382-1ef1-4cab-af5b-26ffecd9354a" (UID: "38ec4382-1ef1-4cab-af5b-26ffecd9354a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.283665 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bskx4\" (UniqueName: \"kubernetes.io/projected/2c7b6238-2ba3-4d7f-a7f2-8f74092ce260-kube-api-access-bskx4\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.283684 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6d7ec58-a5a8-48aa-920d-d7f81e603f76-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.283693 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7s2xc\" (UniqueName: \"kubernetes.io/projected/a9ae59af-54ec-4b1d-8285-a248f0b5f50a-kube-api-access-7s2xc\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.283701 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9ae59af-54ec-4b1d-8285-a248f0b5f50a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.283709 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/38ec4382-1ef1-4cab-af5b-26ffecd9354a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.283717 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pczmg\" (UniqueName: \"kubernetes.io/projected/d6d7ec58-a5a8-48aa-920d-d7f81e603f76-kube-api-access-pczmg\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.285864 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38ec4382-1ef1-4cab-af5b-26ffecd9354a-kube-api-access-qhlb8" (OuterVolumeSpecName: "kube-api-access-qhlb8") pod "38ec4382-1ef1-4cab-af5b-26ffecd9354a" (UID: "38ec4382-1ef1-4cab-af5b-26ffecd9354a"). InnerVolumeSpecName "kube-api-access-qhlb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.385627 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhlb8\" (UniqueName: \"kubernetes.io/projected/38ec4382-1ef1-4cab-af5b-26ffecd9354a-kube-api-access-qhlb8\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.639010 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-jj7jb" event={"ID":"96dba744-9bb8-480c-8498-c8eedd4fdb88","Type":"ContainerDied","Data":"a10c1536ee37010fdb3e97ecc30a20cbced35bb8da4484df19b8a979f1aad9fe"} Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.639341 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a10c1536ee37010fdb3e97ecc30a20cbced35bb8da4484df19b8a979f1aad9fe" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.639414 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jj7jb" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.642227 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-0d1a-account-create-mhvkn" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.642232 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-0d1a-account-create-mhvkn" event={"ID":"2c7b6238-2ba3-4d7f-a7f2-8f74092ce260","Type":"ContainerDied","Data":"671de968679f0027a613b9a0b94cc14ac8ba915c5b8bf31621e7af9c187da49d"} Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.642274 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="671de968679f0027a613b9a0b94cc14ac8ba915c5b8bf31621e7af9c187da49d" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.643962 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-2x8x7" event={"ID":"d6d7ec58-a5a8-48aa-920d-d7f81e603f76","Type":"ContainerDied","Data":"10c18cc07e0636703bcc95f2d17e1ee2bd40a7be6af3967e1ba7870f08d2c635"} Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.643985 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10c18cc07e0636703bcc95f2d17e1ee2bd40a7be6af3967e1ba7870f08d2c635" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.644038 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-2x8x7" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.646542 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-173a-account-create-2vr9z" event={"ID":"a9ae59af-54ec-4b1d-8285-a248f0b5f50a","Type":"ContainerDied","Data":"097edc68524d81ee6dc49c048cce2375296e1ae90a743d94eb0c0f60b8f28096"} Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.646559 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-173a-account-create-2vr9z" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.646572 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="097edc68524d81ee6dc49c048cce2375296e1ae90a743d94eb0c0f60b8f28096" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.648776 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-e3f8-account-create-ht27m" event={"ID":"38ec4382-1ef1-4cab-af5b-26ffecd9354a","Type":"ContainerDied","Data":"779f9556c33f07a6d9b63bbcef9d7db9b107b679e0cbb49289d5394d0f6e1e9c"} Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.648804 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="779f9556c33f07a6d9b63bbcef9d7db9b107b679e0cbb49289d5394d0f6e1e9c" Nov 25 11:03:03 crc kubenswrapper[4776]: I1125 11:03:03.648857 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e3f8-account-create-ht27m" Nov 25 11:03:04 crc kubenswrapper[4776]: I1125 11:03:04.828271 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gpz2x"] Nov 25 11:03:04 crc kubenswrapper[4776]: E1125 11:03:04.828772 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96dba744-9bb8-480c-8498-c8eedd4fdb88" containerName="mariadb-database-create" Nov 25 11:03:04 crc kubenswrapper[4776]: I1125 11:03:04.828787 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="96dba744-9bb8-480c-8498-c8eedd4fdb88" containerName="mariadb-database-create" Nov 25 11:03:04 crc kubenswrapper[4776]: E1125 11:03:04.828804 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38ec4382-1ef1-4cab-af5b-26ffecd9354a" containerName="mariadb-account-create" Nov 25 11:03:04 crc kubenswrapper[4776]: I1125 11:03:04.828812 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="38ec4382-1ef1-4cab-af5b-26ffecd9354a" containerName="mariadb-account-create" Nov 25 11:03:04 crc kubenswrapper[4776]: E1125 11:03:04.828831 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4988892b-e67d-4595-abb3-c097fa0808b5" containerName="mariadb-database-create" Nov 25 11:03:04 crc kubenswrapper[4776]: I1125 11:03:04.828839 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="4988892b-e67d-4595-abb3-c097fa0808b5" containerName="mariadb-database-create" Nov 25 11:03:04 crc kubenswrapper[4776]: E1125 11:03:04.828872 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c7b6238-2ba3-4d7f-a7f2-8f74092ce260" containerName="mariadb-account-create" Nov 25 11:03:04 crc kubenswrapper[4776]: I1125 11:03:04.828880 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c7b6238-2ba3-4d7f-a7f2-8f74092ce260" containerName="mariadb-account-create" Nov 25 11:03:04 crc kubenswrapper[4776]: E1125 11:03:04.828900 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6d7ec58-a5a8-48aa-920d-d7f81e603f76" containerName="mariadb-database-create" Nov 25 11:03:04 crc kubenswrapper[4776]: I1125 11:03:04.828909 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6d7ec58-a5a8-48aa-920d-d7f81e603f76" containerName="mariadb-database-create" Nov 25 11:03:04 crc kubenswrapper[4776]: E1125 11:03:04.828920 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9ae59af-54ec-4b1d-8285-a248f0b5f50a" containerName="mariadb-account-create" Nov 25 11:03:04 crc kubenswrapper[4776]: I1125 11:03:04.828927 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9ae59af-54ec-4b1d-8285-a248f0b5f50a" containerName="mariadb-account-create" Nov 25 11:03:04 crc kubenswrapper[4776]: I1125 11:03:04.829135 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9ae59af-54ec-4b1d-8285-a248f0b5f50a" containerName="mariadb-account-create" Nov 25 11:03:04 crc kubenswrapper[4776]: I1125 11:03:04.829152 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c7b6238-2ba3-4d7f-a7f2-8f74092ce260" containerName="mariadb-account-create" Nov 25 11:03:04 crc kubenswrapper[4776]: I1125 11:03:04.829170 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="96dba744-9bb8-480c-8498-c8eedd4fdb88" containerName="mariadb-database-create" Nov 25 11:03:04 crc kubenswrapper[4776]: I1125 11:03:04.829181 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6d7ec58-a5a8-48aa-920d-d7f81e603f76" containerName="mariadb-database-create" Nov 25 11:03:04 crc kubenswrapper[4776]: I1125 11:03:04.829191 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="4988892b-e67d-4595-abb3-c097fa0808b5" containerName="mariadb-database-create" Nov 25 11:03:04 crc kubenswrapper[4776]: I1125 11:03:04.829206 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="38ec4382-1ef1-4cab-af5b-26ffecd9354a" containerName="mariadb-account-create" Nov 25 11:03:04 crc kubenswrapper[4776]: I1125 11:03:04.829993 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-gpz2x" Nov 25 11:03:04 crc kubenswrapper[4776]: I1125 11:03:04.833233 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-bmg8f" Nov 25 11:03:04 crc kubenswrapper[4776]: I1125 11:03:04.833244 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 25 11:03:04 crc kubenswrapper[4776]: I1125 11:03:04.833244 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 25 11:03:04 crc kubenswrapper[4776]: I1125 11:03:04.844717 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gpz2x"] Nov 25 11:03:05 crc kubenswrapper[4776]: I1125 11:03:05.013437 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad93639-113c-4055-8bc2-d4fb8a35d5bb-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-gpz2x\" (UID: \"5ad93639-113c-4055-8bc2-d4fb8a35d5bb\") " pod="openstack/nova-cell0-conductor-db-sync-gpz2x" Nov 25 11:03:05 crc kubenswrapper[4776]: I1125 11:03:05.013753 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptprm\" (UniqueName: \"kubernetes.io/projected/5ad93639-113c-4055-8bc2-d4fb8a35d5bb-kube-api-access-ptprm\") pod \"nova-cell0-conductor-db-sync-gpz2x\" (UID: \"5ad93639-113c-4055-8bc2-d4fb8a35d5bb\") " pod="openstack/nova-cell0-conductor-db-sync-gpz2x" Nov 25 11:03:05 crc kubenswrapper[4776]: I1125 11:03:05.013908 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ad93639-113c-4055-8bc2-d4fb8a35d5bb-scripts\") pod \"nova-cell0-conductor-db-sync-gpz2x\" (UID: \"5ad93639-113c-4055-8bc2-d4fb8a35d5bb\") " pod="openstack/nova-cell0-conductor-db-sync-gpz2x" Nov 25 11:03:05 crc kubenswrapper[4776]: I1125 11:03:05.013949 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad93639-113c-4055-8bc2-d4fb8a35d5bb-config-data\") pod \"nova-cell0-conductor-db-sync-gpz2x\" (UID: \"5ad93639-113c-4055-8bc2-d4fb8a35d5bb\") " pod="openstack/nova-cell0-conductor-db-sync-gpz2x" Nov 25 11:03:05 crc kubenswrapper[4776]: I1125 11:03:05.115648 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad93639-113c-4055-8bc2-d4fb8a35d5bb-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-gpz2x\" (UID: \"5ad93639-113c-4055-8bc2-d4fb8a35d5bb\") " pod="openstack/nova-cell0-conductor-db-sync-gpz2x" Nov 25 11:03:05 crc kubenswrapper[4776]: I1125 11:03:05.115746 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptprm\" (UniqueName: \"kubernetes.io/projected/5ad93639-113c-4055-8bc2-d4fb8a35d5bb-kube-api-access-ptprm\") pod \"nova-cell0-conductor-db-sync-gpz2x\" (UID: \"5ad93639-113c-4055-8bc2-d4fb8a35d5bb\") " pod="openstack/nova-cell0-conductor-db-sync-gpz2x" Nov 25 11:03:05 crc kubenswrapper[4776]: I1125 11:03:05.115854 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ad93639-113c-4055-8bc2-d4fb8a35d5bb-scripts\") pod \"nova-cell0-conductor-db-sync-gpz2x\" (UID: \"5ad93639-113c-4055-8bc2-d4fb8a35d5bb\") " pod="openstack/nova-cell0-conductor-db-sync-gpz2x" Nov 25 11:03:05 crc kubenswrapper[4776]: I1125 11:03:05.115904 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad93639-113c-4055-8bc2-d4fb8a35d5bb-config-data\") pod \"nova-cell0-conductor-db-sync-gpz2x\" (UID: \"5ad93639-113c-4055-8bc2-d4fb8a35d5bb\") " pod="openstack/nova-cell0-conductor-db-sync-gpz2x" Nov 25 11:03:05 crc kubenswrapper[4776]: I1125 11:03:05.120643 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad93639-113c-4055-8bc2-d4fb8a35d5bb-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-gpz2x\" (UID: \"5ad93639-113c-4055-8bc2-d4fb8a35d5bb\") " pod="openstack/nova-cell0-conductor-db-sync-gpz2x" Nov 25 11:03:05 crc kubenswrapper[4776]: I1125 11:03:05.120721 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad93639-113c-4055-8bc2-d4fb8a35d5bb-config-data\") pod \"nova-cell0-conductor-db-sync-gpz2x\" (UID: \"5ad93639-113c-4055-8bc2-d4fb8a35d5bb\") " pod="openstack/nova-cell0-conductor-db-sync-gpz2x" Nov 25 11:03:05 crc kubenswrapper[4776]: I1125 11:03:05.138037 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptprm\" (UniqueName: \"kubernetes.io/projected/5ad93639-113c-4055-8bc2-d4fb8a35d5bb-kube-api-access-ptprm\") pod \"nova-cell0-conductor-db-sync-gpz2x\" (UID: \"5ad93639-113c-4055-8bc2-d4fb8a35d5bb\") " pod="openstack/nova-cell0-conductor-db-sync-gpz2x" Nov 25 11:03:05 crc kubenswrapper[4776]: I1125 11:03:05.138133 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ad93639-113c-4055-8bc2-d4fb8a35d5bb-scripts\") pod \"nova-cell0-conductor-db-sync-gpz2x\" (UID: \"5ad93639-113c-4055-8bc2-d4fb8a35d5bb\") " pod="openstack/nova-cell0-conductor-db-sync-gpz2x" Nov 25 11:03:05 crc kubenswrapper[4776]: I1125 11:03:05.162474 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-gpz2x" Nov 25 11:03:05 crc kubenswrapper[4776]: I1125 11:03:05.608822 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gpz2x"] Nov 25 11:03:05 crc kubenswrapper[4776]: W1125 11:03:05.613574 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ad93639_113c_4055_8bc2_d4fb8a35d5bb.slice/crio-fde8353aee29cc080230e2ced06e7579d7000602196270d3f984cb328e19be2b WatchSource:0}: Error finding container fde8353aee29cc080230e2ced06e7579d7000602196270d3f984cb328e19be2b: Status 404 returned error can't find the container with id fde8353aee29cc080230e2ced06e7579d7000602196270d3f984cb328e19be2b Nov 25 11:03:05 crc kubenswrapper[4776]: I1125 11:03:05.692410 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-gpz2x" event={"ID":"5ad93639-113c-4055-8bc2-d4fb8a35d5bb","Type":"ContainerStarted","Data":"fde8353aee29cc080230e2ced06e7579d7000602196270d3f984cb328e19be2b"} Nov 25 11:03:06 crc kubenswrapper[4776]: I1125 11:03:06.686055 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-gpz2x" event={"ID":"5ad93639-113c-4055-8bc2-d4fb8a35d5bb","Type":"ContainerStarted","Data":"89b549806d7cd66875b6ebdec26f160b49e682a3bab883acb744e930545b06be"} Nov 25 11:03:06 crc kubenswrapper[4776]: I1125 11:03:06.710554 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-gpz2x" podStartSLOduration=2.710534682 podStartE2EDuration="2.710534682s" podCreationTimestamp="2025-11-25 11:03:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:03:06.707574348 +0000 UTC m=+5931.748633941" watchObservedRunningTime="2025-11-25 11:03:06.710534682 +0000 UTC m=+5931.751594245" Nov 25 11:03:11 crc kubenswrapper[4776]: I1125 11:03:11.749250 4776 generic.go:334] "Generic (PLEG): container finished" podID="5ad93639-113c-4055-8bc2-d4fb8a35d5bb" containerID="89b549806d7cd66875b6ebdec26f160b49e682a3bab883acb744e930545b06be" exitCode=0 Nov 25 11:03:11 crc kubenswrapper[4776]: I1125 11:03:11.749311 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-gpz2x" event={"ID":"5ad93639-113c-4055-8bc2-d4fb8a35d5bb","Type":"ContainerDied","Data":"89b549806d7cd66875b6ebdec26f160b49e682a3bab883acb744e930545b06be"} Nov 25 11:03:13 crc kubenswrapper[4776]: I1125 11:03:13.074981 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-gpz2x" Nov 25 11:03:13 crc kubenswrapper[4776]: I1125 11:03:13.084313 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad93639-113c-4055-8bc2-d4fb8a35d5bb-combined-ca-bundle\") pod \"5ad93639-113c-4055-8bc2-d4fb8a35d5bb\" (UID: \"5ad93639-113c-4055-8bc2-d4fb8a35d5bb\") " Nov 25 11:03:13 crc kubenswrapper[4776]: I1125 11:03:13.084548 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad93639-113c-4055-8bc2-d4fb8a35d5bb-config-data\") pod \"5ad93639-113c-4055-8bc2-d4fb8a35d5bb\" (UID: \"5ad93639-113c-4055-8bc2-d4fb8a35d5bb\") " Nov 25 11:03:13 crc kubenswrapper[4776]: I1125 11:03:13.084712 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptprm\" (UniqueName: \"kubernetes.io/projected/5ad93639-113c-4055-8bc2-d4fb8a35d5bb-kube-api-access-ptprm\") pod \"5ad93639-113c-4055-8bc2-d4fb8a35d5bb\" (UID: \"5ad93639-113c-4055-8bc2-d4fb8a35d5bb\") " Nov 25 11:03:13 crc kubenswrapper[4776]: I1125 11:03:13.084775 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ad93639-113c-4055-8bc2-d4fb8a35d5bb-scripts\") pod \"5ad93639-113c-4055-8bc2-d4fb8a35d5bb\" (UID: \"5ad93639-113c-4055-8bc2-d4fb8a35d5bb\") " Nov 25 11:03:13 crc kubenswrapper[4776]: I1125 11:03:13.090820 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ad93639-113c-4055-8bc2-d4fb8a35d5bb-scripts" (OuterVolumeSpecName: "scripts") pod "5ad93639-113c-4055-8bc2-d4fb8a35d5bb" (UID: "5ad93639-113c-4055-8bc2-d4fb8a35d5bb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:13 crc kubenswrapper[4776]: I1125 11:03:13.092288 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ad93639-113c-4055-8bc2-d4fb8a35d5bb-kube-api-access-ptprm" (OuterVolumeSpecName: "kube-api-access-ptprm") pod "5ad93639-113c-4055-8bc2-d4fb8a35d5bb" (UID: "5ad93639-113c-4055-8bc2-d4fb8a35d5bb"). InnerVolumeSpecName "kube-api-access-ptprm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:03:13 crc kubenswrapper[4776]: I1125 11:03:13.122471 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ad93639-113c-4055-8bc2-d4fb8a35d5bb-config-data" (OuterVolumeSpecName: "config-data") pod "5ad93639-113c-4055-8bc2-d4fb8a35d5bb" (UID: "5ad93639-113c-4055-8bc2-d4fb8a35d5bb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:13 crc kubenswrapper[4776]: I1125 11:03:13.124590 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ad93639-113c-4055-8bc2-d4fb8a35d5bb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ad93639-113c-4055-8bc2-d4fb8a35d5bb" (UID: "5ad93639-113c-4055-8bc2-d4fb8a35d5bb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:13 crc kubenswrapper[4776]: I1125 11:03:13.187358 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ad93639-113c-4055-8bc2-d4fb8a35d5bb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:13 crc kubenswrapper[4776]: I1125 11:03:13.187407 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ad93639-113c-4055-8bc2-d4fb8a35d5bb-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:13 crc kubenswrapper[4776]: I1125 11:03:13.187420 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptprm\" (UniqueName: \"kubernetes.io/projected/5ad93639-113c-4055-8bc2-d4fb8a35d5bb-kube-api-access-ptprm\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:13 crc kubenswrapper[4776]: I1125 11:03:13.187432 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ad93639-113c-4055-8bc2-d4fb8a35d5bb-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:13 crc kubenswrapper[4776]: I1125 11:03:13.770771 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-gpz2x" event={"ID":"5ad93639-113c-4055-8bc2-d4fb8a35d5bb","Type":"ContainerDied","Data":"fde8353aee29cc080230e2ced06e7579d7000602196270d3f984cb328e19be2b"} Nov 25 11:03:13 crc kubenswrapper[4776]: I1125 11:03:13.771033 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fde8353aee29cc080230e2ced06e7579d7000602196270d3f984cb328e19be2b" Nov 25 11:03:13 crc kubenswrapper[4776]: I1125 11:03:13.770870 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-gpz2x" Nov 25 11:03:13 crc kubenswrapper[4776]: I1125 11:03:13.864993 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 11:03:13 crc kubenswrapper[4776]: E1125 11:03:13.865403 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ad93639-113c-4055-8bc2-d4fb8a35d5bb" containerName="nova-cell0-conductor-db-sync" Nov 25 11:03:13 crc kubenswrapper[4776]: I1125 11:03:13.865419 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ad93639-113c-4055-8bc2-d4fb8a35d5bb" containerName="nova-cell0-conductor-db-sync" Nov 25 11:03:13 crc kubenswrapper[4776]: I1125 11:03:13.865591 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ad93639-113c-4055-8bc2-d4fb8a35d5bb" containerName="nova-cell0-conductor-db-sync" Nov 25 11:03:13 crc kubenswrapper[4776]: I1125 11:03:13.866220 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 25 11:03:13 crc kubenswrapper[4776]: I1125 11:03:13.869306 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 25 11:03:13 crc kubenswrapper[4776]: I1125 11:03:13.869353 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-bmg8f" Nov 25 11:03:13 crc kubenswrapper[4776]: I1125 11:03:13.886198 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 11:03:14 crc kubenswrapper[4776]: I1125 11:03:14.020193 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c39e31e-7255-4c8e-90d9-5e0cc5d3910a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"9c39e31e-7255-4c8e-90d9-5e0cc5d3910a\") " pod="openstack/nova-cell0-conductor-0" Nov 25 11:03:14 crc kubenswrapper[4776]: I1125 11:03:14.020545 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bh96c\" (UniqueName: \"kubernetes.io/projected/9c39e31e-7255-4c8e-90d9-5e0cc5d3910a-kube-api-access-bh96c\") pod \"nova-cell0-conductor-0\" (UID: \"9c39e31e-7255-4c8e-90d9-5e0cc5d3910a\") " pod="openstack/nova-cell0-conductor-0" Nov 25 11:03:14 crc kubenswrapper[4776]: I1125 11:03:14.020774 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c39e31e-7255-4c8e-90d9-5e0cc5d3910a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"9c39e31e-7255-4c8e-90d9-5e0cc5d3910a\") " pod="openstack/nova-cell0-conductor-0" Nov 25 11:03:14 crc kubenswrapper[4776]: I1125 11:03:14.124086 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c39e31e-7255-4c8e-90d9-5e0cc5d3910a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"9c39e31e-7255-4c8e-90d9-5e0cc5d3910a\") " pod="openstack/nova-cell0-conductor-0" Nov 25 11:03:14 crc kubenswrapper[4776]: I1125 11:03:14.124348 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c39e31e-7255-4c8e-90d9-5e0cc5d3910a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"9c39e31e-7255-4c8e-90d9-5e0cc5d3910a\") " pod="openstack/nova-cell0-conductor-0" Nov 25 11:03:14 crc kubenswrapper[4776]: I1125 11:03:14.124417 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bh96c\" (UniqueName: \"kubernetes.io/projected/9c39e31e-7255-4c8e-90d9-5e0cc5d3910a-kube-api-access-bh96c\") pod \"nova-cell0-conductor-0\" (UID: \"9c39e31e-7255-4c8e-90d9-5e0cc5d3910a\") " pod="openstack/nova-cell0-conductor-0" Nov 25 11:03:14 crc kubenswrapper[4776]: I1125 11:03:14.131731 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c39e31e-7255-4c8e-90d9-5e0cc5d3910a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"9c39e31e-7255-4c8e-90d9-5e0cc5d3910a\") " pod="openstack/nova-cell0-conductor-0" Nov 25 11:03:14 crc kubenswrapper[4776]: I1125 11:03:14.136733 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c39e31e-7255-4c8e-90d9-5e0cc5d3910a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"9c39e31e-7255-4c8e-90d9-5e0cc5d3910a\") " pod="openstack/nova-cell0-conductor-0" Nov 25 11:03:14 crc kubenswrapper[4776]: I1125 11:03:14.149147 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bh96c\" (UniqueName: \"kubernetes.io/projected/9c39e31e-7255-4c8e-90d9-5e0cc5d3910a-kube-api-access-bh96c\") pod \"nova-cell0-conductor-0\" (UID: \"9c39e31e-7255-4c8e-90d9-5e0cc5d3910a\") " pod="openstack/nova-cell0-conductor-0" Nov 25 11:03:14 crc kubenswrapper[4776]: I1125 11:03:14.190192 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 25 11:03:14 crc kubenswrapper[4776]: I1125 11:03:14.705678 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 11:03:14 crc kubenswrapper[4776]: I1125 11:03:14.780747 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"9c39e31e-7255-4c8e-90d9-5e0cc5d3910a","Type":"ContainerStarted","Data":"77e62d592424186ef952b8e5f3b09300e547a04fc1000b176e7e9dec9ddcfecc"} Nov 25 11:03:15 crc kubenswrapper[4776]: I1125 11:03:15.801745 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"9c39e31e-7255-4c8e-90d9-5e0cc5d3910a","Type":"ContainerStarted","Data":"15c598d4474dfc8c30f350b528152717780cc311dc924d946511163c8d02d979"} Nov 25 11:03:15 crc kubenswrapper[4776]: I1125 11:03:15.802361 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 25 11:03:15 crc kubenswrapper[4776]: I1125 11:03:15.821449 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.821427595 podStartE2EDuration="2.821427595s" podCreationTimestamp="2025-11-25 11:03:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:03:15.815056365 +0000 UTC m=+5940.856115918" watchObservedRunningTime="2025-11-25 11:03:15.821427595 +0000 UTC m=+5940.862487148" Nov 25 11:03:17 crc kubenswrapper[4776]: I1125 11:03:17.819030 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:03:17 crc kubenswrapper[4776]: I1125 11:03:17.819577 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.219260 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.673377 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-k465n"] Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.674461 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-k465n" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.676237 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.676942 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.680566 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-k465n"] Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.781946 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c54defc-dcc0-416c-9d88-fbc5944aa0e3-scripts\") pod \"nova-cell0-cell-mapping-k465n\" (UID: \"9c54defc-dcc0-416c-9d88-fbc5944aa0e3\") " pod="openstack/nova-cell0-cell-mapping-k465n" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.782372 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c54defc-dcc0-416c-9d88-fbc5944aa0e3-config-data\") pod \"nova-cell0-cell-mapping-k465n\" (UID: \"9c54defc-dcc0-416c-9d88-fbc5944aa0e3\") " pod="openstack/nova-cell0-cell-mapping-k465n" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.782414 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c54defc-dcc0-416c-9d88-fbc5944aa0e3-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-k465n\" (UID: \"9c54defc-dcc0-416c-9d88-fbc5944aa0e3\") " pod="openstack/nova-cell0-cell-mapping-k465n" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.782449 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btwj6\" (UniqueName: \"kubernetes.io/projected/9c54defc-dcc0-416c-9d88-fbc5944aa0e3-kube-api-access-btwj6\") pod \"nova-cell0-cell-mapping-k465n\" (UID: \"9c54defc-dcc0-416c-9d88-fbc5944aa0e3\") " pod="openstack/nova-cell0-cell-mapping-k465n" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.805260 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.807030 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.811713 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.835319 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.851672 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.853166 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.857408 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.887914 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c54defc-dcc0-416c-9d88-fbc5944aa0e3-config-data\") pod \"nova-cell0-cell-mapping-k465n\" (UID: \"9c54defc-dcc0-416c-9d88-fbc5944aa0e3\") " pod="openstack/nova-cell0-cell-mapping-k465n" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.887966 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c54defc-dcc0-416c-9d88-fbc5944aa0e3-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-k465n\" (UID: \"9c54defc-dcc0-416c-9d88-fbc5944aa0e3\") " pod="openstack/nova-cell0-cell-mapping-k465n" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.887991 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btwj6\" (UniqueName: \"kubernetes.io/projected/9c54defc-dcc0-416c-9d88-fbc5944aa0e3-kube-api-access-btwj6\") pod \"nova-cell0-cell-mapping-k465n\" (UID: \"9c54defc-dcc0-416c-9d88-fbc5944aa0e3\") " pod="openstack/nova-cell0-cell-mapping-k465n" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.888043 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce80bccc-2b9e-44c9-a1c4-915c7bed4897-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ce80bccc-2b9e-44c9-a1c4-915c7bed4897\") " pod="openstack/nova-api-0" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.888093 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a9de870-885d-4c13-b8f2-1180378b1a79-config-data\") pod \"nova-metadata-0\" (UID: \"3a9de870-885d-4c13-b8f2-1180378b1a79\") " pod="openstack/nova-metadata-0" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.888128 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a9de870-885d-4c13-b8f2-1180378b1a79-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3a9de870-885d-4c13-b8f2-1180378b1a79\") " pod="openstack/nova-metadata-0" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.888159 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g74ns\" (UniqueName: \"kubernetes.io/projected/ce80bccc-2b9e-44c9-a1c4-915c7bed4897-kube-api-access-g74ns\") pod \"nova-api-0\" (UID: \"ce80bccc-2b9e-44c9-a1c4-915c7bed4897\") " pod="openstack/nova-api-0" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.888200 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c54defc-dcc0-416c-9d88-fbc5944aa0e3-scripts\") pod \"nova-cell0-cell-mapping-k465n\" (UID: \"9c54defc-dcc0-416c-9d88-fbc5944aa0e3\") " pod="openstack/nova-cell0-cell-mapping-k465n" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.888238 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce80bccc-2b9e-44c9-a1c4-915c7bed4897-logs\") pod \"nova-api-0\" (UID: \"ce80bccc-2b9e-44c9-a1c4-915c7bed4897\") " pod="openstack/nova-api-0" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.888279 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a9de870-885d-4c13-b8f2-1180378b1a79-logs\") pod \"nova-metadata-0\" (UID: \"3a9de870-885d-4c13-b8f2-1180378b1a79\") " pod="openstack/nova-metadata-0" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.888312 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmthn\" (UniqueName: \"kubernetes.io/projected/3a9de870-885d-4c13-b8f2-1180378b1a79-kube-api-access-rmthn\") pod \"nova-metadata-0\" (UID: \"3a9de870-885d-4c13-b8f2-1180378b1a79\") " pod="openstack/nova-metadata-0" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.888378 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce80bccc-2b9e-44c9-a1c4-915c7bed4897-config-data\") pod \"nova-api-0\" (UID: \"ce80bccc-2b9e-44c9-a1c4-915c7bed4897\") " pod="openstack/nova-api-0" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.890130 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.899303 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c54defc-dcc0-416c-9d88-fbc5944aa0e3-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-k465n\" (UID: \"9c54defc-dcc0-416c-9d88-fbc5944aa0e3\") " pod="openstack/nova-cell0-cell-mapping-k465n" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.899394 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.900670 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.901637 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c54defc-dcc0-416c-9d88-fbc5944aa0e3-config-data\") pod \"nova-cell0-cell-mapping-k465n\" (UID: \"9c54defc-dcc0-416c-9d88-fbc5944aa0e3\") " pod="openstack/nova-cell0-cell-mapping-k465n" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.923627 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c54defc-dcc0-416c-9d88-fbc5944aa0e3-scripts\") pod \"nova-cell0-cell-mapping-k465n\" (UID: \"9c54defc-dcc0-416c-9d88-fbc5944aa0e3\") " pod="openstack/nova-cell0-cell-mapping-k465n" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.924611 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.945736 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btwj6\" (UniqueName: \"kubernetes.io/projected/9c54defc-dcc0-416c-9d88-fbc5944aa0e3-kube-api-access-btwj6\") pod \"nova-cell0-cell-mapping-k465n\" (UID: \"9c54defc-dcc0-416c-9d88-fbc5944aa0e3\") " pod="openstack/nova-cell0-cell-mapping-k465n" Nov 25 11:03:19 crc kubenswrapper[4776]: I1125 11:03:19.968230 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:19.996929 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce80bccc-2b9e-44c9-a1c4-915c7bed4897-logs\") pod \"nova-api-0\" (UID: \"ce80bccc-2b9e-44c9-a1c4-915c7bed4897\") " pod="openstack/nova-api-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.016191 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da7d17e7-2d62-466a-8826-817e197158d2-config-data\") pod \"nova-scheduler-0\" (UID: \"da7d17e7-2d62-466a-8826-817e197158d2\") " pod="openstack/nova-scheduler-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.016298 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a9de870-885d-4c13-b8f2-1180378b1a79-logs\") pod \"nova-metadata-0\" (UID: \"3a9de870-885d-4c13-b8f2-1180378b1a79\") " pod="openstack/nova-metadata-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.016324 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da7d17e7-2d62-466a-8826-817e197158d2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"da7d17e7-2d62-466a-8826-817e197158d2\") " pod="openstack/nova-scheduler-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.016374 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmthn\" (UniqueName: \"kubernetes.io/projected/3a9de870-885d-4c13-b8f2-1180378b1a79-kube-api-access-rmthn\") pod \"nova-metadata-0\" (UID: \"3a9de870-885d-4c13-b8f2-1180378b1a79\") " pod="openstack/nova-metadata-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.016399 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9nkk\" (UniqueName: \"kubernetes.io/projected/da7d17e7-2d62-466a-8826-817e197158d2-kube-api-access-h9nkk\") pod \"nova-scheduler-0\" (UID: \"da7d17e7-2d62-466a-8826-817e197158d2\") " pod="openstack/nova-scheduler-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.016504 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce80bccc-2b9e-44c9-a1c4-915c7bed4897-config-data\") pod \"nova-api-0\" (UID: \"ce80bccc-2b9e-44c9-a1c4-915c7bed4897\") " pod="openstack/nova-api-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.016700 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce80bccc-2b9e-44c9-a1c4-915c7bed4897-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ce80bccc-2b9e-44c9-a1c4-915c7bed4897\") " pod="openstack/nova-api-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.016725 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a9de870-885d-4c13-b8f2-1180378b1a79-config-data\") pod \"nova-metadata-0\" (UID: \"3a9de870-885d-4c13-b8f2-1180378b1a79\") " pod="openstack/nova-metadata-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.016753 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a9de870-885d-4c13-b8f2-1180378b1a79-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3a9de870-885d-4c13-b8f2-1180378b1a79\") " pod="openstack/nova-metadata-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.016794 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g74ns\" (UniqueName: \"kubernetes.io/projected/ce80bccc-2b9e-44c9-a1c4-915c7bed4897-kube-api-access-g74ns\") pod \"nova-api-0\" (UID: \"ce80bccc-2b9e-44c9-a1c4-915c7bed4897\") " pod="openstack/nova-api-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.003361 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-k465n" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.003505 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce80bccc-2b9e-44c9-a1c4-915c7bed4897-logs\") pod \"nova-api-0\" (UID: \"ce80bccc-2b9e-44c9-a1c4-915c7bed4897\") " pod="openstack/nova-api-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.018357 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a9de870-885d-4c13-b8f2-1180378b1a79-logs\") pod \"nova-metadata-0\" (UID: \"3a9de870-885d-4c13-b8f2-1180378b1a79\") " pod="openstack/nova-metadata-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.036883 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce80bccc-2b9e-44c9-a1c4-915c7bed4897-config-data\") pod \"nova-api-0\" (UID: \"ce80bccc-2b9e-44c9-a1c4-915c7bed4897\") " pod="openstack/nova-api-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.044022 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce80bccc-2b9e-44c9-a1c4-915c7bed4897-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ce80bccc-2b9e-44c9-a1c4-915c7bed4897\") " pod="openstack/nova-api-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.045196 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a9de870-885d-4c13-b8f2-1180378b1a79-config-data\") pod \"nova-metadata-0\" (UID: \"3a9de870-885d-4c13-b8f2-1180378b1a79\") " pod="openstack/nova-metadata-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.046188 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g74ns\" (UniqueName: \"kubernetes.io/projected/ce80bccc-2b9e-44c9-a1c4-915c7bed4897-kube-api-access-g74ns\") pod \"nova-api-0\" (UID: \"ce80bccc-2b9e-44c9-a1c4-915c7bed4897\") " pod="openstack/nova-api-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.046244 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.048286 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.050042 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmthn\" (UniqueName: \"kubernetes.io/projected/3a9de870-885d-4c13-b8f2-1180378b1a79-kube-api-access-rmthn\") pod \"nova-metadata-0\" (UID: \"3a9de870-885d-4c13-b8f2-1180378b1a79\") " pod="openstack/nova-metadata-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.054191 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.054812 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a9de870-885d-4c13-b8f2-1180378b1a79-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3a9de870-885d-4c13-b8f2-1180378b1a79\") " pod="openstack/nova-metadata-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.061603 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bdbcf4d67-qklzz"] Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.065298 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bdbcf4d67-qklzz" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.081368 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.110633 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bdbcf4d67-qklzz"] Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.118611 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-dns-svc\") pod \"dnsmasq-dns-5bdbcf4d67-qklzz\" (UID: \"386eb4a1-747e-45d3-b665-fbb1a6c7b06b\") " pod="openstack/dnsmasq-dns-5bdbcf4d67-qklzz" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.118882 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5blf\" (UniqueName: \"kubernetes.io/projected/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-kube-api-access-b5blf\") pod \"dnsmasq-dns-5bdbcf4d67-qklzz\" (UID: \"386eb4a1-747e-45d3-b665-fbb1a6c7b06b\") " pod="openstack/dnsmasq-dns-5bdbcf4d67-qklzz" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.118947 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-ovsdbserver-sb\") pod \"dnsmasq-dns-5bdbcf4d67-qklzz\" (UID: \"386eb4a1-747e-45d3-b665-fbb1a6c7b06b\") " pod="openstack/dnsmasq-dns-5bdbcf4d67-qklzz" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.118981 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-ovsdbserver-nb\") pod \"dnsmasq-dns-5bdbcf4d67-qklzz\" (UID: \"386eb4a1-747e-45d3-b665-fbb1a6c7b06b\") " pod="openstack/dnsmasq-dns-5bdbcf4d67-qklzz" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.119004 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f3d6eb8-752d-45b3-97fb-14be1a6f3e38-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4f3d6eb8-752d-45b3-97fb-14be1a6f3e38\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.119056 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da7d17e7-2d62-466a-8826-817e197158d2-config-data\") pod \"nova-scheduler-0\" (UID: \"da7d17e7-2d62-466a-8826-817e197158d2\") " pod="openstack/nova-scheduler-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.119116 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da7d17e7-2d62-466a-8826-817e197158d2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"da7d17e7-2d62-466a-8826-817e197158d2\") " pod="openstack/nova-scheduler-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.119937 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9nkk\" (UniqueName: \"kubernetes.io/projected/da7d17e7-2d62-466a-8826-817e197158d2-kube-api-access-h9nkk\") pod \"nova-scheduler-0\" (UID: \"da7d17e7-2d62-466a-8826-817e197158d2\") " pod="openstack/nova-scheduler-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.120018 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f3d6eb8-752d-45b3-97fb-14be1a6f3e38-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4f3d6eb8-752d-45b3-97fb-14be1a6f3e38\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.120113 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-config\") pod \"dnsmasq-dns-5bdbcf4d67-qklzz\" (UID: \"386eb4a1-747e-45d3-b665-fbb1a6c7b06b\") " pod="openstack/dnsmasq-dns-5bdbcf4d67-qklzz" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.120183 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g66ff\" (UniqueName: \"kubernetes.io/projected/4f3d6eb8-752d-45b3-97fb-14be1a6f3e38-kube-api-access-g66ff\") pod \"nova-cell1-novncproxy-0\" (UID: \"4f3d6eb8-752d-45b3-97fb-14be1a6f3e38\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.123810 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da7d17e7-2d62-466a-8826-817e197158d2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"da7d17e7-2d62-466a-8826-817e197158d2\") " pod="openstack/nova-scheduler-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.129483 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.132658 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da7d17e7-2d62-466a-8826-817e197158d2-config-data\") pod \"nova-scheduler-0\" (UID: \"da7d17e7-2d62-466a-8826-817e197158d2\") " pod="openstack/nova-scheduler-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.145686 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9nkk\" (UniqueName: \"kubernetes.io/projected/da7d17e7-2d62-466a-8826-817e197158d2-kube-api-access-h9nkk\") pod \"nova-scheduler-0\" (UID: \"da7d17e7-2d62-466a-8826-817e197158d2\") " pod="openstack/nova-scheduler-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.148243 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.179380 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.223346 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-dns-svc\") pod \"dnsmasq-dns-5bdbcf4d67-qklzz\" (UID: \"386eb4a1-747e-45d3-b665-fbb1a6c7b06b\") " pod="openstack/dnsmasq-dns-5bdbcf4d67-qklzz" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.223450 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5blf\" (UniqueName: \"kubernetes.io/projected/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-kube-api-access-b5blf\") pod \"dnsmasq-dns-5bdbcf4d67-qklzz\" (UID: \"386eb4a1-747e-45d3-b665-fbb1a6c7b06b\") " pod="openstack/dnsmasq-dns-5bdbcf4d67-qklzz" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.223512 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-ovsdbserver-sb\") pod \"dnsmasq-dns-5bdbcf4d67-qklzz\" (UID: \"386eb4a1-747e-45d3-b665-fbb1a6c7b06b\") " pod="openstack/dnsmasq-dns-5bdbcf4d67-qklzz" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.223548 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-ovsdbserver-nb\") pod \"dnsmasq-dns-5bdbcf4d67-qklzz\" (UID: \"386eb4a1-747e-45d3-b665-fbb1a6c7b06b\") " pod="openstack/dnsmasq-dns-5bdbcf4d67-qklzz" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.223571 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f3d6eb8-752d-45b3-97fb-14be1a6f3e38-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4f3d6eb8-752d-45b3-97fb-14be1a6f3e38\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.223692 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f3d6eb8-752d-45b3-97fb-14be1a6f3e38-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4f3d6eb8-752d-45b3-97fb-14be1a6f3e38\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.223732 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-config\") pod \"dnsmasq-dns-5bdbcf4d67-qklzz\" (UID: \"386eb4a1-747e-45d3-b665-fbb1a6c7b06b\") " pod="openstack/dnsmasq-dns-5bdbcf4d67-qklzz" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.223772 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g66ff\" (UniqueName: \"kubernetes.io/projected/4f3d6eb8-752d-45b3-97fb-14be1a6f3e38-kube-api-access-g66ff\") pod \"nova-cell1-novncproxy-0\" (UID: \"4f3d6eb8-752d-45b3-97fb-14be1a6f3e38\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.224515 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-dns-svc\") pod \"dnsmasq-dns-5bdbcf4d67-qklzz\" (UID: \"386eb4a1-747e-45d3-b665-fbb1a6c7b06b\") " pod="openstack/dnsmasq-dns-5bdbcf4d67-qklzz" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.225316 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-ovsdbserver-sb\") pod \"dnsmasq-dns-5bdbcf4d67-qklzz\" (UID: \"386eb4a1-747e-45d3-b665-fbb1a6c7b06b\") " pod="openstack/dnsmasq-dns-5bdbcf4d67-qklzz" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.225935 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-config\") pod \"dnsmasq-dns-5bdbcf4d67-qklzz\" (UID: \"386eb4a1-747e-45d3-b665-fbb1a6c7b06b\") " pod="openstack/dnsmasq-dns-5bdbcf4d67-qklzz" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.229439 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f3d6eb8-752d-45b3-97fb-14be1a6f3e38-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"4f3d6eb8-752d-45b3-97fb-14be1a6f3e38\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.229501 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f3d6eb8-752d-45b3-97fb-14be1a6f3e38-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"4f3d6eb8-752d-45b3-97fb-14be1a6f3e38\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.230947 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-ovsdbserver-nb\") pod \"dnsmasq-dns-5bdbcf4d67-qklzz\" (UID: \"386eb4a1-747e-45d3-b665-fbb1a6c7b06b\") " pod="openstack/dnsmasq-dns-5bdbcf4d67-qklzz" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.242761 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g66ff\" (UniqueName: \"kubernetes.io/projected/4f3d6eb8-752d-45b3-97fb-14be1a6f3e38-kube-api-access-g66ff\") pod \"nova-cell1-novncproxy-0\" (UID: \"4f3d6eb8-752d-45b3-97fb-14be1a6f3e38\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.245255 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5blf\" (UniqueName: \"kubernetes.io/projected/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-kube-api-access-b5blf\") pod \"dnsmasq-dns-5bdbcf4d67-qklzz\" (UID: \"386eb4a1-747e-45d3-b665-fbb1a6c7b06b\") " pod="openstack/dnsmasq-dns-5bdbcf4d67-qklzz" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.467659 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.485589 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bdbcf4d67-qklzz" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.659450 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.695551 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5trlq"] Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.700576 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5trlq" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.704146 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.704218 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.716641 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5trlq"] Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.735199 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd2a4d0a-c3e5-47de-86ba-2ff48462f434-config-data\") pod \"nova-cell1-conductor-db-sync-5trlq\" (UID: \"bd2a4d0a-c3e5-47de-86ba-2ff48462f434\") " pod="openstack/nova-cell1-conductor-db-sync-5trlq" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.735349 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd2a4d0a-c3e5-47de-86ba-2ff48462f434-scripts\") pod \"nova-cell1-conductor-db-sync-5trlq\" (UID: \"bd2a4d0a-c3e5-47de-86ba-2ff48462f434\") " pod="openstack/nova-cell1-conductor-db-sync-5trlq" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.736437 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mm85f\" (UniqueName: \"kubernetes.io/projected/bd2a4d0a-c3e5-47de-86ba-2ff48462f434-kube-api-access-mm85f\") pod \"nova-cell1-conductor-db-sync-5trlq\" (UID: \"bd2a4d0a-c3e5-47de-86ba-2ff48462f434\") " pod="openstack/nova-cell1-conductor-db-sync-5trlq" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.736585 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd2a4d0a-c3e5-47de-86ba-2ff48462f434-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-5trlq\" (UID: \"bd2a4d0a-c3e5-47de-86ba-2ff48462f434\") " pod="openstack/nova-cell1-conductor-db-sync-5trlq" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.767297 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-k465n"] Nov 25 11:03:20 crc kubenswrapper[4776]: W1125 11:03:20.771371 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda7d17e7_2d62_466a_8826_817e197158d2.slice/crio-d28dcebf7005acb1f41f918e058acdb97c0ae95bb57d33c71f4310dcd494b739 WatchSource:0}: Error finding container d28dcebf7005acb1f41f918e058acdb97c0ae95bb57d33c71f4310dcd494b739: Status 404 returned error can't find the container with id d28dcebf7005acb1f41f918e058acdb97c0ae95bb57d33c71f4310dcd494b739 Nov 25 11:03:20 crc kubenswrapper[4776]: W1125 11:03:20.772670 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c54defc_dcc0_416c_9d88_fbc5944aa0e3.slice/crio-b4f4cb9dc4ebf7a4a8453a08fd66bad7d0e81829312e1fe66fbaaf0632fb3dd3 WatchSource:0}: Error finding container b4f4cb9dc4ebf7a4a8453a08fd66bad7d0e81829312e1fe66fbaaf0632fb3dd3: Status 404 returned error can't find the container with id b4f4cb9dc4ebf7a4a8453a08fd66bad7d0e81829312e1fe66fbaaf0632fb3dd3 Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.779312 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.838436 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd2a4d0a-c3e5-47de-86ba-2ff48462f434-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-5trlq\" (UID: \"bd2a4d0a-c3e5-47de-86ba-2ff48462f434\") " pod="openstack/nova-cell1-conductor-db-sync-5trlq" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.838546 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd2a4d0a-c3e5-47de-86ba-2ff48462f434-config-data\") pod \"nova-cell1-conductor-db-sync-5trlq\" (UID: \"bd2a4d0a-c3e5-47de-86ba-2ff48462f434\") " pod="openstack/nova-cell1-conductor-db-sync-5trlq" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.838634 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd2a4d0a-c3e5-47de-86ba-2ff48462f434-scripts\") pod \"nova-cell1-conductor-db-sync-5trlq\" (UID: \"bd2a4d0a-c3e5-47de-86ba-2ff48462f434\") " pod="openstack/nova-cell1-conductor-db-sync-5trlq" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.838685 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mm85f\" (UniqueName: \"kubernetes.io/projected/bd2a4d0a-c3e5-47de-86ba-2ff48462f434-kube-api-access-mm85f\") pod \"nova-cell1-conductor-db-sync-5trlq\" (UID: \"bd2a4d0a-c3e5-47de-86ba-2ff48462f434\") " pod="openstack/nova-cell1-conductor-db-sync-5trlq" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.842697 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd2a4d0a-c3e5-47de-86ba-2ff48462f434-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-5trlq\" (UID: \"bd2a4d0a-c3e5-47de-86ba-2ff48462f434\") " pod="openstack/nova-cell1-conductor-db-sync-5trlq" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.846197 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd2a4d0a-c3e5-47de-86ba-2ff48462f434-config-data\") pod \"nova-cell1-conductor-db-sync-5trlq\" (UID: \"bd2a4d0a-c3e5-47de-86ba-2ff48462f434\") " pod="openstack/nova-cell1-conductor-db-sync-5trlq" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.848544 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd2a4d0a-c3e5-47de-86ba-2ff48462f434-scripts\") pod \"nova-cell1-conductor-db-sync-5trlq\" (UID: \"bd2a4d0a-c3e5-47de-86ba-2ff48462f434\") " pod="openstack/nova-cell1-conductor-db-sync-5trlq" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.857359 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mm85f\" (UniqueName: \"kubernetes.io/projected/bd2a4d0a-c3e5-47de-86ba-2ff48462f434-kube-api-access-mm85f\") pod \"nova-cell1-conductor-db-sync-5trlq\" (UID: \"bd2a4d0a-c3e5-47de-86ba-2ff48462f434\") " pod="openstack/nova-cell1-conductor-db-sync-5trlq" Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.858004 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.861304 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce80bccc-2b9e-44c9-a1c4-915c7bed4897","Type":"ContainerStarted","Data":"a6ce6d5e5b35cc668367146b3c9db71ca8015268b8e8def7afbdee2150a92503"} Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.864298 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"da7d17e7-2d62-466a-8826-817e197158d2","Type":"ContainerStarted","Data":"d28dcebf7005acb1f41f918e058acdb97c0ae95bb57d33c71f4310dcd494b739"} Nov 25 11:03:20 crc kubenswrapper[4776]: I1125 11:03:20.865686 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-k465n" event={"ID":"9c54defc-dcc0-416c-9d88-fbc5944aa0e3","Type":"ContainerStarted","Data":"b4f4cb9dc4ebf7a4a8453a08fd66bad7d0e81829312e1fe66fbaaf0632fb3dd3"} Nov 25 11:03:21 crc kubenswrapper[4776]: I1125 11:03:21.023722 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5trlq" Nov 25 11:03:21 crc kubenswrapper[4776]: I1125 11:03:21.059025 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bdbcf4d67-qklzz"] Nov 25 11:03:21 crc kubenswrapper[4776]: I1125 11:03:21.069352 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 11:03:21 crc kubenswrapper[4776]: W1125 11:03:21.077063 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f3d6eb8_752d_45b3_97fb_14be1a6f3e38.slice/crio-67578796b556971fc2bc41c3ee9401c0664fb98ac7b5b63ff67ca95ee5d3bce0 WatchSource:0}: Error finding container 67578796b556971fc2bc41c3ee9401c0664fb98ac7b5b63ff67ca95ee5d3bce0: Status 404 returned error can't find the container with id 67578796b556971fc2bc41c3ee9401c0664fb98ac7b5b63ff67ca95ee5d3bce0 Nov 25 11:03:21 crc kubenswrapper[4776]: I1125 11:03:21.517728 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5trlq"] Nov 25 11:03:21 crc kubenswrapper[4776]: I1125 11:03:21.876796 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3a9de870-885d-4c13-b8f2-1180378b1a79","Type":"ContainerStarted","Data":"87f1f9e32986f80ab812da14b020b64c544b330e6ffb20a8522319e9b4c2526d"} Nov 25 11:03:21 crc kubenswrapper[4776]: I1125 11:03:21.877262 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3a9de870-885d-4c13-b8f2-1180378b1a79","Type":"ContainerStarted","Data":"6c1f03714242cc42c8fa4a89f7a288ff8175b50f90198e6c57a182feb519be6f"} Nov 25 11:03:21 crc kubenswrapper[4776]: I1125 11:03:21.877274 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3a9de870-885d-4c13-b8f2-1180378b1a79","Type":"ContainerStarted","Data":"0e740d0a5aecb96bdd62d3a47f3ddbf490a209e25501b97ce6f503d379f2ebfc"} Nov 25 11:03:21 crc kubenswrapper[4776]: I1125 11:03:21.879277 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5trlq" event={"ID":"bd2a4d0a-c3e5-47de-86ba-2ff48462f434","Type":"ContainerStarted","Data":"0de30baae2bdfa0611870b2de52c24d2c88d042214876580178e3cef4f812d78"} Nov 25 11:03:21 crc kubenswrapper[4776]: I1125 11:03:21.879347 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5trlq" event={"ID":"bd2a4d0a-c3e5-47de-86ba-2ff48462f434","Type":"ContainerStarted","Data":"7c3dc699885177c5794b741c87b314688b8720f638deae1d727eed5e55db9261"} Nov 25 11:03:21 crc kubenswrapper[4776]: I1125 11:03:21.882963 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce80bccc-2b9e-44c9-a1c4-915c7bed4897","Type":"ContainerStarted","Data":"1e3ccbb2372f3500a05c901e101261561c662b2dfca0cd1b849963390eb2b1f0"} Nov 25 11:03:21 crc kubenswrapper[4776]: I1125 11:03:21.883004 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce80bccc-2b9e-44c9-a1c4-915c7bed4897","Type":"ContainerStarted","Data":"61288b2a6915a828bb1b74bb60f6c0c96c552095ba338733f0a9e8617eecec3d"} Nov 25 11:03:21 crc kubenswrapper[4776]: I1125 11:03:21.884624 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"da7d17e7-2d62-466a-8826-817e197158d2","Type":"ContainerStarted","Data":"9e210dd879b5818e9d42a43ebabfe457a1b3f8c12a741c85c61fc945bfd23aa0"} Nov 25 11:03:21 crc kubenswrapper[4776]: I1125 11:03:21.888866 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4f3d6eb8-752d-45b3-97fb-14be1a6f3e38","Type":"ContainerStarted","Data":"ed7642bc389ab7773ecf072488877b11eeb13b77d9371805fa4ae2852fb5157b"} Nov 25 11:03:21 crc kubenswrapper[4776]: I1125 11:03:21.888903 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4f3d6eb8-752d-45b3-97fb-14be1a6f3e38","Type":"ContainerStarted","Data":"67578796b556971fc2bc41c3ee9401c0664fb98ac7b5b63ff67ca95ee5d3bce0"} Nov 25 11:03:21 crc kubenswrapper[4776]: I1125 11:03:21.892246 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-k465n" event={"ID":"9c54defc-dcc0-416c-9d88-fbc5944aa0e3","Type":"ContainerStarted","Data":"a80cec02848f3a1706589c3f1ffe0c08307a0c0588d9539181831f6e8d7221d2"} Nov 25 11:03:21 crc kubenswrapper[4776]: I1125 11:03:21.901439 4776 generic.go:334] "Generic (PLEG): container finished" podID="386eb4a1-747e-45d3-b665-fbb1a6c7b06b" containerID="bc26d14a9a86b5d052663b9b6e0df405a57c8fd4a7e1a22acae4fa55c7eccd22" exitCode=0 Nov 25 11:03:21 crc kubenswrapper[4776]: I1125 11:03:21.901521 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bdbcf4d67-qklzz" event={"ID":"386eb4a1-747e-45d3-b665-fbb1a6c7b06b","Type":"ContainerDied","Data":"bc26d14a9a86b5d052663b9b6e0df405a57c8fd4a7e1a22acae4fa55c7eccd22"} Nov 25 11:03:21 crc kubenswrapper[4776]: I1125 11:03:21.901550 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bdbcf4d67-qklzz" event={"ID":"386eb4a1-747e-45d3-b665-fbb1a6c7b06b","Type":"ContainerStarted","Data":"4fb31ebee5814aaa27e8e537b779385a3308b339fe6d5fda34f50880de3e44a7"} Nov 25 11:03:21 crc kubenswrapper[4776]: I1125 11:03:21.905756 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.90573724 podStartE2EDuration="2.90573724s" podCreationTimestamp="2025-11-25 11:03:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:03:21.896259553 +0000 UTC m=+5946.937319106" watchObservedRunningTime="2025-11-25 11:03:21.90573724 +0000 UTC m=+5946.946796793" Nov 25 11:03:21 crc kubenswrapper[4776]: I1125 11:03:21.919717 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.919701051 podStartE2EDuration="2.919701051s" podCreationTimestamp="2025-11-25 11:03:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:03:21.915438044 +0000 UTC m=+5946.956497587" watchObservedRunningTime="2025-11-25 11:03:21.919701051 +0000 UTC m=+5946.960760604" Nov 25 11:03:21 crc kubenswrapper[4776]: I1125 11:03:21.936573 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.936531283 podStartE2EDuration="2.936531283s" podCreationTimestamp="2025-11-25 11:03:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:03:21.93519102 +0000 UTC m=+5946.976250593" watchObservedRunningTime="2025-11-25 11:03:21.936531283 +0000 UTC m=+5946.977590846" Nov 25 11:03:21 crc kubenswrapper[4776]: I1125 11:03:21.965105 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.96508634 podStartE2EDuration="2.96508634s" podCreationTimestamp="2025-11-25 11:03:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:03:21.963678075 +0000 UTC m=+5947.004737618" watchObservedRunningTime="2025-11-25 11:03:21.96508634 +0000 UTC m=+5947.006145893" Nov 25 11:03:21 crc kubenswrapper[4776]: I1125 11:03:21.984376 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-k465n" podStartSLOduration=2.984361394 podStartE2EDuration="2.984361394s" podCreationTimestamp="2025-11-25 11:03:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:03:21.982190289 +0000 UTC m=+5947.023249842" watchObservedRunningTime="2025-11-25 11:03:21.984361394 +0000 UTC m=+5947.025420947" Nov 25 11:03:22 crc kubenswrapper[4776]: I1125 11:03:22.011339 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-5trlq" podStartSLOduration=2.01131557 podStartE2EDuration="2.01131557s" podCreationTimestamp="2025-11-25 11:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:03:22.003544675 +0000 UTC m=+5947.044604228" watchObservedRunningTime="2025-11-25 11:03:22.01131557 +0000 UTC m=+5947.052375123" Nov 25 11:03:22 crc kubenswrapper[4776]: I1125 11:03:22.912718 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bdbcf4d67-qklzz" event={"ID":"386eb4a1-747e-45d3-b665-fbb1a6c7b06b","Type":"ContainerStarted","Data":"9b819bc49d6e3e198bb8825480a3e5ed1756ef83efdf29cd7e01996b5f25e992"} Nov 25 11:03:22 crc kubenswrapper[4776]: I1125 11:03:22.959192 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5bdbcf4d67-qklzz" podStartSLOduration=3.959178509 podStartE2EDuration="3.959178509s" podCreationTimestamp="2025-11-25 11:03:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:03:22.958398939 +0000 UTC m=+5947.999458492" watchObservedRunningTime="2025-11-25 11:03:22.959178509 +0000 UTC m=+5948.000238062" Nov 25 11:03:23 crc kubenswrapper[4776]: I1125 11:03:23.592876 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 11:03:23 crc kubenswrapper[4776]: I1125 11:03:23.604522 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 11:03:23 crc kubenswrapper[4776]: I1125 11:03:23.919934 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3a9de870-885d-4c13-b8f2-1180378b1a79" containerName="nova-metadata-log" containerID="cri-o://6c1f03714242cc42c8fa4a89f7a288ff8175b50f90198e6c57a182feb519be6f" gracePeriod=30 Nov 25 11:03:23 crc kubenswrapper[4776]: I1125 11:03:23.919947 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="4f3d6eb8-752d-45b3-97fb-14be1a6f3e38" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://ed7642bc389ab7773ecf072488877b11eeb13b77d9371805fa4ae2852fb5157b" gracePeriod=30 Nov 25 11:03:23 crc kubenswrapper[4776]: I1125 11:03:23.919947 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3a9de870-885d-4c13-b8f2-1180378b1a79" containerName="nova-metadata-metadata" containerID="cri-o://87f1f9e32986f80ab812da14b020b64c544b330e6ffb20a8522319e9b4c2526d" gracePeriod=30 Nov 25 11:03:23 crc kubenswrapper[4776]: I1125 11:03:23.920220 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5bdbcf4d67-qklzz" Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.507247 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.631645 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.648437 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmthn\" (UniqueName: \"kubernetes.io/projected/3a9de870-885d-4c13-b8f2-1180378b1a79-kube-api-access-rmthn\") pod \"3a9de870-885d-4c13-b8f2-1180378b1a79\" (UID: \"3a9de870-885d-4c13-b8f2-1180378b1a79\") " Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.648519 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a9de870-885d-4c13-b8f2-1180378b1a79-config-data\") pod \"3a9de870-885d-4c13-b8f2-1180378b1a79\" (UID: \"3a9de870-885d-4c13-b8f2-1180378b1a79\") " Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.648654 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a9de870-885d-4c13-b8f2-1180378b1a79-combined-ca-bundle\") pod \"3a9de870-885d-4c13-b8f2-1180378b1a79\" (UID: \"3a9de870-885d-4c13-b8f2-1180378b1a79\") " Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.648741 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a9de870-885d-4c13-b8f2-1180378b1a79-logs\") pod \"3a9de870-885d-4c13-b8f2-1180378b1a79\" (UID: \"3a9de870-885d-4c13-b8f2-1180378b1a79\") " Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.650446 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a9de870-885d-4c13-b8f2-1180378b1a79-logs" (OuterVolumeSpecName: "logs") pod "3a9de870-885d-4c13-b8f2-1180378b1a79" (UID: "3a9de870-885d-4c13-b8f2-1180378b1a79"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.656348 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a9de870-885d-4c13-b8f2-1180378b1a79-kube-api-access-rmthn" (OuterVolumeSpecName: "kube-api-access-rmthn") pod "3a9de870-885d-4c13-b8f2-1180378b1a79" (UID: "3a9de870-885d-4c13-b8f2-1180378b1a79"). InnerVolumeSpecName "kube-api-access-rmthn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.685085 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a9de870-885d-4c13-b8f2-1180378b1a79-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a9de870-885d-4c13-b8f2-1180378b1a79" (UID: "3a9de870-885d-4c13-b8f2-1180378b1a79"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.687325 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a9de870-885d-4c13-b8f2-1180378b1a79-config-data" (OuterVolumeSpecName: "config-data") pod "3a9de870-885d-4c13-b8f2-1180378b1a79" (UID: "3a9de870-885d-4c13-b8f2-1180378b1a79"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.750903 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f3d6eb8-752d-45b3-97fb-14be1a6f3e38-combined-ca-bundle\") pod \"4f3d6eb8-752d-45b3-97fb-14be1a6f3e38\" (UID: \"4f3d6eb8-752d-45b3-97fb-14be1a6f3e38\") " Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.751519 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g66ff\" (UniqueName: \"kubernetes.io/projected/4f3d6eb8-752d-45b3-97fb-14be1a6f3e38-kube-api-access-g66ff\") pod \"4f3d6eb8-752d-45b3-97fb-14be1a6f3e38\" (UID: \"4f3d6eb8-752d-45b3-97fb-14be1a6f3e38\") " Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.751580 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f3d6eb8-752d-45b3-97fb-14be1a6f3e38-config-data\") pod \"4f3d6eb8-752d-45b3-97fb-14be1a6f3e38\" (UID: \"4f3d6eb8-752d-45b3-97fb-14be1a6f3e38\") " Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.752269 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmthn\" (UniqueName: \"kubernetes.io/projected/3a9de870-885d-4c13-b8f2-1180378b1a79-kube-api-access-rmthn\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.752293 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a9de870-885d-4c13-b8f2-1180378b1a79-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.752306 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a9de870-885d-4c13-b8f2-1180378b1a79-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.752317 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a9de870-885d-4c13-b8f2-1180378b1a79-logs\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.757904 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f3d6eb8-752d-45b3-97fb-14be1a6f3e38-kube-api-access-g66ff" (OuterVolumeSpecName: "kube-api-access-g66ff") pod "4f3d6eb8-752d-45b3-97fb-14be1a6f3e38" (UID: "4f3d6eb8-752d-45b3-97fb-14be1a6f3e38"). InnerVolumeSpecName "kube-api-access-g66ff". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.774026 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f3d6eb8-752d-45b3-97fb-14be1a6f3e38-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4f3d6eb8-752d-45b3-97fb-14be1a6f3e38" (UID: "4f3d6eb8-752d-45b3-97fb-14be1a6f3e38"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.781518 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f3d6eb8-752d-45b3-97fb-14be1a6f3e38-config-data" (OuterVolumeSpecName: "config-data") pod "4f3d6eb8-752d-45b3-97fb-14be1a6f3e38" (UID: "4f3d6eb8-752d-45b3-97fb-14be1a6f3e38"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.854053 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g66ff\" (UniqueName: \"kubernetes.io/projected/4f3d6eb8-752d-45b3-97fb-14be1a6f3e38-kube-api-access-g66ff\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.854455 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f3d6eb8-752d-45b3-97fb-14be1a6f3e38-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.854469 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f3d6eb8-752d-45b3-97fb-14be1a6f3e38-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.932049 4776 generic.go:334] "Generic (PLEG): container finished" podID="4f3d6eb8-752d-45b3-97fb-14be1a6f3e38" containerID="ed7642bc389ab7773ecf072488877b11eeb13b77d9371805fa4ae2852fb5157b" exitCode=0 Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.932141 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4f3d6eb8-752d-45b3-97fb-14be1a6f3e38","Type":"ContainerDied","Data":"ed7642bc389ab7773ecf072488877b11eeb13b77d9371805fa4ae2852fb5157b"} Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.932169 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"4f3d6eb8-752d-45b3-97fb-14be1a6f3e38","Type":"ContainerDied","Data":"67578796b556971fc2bc41c3ee9401c0664fb98ac7b5b63ff67ca95ee5d3bce0"} Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.932184 4776 scope.go:117] "RemoveContainer" containerID="ed7642bc389ab7773ecf072488877b11eeb13b77d9371805fa4ae2852fb5157b" Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.932292 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.944522 4776 generic.go:334] "Generic (PLEG): container finished" podID="3a9de870-885d-4c13-b8f2-1180378b1a79" containerID="87f1f9e32986f80ab812da14b020b64c544b330e6ffb20a8522319e9b4c2526d" exitCode=0 Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.944547 4776 generic.go:334] "Generic (PLEG): container finished" podID="3a9de870-885d-4c13-b8f2-1180378b1a79" containerID="6c1f03714242cc42c8fa4a89f7a288ff8175b50f90198e6c57a182feb519be6f" exitCode=143 Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.944583 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3a9de870-885d-4c13-b8f2-1180378b1a79","Type":"ContainerDied","Data":"87f1f9e32986f80ab812da14b020b64c544b330e6ffb20a8522319e9b4c2526d"} Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.944607 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3a9de870-885d-4c13-b8f2-1180378b1a79","Type":"ContainerDied","Data":"6c1f03714242cc42c8fa4a89f7a288ff8175b50f90198e6c57a182feb519be6f"} Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.944617 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3a9de870-885d-4c13-b8f2-1180378b1a79","Type":"ContainerDied","Data":"0e740d0a5aecb96bdd62d3a47f3ddbf490a209e25501b97ce6f503d379f2ebfc"} Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.944678 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.948311 4776 generic.go:334] "Generic (PLEG): container finished" podID="bd2a4d0a-c3e5-47de-86ba-2ff48462f434" containerID="0de30baae2bdfa0611870b2de52c24d2c88d042214876580178e3cef4f812d78" exitCode=0 Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.948432 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5trlq" event={"ID":"bd2a4d0a-c3e5-47de-86ba-2ff48462f434","Type":"ContainerDied","Data":"0de30baae2bdfa0611870b2de52c24d2c88d042214876580178e3cef4f812d78"} Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.969429 4776 scope.go:117] "RemoveContainer" containerID="ed7642bc389ab7773ecf072488877b11eeb13b77d9371805fa4ae2852fb5157b" Nov 25 11:03:24 crc kubenswrapper[4776]: E1125 11:03:24.972054 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed7642bc389ab7773ecf072488877b11eeb13b77d9371805fa4ae2852fb5157b\": container with ID starting with ed7642bc389ab7773ecf072488877b11eeb13b77d9371805fa4ae2852fb5157b not found: ID does not exist" containerID="ed7642bc389ab7773ecf072488877b11eeb13b77d9371805fa4ae2852fb5157b" Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.972124 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed7642bc389ab7773ecf072488877b11eeb13b77d9371805fa4ae2852fb5157b"} err="failed to get container status \"ed7642bc389ab7773ecf072488877b11eeb13b77d9371805fa4ae2852fb5157b\": rpc error: code = NotFound desc = could not find container \"ed7642bc389ab7773ecf072488877b11eeb13b77d9371805fa4ae2852fb5157b\": container with ID starting with ed7642bc389ab7773ecf072488877b11eeb13b77d9371805fa4ae2852fb5157b not found: ID does not exist" Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.972157 4776 scope.go:117] "RemoveContainer" containerID="87f1f9e32986f80ab812da14b020b64c544b330e6ffb20a8522319e9b4c2526d" Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.982624 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 11:03:24 crc kubenswrapper[4776]: I1125 11:03:24.991799 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.008535 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 11:03:25 crc kubenswrapper[4776]: E1125 11:03:25.009948 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a9de870-885d-4c13-b8f2-1180378b1a79" containerName="nova-metadata-metadata" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.010090 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a9de870-885d-4c13-b8f2-1180378b1a79" containerName="nova-metadata-metadata" Nov 25 11:03:25 crc kubenswrapper[4776]: E1125 11:03:25.010174 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f3d6eb8-752d-45b3-97fb-14be1a6f3e38" containerName="nova-cell1-novncproxy-novncproxy" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.010266 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f3d6eb8-752d-45b3-97fb-14be1a6f3e38" containerName="nova-cell1-novncproxy-novncproxy" Nov 25 11:03:25 crc kubenswrapper[4776]: E1125 11:03:25.010367 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a9de870-885d-4c13-b8f2-1180378b1a79" containerName="nova-metadata-log" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.010448 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a9de870-885d-4c13-b8f2-1180378b1a79" containerName="nova-metadata-log" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.010768 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f3d6eb8-752d-45b3-97fb-14be1a6f3e38" containerName="nova-cell1-novncproxy-novncproxy" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.010864 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a9de870-885d-4c13-b8f2-1180378b1a79" containerName="nova-metadata-metadata" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.010986 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a9de870-885d-4c13-b8f2-1180378b1a79" containerName="nova-metadata-log" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.011889 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.014938 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.015276 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.016018 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.030583 4776 scope.go:117] "RemoveContainer" containerID="6c1f03714242cc42c8fa4a89f7a288ff8175b50f90198e6c57a182feb519be6f" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.047125 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.079214 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.087582 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.145708 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.151210 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.152119 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.159773 4776 scope.go:117] "RemoveContainer" containerID="87f1f9e32986f80ab812da14b020b64c544b330e6ffb20a8522319e9b4c2526d" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.160490 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.160565 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.161015 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a40b8b0e-76a4-402b-962d-46e22f5bbaef-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a40b8b0e-76a4-402b-962d-46e22f5bbaef\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.161212 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/a40b8b0e-76a4-402b-962d-46e22f5bbaef-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a40b8b0e-76a4-402b-962d-46e22f5bbaef\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.161264 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/a40b8b0e-76a4-402b-962d-46e22f5bbaef-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a40b8b0e-76a4-402b-962d-46e22f5bbaef\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.161418 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nn6wq\" (UniqueName: \"kubernetes.io/projected/a40b8b0e-76a4-402b-962d-46e22f5bbaef-kube-api-access-nn6wq\") pod \"nova-cell1-novncproxy-0\" (UID: \"a40b8b0e-76a4-402b-962d-46e22f5bbaef\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.161642 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a40b8b0e-76a4-402b-962d-46e22f5bbaef-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a40b8b0e-76a4-402b-962d-46e22f5bbaef\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:25 crc kubenswrapper[4776]: E1125 11:03:25.161968 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87f1f9e32986f80ab812da14b020b64c544b330e6ffb20a8522319e9b4c2526d\": container with ID starting with 87f1f9e32986f80ab812da14b020b64c544b330e6ffb20a8522319e9b4c2526d not found: ID does not exist" containerID="87f1f9e32986f80ab812da14b020b64c544b330e6ffb20a8522319e9b4c2526d" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.162006 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87f1f9e32986f80ab812da14b020b64c544b330e6ffb20a8522319e9b4c2526d"} err="failed to get container status \"87f1f9e32986f80ab812da14b020b64c544b330e6ffb20a8522319e9b4c2526d\": rpc error: code = NotFound desc = could not find container \"87f1f9e32986f80ab812da14b020b64c544b330e6ffb20a8522319e9b4c2526d\": container with ID starting with 87f1f9e32986f80ab812da14b020b64c544b330e6ffb20a8522319e9b4c2526d not found: ID does not exist" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.162032 4776 scope.go:117] "RemoveContainer" containerID="6c1f03714242cc42c8fa4a89f7a288ff8175b50f90198e6c57a182feb519be6f" Nov 25 11:03:25 crc kubenswrapper[4776]: E1125 11:03:25.163438 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c1f03714242cc42c8fa4a89f7a288ff8175b50f90198e6c57a182feb519be6f\": container with ID starting with 6c1f03714242cc42c8fa4a89f7a288ff8175b50f90198e6c57a182feb519be6f not found: ID does not exist" containerID="6c1f03714242cc42c8fa4a89f7a288ff8175b50f90198e6c57a182feb519be6f" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.163482 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c1f03714242cc42c8fa4a89f7a288ff8175b50f90198e6c57a182feb519be6f"} err="failed to get container status \"6c1f03714242cc42c8fa4a89f7a288ff8175b50f90198e6c57a182feb519be6f\": rpc error: code = NotFound desc = could not find container \"6c1f03714242cc42c8fa4a89f7a288ff8175b50f90198e6c57a182feb519be6f\": container with ID starting with 6c1f03714242cc42c8fa4a89f7a288ff8175b50f90198e6c57a182feb519be6f not found: ID does not exist" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.163516 4776 scope.go:117] "RemoveContainer" containerID="87f1f9e32986f80ab812da14b020b64c544b330e6ffb20a8522319e9b4c2526d" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.163808 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87f1f9e32986f80ab812da14b020b64c544b330e6ffb20a8522319e9b4c2526d"} err="failed to get container status \"87f1f9e32986f80ab812da14b020b64c544b330e6ffb20a8522319e9b4c2526d\": rpc error: code = NotFound desc = could not find container \"87f1f9e32986f80ab812da14b020b64c544b330e6ffb20a8522319e9b4c2526d\": container with ID starting with 87f1f9e32986f80ab812da14b020b64c544b330e6ffb20a8522319e9b4c2526d not found: ID does not exist" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.163838 4776 scope.go:117] "RemoveContainer" containerID="6c1f03714242cc42c8fa4a89f7a288ff8175b50f90198e6c57a182feb519be6f" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.164145 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c1f03714242cc42c8fa4a89f7a288ff8175b50f90198e6c57a182feb519be6f"} err="failed to get container status \"6c1f03714242cc42c8fa4a89f7a288ff8175b50f90198e6c57a182feb519be6f\": rpc error: code = NotFound desc = could not find container \"6c1f03714242cc42c8fa4a89f7a288ff8175b50f90198e6c57a182feb519be6f\": container with ID starting with 6c1f03714242cc42c8fa4a89f7a288ff8175b50f90198e6c57a182feb519be6f not found: ID does not exist" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.165196 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.263835 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a40b8b0e-76a4-402b-962d-46e22f5bbaef-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a40b8b0e-76a4-402b-962d-46e22f5bbaef\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.263986 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50aa9cec-96b4-4436-b03f-5ccce4696e92-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"50aa9cec-96b4-4436-b03f-5ccce4696e92\") " pod="openstack/nova-metadata-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.264087 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50aa9cec-96b4-4436-b03f-5ccce4696e92-logs\") pod \"nova-metadata-0\" (UID: \"50aa9cec-96b4-4436-b03f-5ccce4696e92\") " pod="openstack/nova-metadata-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.264124 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50aa9cec-96b4-4436-b03f-5ccce4696e92-config-data\") pod \"nova-metadata-0\" (UID: \"50aa9cec-96b4-4436-b03f-5ccce4696e92\") " pod="openstack/nova-metadata-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.264179 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/a40b8b0e-76a4-402b-962d-46e22f5bbaef-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a40b8b0e-76a4-402b-962d-46e22f5bbaef\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.264355 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/a40b8b0e-76a4-402b-962d-46e22f5bbaef-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a40b8b0e-76a4-402b-962d-46e22f5bbaef\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.264537 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nn6wq\" (UniqueName: \"kubernetes.io/projected/a40b8b0e-76a4-402b-962d-46e22f5bbaef-kube-api-access-nn6wq\") pod \"nova-cell1-novncproxy-0\" (UID: \"a40b8b0e-76a4-402b-962d-46e22f5bbaef\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.264637 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n5bn\" (UniqueName: \"kubernetes.io/projected/50aa9cec-96b4-4436-b03f-5ccce4696e92-kube-api-access-9n5bn\") pod \"nova-metadata-0\" (UID: \"50aa9cec-96b4-4436-b03f-5ccce4696e92\") " pod="openstack/nova-metadata-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.264773 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a40b8b0e-76a4-402b-962d-46e22f5bbaef-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a40b8b0e-76a4-402b-962d-46e22f5bbaef\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.264865 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/50aa9cec-96b4-4436-b03f-5ccce4696e92-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"50aa9cec-96b4-4436-b03f-5ccce4696e92\") " pod="openstack/nova-metadata-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.283428 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/a40b8b0e-76a4-402b-962d-46e22f5bbaef-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a40b8b0e-76a4-402b-962d-46e22f5bbaef\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.283468 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/a40b8b0e-76a4-402b-962d-46e22f5bbaef-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a40b8b0e-76a4-402b-962d-46e22f5bbaef\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.285228 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a40b8b0e-76a4-402b-962d-46e22f5bbaef-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a40b8b0e-76a4-402b-962d-46e22f5bbaef\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.286397 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nn6wq\" (UniqueName: \"kubernetes.io/projected/a40b8b0e-76a4-402b-962d-46e22f5bbaef-kube-api-access-nn6wq\") pod \"nova-cell1-novncproxy-0\" (UID: \"a40b8b0e-76a4-402b-962d-46e22f5bbaef\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.288592 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a40b8b0e-76a4-402b-962d-46e22f5bbaef-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a40b8b0e-76a4-402b-962d-46e22f5bbaef\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.366682 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50aa9cec-96b4-4436-b03f-5ccce4696e92-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"50aa9cec-96b4-4436-b03f-5ccce4696e92\") " pod="openstack/nova-metadata-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.366774 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50aa9cec-96b4-4436-b03f-5ccce4696e92-logs\") pod \"nova-metadata-0\" (UID: \"50aa9cec-96b4-4436-b03f-5ccce4696e92\") " pod="openstack/nova-metadata-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.366797 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50aa9cec-96b4-4436-b03f-5ccce4696e92-config-data\") pod \"nova-metadata-0\" (UID: \"50aa9cec-96b4-4436-b03f-5ccce4696e92\") " pod="openstack/nova-metadata-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.366850 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n5bn\" (UniqueName: \"kubernetes.io/projected/50aa9cec-96b4-4436-b03f-5ccce4696e92-kube-api-access-9n5bn\") pod \"nova-metadata-0\" (UID: \"50aa9cec-96b4-4436-b03f-5ccce4696e92\") " pod="openstack/nova-metadata-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.366891 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/50aa9cec-96b4-4436-b03f-5ccce4696e92-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"50aa9cec-96b4-4436-b03f-5ccce4696e92\") " pod="openstack/nova-metadata-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.367712 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50aa9cec-96b4-4436-b03f-5ccce4696e92-logs\") pod \"nova-metadata-0\" (UID: \"50aa9cec-96b4-4436-b03f-5ccce4696e92\") " pod="openstack/nova-metadata-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.369977 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/50aa9cec-96b4-4436-b03f-5ccce4696e92-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"50aa9cec-96b4-4436-b03f-5ccce4696e92\") " pod="openstack/nova-metadata-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.370203 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50aa9cec-96b4-4436-b03f-5ccce4696e92-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"50aa9cec-96b4-4436-b03f-5ccce4696e92\") " pod="openstack/nova-metadata-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.371836 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50aa9cec-96b4-4436-b03f-5ccce4696e92-config-data\") pod \"nova-metadata-0\" (UID: \"50aa9cec-96b4-4436-b03f-5ccce4696e92\") " pod="openstack/nova-metadata-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.382980 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n5bn\" (UniqueName: \"kubernetes.io/projected/50aa9cec-96b4-4436-b03f-5ccce4696e92-kube-api-access-9n5bn\") pod \"nova-metadata-0\" (UID: \"50aa9cec-96b4-4436-b03f-5ccce4696e92\") " pod="openstack/nova-metadata-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.425022 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.477903 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.678718 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a9de870-885d-4c13-b8f2-1180378b1a79" path="/var/lib/kubelet/pods/3a9de870-885d-4c13-b8f2-1180378b1a79/volumes" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.679781 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f3d6eb8-752d-45b3-97fb-14be1a6f3e38" path="/var/lib/kubelet/pods/4f3d6eb8-752d-45b3-97fb-14be1a6f3e38/volumes" Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.847905 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 11:03:25 crc kubenswrapper[4776]: W1125 11:03:25.854836 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda40b8b0e_76a4_402b_962d_46e22f5bbaef.slice/crio-5ec718af3606cd6b9bbca0b425a2f41ab2c0cf0496733f19255cab12d7469165 WatchSource:0}: Error finding container 5ec718af3606cd6b9bbca0b425a2f41ab2c0cf0496733f19255cab12d7469165: Status 404 returned error can't find the container with id 5ec718af3606cd6b9bbca0b425a2f41ab2c0cf0496733f19255cab12d7469165 Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.943385 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 11:03:25 crc kubenswrapper[4776]: I1125 11:03:25.968280 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a40b8b0e-76a4-402b-962d-46e22f5bbaef","Type":"ContainerStarted","Data":"5ec718af3606cd6b9bbca0b425a2f41ab2c0cf0496733f19255cab12d7469165"} Nov 25 11:03:26 crc kubenswrapper[4776]: I1125 11:03:26.421259 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5trlq" Nov 25 11:03:26 crc kubenswrapper[4776]: I1125 11:03:26.610968 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd2a4d0a-c3e5-47de-86ba-2ff48462f434-scripts\") pod \"bd2a4d0a-c3e5-47de-86ba-2ff48462f434\" (UID: \"bd2a4d0a-c3e5-47de-86ba-2ff48462f434\") " Nov 25 11:03:26 crc kubenswrapper[4776]: I1125 11:03:26.611187 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd2a4d0a-c3e5-47de-86ba-2ff48462f434-combined-ca-bundle\") pod \"bd2a4d0a-c3e5-47de-86ba-2ff48462f434\" (UID: \"bd2a4d0a-c3e5-47de-86ba-2ff48462f434\") " Nov 25 11:03:26 crc kubenswrapper[4776]: I1125 11:03:26.612001 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mm85f\" (UniqueName: \"kubernetes.io/projected/bd2a4d0a-c3e5-47de-86ba-2ff48462f434-kube-api-access-mm85f\") pod \"bd2a4d0a-c3e5-47de-86ba-2ff48462f434\" (UID: \"bd2a4d0a-c3e5-47de-86ba-2ff48462f434\") " Nov 25 11:03:26 crc kubenswrapper[4776]: I1125 11:03:26.612143 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd2a4d0a-c3e5-47de-86ba-2ff48462f434-config-data\") pod \"bd2a4d0a-c3e5-47de-86ba-2ff48462f434\" (UID: \"bd2a4d0a-c3e5-47de-86ba-2ff48462f434\") " Nov 25 11:03:26 crc kubenswrapper[4776]: I1125 11:03:26.615652 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd2a4d0a-c3e5-47de-86ba-2ff48462f434-kube-api-access-mm85f" (OuterVolumeSpecName: "kube-api-access-mm85f") pod "bd2a4d0a-c3e5-47de-86ba-2ff48462f434" (UID: "bd2a4d0a-c3e5-47de-86ba-2ff48462f434"). InnerVolumeSpecName "kube-api-access-mm85f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:03:26 crc kubenswrapper[4776]: I1125 11:03:26.627054 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd2a4d0a-c3e5-47de-86ba-2ff48462f434-scripts" (OuterVolumeSpecName: "scripts") pod "bd2a4d0a-c3e5-47de-86ba-2ff48462f434" (UID: "bd2a4d0a-c3e5-47de-86ba-2ff48462f434"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:26 crc kubenswrapper[4776]: I1125 11:03:26.652714 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd2a4d0a-c3e5-47de-86ba-2ff48462f434-config-data" (OuterVolumeSpecName: "config-data") pod "bd2a4d0a-c3e5-47de-86ba-2ff48462f434" (UID: "bd2a4d0a-c3e5-47de-86ba-2ff48462f434"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:26 crc kubenswrapper[4776]: I1125 11:03:26.656497 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd2a4d0a-c3e5-47de-86ba-2ff48462f434-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bd2a4d0a-c3e5-47de-86ba-2ff48462f434" (UID: "bd2a4d0a-c3e5-47de-86ba-2ff48462f434"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:26 crc kubenswrapper[4776]: I1125 11:03:26.714683 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd2a4d0a-c3e5-47de-86ba-2ff48462f434-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:26 crc kubenswrapper[4776]: I1125 11:03:26.714724 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd2a4d0a-c3e5-47de-86ba-2ff48462f434-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:26 crc kubenswrapper[4776]: I1125 11:03:26.714736 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mm85f\" (UniqueName: \"kubernetes.io/projected/bd2a4d0a-c3e5-47de-86ba-2ff48462f434-kube-api-access-mm85f\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:26 crc kubenswrapper[4776]: I1125 11:03:26.714747 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd2a4d0a-c3e5-47de-86ba-2ff48462f434-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:26 crc kubenswrapper[4776]: I1125 11:03:26.984005 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-5trlq" event={"ID":"bd2a4d0a-c3e5-47de-86ba-2ff48462f434","Type":"ContainerDied","Data":"7c3dc699885177c5794b741c87b314688b8720f638deae1d727eed5e55db9261"} Nov 25 11:03:26 crc kubenswrapper[4776]: I1125 11:03:26.984321 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c3dc699885177c5794b741c87b314688b8720f638deae1d727eed5e55db9261" Nov 25 11:03:26 crc kubenswrapper[4776]: I1125 11:03:26.984113 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-5trlq" Nov 25 11:03:26 crc kubenswrapper[4776]: I1125 11:03:26.996558 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a40b8b0e-76a4-402b-962d-46e22f5bbaef","Type":"ContainerStarted","Data":"f04a29d2960eb106a8ab8ee62a13eeb3962f3a7dd813c87e07a819e00f46cc44"} Nov 25 11:03:27 crc kubenswrapper[4776]: I1125 11:03:27.003329 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"50aa9cec-96b4-4436-b03f-5ccce4696e92","Type":"ContainerStarted","Data":"67af669ffc3149bdf39db53aa6851228ca280928ddfac783cb96cf5149471c4d"} Nov 25 11:03:27 crc kubenswrapper[4776]: I1125 11:03:27.003379 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"50aa9cec-96b4-4436-b03f-5ccce4696e92","Type":"ContainerStarted","Data":"7ad81c61e7dd17e4d3af796fbf69072097562eeba0a7fd2f043bfbeb49274c52"} Nov 25 11:03:27 crc kubenswrapper[4776]: I1125 11:03:27.003393 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"50aa9cec-96b4-4436-b03f-5ccce4696e92","Type":"ContainerStarted","Data":"fbed29dab6b970b1c5f5fa4f926c5a9afc57bd49e9ebc5b5be074bad7f78bddb"} Nov 25 11:03:27 crc kubenswrapper[4776]: I1125 11:03:27.005825 4776 generic.go:334] "Generic (PLEG): container finished" podID="9c54defc-dcc0-416c-9d88-fbc5944aa0e3" containerID="a80cec02848f3a1706589c3f1ffe0c08307a0c0588d9539181831f6e8d7221d2" exitCode=0 Nov 25 11:03:27 crc kubenswrapper[4776]: I1125 11:03:27.005863 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-k465n" event={"ID":"9c54defc-dcc0-416c-9d88-fbc5944aa0e3","Type":"ContainerDied","Data":"a80cec02848f3a1706589c3f1ffe0c08307a0c0588d9539181831f6e8d7221d2"} Nov 25 11:03:27 crc kubenswrapper[4776]: I1125 11:03:27.031726 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.031700265 podStartE2EDuration="3.031700265s" podCreationTimestamp="2025-11-25 11:03:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:03:27.021997552 +0000 UTC m=+5952.063057105" watchObservedRunningTime="2025-11-25 11:03:27.031700265 +0000 UTC m=+5952.072759818" Nov 25 11:03:27 crc kubenswrapper[4776]: I1125 11:03:27.057048 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.057026021 podStartE2EDuration="2.057026021s" podCreationTimestamp="2025-11-25 11:03:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:03:27.042106907 +0000 UTC m=+5952.083166480" watchObservedRunningTime="2025-11-25 11:03:27.057026021 +0000 UTC m=+5952.098085574" Nov 25 11:03:27 crc kubenswrapper[4776]: I1125 11:03:27.120921 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 11:03:27 crc kubenswrapper[4776]: E1125 11:03:27.121456 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd2a4d0a-c3e5-47de-86ba-2ff48462f434" containerName="nova-cell1-conductor-db-sync" Nov 25 11:03:27 crc kubenswrapper[4776]: I1125 11:03:27.121481 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd2a4d0a-c3e5-47de-86ba-2ff48462f434" containerName="nova-cell1-conductor-db-sync" Nov 25 11:03:27 crc kubenswrapper[4776]: I1125 11:03:27.121716 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd2a4d0a-c3e5-47de-86ba-2ff48462f434" containerName="nova-cell1-conductor-db-sync" Nov 25 11:03:27 crc kubenswrapper[4776]: I1125 11:03:27.122937 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 25 11:03:27 crc kubenswrapper[4776]: I1125 11:03:27.124973 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 25 11:03:27 crc kubenswrapper[4776]: I1125 11:03:27.132800 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 11:03:27 crc kubenswrapper[4776]: I1125 11:03:27.325775 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fgbm\" (UniqueName: \"kubernetes.io/projected/ec8114a1-e74a-498c-a180-5d20733326f5-kube-api-access-9fgbm\") pod \"nova-cell1-conductor-0\" (UID: \"ec8114a1-e74a-498c-a180-5d20733326f5\") " pod="openstack/nova-cell1-conductor-0" Nov 25 11:03:27 crc kubenswrapper[4776]: I1125 11:03:27.325859 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec8114a1-e74a-498c-a180-5d20733326f5-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ec8114a1-e74a-498c-a180-5d20733326f5\") " pod="openstack/nova-cell1-conductor-0" Nov 25 11:03:27 crc kubenswrapper[4776]: I1125 11:03:27.325918 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec8114a1-e74a-498c-a180-5d20733326f5-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ec8114a1-e74a-498c-a180-5d20733326f5\") " pod="openstack/nova-cell1-conductor-0" Nov 25 11:03:27 crc kubenswrapper[4776]: I1125 11:03:27.427545 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec8114a1-e74a-498c-a180-5d20733326f5-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ec8114a1-e74a-498c-a180-5d20733326f5\") " pod="openstack/nova-cell1-conductor-0" Nov 25 11:03:27 crc kubenswrapper[4776]: I1125 11:03:27.427632 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec8114a1-e74a-498c-a180-5d20733326f5-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ec8114a1-e74a-498c-a180-5d20733326f5\") " pod="openstack/nova-cell1-conductor-0" Nov 25 11:03:27 crc kubenswrapper[4776]: I1125 11:03:27.427785 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fgbm\" (UniqueName: \"kubernetes.io/projected/ec8114a1-e74a-498c-a180-5d20733326f5-kube-api-access-9fgbm\") pod \"nova-cell1-conductor-0\" (UID: \"ec8114a1-e74a-498c-a180-5d20733326f5\") " pod="openstack/nova-cell1-conductor-0" Nov 25 11:03:27 crc kubenswrapper[4776]: I1125 11:03:27.450810 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec8114a1-e74a-498c-a180-5d20733326f5-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ec8114a1-e74a-498c-a180-5d20733326f5\") " pod="openstack/nova-cell1-conductor-0" Nov 25 11:03:27 crc kubenswrapper[4776]: I1125 11:03:27.451870 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec8114a1-e74a-498c-a180-5d20733326f5-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ec8114a1-e74a-498c-a180-5d20733326f5\") " pod="openstack/nova-cell1-conductor-0" Nov 25 11:03:27 crc kubenswrapper[4776]: I1125 11:03:27.455612 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fgbm\" (UniqueName: \"kubernetes.io/projected/ec8114a1-e74a-498c-a180-5d20733326f5-kube-api-access-9fgbm\") pod \"nova-cell1-conductor-0\" (UID: \"ec8114a1-e74a-498c-a180-5d20733326f5\") " pod="openstack/nova-cell1-conductor-0" Nov 25 11:03:27 crc kubenswrapper[4776]: I1125 11:03:27.749300 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 25 11:03:28 crc kubenswrapper[4776]: I1125 11:03:28.195300 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 11:03:28 crc kubenswrapper[4776]: W1125 11:03:28.198646 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec8114a1_e74a_498c_a180_5d20733326f5.slice/crio-6bc8806f176397f277181f32f6b7f9283d79e4f837faf4e59ebf91978e984ff7 WatchSource:0}: Error finding container 6bc8806f176397f277181f32f6b7f9283d79e4f837faf4e59ebf91978e984ff7: Status 404 returned error can't find the container with id 6bc8806f176397f277181f32f6b7f9283d79e4f837faf4e59ebf91978e984ff7 Nov 25 11:03:28 crc kubenswrapper[4776]: I1125 11:03:28.361708 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-k465n" Nov 25 11:03:28 crc kubenswrapper[4776]: I1125 11:03:28.551490 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c54defc-dcc0-416c-9d88-fbc5944aa0e3-combined-ca-bundle\") pod \"9c54defc-dcc0-416c-9d88-fbc5944aa0e3\" (UID: \"9c54defc-dcc0-416c-9d88-fbc5944aa0e3\") " Nov 25 11:03:28 crc kubenswrapper[4776]: I1125 11:03:28.551763 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c54defc-dcc0-416c-9d88-fbc5944aa0e3-config-data\") pod \"9c54defc-dcc0-416c-9d88-fbc5944aa0e3\" (UID: \"9c54defc-dcc0-416c-9d88-fbc5944aa0e3\") " Nov 25 11:03:28 crc kubenswrapper[4776]: I1125 11:03:28.551892 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c54defc-dcc0-416c-9d88-fbc5944aa0e3-scripts\") pod \"9c54defc-dcc0-416c-9d88-fbc5944aa0e3\" (UID: \"9c54defc-dcc0-416c-9d88-fbc5944aa0e3\") " Nov 25 11:03:28 crc kubenswrapper[4776]: I1125 11:03:28.551916 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btwj6\" (UniqueName: \"kubernetes.io/projected/9c54defc-dcc0-416c-9d88-fbc5944aa0e3-kube-api-access-btwj6\") pod \"9c54defc-dcc0-416c-9d88-fbc5944aa0e3\" (UID: \"9c54defc-dcc0-416c-9d88-fbc5944aa0e3\") " Nov 25 11:03:28 crc kubenswrapper[4776]: I1125 11:03:28.557261 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c54defc-dcc0-416c-9d88-fbc5944aa0e3-scripts" (OuterVolumeSpecName: "scripts") pod "9c54defc-dcc0-416c-9d88-fbc5944aa0e3" (UID: "9c54defc-dcc0-416c-9d88-fbc5944aa0e3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:28 crc kubenswrapper[4776]: I1125 11:03:28.557861 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c54defc-dcc0-416c-9d88-fbc5944aa0e3-kube-api-access-btwj6" (OuterVolumeSpecName: "kube-api-access-btwj6") pod "9c54defc-dcc0-416c-9d88-fbc5944aa0e3" (UID: "9c54defc-dcc0-416c-9d88-fbc5944aa0e3"). InnerVolumeSpecName "kube-api-access-btwj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:03:28 crc kubenswrapper[4776]: I1125 11:03:28.583345 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c54defc-dcc0-416c-9d88-fbc5944aa0e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c54defc-dcc0-416c-9d88-fbc5944aa0e3" (UID: "9c54defc-dcc0-416c-9d88-fbc5944aa0e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:28 crc kubenswrapper[4776]: I1125 11:03:28.587416 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c54defc-dcc0-416c-9d88-fbc5944aa0e3-config-data" (OuterVolumeSpecName: "config-data") pod "9c54defc-dcc0-416c-9d88-fbc5944aa0e3" (UID: "9c54defc-dcc0-416c-9d88-fbc5944aa0e3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:28 crc kubenswrapper[4776]: I1125 11:03:28.654730 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c54defc-dcc0-416c-9d88-fbc5944aa0e3-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:28 crc kubenswrapper[4776]: I1125 11:03:28.654770 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c54defc-dcc0-416c-9d88-fbc5944aa0e3-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:28 crc kubenswrapper[4776]: I1125 11:03:28.654781 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btwj6\" (UniqueName: \"kubernetes.io/projected/9c54defc-dcc0-416c-9d88-fbc5944aa0e3-kube-api-access-btwj6\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:28 crc kubenswrapper[4776]: I1125 11:03:28.654794 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c54defc-dcc0-416c-9d88-fbc5944aa0e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:29 crc kubenswrapper[4776]: I1125 11:03:29.022922 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-k465n" Nov 25 11:03:29 crc kubenswrapper[4776]: I1125 11:03:29.022928 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-k465n" event={"ID":"9c54defc-dcc0-416c-9d88-fbc5944aa0e3","Type":"ContainerDied","Data":"b4f4cb9dc4ebf7a4a8453a08fd66bad7d0e81829312e1fe66fbaaf0632fb3dd3"} Nov 25 11:03:29 crc kubenswrapper[4776]: I1125 11:03:29.023838 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4f4cb9dc4ebf7a4a8453a08fd66bad7d0e81829312e1fe66fbaaf0632fb3dd3" Nov 25 11:03:29 crc kubenswrapper[4776]: I1125 11:03:29.025285 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ec8114a1-e74a-498c-a180-5d20733326f5","Type":"ContainerStarted","Data":"45a9c50d84c62aa7c775647665d279cdfcddd2f980985e31b8d45ecaf948d334"} Nov 25 11:03:29 crc kubenswrapper[4776]: I1125 11:03:29.025326 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ec8114a1-e74a-498c-a180-5d20733326f5","Type":"ContainerStarted","Data":"6bc8806f176397f277181f32f6b7f9283d79e4f837faf4e59ebf91978e984ff7"} Nov 25 11:03:29 crc kubenswrapper[4776]: I1125 11:03:29.025391 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 25 11:03:29 crc kubenswrapper[4776]: I1125 11:03:29.052030 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.052011278 podStartE2EDuration="2.052011278s" podCreationTimestamp="2025-11-25 11:03:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:03:29.042406417 +0000 UTC m=+5954.083465970" watchObservedRunningTime="2025-11-25 11:03:29.052011278 +0000 UTC m=+5954.093070831" Nov 25 11:03:29 crc kubenswrapper[4776]: I1125 11:03:29.247564 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 11:03:29 crc kubenswrapper[4776]: I1125 11:03:29.247841 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="da7d17e7-2d62-466a-8826-817e197158d2" containerName="nova-scheduler-scheduler" containerID="cri-o://9e210dd879b5818e9d42a43ebabfe457a1b3f8c12a741c85c61fc945bfd23aa0" gracePeriod=30 Nov 25 11:03:29 crc kubenswrapper[4776]: I1125 11:03:29.256679 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 11:03:29 crc kubenswrapper[4776]: I1125 11:03:29.256989 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ce80bccc-2b9e-44c9-a1c4-915c7bed4897" containerName="nova-api-log" containerID="cri-o://61288b2a6915a828bb1b74bb60f6c0c96c552095ba338733f0a9e8617eecec3d" gracePeriod=30 Nov 25 11:03:29 crc kubenswrapper[4776]: I1125 11:03:29.257087 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ce80bccc-2b9e-44c9-a1c4-915c7bed4897" containerName="nova-api-api" containerID="cri-o://1e3ccbb2372f3500a05c901e101261561c662b2dfca0cd1b849963390eb2b1f0" gracePeriod=30 Nov 25 11:03:29 crc kubenswrapper[4776]: I1125 11:03:29.264250 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 11:03:29 crc kubenswrapper[4776]: I1125 11:03:29.264465 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="50aa9cec-96b4-4436-b03f-5ccce4696e92" containerName="nova-metadata-log" containerID="cri-o://7ad81c61e7dd17e4d3af796fbf69072097562eeba0a7fd2f043bfbeb49274c52" gracePeriod=30 Nov 25 11:03:29 crc kubenswrapper[4776]: I1125 11:03:29.264607 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="50aa9cec-96b4-4436-b03f-5ccce4696e92" containerName="nova-metadata-metadata" containerID="cri-o://67af669ffc3149bdf39db53aa6851228ca280928ddfac783cb96cf5149471c4d" gracePeriod=30 Nov 25 11:03:29 crc kubenswrapper[4776]: I1125 11:03:29.910900 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 11:03:29 crc kubenswrapper[4776]: I1125 11:03:29.917481 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.040327 4776 generic.go:334] "Generic (PLEG): container finished" podID="50aa9cec-96b4-4436-b03f-5ccce4696e92" containerID="67af669ffc3149bdf39db53aa6851228ca280928ddfac783cb96cf5149471c4d" exitCode=0 Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.040369 4776 generic.go:334] "Generic (PLEG): container finished" podID="50aa9cec-96b4-4436-b03f-5ccce4696e92" containerID="7ad81c61e7dd17e4d3af796fbf69072097562eeba0a7fd2f043bfbeb49274c52" exitCode=143 Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.040418 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"50aa9cec-96b4-4436-b03f-5ccce4696e92","Type":"ContainerDied","Data":"67af669ffc3149bdf39db53aa6851228ca280928ddfac783cb96cf5149471c4d"} Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.040450 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"50aa9cec-96b4-4436-b03f-5ccce4696e92","Type":"ContainerDied","Data":"7ad81c61e7dd17e4d3af796fbf69072097562eeba0a7fd2f043bfbeb49274c52"} Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.040460 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"50aa9cec-96b4-4436-b03f-5ccce4696e92","Type":"ContainerDied","Data":"fbed29dab6b970b1c5f5fa4f926c5a9afc57bd49e9ebc5b5be074bad7f78bddb"} Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.040475 4776 scope.go:117] "RemoveContainer" containerID="67af669ffc3149bdf39db53aa6851228ca280928ddfac783cb96cf5149471c4d" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.040657 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.045132 4776 generic.go:334] "Generic (PLEG): container finished" podID="ce80bccc-2b9e-44c9-a1c4-915c7bed4897" containerID="1e3ccbb2372f3500a05c901e101261561c662b2dfca0cd1b849963390eb2b1f0" exitCode=0 Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.045202 4776 generic.go:334] "Generic (PLEG): container finished" podID="ce80bccc-2b9e-44c9-a1c4-915c7bed4897" containerID="61288b2a6915a828bb1b74bb60f6c0c96c552095ba338733f0a9e8617eecec3d" exitCode=143 Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.046041 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.047235 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce80bccc-2b9e-44c9-a1c4-915c7bed4897","Type":"ContainerDied","Data":"1e3ccbb2372f3500a05c901e101261561c662b2dfca0cd1b849963390eb2b1f0"} Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.047307 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce80bccc-2b9e-44c9-a1c4-915c7bed4897","Type":"ContainerDied","Data":"61288b2a6915a828bb1b74bb60f6c0c96c552095ba338733f0a9e8617eecec3d"} Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.047326 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ce80bccc-2b9e-44c9-a1c4-915c7bed4897","Type":"ContainerDied","Data":"a6ce6d5e5b35cc668367146b3c9db71ca8015268b8e8def7afbdee2150a92503"} Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.070732 4776 scope.go:117] "RemoveContainer" containerID="7ad81c61e7dd17e4d3af796fbf69072097562eeba0a7fd2f043bfbeb49274c52" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.082214 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50aa9cec-96b4-4436-b03f-5ccce4696e92-logs\") pod \"50aa9cec-96b4-4436-b03f-5ccce4696e92\" (UID: \"50aa9cec-96b4-4436-b03f-5ccce4696e92\") " Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.082296 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce80bccc-2b9e-44c9-a1c4-915c7bed4897-combined-ca-bundle\") pod \"ce80bccc-2b9e-44c9-a1c4-915c7bed4897\" (UID: \"ce80bccc-2b9e-44c9-a1c4-915c7bed4897\") " Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.082340 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g74ns\" (UniqueName: \"kubernetes.io/projected/ce80bccc-2b9e-44c9-a1c4-915c7bed4897-kube-api-access-g74ns\") pod \"ce80bccc-2b9e-44c9-a1c4-915c7bed4897\" (UID: \"ce80bccc-2b9e-44c9-a1c4-915c7bed4897\") " Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.082364 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50aa9cec-96b4-4436-b03f-5ccce4696e92-combined-ca-bundle\") pod \"50aa9cec-96b4-4436-b03f-5ccce4696e92\" (UID: \"50aa9cec-96b4-4436-b03f-5ccce4696e92\") " Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.082391 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/50aa9cec-96b4-4436-b03f-5ccce4696e92-nova-metadata-tls-certs\") pod \"50aa9cec-96b4-4436-b03f-5ccce4696e92\" (UID: \"50aa9cec-96b4-4436-b03f-5ccce4696e92\") " Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.082531 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce80bccc-2b9e-44c9-a1c4-915c7bed4897-config-data\") pod \"ce80bccc-2b9e-44c9-a1c4-915c7bed4897\" (UID: \"ce80bccc-2b9e-44c9-a1c4-915c7bed4897\") " Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.082548 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce80bccc-2b9e-44c9-a1c4-915c7bed4897-logs\") pod \"ce80bccc-2b9e-44c9-a1c4-915c7bed4897\" (UID: \"ce80bccc-2b9e-44c9-a1c4-915c7bed4897\") " Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.082588 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50aa9cec-96b4-4436-b03f-5ccce4696e92-config-data\") pod \"50aa9cec-96b4-4436-b03f-5ccce4696e92\" (UID: \"50aa9cec-96b4-4436-b03f-5ccce4696e92\") " Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.082639 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9n5bn\" (UniqueName: \"kubernetes.io/projected/50aa9cec-96b4-4436-b03f-5ccce4696e92-kube-api-access-9n5bn\") pod \"50aa9cec-96b4-4436-b03f-5ccce4696e92\" (UID: \"50aa9cec-96b4-4436-b03f-5ccce4696e92\") " Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.084266 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50aa9cec-96b4-4436-b03f-5ccce4696e92-logs" (OuterVolumeSpecName: "logs") pod "50aa9cec-96b4-4436-b03f-5ccce4696e92" (UID: "50aa9cec-96b4-4436-b03f-5ccce4696e92"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.084858 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce80bccc-2b9e-44c9-a1c4-915c7bed4897-logs" (OuterVolumeSpecName: "logs") pod "ce80bccc-2b9e-44c9-a1c4-915c7bed4897" (UID: "ce80bccc-2b9e-44c9-a1c4-915c7bed4897"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.091290 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50aa9cec-96b4-4436-b03f-5ccce4696e92-kube-api-access-9n5bn" (OuterVolumeSpecName: "kube-api-access-9n5bn") pod "50aa9cec-96b4-4436-b03f-5ccce4696e92" (UID: "50aa9cec-96b4-4436-b03f-5ccce4696e92"). InnerVolumeSpecName "kube-api-access-9n5bn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.091404 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce80bccc-2b9e-44c9-a1c4-915c7bed4897-kube-api-access-g74ns" (OuterVolumeSpecName: "kube-api-access-g74ns") pod "ce80bccc-2b9e-44c9-a1c4-915c7bed4897" (UID: "ce80bccc-2b9e-44c9-a1c4-915c7bed4897"). InnerVolumeSpecName "kube-api-access-g74ns". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.095774 4776 scope.go:117] "RemoveContainer" containerID="67af669ffc3149bdf39db53aa6851228ca280928ddfac783cb96cf5149471c4d" Nov 25 11:03:30 crc kubenswrapper[4776]: E1125 11:03:30.096128 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67af669ffc3149bdf39db53aa6851228ca280928ddfac783cb96cf5149471c4d\": container with ID starting with 67af669ffc3149bdf39db53aa6851228ca280928ddfac783cb96cf5149471c4d not found: ID does not exist" containerID="67af669ffc3149bdf39db53aa6851228ca280928ddfac783cb96cf5149471c4d" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.096175 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67af669ffc3149bdf39db53aa6851228ca280928ddfac783cb96cf5149471c4d"} err="failed to get container status \"67af669ffc3149bdf39db53aa6851228ca280928ddfac783cb96cf5149471c4d\": rpc error: code = NotFound desc = could not find container \"67af669ffc3149bdf39db53aa6851228ca280928ddfac783cb96cf5149471c4d\": container with ID starting with 67af669ffc3149bdf39db53aa6851228ca280928ddfac783cb96cf5149471c4d not found: ID does not exist" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.096201 4776 scope.go:117] "RemoveContainer" containerID="7ad81c61e7dd17e4d3af796fbf69072097562eeba0a7fd2f043bfbeb49274c52" Nov 25 11:03:30 crc kubenswrapper[4776]: E1125 11:03:30.096451 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ad81c61e7dd17e4d3af796fbf69072097562eeba0a7fd2f043bfbeb49274c52\": container with ID starting with 7ad81c61e7dd17e4d3af796fbf69072097562eeba0a7fd2f043bfbeb49274c52 not found: ID does not exist" containerID="7ad81c61e7dd17e4d3af796fbf69072097562eeba0a7fd2f043bfbeb49274c52" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.096480 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ad81c61e7dd17e4d3af796fbf69072097562eeba0a7fd2f043bfbeb49274c52"} err="failed to get container status \"7ad81c61e7dd17e4d3af796fbf69072097562eeba0a7fd2f043bfbeb49274c52\": rpc error: code = NotFound desc = could not find container \"7ad81c61e7dd17e4d3af796fbf69072097562eeba0a7fd2f043bfbeb49274c52\": container with ID starting with 7ad81c61e7dd17e4d3af796fbf69072097562eeba0a7fd2f043bfbeb49274c52 not found: ID does not exist" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.096503 4776 scope.go:117] "RemoveContainer" containerID="67af669ffc3149bdf39db53aa6851228ca280928ddfac783cb96cf5149471c4d" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.096719 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67af669ffc3149bdf39db53aa6851228ca280928ddfac783cb96cf5149471c4d"} err="failed to get container status \"67af669ffc3149bdf39db53aa6851228ca280928ddfac783cb96cf5149471c4d\": rpc error: code = NotFound desc = could not find container \"67af669ffc3149bdf39db53aa6851228ca280928ddfac783cb96cf5149471c4d\": container with ID starting with 67af669ffc3149bdf39db53aa6851228ca280928ddfac783cb96cf5149471c4d not found: ID does not exist" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.096747 4776 scope.go:117] "RemoveContainer" containerID="7ad81c61e7dd17e4d3af796fbf69072097562eeba0a7fd2f043bfbeb49274c52" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.096966 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ad81c61e7dd17e4d3af796fbf69072097562eeba0a7fd2f043bfbeb49274c52"} err="failed to get container status \"7ad81c61e7dd17e4d3af796fbf69072097562eeba0a7fd2f043bfbeb49274c52\": rpc error: code = NotFound desc = could not find container \"7ad81c61e7dd17e4d3af796fbf69072097562eeba0a7fd2f043bfbeb49274c52\": container with ID starting with 7ad81c61e7dd17e4d3af796fbf69072097562eeba0a7fd2f043bfbeb49274c52 not found: ID does not exist" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.096991 4776 scope.go:117] "RemoveContainer" containerID="1e3ccbb2372f3500a05c901e101261561c662b2dfca0cd1b849963390eb2b1f0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.111486 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce80bccc-2b9e-44c9-a1c4-915c7bed4897-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ce80bccc-2b9e-44c9-a1c4-915c7bed4897" (UID: "ce80bccc-2b9e-44c9-a1c4-915c7bed4897"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.112021 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50aa9cec-96b4-4436-b03f-5ccce4696e92-config-data" (OuterVolumeSpecName: "config-data") pod "50aa9cec-96b4-4436-b03f-5ccce4696e92" (UID: "50aa9cec-96b4-4436-b03f-5ccce4696e92"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.113869 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50aa9cec-96b4-4436-b03f-5ccce4696e92-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "50aa9cec-96b4-4436-b03f-5ccce4696e92" (UID: "50aa9cec-96b4-4436-b03f-5ccce4696e92"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.119271 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce80bccc-2b9e-44c9-a1c4-915c7bed4897-config-data" (OuterVolumeSpecName: "config-data") pod "ce80bccc-2b9e-44c9-a1c4-915c7bed4897" (UID: "ce80bccc-2b9e-44c9-a1c4-915c7bed4897"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.120962 4776 scope.go:117] "RemoveContainer" containerID="61288b2a6915a828bb1b74bb60f6c0c96c552095ba338733f0a9e8617eecec3d" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.137775 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50aa9cec-96b4-4436-b03f-5ccce4696e92-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "50aa9cec-96b4-4436-b03f-5ccce4696e92" (UID: "50aa9cec-96b4-4436-b03f-5ccce4696e92"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.137839 4776 scope.go:117] "RemoveContainer" containerID="1e3ccbb2372f3500a05c901e101261561c662b2dfca0cd1b849963390eb2b1f0" Nov 25 11:03:30 crc kubenswrapper[4776]: E1125 11:03:30.138280 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e3ccbb2372f3500a05c901e101261561c662b2dfca0cd1b849963390eb2b1f0\": container with ID starting with 1e3ccbb2372f3500a05c901e101261561c662b2dfca0cd1b849963390eb2b1f0 not found: ID does not exist" containerID="1e3ccbb2372f3500a05c901e101261561c662b2dfca0cd1b849963390eb2b1f0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.138315 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e3ccbb2372f3500a05c901e101261561c662b2dfca0cd1b849963390eb2b1f0"} err="failed to get container status \"1e3ccbb2372f3500a05c901e101261561c662b2dfca0cd1b849963390eb2b1f0\": rpc error: code = NotFound desc = could not find container \"1e3ccbb2372f3500a05c901e101261561c662b2dfca0cd1b849963390eb2b1f0\": container with ID starting with 1e3ccbb2372f3500a05c901e101261561c662b2dfca0cd1b849963390eb2b1f0 not found: ID does not exist" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.138341 4776 scope.go:117] "RemoveContainer" containerID="61288b2a6915a828bb1b74bb60f6c0c96c552095ba338733f0a9e8617eecec3d" Nov 25 11:03:30 crc kubenswrapper[4776]: E1125 11:03:30.138736 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61288b2a6915a828bb1b74bb60f6c0c96c552095ba338733f0a9e8617eecec3d\": container with ID starting with 61288b2a6915a828bb1b74bb60f6c0c96c552095ba338733f0a9e8617eecec3d not found: ID does not exist" containerID="61288b2a6915a828bb1b74bb60f6c0c96c552095ba338733f0a9e8617eecec3d" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.138771 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61288b2a6915a828bb1b74bb60f6c0c96c552095ba338733f0a9e8617eecec3d"} err="failed to get container status \"61288b2a6915a828bb1b74bb60f6c0c96c552095ba338733f0a9e8617eecec3d\": rpc error: code = NotFound desc = could not find container \"61288b2a6915a828bb1b74bb60f6c0c96c552095ba338733f0a9e8617eecec3d\": container with ID starting with 61288b2a6915a828bb1b74bb60f6c0c96c552095ba338733f0a9e8617eecec3d not found: ID does not exist" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.138791 4776 scope.go:117] "RemoveContainer" containerID="1e3ccbb2372f3500a05c901e101261561c662b2dfca0cd1b849963390eb2b1f0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.139075 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e3ccbb2372f3500a05c901e101261561c662b2dfca0cd1b849963390eb2b1f0"} err="failed to get container status \"1e3ccbb2372f3500a05c901e101261561c662b2dfca0cd1b849963390eb2b1f0\": rpc error: code = NotFound desc = could not find container \"1e3ccbb2372f3500a05c901e101261561c662b2dfca0cd1b849963390eb2b1f0\": container with ID starting with 1e3ccbb2372f3500a05c901e101261561c662b2dfca0cd1b849963390eb2b1f0 not found: ID does not exist" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.139094 4776 scope.go:117] "RemoveContainer" containerID="61288b2a6915a828bb1b74bb60f6c0c96c552095ba338733f0a9e8617eecec3d" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.139364 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61288b2a6915a828bb1b74bb60f6c0c96c552095ba338733f0a9e8617eecec3d"} err="failed to get container status \"61288b2a6915a828bb1b74bb60f6c0c96c552095ba338733f0a9e8617eecec3d\": rpc error: code = NotFound desc = could not find container \"61288b2a6915a828bb1b74bb60f6c0c96c552095ba338733f0a9e8617eecec3d\": container with ID starting with 61288b2a6915a828bb1b74bb60f6c0c96c552095ba338733f0a9e8617eecec3d not found: ID does not exist" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.185982 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50aa9cec-96b4-4436-b03f-5ccce4696e92-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.186031 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9n5bn\" (UniqueName: \"kubernetes.io/projected/50aa9cec-96b4-4436-b03f-5ccce4696e92-kube-api-access-9n5bn\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.186052 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50aa9cec-96b4-4436-b03f-5ccce4696e92-logs\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.186092 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce80bccc-2b9e-44c9-a1c4-915c7bed4897-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.186109 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g74ns\" (UniqueName: \"kubernetes.io/projected/ce80bccc-2b9e-44c9-a1c4-915c7bed4897-kube-api-access-g74ns\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.186126 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50aa9cec-96b4-4436-b03f-5ccce4696e92-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.186143 4776 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/50aa9cec-96b4-4436-b03f-5ccce4696e92-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.186162 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce80bccc-2b9e-44c9-a1c4-915c7bed4897-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.186178 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce80bccc-2b9e-44c9-a1c4-915c7bed4897-logs\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.376533 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.386711 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.408599 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 25 11:03:30 crc kubenswrapper[4776]: E1125 11:03:30.409215 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50aa9cec-96b4-4436-b03f-5ccce4696e92" containerName="nova-metadata-metadata" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.409320 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="50aa9cec-96b4-4436-b03f-5ccce4696e92" containerName="nova-metadata-metadata" Nov 25 11:03:30 crc kubenswrapper[4776]: E1125 11:03:30.409384 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c54defc-dcc0-416c-9d88-fbc5944aa0e3" containerName="nova-manage" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.409469 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c54defc-dcc0-416c-9d88-fbc5944aa0e3" containerName="nova-manage" Nov 25 11:03:30 crc kubenswrapper[4776]: E1125 11:03:30.409527 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50aa9cec-96b4-4436-b03f-5ccce4696e92" containerName="nova-metadata-log" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.409586 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="50aa9cec-96b4-4436-b03f-5ccce4696e92" containerName="nova-metadata-log" Nov 25 11:03:30 crc kubenswrapper[4776]: E1125 11:03:30.409658 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce80bccc-2b9e-44c9-a1c4-915c7bed4897" containerName="nova-api-api" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.409707 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce80bccc-2b9e-44c9-a1c4-915c7bed4897" containerName="nova-api-api" Nov 25 11:03:30 crc kubenswrapper[4776]: E1125 11:03:30.409939 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce80bccc-2b9e-44c9-a1c4-915c7bed4897" containerName="nova-api-log" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.409996 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce80bccc-2b9e-44c9-a1c4-915c7bed4897" containerName="nova-api-log" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.410227 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce80bccc-2b9e-44c9-a1c4-915c7bed4897" containerName="nova-api-log" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.410292 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="50aa9cec-96b4-4436-b03f-5ccce4696e92" containerName="nova-metadata-log" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.410349 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c54defc-dcc0-416c-9d88-fbc5944aa0e3" containerName="nova-manage" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.410408 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce80bccc-2b9e-44c9-a1c4-915c7bed4897" containerName="nova-api-api" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.410472 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="50aa9cec-96b4-4436-b03f-5ccce4696e92" containerName="nova-metadata-metadata" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.411664 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.415572 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.416148 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.426221 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.437682 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.482020 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.488249 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5bdbcf4d67-qklzz" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.490907 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-227gq\" (UniqueName: \"kubernetes.io/projected/49ae1610-2c5a-47fa-8806-053213b43512-kube-api-access-227gq\") pod \"nova-metadata-0\" (UID: \"49ae1610-2c5a-47fa-8806-053213b43512\") " pod="openstack/nova-metadata-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.491036 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49ae1610-2c5a-47fa-8806-053213b43512-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"49ae1610-2c5a-47fa-8806-053213b43512\") " pod="openstack/nova-metadata-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.491189 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49ae1610-2c5a-47fa-8806-053213b43512-logs\") pod \"nova-metadata-0\" (UID: \"49ae1610-2c5a-47fa-8806-053213b43512\") " pod="openstack/nova-metadata-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.491455 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49ae1610-2c5a-47fa-8806-053213b43512-config-data\") pod \"nova-metadata-0\" (UID: \"49ae1610-2c5a-47fa-8806-053213b43512\") " pod="openstack/nova-metadata-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.491575 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/49ae1610-2c5a-47fa-8806-053213b43512-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"49ae1610-2c5a-47fa-8806-053213b43512\") " pod="openstack/nova-metadata-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.496149 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.509788 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.511753 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.516111 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.527208 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.579109 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-674c55699c-vp7jm"] Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.583641 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-674c55699c-vp7jm" podUID="6bdfe28a-e199-4082-9f5a-3d8b38131d49" containerName="dnsmasq-dns" containerID="cri-o://b3a7dd9ebc25e0f457dc46848f9e46d2f30504b2d54a45e2705416e9ef580e93" gracePeriod=10 Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.596477 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-227gq\" (UniqueName: \"kubernetes.io/projected/49ae1610-2c5a-47fa-8806-053213b43512-kube-api-access-227gq\") pod \"nova-metadata-0\" (UID: \"49ae1610-2c5a-47fa-8806-053213b43512\") " pod="openstack/nova-metadata-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.596545 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49ae1610-2c5a-47fa-8806-053213b43512-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"49ae1610-2c5a-47fa-8806-053213b43512\") " pod="openstack/nova-metadata-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.596587 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49ae1610-2c5a-47fa-8806-053213b43512-logs\") pod \"nova-metadata-0\" (UID: \"49ae1610-2c5a-47fa-8806-053213b43512\") " pod="openstack/nova-metadata-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.596697 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49ae1610-2c5a-47fa-8806-053213b43512-config-data\") pod \"nova-metadata-0\" (UID: \"49ae1610-2c5a-47fa-8806-053213b43512\") " pod="openstack/nova-metadata-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.596729 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/49ae1610-2c5a-47fa-8806-053213b43512-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"49ae1610-2c5a-47fa-8806-053213b43512\") " pod="openstack/nova-metadata-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.611814 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49ae1610-2c5a-47fa-8806-053213b43512-logs\") pod \"nova-metadata-0\" (UID: \"49ae1610-2c5a-47fa-8806-053213b43512\") " pod="openstack/nova-metadata-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.612602 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/49ae1610-2c5a-47fa-8806-053213b43512-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"49ae1610-2c5a-47fa-8806-053213b43512\") " pod="openstack/nova-metadata-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.615850 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49ae1610-2c5a-47fa-8806-053213b43512-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"49ae1610-2c5a-47fa-8806-053213b43512\") " pod="openstack/nova-metadata-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.628821 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49ae1610-2c5a-47fa-8806-053213b43512-config-data\") pod \"nova-metadata-0\" (UID: \"49ae1610-2c5a-47fa-8806-053213b43512\") " pod="openstack/nova-metadata-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.633486 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-227gq\" (UniqueName: \"kubernetes.io/projected/49ae1610-2c5a-47fa-8806-053213b43512-kube-api-access-227gq\") pod \"nova-metadata-0\" (UID: \"49ae1610-2c5a-47fa-8806-053213b43512\") " pod="openstack/nova-metadata-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.698145 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/018ae494-f813-4803-8bbe-515e8ae74af1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"018ae494-f813-4803-8bbe-515e8ae74af1\") " pod="openstack/nova-api-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.698400 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97w7r\" (UniqueName: \"kubernetes.io/projected/018ae494-f813-4803-8bbe-515e8ae74af1-kube-api-access-97w7r\") pod \"nova-api-0\" (UID: \"018ae494-f813-4803-8bbe-515e8ae74af1\") " pod="openstack/nova-api-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.698559 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/018ae494-f813-4803-8bbe-515e8ae74af1-logs\") pod \"nova-api-0\" (UID: \"018ae494-f813-4803-8bbe-515e8ae74af1\") " pod="openstack/nova-api-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.698655 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/018ae494-f813-4803-8bbe-515e8ae74af1-config-data\") pod \"nova-api-0\" (UID: \"018ae494-f813-4803-8bbe-515e8ae74af1\") " pod="openstack/nova-api-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.787708 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.800420 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/018ae494-f813-4803-8bbe-515e8ae74af1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"018ae494-f813-4803-8bbe-515e8ae74af1\") " pod="openstack/nova-api-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.800504 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97w7r\" (UniqueName: \"kubernetes.io/projected/018ae494-f813-4803-8bbe-515e8ae74af1-kube-api-access-97w7r\") pod \"nova-api-0\" (UID: \"018ae494-f813-4803-8bbe-515e8ae74af1\") " pod="openstack/nova-api-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.800655 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/018ae494-f813-4803-8bbe-515e8ae74af1-logs\") pod \"nova-api-0\" (UID: \"018ae494-f813-4803-8bbe-515e8ae74af1\") " pod="openstack/nova-api-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.800695 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/018ae494-f813-4803-8bbe-515e8ae74af1-config-data\") pod \"nova-api-0\" (UID: \"018ae494-f813-4803-8bbe-515e8ae74af1\") " pod="openstack/nova-api-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.801675 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/018ae494-f813-4803-8bbe-515e8ae74af1-logs\") pod \"nova-api-0\" (UID: \"018ae494-f813-4803-8bbe-515e8ae74af1\") " pod="openstack/nova-api-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.805317 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/018ae494-f813-4803-8bbe-515e8ae74af1-config-data\") pod \"nova-api-0\" (UID: \"018ae494-f813-4803-8bbe-515e8ae74af1\") " pod="openstack/nova-api-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.805888 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/018ae494-f813-4803-8bbe-515e8ae74af1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"018ae494-f813-4803-8bbe-515e8ae74af1\") " pod="openstack/nova-api-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.818485 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97w7r\" (UniqueName: \"kubernetes.io/projected/018ae494-f813-4803-8bbe-515e8ae74af1-kube-api-access-97w7r\") pod \"nova-api-0\" (UID: \"018ae494-f813-4803-8bbe-515e8ae74af1\") " pod="openstack/nova-api-0" Nov 25 11:03:30 crc kubenswrapper[4776]: I1125 11:03:30.834996 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.029158 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-674c55699c-vp7jm" Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.063908 4776 generic.go:334] "Generic (PLEG): container finished" podID="6bdfe28a-e199-4082-9f5a-3d8b38131d49" containerID="b3a7dd9ebc25e0f457dc46848f9e46d2f30504b2d54a45e2705416e9ef580e93" exitCode=0 Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.064003 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-674c55699c-vp7jm" Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.064014 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-674c55699c-vp7jm" event={"ID":"6bdfe28a-e199-4082-9f5a-3d8b38131d49","Type":"ContainerDied","Data":"b3a7dd9ebc25e0f457dc46848f9e46d2f30504b2d54a45e2705416e9ef580e93"} Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.064060 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-674c55699c-vp7jm" event={"ID":"6bdfe28a-e199-4082-9f5a-3d8b38131d49","Type":"ContainerDied","Data":"0614e906961a923b101c6dd5c1435b3b8db6826744ab370b43722e5181168de5"} Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.064090 4776 scope.go:117] "RemoveContainer" containerID="b3a7dd9ebc25e0f457dc46848f9e46d2f30504b2d54a45e2705416e9ef580e93" Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.087368 4776 scope.go:117] "RemoveContainer" containerID="571a5c8b23e91c84b58876c8c205e770d328659cd087426a20c82f701aec677c" Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.119592 4776 scope.go:117] "RemoveContainer" containerID="b3a7dd9ebc25e0f457dc46848f9e46d2f30504b2d54a45e2705416e9ef580e93" Nov 25 11:03:31 crc kubenswrapper[4776]: E1125 11:03:31.120513 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3a7dd9ebc25e0f457dc46848f9e46d2f30504b2d54a45e2705416e9ef580e93\": container with ID starting with b3a7dd9ebc25e0f457dc46848f9e46d2f30504b2d54a45e2705416e9ef580e93 not found: ID does not exist" containerID="b3a7dd9ebc25e0f457dc46848f9e46d2f30504b2d54a45e2705416e9ef580e93" Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.120549 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3a7dd9ebc25e0f457dc46848f9e46d2f30504b2d54a45e2705416e9ef580e93"} err="failed to get container status \"b3a7dd9ebc25e0f457dc46848f9e46d2f30504b2d54a45e2705416e9ef580e93\": rpc error: code = NotFound desc = could not find container \"b3a7dd9ebc25e0f457dc46848f9e46d2f30504b2d54a45e2705416e9ef580e93\": container with ID starting with b3a7dd9ebc25e0f457dc46848f9e46d2f30504b2d54a45e2705416e9ef580e93 not found: ID does not exist" Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.120579 4776 scope.go:117] "RemoveContainer" containerID="571a5c8b23e91c84b58876c8c205e770d328659cd087426a20c82f701aec677c" Nov 25 11:03:31 crc kubenswrapper[4776]: E1125 11:03:31.120818 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"571a5c8b23e91c84b58876c8c205e770d328659cd087426a20c82f701aec677c\": container with ID starting with 571a5c8b23e91c84b58876c8c205e770d328659cd087426a20c82f701aec677c not found: ID does not exist" containerID="571a5c8b23e91c84b58876c8c205e770d328659cd087426a20c82f701aec677c" Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.120838 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"571a5c8b23e91c84b58876c8c205e770d328659cd087426a20c82f701aec677c"} err="failed to get container status \"571a5c8b23e91c84b58876c8c205e770d328659cd087426a20c82f701aec677c\": rpc error: code = NotFound desc = could not find container \"571a5c8b23e91c84b58876c8c205e770d328659cd087426a20c82f701aec677c\": container with ID starting with 571a5c8b23e91c84b58876c8c205e770d328659cd087426a20c82f701aec677c not found: ID does not exist" Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.211680 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfxx7\" (UniqueName: \"kubernetes.io/projected/6bdfe28a-e199-4082-9f5a-3d8b38131d49-kube-api-access-zfxx7\") pod \"6bdfe28a-e199-4082-9f5a-3d8b38131d49\" (UID: \"6bdfe28a-e199-4082-9f5a-3d8b38131d49\") " Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.211737 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6bdfe28a-e199-4082-9f5a-3d8b38131d49-ovsdbserver-sb\") pod \"6bdfe28a-e199-4082-9f5a-3d8b38131d49\" (UID: \"6bdfe28a-e199-4082-9f5a-3d8b38131d49\") " Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.211821 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6bdfe28a-e199-4082-9f5a-3d8b38131d49-ovsdbserver-nb\") pod \"6bdfe28a-e199-4082-9f5a-3d8b38131d49\" (UID: \"6bdfe28a-e199-4082-9f5a-3d8b38131d49\") " Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.211949 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bdfe28a-e199-4082-9f5a-3d8b38131d49-config\") pod \"6bdfe28a-e199-4082-9f5a-3d8b38131d49\" (UID: \"6bdfe28a-e199-4082-9f5a-3d8b38131d49\") " Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.211982 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6bdfe28a-e199-4082-9f5a-3d8b38131d49-dns-svc\") pod \"6bdfe28a-e199-4082-9f5a-3d8b38131d49\" (UID: \"6bdfe28a-e199-4082-9f5a-3d8b38131d49\") " Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.221424 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bdfe28a-e199-4082-9f5a-3d8b38131d49-kube-api-access-zfxx7" (OuterVolumeSpecName: "kube-api-access-zfxx7") pod "6bdfe28a-e199-4082-9f5a-3d8b38131d49" (UID: "6bdfe28a-e199-4082-9f5a-3d8b38131d49"). InnerVolumeSpecName "kube-api-access-zfxx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.265044 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6bdfe28a-e199-4082-9f5a-3d8b38131d49-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6bdfe28a-e199-4082-9f5a-3d8b38131d49" (UID: "6bdfe28a-e199-4082-9f5a-3d8b38131d49"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.266461 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6bdfe28a-e199-4082-9f5a-3d8b38131d49-config" (OuterVolumeSpecName: "config") pod "6bdfe28a-e199-4082-9f5a-3d8b38131d49" (UID: "6bdfe28a-e199-4082-9f5a-3d8b38131d49"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.277872 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6bdfe28a-e199-4082-9f5a-3d8b38131d49-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6bdfe28a-e199-4082-9f5a-3d8b38131d49" (UID: "6bdfe28a-e199-4082-9f5a-3d8b38131d49"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.284181 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6bdfe28a-e199-4082-9f5a-3d8b38131d49-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6bdfe28a-e199-4082-9f5a-3d8b38131d49" (UID: "6bdfe28a-e199-4082-9f5a-3d8b38131d49"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.285036 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 11:03:31 crc kubenswrapper[4776]: W1125 11:03:31.305464 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49ae1610_2c5a_47fa_8806_053213b43512.slice/crio-c77785e4b130acb377a84c2d1a0c30bc158a409c98fef6ed6185a0e96c1bfb02 WatchSource:0}: Error finding container c77785e4b130acb377a84c2d1a0c30bc158a409c98fef6ed6185a0e96c1bfb02: Status 404 returned error can't find the container with id c77785e4b130acb377a84c2d1a0c30bc158a409c98fef6ed6185a0e96c1bfb02 Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.306719 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.314553 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfxx7\" (UniqueName: \"kubernetes.io/projected/6bdfe28a-e199-4082-9f5a-3d8b38131d49-kube-api-access-zfxx7\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.314592 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6bdfe28a-e199-4082-9f5a-3d8b38131d49-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.314606 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6bdfe28a-e199-4082-9f5a-3d8b38131d49-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.314636 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6bdfe28a-e199-4082-9f5a-3d8b38131d49-config\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.314647 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6bdfe28a-e199-4082-9f5a-3d8b38131d49-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.594913 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-674c55699c-vp7jm"] Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.606246 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-674c55699c-vp7jm"] Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.680351 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50aa9cec-96b4-4436-b03f-5ccce4696e92" path="/var/lib/kubelet/pods/50aa9cec-96b4-4436-b03f-5ccce4696e92/volumes" Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.681141 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bdfe28a-e199-4082-9f5a-3d8b38131d49" path="/var/lib/kubelet/pods/6bdfe28a-e199-4082-9f5a-3d8b38131d49/volumes" Nov 25 11:03:31 crc kubenswrapper[4776]: I1125 11:03:31.681869 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce80bccc-2b9e-44c9-a1c4-915c7bed4897" path="/var/lib/kubelet/pods/ce80bccc-2b9e-44c9-a1c4-915c7bed4897/volumes" Nov 25 11:03:32 crc kubenswrapper[4776]: I1125 11:03:32.078283 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"018ae494-f813-4803-8bbe-515e8ae74af1","Type":"ContainerStarted","Data":"f330afce9c3a75a17964d25dea2be651f6f040bce1c97ffa756c9c32e6a18473"} Nov 25 11:03:32 crc kubenswrapper[4776]: I1125 11:03:32.078638 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"018ae494-f813-4803-8bbe-515e8ae74af1","Type":"ContainerStarted","Data":"23be4dfb14612aba3f9150002584d02c48fca13a6c0249da4bfb3ce475a48bec"} Nov 25 11:03:32 crc kubenswrapper[4776]: I1125 11:03:32.078652 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"018ae494-f813-4803-8bbe-515e8ae74af1","Type":"ContainerStarted","Data":"813ac7f71b8df1a874a0c553d22d6c5af5bf83e33a45b946b8b8d59d68bedbb2"} Nov 25 11:03:32 crc kubenswrapper[4776]: I1125 11:03:32.082349 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"49ae1610-2c5a-47fa-8806-053213b43512","Type":"ContainerStarted","Data":"469d6c1bdf51c34935a2f6bfeea1b7384e2b77d7f08bc3a6532f05eb0a1693e6"} Nov 25 11:03:32 crc kubenswrapper[4776]: I1125 11:03:32.082379 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"49ae1610-2c5a-47fa-8806-053213b43512","Type":"ContainerStarted","Data":"b1a316733f1576dad76f2099e314f361d249d86fb6d9e477499650ff9442bcf6"} Nov 25 11:03:32 crc kubenswrapper[4776]: I1125 11:03:32.082392 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"49ae1610-2c5a-47fa-8806-053213b43512","Type":"ContainerStarted","Data":"c77785e4b130acb377a84c2d1a0c30bc158a409c98fef6ed6185a0e96c1bfb02"} Nov 25 11:03:32 crc kubenswrapper[4776]: I1125 11:03:32.104676 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.104658409 podStartE2EDuration="2.104658409s" podCreationTimestamp="2025-11-25 11:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:03:32.096471594 +0000 UTC m=+5957.137531147" watchObservedRunningTime="2025-11-25 11:03:32.104658409 +0000 UTC m=+5957.145717962" Nov 25 11:03:32 crc kubenswrapper[4776]: I1125 11:03:32.115995 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.115969533 podStartE2EDuration="2.115969533s" podCreationTimestamp="2025-11-25 11:03:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:03:32.111214904 +0000 UTC m=+5957.152274457" watchObservedRunningTime="2025-11-25 11:03:32.115969533 +0000 UTC m=+5957.157029086" Nov 25 11:03:32 crc kubenswrapper[4776]: I1125 11:03:32.931811 4776 scope.go:117] "RemoveContainer" containerID="c2625123e156e2f6f7330db6834dc55ed01bc5ba1d5281595e746092ee6edb11" Nov 25 11:03:32 crc kubenswrapper[4776]: I1125 11:03:32.962001 4776 scope.go:117] "RemoveContainer" containerID="2c8dbdb4d3af73e3946b030d348bf16e708d07c3eba781cced54fe765b0839b9" Nov 25 11:03:35 crc kubenswrapper[4776]: I1125 11:03:35.426165 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:35 crc kubenswrapper[4776]: I1125 11:03:35.443695 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:35 crc kubenswrapper[4776]: I1125 11:03:35.788823 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 11:03:35 crc kubenswrapper[4776]: I1125 11:03:35.789163 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 11:03:36 crc kubenswrapper[4776]: I1125 11:03:36.139250 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 25 11:03:37 crc kubenswrapper[4776]: I1125 11:03:37.801949 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 25 11:03:38 crc kubenswrapper[4776]: I1125 11:03:38.308311 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-bz5ht"] Nov 25 11:03:38 crc kubenswrapper[4776]: E1125 11:03:38.308697 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bdfe28a-e199-4082-9f5a-3d8b38131d49" containerName="dnsmasq-dns" Nov 25 11:03:38 crc kubenswrapper[4776]: I1125 11:03:38.308714 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bdfe28a-e199-4082-9f5a-3d8b38131d49" containerName="dnsmasq-dns" Nov 25 11:03:38 crc kubenswrapper[4776]: E1125 11:03:38.308753 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bdfe28a-e199-4082-9f5a-3d8b38131d49" containerName="init" Nov 25 11:03:38 crc kubenswrapper[4776]: I1125 11:03:38.308759 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bdfe28a-e199-4082-9f5a-3d8b38131d49" containerName="init" Nov 25 11:03:38 crc kubenswrapper[4776]: I1125 11:03:38.308951 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bdfe28a-e199-4082-9f5a-3d8b38131d49" containerName="dnsmasq-dns" Nov 25 11:03:38 crc kubenswrapper[4776]: I1125 11:03:38.309583 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-bz5ht" Nov 25 11:03:38 crc kubenswrapper[4776]: I1125 11:03:38.311402 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 25 11:03:38 crc kubenswrapper[4776]: I1125 11:03:38.311626 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 25 11:03:38 crc kubenswrapper[4776]: I1125 11:03:38.318384 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-bz5ht"] Nov 25 11:03:38 crc kubenswrapper[4776]: I1125 11:03:38.364367 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430-config-data\") pod \"nova-cell1-cell-mapping-bz5ht\" (UID: \"f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430\") " pod="openstack/nova-cell1-cell-mapping-bz5ht" Nov 25 11:03:38 crc kubenswrapper[4776]: I1125 11:03:38.364470 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-bz5ht\" (UID: \"f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430\") " pod="openstack/nova-cell1-cell-mapping-bz5ht" Nov 25 11:03:38 crc kubenswrapper[4776]: I1125 11:03:38.364733 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rsjn\" (UniqueName: \"kubernetes.io/projected/f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430-kube-api-access-8rsjn\") pod \"nova-cell1-cell-mapping-bz5ht\" (UID: \"f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430\") " pod="openstack/nova-cell1-cell-mapping-bz5ht" Nov 25 11:03:38 crc kubenswrapper[4776]: I1125 11:03:38.364889 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430-scripts\") pod \"nova-cell1-cell-mapping-bz5ht\" (UID: \"f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430\") " pod="openstack/nova-cell1-cell-mapping-bz5ht" Nov 25 11:03:38 crc kubenswrapper[4776]: I1125 11:03:38.466192 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430-config-data\") pod \"nova-cell1-cell-mapping-bz5ht\" (UID: \"f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430\") " pod="openstack/nova-cell1-cell-mapping-bz5ht" Nov 25 11:03:38 crc kubenswrapper[4776]: I1125 11:03:38.466268 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-bz5ht\" (UID: \"f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430\") " pod="openstack/nova-cell1-cell-mapping-bz5ht" Nov 25 11:03:38 crc kubenswrapper[4776]: I1125 11:03:38.466330 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rsjn\" (UniqueName: \"kubernetes.io/projected/f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430-kube-api-access-8rsjn\") pod \"nova-cell1-cell-mapping-bz5ht\" (UID: \"f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430\") " pod="openstack/nova-cell1-cell-mapping-bz5ht" Nov 25 11:03:38 crc kubenswrapper[4776]: I1125 11:03:38.466372 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430-scripts\") pod \"nova-cell1-cell-mapping-bz5ht\" (UID: \"f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430\") " pod="openstack/nova-cell1-cell-mapping-bz5ht" Nov 25 11:03:38 crc kubenswrapper[4776]: I1125 11:03:38.485085 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-bz5ht\" (UID: \"f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430\") " pod="openstack/nova-cell1-cell-mapping-bz5ht" Nov 25 11:03:38 crc kubenswrapper[4776]: I1125 11:03:38.485132 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430-config-data\") pod \"nova-cell1-cell-mapping-bz5ht\" (UID: \"f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430\") " pod="openstack/nova-cell1-cell-mapping-bz5ht" Nov 25 11:03:38 crc kubenswrapper[4776]: I1125 11:03:38.489830 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rsjn\" (UniqueName: \"kubernetes.io/projected/f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430-kube-api-access-8rsjn\") pod \"nova-cell1-cell-mapping-bz5ht\" (UID: \"f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430\") " pod="openstack/nova-cell1-cell-mapping-bz5ht" Nov 25 11:03:38 crc kubenswrapper[4776]: I1125 11:03:38.489822 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430-scripts\") pod \"nova-cell1-cell-mapping-bz5ht\" (UID: \"f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430\") " pod="openstack/nova-cell1-cell-mapping-bz5ht" Nov 25 11:03:38 crc kubenswrapper[4776]: I1125 11:03:38.630452 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-bz5ht" Nov 25 11:03:39 crc kubenswrapper[4776]: I1125 11:03:39.113937 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-bz5ht"] Nov 25 11:03:39 crc kubenswrapper[4776]: I1125 11:03:39.157293 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-bz5ht" event={"ID":"f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430","Type":"ContainerStarted","Data":"25e6026936b0b69b50ec271f1c1d3eeb678c0f825b92ab35a9c782ac2b7996ea"} Nov 25 11:03:40 crc kubenswrapper[4776]: I1125 11:03:40.165019 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-bz5ht" event={"ID":"f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430","Type":"ContainerStarted","Data":"c8decff072d39c84cc2226421867809629b8654d98030ccb6a1bbc5bbc277e53"} Nov 25 11:03:40 crc kubenswrapper[4776]: I1125 11:03:40.185237 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-bz5ht" podStartSLOduration=2.185220015 podStartE2EDuration="2.185220015s" podCreationTimestamp="2025-11-25 11:03:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:03:40.182896686 +0000 UTC m=+5965.223956249" watchObservedRunningTime="2025-11-25 11:03:40.185220015 +0000 UTC m=+5965.226279568" Nov 25 11:03:40 crc kubenswrapper[4776]: I1125 11:03:40.788204 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 11:03:40 crc kubenswrapper[4776]: I1125 11:03:40.788691 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 11:03:40 crc kubenswrapper[4776]: I1125 11:03:40.836249 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 11:03:40 crc kubenswrapper[4776]: I1125 11:03:40.836306 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 11:03:41 crc kubenswrapper[4776]: I1125 11:03:41.800538 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="49ae1610-2c5a-47fa-8806-053213b43512" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.94:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 11:03:41 crc kubenswrapper[4776]: I1125 11:03:41.800944 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="49ae1610-2c5a-47fa-8806-053213b43512" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.94:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 11:03:41 crc kubenswrapper[4776]: I1125 11:03:41.919614 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="018ae494-f813-4803-8bbe-515e8ae74af1" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.95:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 11:03:41 crc kubenswrapper[4776]: I1125 11:03:41.919614 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="018ae494-f813-4803-8bbe-515e8ae74af1" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.95:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 11:03:44 crc kubenswrapper[4776]: I1125 11:03:44.229839 4776 generic.go:334] "Generic (PLEG): container finished" podID="f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430" containerID="c8decff072d39c84cc2226421867809629b8654d98030ccb6a1bbc5bbc277e53" exitCode=0 Nov 25 11:03:44 crc kubenswrapper[4776]: I1125 11:03:44.229903 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-bz5ht" event={"ID":"f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430","Type":"ContainerDied","Data":"c8decff072d39c84cc2226421867809629b8654d98030ccb6a1bbc5bbc277e53"} Nov 25 11:03:45 crc kubenswrapper[4776]: I1125 11:03:45.561565 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-bz5ht" Nov 25 11:03:45 crc kubenswrapper[4776]: I1125 11:03:45.726247 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rsjn\" (UniqueName: \"kubernetes.io/projected/f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430-kube-api-access-8rsjn\") pod \"f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430\" (UID: \"f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430\") " Nov 25 11:03:45 crc kubenswrapper[4776]: I1125 11:03:45.726808 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430-config-data\") pod \"f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430\" (UID: \"f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430\") " Nov 25 11:03:45 crc kubenswrapper[4776]: I1125 11:03:45.727526 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430-scripts\") pod \"f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430\" (UID: \"f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430\") " Nov 25 11:03:45 crc kubenswrapper[4776]: I1125 11:03:45.727574 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430-combined-ca-bundle\") pod \"f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430\" (UID: \"f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430\") " Nov 25 11:03:45 crc kubenswrapper[4776]: I1125 11:03:45.732711 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430-scripts" (OuterVolumeSpecName: "scripts") pod "f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430" (UID: "f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:45 crc kubenswrapper[4776]: I1125 11:03:45.742016 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430-kube-api-access-8rsjn" (OuterVolumeSpecName: "kube-api-access-8rsjn") pod "f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430" (UID: "f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430"). InnerVolumeSpecName "kube-api-access-8rsjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:03:45 crc kubenswrapper[4776]: I1125 11:03:45.755459 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430-config-data" (OuterVolumeSpecName: "config-data") pod "f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430" (UID: "f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:45 crc kubenswrapper[4776]: I1125 11:03:45.756263 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430" (UID: "f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:45 crc kubenswrapper[4776]: I1125 11:03:45.830168 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rsjn\" (UniqueName: \"kubernetes.io/projected/f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430-kube-api-access-8rsjn\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:45 crc kubenswrapper[4776]: I1125 11:03:45.830219 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:45 crc kubenswrapper[4776]: I1125 11:03:45.830228 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:45 crc kubenswrapper[4776]: I1125 11:03:45.830237 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:46 crc kubenswrapper[4776]: I1125 11:03:46.250945 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-bz5ht" event={"ID":"f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430","Type":"ContainerDied","Data":"25e6026936b0b69b50ec271f1c1d3eeb678c0f825b92ab35a9c782ac2b7996ea"} Nov 25 11:03:46 crc kubenswrapper[4776]: I1125 11:03:46.250998 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25e6026936b0b69b50ec271f1c1d3eeb678c0f825b92ab35a9c782ac2b7996ea" Nov 25 11:03:46 crc kubenswrapper[4776]: I1125 11:03:46.251020 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-bz5ht" Nov 25 11:03:46 crc kubenswrapper[4776]: I1125 11:03:46.441568 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 11:03:46 crc kubenswrapper[4776]: I1125 11:03:46.441848 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="018ae494-f813-4803-8bbe-515e8ae74af1" containerName="nova-api-log" containerID="cri-o://23be4dfb14612aba3f9150002584d02c48fca13a6c0249da4bfb3ce475a48bec" gracePeriod=30 Nov 25 11:03:46 crc kubenswrapper[4776]: I1125 11:03:46.441966 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="018ae494-f813-4803-8bbe-515e8ae74af1" containerName="nova-api-api" containerID="cri-o://f330afce9c3a75a17964d25dea2be651f6f040bce1c97ffa756c9c32e6a18473" gracePeriod=30 Nov 25 11:03:46 crc kubenswrapper[4776]: I1125 11:03:46.502309 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 11:03:46 crc kubenswrapper[4776]: I1125 11:03:46.502562 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="49ae1610-2c5a-47fa-8806-053213b43512" containerName="nova-metadata-log" containerID="cri-o://b1a316733f1576dad76f2099e314f361d249d86fb6d9e477499650ff9442bcf6" gracePeriod=30 Nov 25 11:03:46 crc kubenswrapper[4776]: I1125 11:03:46.502703 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="49ae1610-2c5a-47fa-8806-053213b43512" containerName="nova-metadata-metadata" containerID="cri-o://469d6c1bdf51c34935a2f6bfeea1b7384e2b77d7f08bc3a6532f05eb0a1693e6" gracePeriod=30 Nov 25 11:03:47 crc kubenswrapper[4776]: I1125 11:03:47.261983 4776 generic.go:334] "Generic (PLEG): container finished" podID="018ae494-f813-4803-8bbe-515e8ae74af1" containerID="23be4dfb14612aba3f9150002584d02c48fca13a6c0249da4bfb3ce475a48bec" exitCode=143 Nov 25 11:03:47 crc kubenswrapper[4776]: I1125 11:03:47.262101 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"018ae494-f813-4803-8bbe-515e8ae74af1","Type":"ContainerDied","Data":"23be4dfb14612aba3f9150002584d02c48fca13a6c0249da4bfb3ce475a48bec"} Nov 25 11:03:47 crc kubenswrapper[4776]: I1125 11:03:47.264151 4776 generic.go:334] "Generic (PLEG): container finished" podID="49ae1610-2c5a-47fa-8806-053213b43512" containerID="b1a316733f1576dad76f2099e314f361d249d86fb6d9e477499650ff9442bcf6" exitCode=143 Nov 25 11:03:47 crc kubenswrapper[4776]: I1125 11:03:47.264185 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"49ae1610-2c5a-47fa-8806-053213b43512","Type":"ContainerDied","Data":"b1a316733f1576dad76f2099e314f361d249d86fb6d9e477499650ff9442bcf6"} Nov 25 11:03:47 crc kubenswrapper[4776]: I1125 11:03:47.819422 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:03:47 crc kubenswrapper[4776]: I1125 11:03:47.819522 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:03:59 crc kubenswrapper[4776]: I1125 11:03:59.382927 4776 generic.go:334] "Generic (PLEG): container finished" podID="da7d17e7-2d62-466a-8826-817e197158d2" containerID="9e210dd879b5818e9d42a43ebabfe457a1b3f8c12a741c85c61fc945bfd23aa0" exitCode=137 Nov 25 11:03:59 crc kubenswrapper[4776]: I1125 11:03:59.383005 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"da7d17e7-2d62-466a-8826-817e197158d2","Type":"ContainerDied","Data":"9e210dd879b5818e9d42a43ebabfe457a1b3f8c12a741c85c61fc945bfd23aa0"} Nov 25 11:03:59 crc kubenswrapper[4776]: I1125 11:03:59.690816 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 11:03:59 crc kubenswrapper[4776]: I1125 11:03:59.837989 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9nkk\" (UniqueName: \"kubernetes.io/projected/da7d17e7-2d62-466a-8826-817e197158d2-kube-api-access-h9nkk\") pod \"da7d17e7-2d62-466a-8826-817e197158d2\" (UID: \"da7d17e7-2d62-466a-8826-817e197158d2\") " Nov 25 11:03:59 crc kubenswrapper[4776]: I1125 11:03:59.838064 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da7d17e7-2d62-466a-8826-817e197158d2-combined-ca-bundle\") pod \"da7d17e7-2d62-466a-8826-817e197158d2\" (UID: \"da7d17e7-2d62-466a-8826-817e197158d2\") " Nov 25 11:03:59 crc kubenswrapper[4776]: I1125 11:03:59.838462 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da7d17e7-2d62-466a-8826-817e197158d2-config-data\") pod \"da7d17e7-2d62-466a-8826-817e197158d2\" (UID: \"da7d17e7-2d62-466a-8826-817e197158d2\") " Nov 25 11:03:59 crc kubenswrapper[4776]: I1125 11:03:59.844737 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da7d17e7-2d62-466a-8826-817e197158d2-kube-api-access-h9nkk" (OuterVolumeSpecName: "kube-api-access-h9nkk") pod "da7d17e7-2d62-466a-8826-817e197158d2" (UID: "da7d17e7-2d62-466a-8826-817e197158d2"). InnerVolumeSpecName "kube-api-access-h9nkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:03:59 crc kubenswrapper[4776]: I1125 11:03:59.865036 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da7d17e7-2d62-466a-8826-817e197158d2-config-data" (OuterVolumeSpecName: "config-data") pod "da7d17e7-2d62-466a-8826-817e197158d2" (UID: "da7d17e7-2d62-466a-8826-817e197158d2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:59 crc kubenswrapper[4776]: I1125 11:03:59.872698 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da7d17e7-2d62-466a-8826-817e197158d2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "da7d17e7-2d62-466a-8826-817e197158d2" (UID: "da7d17e7-2d62-466a-8826-817e197158d2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:03:59 crc kubenswrapper[4776]: I1125 11:03:59.940918 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da7d17e7-2d62-466a-8826-817e197158d2-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:59 crc kubenswrapper[4776]: I1125 11:03:59.940957 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9nkk\" (UniqueName: \"kubernetes.io/projected/da7d17e7-2d62-466a-8826-817e197158d2-kube-api-access-h9nkk\") on node \"crc\" DevicePath \"\"" Nov 25 11:03:59 crc kubenswrapper[4776]: I1125 11:03:59.940971 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da7d17e7-2d62-466a-8826-817e197158d2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.184229 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.246517 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.347255 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49ae1610-2c5a-47fa-8806-053213b43512-logs\") pod \"49ae1610-2c5a-47fa-8806-053213b43512\" (UID: \"49ae1610-2c5a-47fa-8806-053213b43512\") " Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.347307 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-227gq\" (UniqueName: \"kubernetes.io/projected/49ae1610-2c5a-47fa-8806-053213b43512-kube-api-access-227gq\") pod \"49ae1610-2c5a-47fa-8806-053213b43512\" (UID: \"49ae1610-2c5a-47fa-8806-053213b43512\") " Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.347360 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97w7r\" (UniqueName: \"kubernetes.io/projected/018ae494-f813-4803-8bbe-515e8ae74af1-kube-api-access-97w7r\") pod \"018ae494-f813-4803-8bbe-515e8ae74af1\" (UID: \"018ae494-f813-4803-8bbe-515e8ae74af1\") " Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.347488 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/49ae1610-2c5a-47fa-8806-053213b43512-nova-metadata-tls-certs\") pod \"49ae1610-2c5a-47fa-8806-053213b43512\" (UID: \"49ae1610-2c5a-47fa-8806-053213b43512\") " Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.347537 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49ae1610-2c5a-47fa-8806-053213b43512-config-data\") pod \"49ae1610-2c5a-47fa-8806-053213b43512\" (UID: \"49ae1610-2c5a-47fa-8806-053213b43512\") " Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.347579 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/018ae494-f813-4803-8bbe-515e8ae74af1-config-data\") pod \"018ae494-f813-4803-8bbe-515e8ae74af1\" (UID: \"018ae494-f813-4803-8bbe-515e8ae74af1\") " Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.347614 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/018ae494-f813-4803-8bbe-515e8ae74af1-combined-ca-bundle\") pod \"018ae494-f813-4803-8bbe-515e8ae74af1\" (UID: \"018ae494-f813-4803-8bbe-515e8ae74af1\") " Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.347660 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/018ae494-f813-4803-8bbe-515e8ae74af1-logs\") pod \"018ae494-f813-4803-8bbe-515e8ae74af1\" (UID: \"018ae494-f813-4803-8bbe-515e8ae74af1\") " Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.347700 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49ae1610-2c5a-47fa-8806-053213b43512-combined-ca-bundle\") pod \"49ae1610-2c5a-47fa-8806-053213b43512\" (UID: \"49ae1610-2c5a-47fa-8806-053213b43512\") " Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.347994 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49ae1610-2c5a-47fa-8806-053213b43512-logs" (OuterVolumeSpecName: "logs") pod "49ae1610-2c5a-47fa-8806-053213b43512" (UID: "49ae1610-2c5a-47fa-8806-053213b43512"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.348257 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/49ae1610-2c5a-47fa-8806-053213b43512-logs\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.349411 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/018ae494-f813-4803-8bbe-515e8ae74af1-logs" (OuterVolumeSpecName: "logs") pod "018ae494-f813-4803-8bbe-515e8ae74af1" (UID: "018ae494-f813-4803-8bbe-515e8ae74af1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.353114 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/018ae494-f813-4803-8bbe-515e8ae74af1-kube-api-access-97w7r" (OuterVolumeSpecName: "kube-api-access-97w7r") pod "018ae494-f813-4803-8bbe-515e8ae74af1" (UID: "018ae494-f813-4803-8bbe-515e8ae74af1"). InnerVolumeSpecName "kube-api-access-97w7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.353362 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ae1610-2c5a-47fa-8806-053213b43512-kube-api-access-227gq" (OuterVolumeSpecName: "kube-api-access-227gq") pod "49ae1610-2c5a-47fa-8806-053213b43512" (UID: "49ae1610-2c5a-47fa-8806-053213b43512"). InnerVolumeSpecName "kube-api-access-227gq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.377332 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/018ae494-f813-4803-8bbe-515e8ae74af1-config-data" (OuterVolumeSpecName: "config-data") pod "018ae494-f813-4803-8bbe-515e8ae74af1" (UID: "018ae494-f813-4803-8bbe-515e8ae74af1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.378277 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49ae1610-2c5a-47fa-8806-053213b43512-config-data" (OuterVolumeSpecName: "config-data") pod "49ae1610-2c5a-47fa-8806-053213b43512" (UID: "49ae1610-2c5a-47fa-8806-053213b43512"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.387422 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/018ae494-f813-4803-8bbe-515e8ae74af1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "018ae494-f813-4803-8bbe-515e8ae74af1" (UID: "018ae494-f813-4803-8bbe-515e8ae74af1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.394950 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49ae1610-2c5a-47fa-8806-053213b43512-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "49ae1610-2c5a-47fa-8806-053213b43512" (UID: "49ae1610-2c5a-47fa-8806-053213b43512"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.396388 4776 generic.go:334] "Generic (PLEG): container finished" podID="49ae1610-2c5a-47fa-8806-053213b43512" containerID="469d6c1bdf51c34935a2f6bfeea1b7384e2b77d7f08bc3a6532f05eb0a1693e6" exitCode=0 Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.396501 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.397325 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"49ae1610-2c5a-47fa-8806-053213b43512","Type":"ContainerDied","Data":"469d6c1bdf51c34935a2f6bfeea1b7384e2b77d7f08bc3a6532f05eb0a1693e6"} Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.397356 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"49ae1610-2c5a-47fa-8806-053213b43512","Type":"ContainerDied","Data":"c77785e4b130acb377a84c2d1a0c30bc158a409c98fef6ed6185a0e96c1bfb02"} Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.397398 4776 scope.go:117] "RemoveContainer" containerID="469d6c1bdf51c34935a2f6bfeea1b7384e2b77d7f08bc3a6532f05eb0a1693e6" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.401761 4776 generic.go:334] "Generic (PLEG): container finished" podID="018ae494-f813-4803-8bbe-515e8ae74af1" containerID="f330afce9c3a75a17964d25dea2be651f6f040bce1c97ffa756c9c32e6a18473" exitCode=0 Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.401809 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"018ae494-f813-4803-8bbe-515e8ae74af1","Type":"ContainerDied","Data":"f330afce9c3a75a17964d25dea2be651f6f040bce1c97ffa756c9c32e6a18473"} Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.401826 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"018ae494-f813-4803-8bbe-515e8ae74af1","Type":"ContainerDied","Data":"813ac7f71b8df1a874a0c553d22d6c5af5bf83e33a45b946b8b8d59d68bedbb2"} Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.401828 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.406838 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"da7d17e7-2d62-466a-8826-817e197158d2","Type":"ContainerDied","Data":"d28dcebf7005acb1f41f918e058acdb97c0ae95bb57d33c71f4310dcd494b739"} Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.406955 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.411940 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49ae1610-2c5a-47fa-8806-053213b43512-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "49ae1610-2c5a-47fa-8806-053213b43512" (UID: "49ae1610-2c5a-47fa-8806-053213b43512"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.459103 4776 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/49ae1610-2c5a-47fa-8806-053213b43512-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.459445 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/49ae1610-2c5a-47fa-8806-053213b43512-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.459460 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/018ae494-f813-4803-8bbe-515e8ae74af1-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.459472 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/018ae494-f813-4803-8bbe-515e8ae74af1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.459484 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/018ae494-f813-4803-8bbe-515e8ae74af1-logs\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.459512 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49ae1610-2c5a-47fa-8806-053213b43512-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.459524 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-227gq\" (UniqueName: \"kubernetes.io/projected/49ae1610-2c5a-47fa-8806-053213b43512-kube-api-access-227gq\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.459535 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97w7r\" (UniqueName: \"kubernetes.io/projected/018ae494-f813-4803-8bbe-515e8ae74af1-kube-api-access-97w7r\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.512866 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.532122 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.544111 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.545366 4776 scope.go:117] "RemoveContainer" containerID="b1a316733f1576dad76f2099e314f361d249d86fb6d9e477499650ff9442bcf6" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.563826 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.574630 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 11:04:00 crc kubenswrapper[4776]: E1125 11:04:00.575030 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49ae1610-2c5a-47fa-8806-053213b43512" containerName="nova-metadata-metadata" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.575042 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="49ae1610-2c5a-47fa-8806-053213b43512" containerName="nova-metadata-metadata" Nov 25 11:04:00 crc kubenswrapper[4776]: E1125 11:04:00.575053 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430" containerName="nova-manage" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.575059 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430" containerName="nova-manage" Nov 25 11:04:00 crc kubenswrapper[4776]: E1125 11:04:00.575085 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49ae1610-2c5a-47fa-8806-053213b43512" containerName="nova-metadata-log" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.575092 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="49ae1610-2c5a-47fa-8806-053213b43512" containerName="nova-metadata-log" Nov 25 11:04:00 crc kubenswrapper[4776]: E1125 11:04:00.575103 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="018ae494-f813-4803-8bbe-515e8ae74af1" containerName="nova-api-api" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.575110 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="018ae494-f813-4803-8bbe-515e8ae74af1" containerName="nova-api-api" Nov 25 11:04:00 crc kubenswrapper[4776]: E1125 11:04:00.575134 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da7d17e7-2d62-466a-8826-817e197158d2" containerName="nova-scheduler-scheduler" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.575165 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="da7d17e7-2d62-466a-8826-817e197158d2" containerName="nova-scheduler-scheduler" Nov 25 11:04:00 crc kubenswrapper[4776]: E1125 11:04:00.575174 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="018ae494-f813-4803-8bbe-515e8ae74af1" containerName="nova-api-log" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.575180 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="018ae494-f813-4803-8bbe-515e8ae74af1" containerName="nova-api-log" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.575657 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="49ae1610-2c5a-47fa-8806-053213b43512" containerName="nova-metadata-metadata" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.575681 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="018ae494-f813-4803-8bbe-515e8ae74af1" containerName="nova-api-log" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.575693 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="018ae494-f813-4803-8bbe-515e8ae74af1" containerName="nova-api-api" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.575702 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="da7d17e7-2d62-466a-8826-817e197158d2" containerName="nova-scheduler-scheduler" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.575713 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="49ae1610-2c5a-47fa-8806-053213b43512" containerName="nova-metadata-log" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.575727 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430" containerName="nova-manage" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.577605 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.580392 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.587183 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.600408 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.602063 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.603454 4776 scope.go:117] "RemoveContainer" containerID="469d6c1bdf51c34935a2f6bfeea1b7384e2b77d7f08bc3a6532f05eb0a1693e6" Nov 25 11:04:00 crc kubenswrapper[4776]: E1125 11:04:00.603930 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"469d6c1bdf51c34935a2f6bfeea1b7384e2b77d7f08bc3a6532f05eb0a1693e6\": container with ID starting with 469d6c1bdf51c34935a2f6bfeea1b7384e2b77d7f08bc3a6532f05eb0a1693e6 not found: ID does not exist" containerID="469d6c1bdf51c34935a2f6bfeea1b7384e2b77d7f08bc3a6532f05eb0a1693e6" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.603966 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"469d6c1bdf51c34935a2f6bfeea1b7384e2b77d7f08bc3a6532f05eb0a1693e6"} err="failed to get container status \"469d6c1bdf51c34935a2f6bfeea1b7384e2b77d7f08bc3a6532f05eb0a1693e6\": rpc error: code = NotFound desc = could not find container \"469d6c1bdf51c34935a2f6bfeea1b7384e2b77d7f08bc3a6532f05eb0a1693e6\": container with ID starting with 469d6c1bdf51c34935a2f6bfeea1b7384e2b77d7f08bc3a6532f05eb0a1693e6 not found: ID does not exist" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.603992 4776 scope.go:117] "RemoveContainer" containerID="b1a316733f1576dad76f2099e314f361d249d86fb6d9e477499650ff9442bcf6" Nov 25 11:04:00 crc kubenswrapper[4776]: E1125 11:04:00.604420 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1a316733f1576dad76f2099e314f361d249d86fb6d9e477499650ff9442bcf6\": container with ID starting with b1a316733f1576dad76f2099e314f361d249d86fb6d9e477499650ff9442bcf6 not found: ID does not exist" containerID="b1a316733f1576dad76f2099e314f361d249d86fb6d9e477499650ff9442bcf6" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.604448 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1a316733f1576dad76f2099e314f361d249d86fb6d9e477499650ff9442bcf6"} err="failed to get container status \"b1a316733f1576dad76f2099e314f361d249d86fb6d9e477499650ff9442bcf6\": rpc error: code = NotFound desc = could not find container \"b1a316733f1576dad76f2099e314f361d249d86fb6d9e477499650ff9442bcf6\": container with ID starting with b1a316733f1576dad76f2099e314f361d249d86fb6d9e477499650ff9442bcf6 not found: ID does not exist" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.604466 4776 scope.go:117] "RemoveContainer" containerID="f330afce9c3a75a17964d25dea2be651f6f040bce1c97ffa756c9c32e6a18473" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.605539 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.608442 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.629341 4776 scope.go:117] "RemoveContainer" containerID="23be4dfb14612aba3f9150002584d02c48fca13a6c0249da4bfb3ce475a48bec" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.651651 4776 scope.go:117] "RemoveContainer" containerID="f330afce9c3a75a17964d25dea2be651f6f040bce1c97ffa756c9c32e6a18473" Nov 25 11:04:00 crc kubenswrapper[4776]: E1125 11:04:00.652229 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f330afce9c3a75a17964d25dea2be651f6f040bce1c97ffa756c9c32e6a18473\": container with ID starting with f330afce9c3a75a17964d25dea2be651f6f040bce1c97ffa756c9c32e6a18473 not found: ID does not exist" containerID="f330afce9c3a75a17964d25dea2be651f6f040bce1c97ffa756c9c32e6a18473" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.652263 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f330afce9c3a75a17964d25dea2be651f6f040bce1c97ffa756c9c32e6a18473"} err="failed to get container status \"f330afce9c3a75a17964d25dea2be651f6f040bce1c97ffa756c9c32e6a18473\": rpc error: code = NotFound desc = could not find container \"f330afce9c3a75a17964d25dea2be651f6f040bce1c97ffa756c9c32e6a18473\": container with ID starting with f330afce9c3a75a17964d25dea2be651f6f040bce1c97ffa756c9c32e6a18473 not found: ID does not exist" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.652289 4776 scope.go:117] "RemoveContainer" containerID="23be4dfb14612aba3f9150002584d02c48fca13a6c0249da4bfb3ce475a48bec" Nov 25 11:04:00 crc kubenswrapper[4776]: E1125 11:04:00.652659 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23be4dfb14612aba3f9150002584d02c48fca13a6c0249da4bfb3ce475a48bec\": container with ID starting with 23be4dfb14612aba3f9150002584d02c48fca13a6c0249da4bfb3ce475a48bec not found: ID does not exist" containerID="23be4dfb14612aba3f9150002584d02c48fca13a6c0249da4bfb3ce475a48bec" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.652684 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23be4dfb14612aba3f9150002584d02c48fca13a6c0249da4bfb3ce475a48bec"} err="failed to get container status \"23be4dfb14612aba3f9150002584d02c48fca13a6c0249da4bfb3ce475a48bec\": rpc error: code = NotFound desc = could not find container \"23be4dfb14612aba3f9150002584d02c48fca13a6c0249da4bfb3ce475a48bec\": container with ID starting with 23be4dfb14612aba3f9150002584d02c48fca13a6c0249da4bfb3ce475a48bec not found: ID does not exist" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.652698 4776 scope.go:117] "RemoveContainer" containerID="9e210dd879b5818e9d42a43ebabfe457a1b3f8c12a741c85c61fc945bfd23aa0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.663376 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srpkz\" (UniqueName: \"kubernetes.io/projected/58a77023-a881-4cb5-b38b-b7b7df6386e9-kube-api-access-srpkz\") pod \"nova-api-0\" (UID: \"58a77023-a881-4cb5-b38b-b7b7df6386e9\") " pod="openstack/nova-api-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.663459 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58a77023-a881-4cb5-b38b-b7b7df6386e9-config-data\") pod \"nova-api-0\" (UID: \"58a77023-a881-4cb5-b38b-b7b7df6386e9\") " pod="openstack/nova-api-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.663525 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3\") " pod="openstack/nova-scheduler-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.663583 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58a77023-a881-4cb5-b38b-b7b7df6386e9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"58a77023-a881-4cb5-b38b-b7b7df6386e9\") " pod="openstack/nova-api-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.663691 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jctt\" (UniqueName: \"kubernetes.io/projected/f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3-kube-api-access-8jctt\") pod \"nova-scheduler-0\" (UID: \"f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3\") " pod="openstack/nova-scheduler-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.663925 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3-config-data\") pod \"nova-scheduler-0\" (UID: \"f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3\") " pod="openstack/nova-scheduler-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.663997 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58a77023-a881-4cb5-b38b-b7b7df6386e9-logs\") pod \"nova-api-0\" (UID: \"58a77023-a881-4cb5-b38b-b7b7df6386e9\") " pod="openstack/nova-api-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.725681 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.751541 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.783214 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srpkz\" (UniqueName: \"kubernetes.io/projected/58a77023-a881-4cb5-b38b-b7b7df6386e9-kube-api-access-srpkz\") pod \"nova-api-0\" (UID: \"58a77023-a881-4cb5-b38b-b7b7df6386e9\") " pod="openstack/nova-api-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.783288 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58a77023-a881-4cb5-b38b-b7b7df6386e9-config-data\") pod \"nova-api-0\" (UID: \"58a77023-a881-4cb5-b38b-b7b7df6386e9\") " pod="openstack/nova-api-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.783345 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3\") " pod="openstack/nova-scheduler-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.783414 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58a77023-a881-4cb5-b38b-b7b7df6386e9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"58a77023-a881-4cb5-b38b-b7b7df6386e9\") " pod="openstack/nova-api-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.783491 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jctt\" (UniqueName: \"kubernetes.io/projected/f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3-kube-api-access-8jctt\") pod \"nova-scheduler-0\" (UID: \"f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3\") " pod="openstack/nova-scheduler-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.783603 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3-config-data\") pod \"nova-scheduler-0\" (UID: \"f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3\") " pod="openstack/nova-scheduler-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.783714 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58a77023-a881-4cb5-b38b-b7b7df6386e9-logs\") pod \"nova-api-0\" (UID: \"58a77023-a881-4cb5-b38b-b7b7df6386e9\") " pod="openstack/nova-api-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.791280 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58a77023-a881-4cb5-b38b-b7b7df6386e9-logs\") pod \"nova-api-0\" (UID: \"58a77023-a881-4cb5-b38b-b7b7df6386e9\") " pod="openstack/nova-api-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.793650 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3\") " pod="openstack/nova-scheduler-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.793004 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58a77023-a881-4cb5-b38b-b7b7df6386e9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"58a77023-a881-4cb5-b38b-b7b7df6386e9\") " pod="openstack/nova-api-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.794470 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58a77023-a881-4cb5-b38b-b7b7df6386e9-config-data\") pod \"nova-api-0\" (UID: \"58a77023-a881-4cb5-b38b-b7b7df6386e9\") " pod="openstack/nova-api-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.798495 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3-config-data\") pod \"nova-scheduler-0\" (UID: \"f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3\") " pod="openstack/nova-scheduler-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.799877 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.803018 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.805238 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.805417 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.810735 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.812970 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jctt\" (UniqueName: \"kubernetes.io/projected/f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3-kube-api-access-8jctt\") pod \"nova-scheduler-0\" (UID: \"f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3\") " pod="openstack/nova-scheduler-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.813216 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srpkz\" (UniqueName: \"kubernetes.io/projected/58a77023-a881-4cb5-b38b-b7b7df6386e9-kube-api-access-srpkz\") pod \"nova-api-0\" (UID: \"58a77023-a881-4cb5-b38b-b7b7df6386e9\") " pod="openstack/nova-api-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.891267 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-config-data\") pod \"nova-metadata-0\" (UID: \"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14\") " pod="openstack/nova-metadata-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.891301 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plwlp\" (UniqueName: \"kubernetes.io/projected/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-kube-api-access-plwlp\") pod \"nova-metadata-0\" (UID: \"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14\") " pod="openstack/nova-metadata-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.891404 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-logs\") pod \"nova-metadata-0\" (UID: \"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14\") " pod="openstack/nova-metadata-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.891493 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14\") " pod="openstack/nova-metadata-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.891685 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14\") " pod="openstack/nova-metadata-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.909051 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.923996 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.992892 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14\") " pod="openstack/nova-metadata-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.993325 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-config-data\") pod \"nova-metadata-0\" (UID: \"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14\") " pod="openstack/nova-metadata-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.993345 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plwlp\" (UniqueName: \"kubernetes.io/projected/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-kube-api-access-plwlp\") pod \"nova-metadata-0\" (UID: \"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14\") " pod="openstack/nova-metadata-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.993366 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-logs\") pod \"nova-metadata-0\" (UID: \"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14\") " pod="openstack/nova-metadata-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.993386 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14\") " pod="openstack/nova-metadata-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.994596 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-logs\") pod \"nova-metadata-0\" (UID: \"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14\") " pod="openstack/nova-metadata-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.996754 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-config-data\") pod \"nova-metadata-0\" (UID: \"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14\") " pod="openstack/nova-metadata-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.996941 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14\") " pod="openstack/nova-metadata-0" Nov 25 11:04:00 crc kubenswrapper[4776]: I1125 11:04:00.998031 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14\") " pod="openstack/nova-metadata-0" Nov 25 11:04:01 crc kubenswrapper[4776]: I1125 11:04:01.014369 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plwlp\" (UniqueName: \"kubernetes.io/projected/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-kube-api-access-plwlp\") pod \"nova-metadata-0\" (UID: \"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14\") " pod="openstack/nova-metadata-0" Nov 25 11:04:01 crc kubenswrapper[4776]: I1125 11:04:01.194789 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 11:04:01 crc kubenswrapper[4776]: I1125 11:04:01.383526 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 11:04:01 crc kubenswrapper[4776]: W1125 11:04:01.392655 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6277e15_c6b7_4f7d_af0a_4ef6d362c2a3.slice/crio-0a0f498048728567e4f4279aa87099f59624abc65b7d75b4c49de048c13e9b2d WatchSource:0}: Error finding container 0a0f498048728567e4f4279aa87099f59624abc65b7d75b4c49de048c13e9b2d: Status 404 returned error can't find the container with id 0a0f498048728567e4f4279aa87099f59624abc65b7d75b4c49de048c13e9b2d Nov 25 11:04:01 crc kubenswrapper[4776]: I1125 11:04:01.430399 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3","Type":"ContainerStarted","Data":"0a0f498048728567e4f4279aa87099f59624abc65b7d75b4c49de048c13e9b2d"} Nov 25 11:04:01 crc kubenswrapper[4776]: I1125 11:04:01.430959 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 11:04:01 crc kubenswrapper[4776]: I1125 11:04:01.470469 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 11:04:01 crc kubenswrapper[4776]: W1125 11:04:01.501615 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58a77023_a881_4cb5_b38b_b7b7df6386e9.slice/crio-155ca79111445865d70046dc9605955be506ebc9abff695f04f085ce201996a0 WatchSource:0}: Error finding container 155ca79111445865d70046dc9605955be506ebc9abff695f04f085ce201996a0: Status 404 returned error can't find the container with id 155ca79111445865d70046dc9605955be506ebc9abff695f04f085ce201996a0 Nov 25 11:04:01 crc kubenswrapper[4776]: I1125 11:04:01.672020 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="018ae494-f813-4803-8bbe-515e8ae74af1" path="/var/lib/kubelet/pods/018ae494-f813-4803-8bbe-515e8ae74af1/volumes" Nov 25 11:04:01 crc kubenswrapper[4776]: I1125 11:04:01.672732 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ae1610-2c5a-47fa-8806-053213b43512" path="/var/lib/kubelet/pods/49ae1610-2c5a-47fa-8806-053213b43512/volumes" Nov 25 11:04:01 crc kubenswrapper[4776]: I1125 11:04:01.673370 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da7d17e7-2d62-466a-8826-817e197158d2" path="/var/lib/kubelet/pods/da7d17e7-2d62-466a-8826-817e197158d2/volumes" Nov 25 11:04:02 crc kubenswrapper[4776]: I1125 11:04:02.450389 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14","Type":"ContainerStarted","Data":"235b27dba37c09496e9794f40807ee09d92277a36d96b376f20e24a9121f25d0"} Nov 25 11:04:02 crc kubenswrapper[4776]: I1125 11:04:02.450748 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14","Type":"ContainerStarted","Data":"d691f3e8528edb91b33ba0d8b46b7550b59a4896095da1369db835bf1ea9af57"} Nov 25 11:04:02 crc kubenswrapper[4776]: I1125 11:04:02.450772 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14","Type":"ContainerStarted","Data":"621286f1ba37497c9f2b5734b368462b55e27bdd382dc8a4ed2c4cf1dadf295c"} Nov 25 11:04:02 crc kubenswrapper[4776]: I1125 11:04:02.452247 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"58a77023-a881-4cb5-b38b-b7b7df6386e9","Type":"ContainerStarted","Data":"445bbbbf0f159b26fa7f1babf13213a21dec119fa98091968a8634be31a611cc"} Nov 25 11:04:02 crc kubenswrapper[4776]: I1125 11:04:02.452312 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"58a77023-a881-4cb5-b38b-b7b7df6386e9","Type":"ContainerStarted","Data":"236b32b9b55cd4563a526e3d631330c2d72821fd8bb7821ee92ac98d11e08f3c"} Nov 25 11:04:02 crc kubenswrapper[4776]: I1125 11:04:02.452323 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"58a77023-a881-4cb5-b38b-b7b7df6386e9","Type":"ContainerStarted","Data":"155ca79111445865d70046dc9605955be506ebc9abff695f04f085ce201996a0"} Nov 25 11:04:02 crc kubenswrapper[4776]: I1125 11:04:02.453289 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3","Type":"ContainerStarted","Data":"457893b901e04fac4aa2a523915fbb0b5c6a60a040f98930992bfbfa4f443310"} Nov 25 11:04:02 crc kubenswrapper[4776]: I1125 11:04:02.550588 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.550564782 podStartE2EDuration="2.550564782s" podCreationTimestamp="2025-11-25 11:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:04:02.542907219 +0000 UTC m=+5987.583966772" watchObservedRunningTime="2025-11-25 11:04:02.550564782 +0000 UTC m=+5987.591624335" Nov 25 11:04:02 crc kubenswrapper[4776]: I1125 11:04:02.552014 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.552004308 podStartE2EDuration="2.552004308s" podCreationTimestamp="2025-11-25 11:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:04:02.522400895 +0000 UTC m=+5987.563460448" watchObservedRunningTime="2025-11-25 11:04:02.552004308 +0000 UTC m=+5987.593063861" Nov 25 11:04:05 crc kubenswrapper[4776]: I1125 11:04:05.924179 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 25 11:04:06 crc kubenswrapper[4776]: I1125 11:04:06.195061 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 11:04:06 crc kubenswrapper[4776]: I1125 11:04:06.195168 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 11:04:10 crc kubenswrapper[4776]: I1125 11:04:10.909356 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 11:04:10 crc kubenswrapper[4776]: I1125 11:04:10.911297 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 11:04:10 crc kubenswrapper[4776]: I1125 11:04:10.924786 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 25 11:04:10 crc kubenswrapper[4776]: I1125 11:04:10.976662 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 25 11:04:11 crc kubenswrapper[4776]: I1125 11:04:11.002575 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=11.002542668 podStartE2EDuration="11.002542668s" podCreationTimestamp="2025-11-25 11:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:04:02.570226375 +0000 UTC m=+5987.611285948" watchObservedRunningTime="2025-11-25 11:04:11.002542668 +0000 UTC m=+5996.043602261" Nov 25 11:04:11 crc kubenswrapper[4776]: I1125 11:04:11.195601 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 11:04:11 crc kubenswrapper[4776]: I1125 11:04:11.195647 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 11:04:11 crc kubenswrapper[4776]: I1125 11:04:11.592553 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 25 11:04:11 crc kubenswrapper[4776]: I1125 11:04:11.992446 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="58a77023-a881-4cb5-b38b-b7b7df6386e9" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.97:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 11:04:11 crc kubenswrapper[4776]: I1125 11:04:11.992493 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="58a77023-a881-4cb5-b38b-b7b7df6386e9" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.97:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 11:04:12 crc kubenswrapper[4776]: I1125 11:04:12.243236 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.99:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 11:04:12 crc kubenswrapper[4776]: I1125 11:04:12.243266 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.99:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 11:04:17 crc kubenswrapper[4776]: I1125 11:04:17.818472 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:04:17 crc kubenswrapper[4776]: I1125 11:04:17.819228 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:04:17 crc kubenswrapper[4776]: I1125 11:04:17.819299 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 11:04:18 crc kubenswrapper[4776]: I1125 11:04:18.649609 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 11:04:18 crc kubenswrapper[4776]: I1125 11:04:18.649704 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698" gracePeriod=600 Nov 25 11:04:18 crc kubenswrapper[4776]: E1125 11:04:18.778791 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:04:19 crc kubenswrapper[4776]: I1125 11:04:19.666496 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698" exitCode=0 Nov 25 11:04:19 crc kubenswrapper[4776]: I1125 11:04:19.679493 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698"} Nov 25 11:04:19 crc kubenswrapper[4776]: I1125 11:04:19.679570 4776 scope.go:117] "RemoveContainer" containerID="8fb1377358b27002d683fed0c035aa47a3bd9a7c30a5ee14f1b728159b2160e7" Nov 25 11:04:19 crc kubenswrapper[4776]: I1125 11:04:19.680304 4776 scope.go:117] "RemoveContainer" containerID="8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698" Nov 25 11:04:19 crc kubenswrapper[4776]: E1125 11:04:19.680597 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:04:20 crc kubenswrapper[4776]: I1125 11:04:20.914577 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 11:04:20 crc kubenswrapper[4776]: I1125 11:04:20.915595 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 11:04:20 crc kubenswrapper[4776]: I1125 11:04:20.915651 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 11:04:20 crc kubenswrapper[4776]: I1125 11:04:20.919565 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 11:04:21 crc kubenswrapper[4776]: I1125 11:04:21.202691 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 25 11:04:21 crc kubenswrapper[4776]: I1125 11:04:21.205494 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 25 11:04:21 crc kubenswrapper[4776]: I1125 11:04:21.211401 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 25 11:04:21 crc kubenswrapper[4776]: I1125 11:04:21.691267 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 11:04:21 crc kubenswrapper[4776]: I1125 11:04:21.696001 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 25 11:04:21 crc kubenswrapper[4776]: I1125 11:04:21.696587 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 11:04:21 crc kubenswrapper[4776]: I1125 11:04:21.933810 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86dbcbdfc9-dl692"] Nov 25 11:04:21 crc kubenswrapper[4776]: I1125 11:04:21.935867 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86dbcbdfc9-dl692" Nov 25 11:04:21 crc kubenswrapper[4776]: I1125 11:04:21.950550 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86dbcbdfc9-dl692"] Nov 25 11:04:22 crc kubenswrapper[4776]: I1125 11:04:22.053101 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aeccbc20-2ea2-4507-924d-f81aab9d947a-config\") pod \"dnsmasq-dns-86dbcbdfc9-dl692\" (UID: \"aeccbc20-2ea2-4507-924d-f81aab9d947a\") " pod="openstack/dnsmasq-dns-86dbcbdfc9-dl692" Nov 25 11:04:22 crc kubenswrapper[4776]: I1125 11:04:22.053168 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aeccbc20-2ea2-4507-924d-f81aab9d947a-dns-svc\") pod \"dnsmasq-dns-86dbcbdfc9-dl692\" (UID: \"aeccbc20-2ea2-4507-924d-f81aab9d947a\") " pod="openstack/dnsmasq-dns-86dbcbdfc9-dl692" Nov 25 11:04:22 crc kubenswrapper[4776]: I1125 11:04:22.053413 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29x8k\" (UniqueName: \"kubernetes.io/projected/aeccbc20-2ea2-4507-924d-f81aab9d947a-kube-api-access-29x8k\") pod \"dnsmasq-dns-86dbcbdfc9-dl692\" (UID: \"aeccbc20-2ea2-4507-924d-f81aab9d947a\") " pod="openstack/dnsmasq-dns-86dbcbdfc9-dl692" Nov 25 11:04:22 crc kubenswrapper[4776]: I1125 11:04:22.053865 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aeccbc20-2ea2-4507-924d-f81aab9d947a-ovsdbserver-sb\") pod \"dnsmasq-dns-86dbcbdfc9-dl692\" (UID: \"aeccbc20-2ea2-4507-924d-f81aab9d947a\") " pod="openstack/dnsmasq-dns-86dbcbdfc9-dl692" Nov 25 11:04:22 crc kubenswrapper[4776]: I1125 11:04:22.053947 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aeccbc20-2ea2-4507-924d-f81aab9d947a-ovsdbserver-nb\") pod \"dnsmasq-dns-86dbcbdfc9-dl692\" (UID: \"aeccbc20-2ea2-4507-924d-f81aab9d947a\") " pod="openstack/dnsmasq-dns-86dbcbdfc9-dl692" Nov 25 11:04:22 crc kubenswrapper[4776]: I1125 11:04:22.155277 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aeccbc20-2ea2-4507-924d-f81aab9d947a-config\") pod \"dnsmasq-dns-86dbcbdfc9-dl692\" (UID: \"aeccbc20-2ea2-4507-924d-f81aab9d947a\") " pod="openstack/dnsmasq-dns-86dbcbdfc9-dl692" Nov 25 11:04:22 crc kubenswrapper[4776]: I1125 11:04:22.155328 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aeccbc20-2ea2-4507-924d-f81aab9d947a-dns-svc\") pod \"dnsmasq-dns-86dbcbdfc9-dl692\" (UID: \"aeccbc20-2ea2-4507-924d-f81aab9d947a\") " pod="openstack/dnsmasq-dns-86dbcbdfc9-dl692" Nov 25 11:04:22 crc kubenswrapper[4776]: I1125 11:04:22.155375 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29x8k\" (UniqueName: \"kubernetes.io/projected/aeccbc20-2ea2-4507-924d-f81aab9d947a-kube-api-access-29x8k\") pod \"dnsmasq-dns-86dbcbdfc9-dl692\" (UID: \"aeccbc20-2ea2-4507-924d-f81aab9d947a\") " pod="openstack/dnsmasq-dns-86dbcbdfc9-dl692" Nov 25 11:04:22 crc kubenswrapper[4776]: I1125 11:04:22.155406 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aeccbc20-2ea2-4507-924d-f81aab9d947a-ovsdbserver-sb\") pod \"dnsmasq-dns-86dbcbdfc9-dl692\" (UID: \"aeccbc20-2ea2-4507-924d-f81aab9d947a\") " pod="openstack/dnsmasq-dns-86dbcbdfc9-dl692" Nov 25 11:04:22 crc kubenswrapper[4776]: I1125 11:04:22.155433 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aeccbc20-2ea2-4507-924d-f81aab9d947a-ovsdbserver-nb\") pod \"dnsmasq-dns-86dbcbdfc9-dl692\" (UID: \"aeccbc20-2ea2-4507-924d-f81aab9d947a\") " pod="openstack/dnsmasq-dns-86dbcbdfc9-dl692" Nov 25 11:04:22 crc kubenswrapper[4776]: I1125 11:04:22.156550 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aeccbc20-2ea2-4507-924d-f81aab9d947a-config\") pod \"dnsmasq-dns-86dbcbdfc9-dl692\" (UID: \"aeccbc20-2ea2-4507-924d-f81aab9d947a\") " pod="openstack/dnsmasq-dns-86dbcbdfc9-dl692" Nov 25 11:04:22 crc kubenswrapper[4776]: I1125 11:04:22.156573 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aeccbc20-2ea2-4507-924d-f81aab9d947a-dns-svc\") pod \"dnsmasq-dns-86dbcbdfc9-dl692\" (UID: \"aeccbc20-2ea2-4507-924d-f81aab9d947a\") " pod="openstack/dnsmasq-dns-86dbcbdfc9-dl692" Nov 25 11:04:22 crc kubenswrapper[4776]: I1125 11:04:22.156808 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aeccbc20-2ea2-4507-924d-f81aab9d947a-ovsdbserver-nb\") pod \"dnsmasq-dns-86dbcbdfc9-dl692\" (UID: \"aeccbc20-2ea2-4507-924d-f81aab9d947a\") " pod="openstack/dnsmasq-dns-86dbcbdfc9-dl692" Nov 25 11:04:22 crc kubenswrapper[4776]: I1125 11:04:22.156937 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aeccbc20-2ea2-4507-924d-f81aab9d947a-ovsdbserver-sb\") pod \"dnsmasq-dns-86dbcbdfc9-dl692\" (UID: \"aeccbc20-2ea2-4507-924d-f81aab9d947a\") " pod="openstack/dnsmasq-dns-86dbcbdfc9-dl692" Nov 25 11:04:22 crc kubenswrapper[4776]: I1125 11:04:22.190265 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29x8k\" (UniqueName: \"kubernetes.io/projected/aeccbc20-2ea2-4507-924d-f81aab9d947a-kube-api-access-29x8k\") pod \"dnsmasq-dns-86dbcbdfc9-dl692\" (UID: \"aeccbc20-2ea2-4507-924d-f81aab9d947a\") " pod="openstack/dnsmasq-dns-86dbcbdfc9-dl692" Nov 25 11:04:22 crc kubenswrapper[4776]: I1125 11:04:22.280171 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86dbcbdfc9-dl692" Nov 25 11:04:22 crc kubenswrapper[4776]: W1125 11:04:22.785778 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaeccbc20_2ea2_4507_924d_f81aab9d947a.slice/crio-acd5be03ece01a4f28a5d1188113522f13fdb5ff3baf71cd74763b729c534b3c WatchSource:0}: Error finding container acd5be03ece01a4f28a5d1188113522f13fdb5ff3baf71cd74763b729c534b3c: Status 404 returned error can't find the container with id acd5be03ece01a4f28a5d1188113522f13fdb5ff3baf71cd74763b729c534b3c Nov 25 11:04:22 crc kubenswrapper[4776]: I1125 11:04:22.793267 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86dbcbdfc9-dl692"] Nov 25 11:04:23 crc kubenswrapper[4776]: I1125 11:04:23.708363 4776 generic.go:334] "Generic (PLEG): container finished" podID="aeccbc20-2ea2-4507-924d-f81aab9d947a" containerID="3648b1978ec3bab3746f507d49deed95c584cb4515abd3c4a274b73b0453f620" exitCode=0 Nov 25 11:04:23 crc kubenswrapper[4776]: I1125 11:04:23.708581 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86dbcbdfc9-dl692" event={"ID":"aeccbc20-2ea2-4507-924d-f81aab9d947a","Type":"ContainerDied","Data":"3648b1978ec3bab3746f507d49deed95c584cb4515abd3c4a274b73b0453f620"} Nov 25 11:04:23 crc kubenswrapper[4776]: I1125 11:04:23.708805 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86dbcbdfc9-dl692" event={"ID":"aeccbc20-2ea2-4507-924d-f81aab9d947a","Type":"ContainerStarted","Data":"acd5be03ece01a4f28a5d1188113522f13fdb5ff3baf71cd74763b729c534b3c"} Nov 25 11:04:24 crc kubenswrapper[4776]: I1125 11:04:24.617184 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 11:04:24 crc kubenswrapper[4776]: I1125 11:04:24.722979 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="58a77023-a881-4cb5-b38b-b7b7df6386e9" containerName="nova-api-log" containerID="cri-o://236b32b9b55cd4563a526e3d631330c2d72821fd8bb7821ee92ac98d11e08f3c" gracePeriod=30 Nov 25 11:04:24 crc kubenswrapper[4776]: I1125 11:04:24.723892 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86dbcbdfc9-dl692" event={"ID":"aeccbc20-2ea2-4507-924d-f81aab9d947a","Type":"ContainerStarted","Data":"08b2752014bddb8246b5b237fdd56860d953bfc26af95426502f062d9c002b79"} Nov 25 11:04:24 crc kubenswrapper[4776]: I1125 11:04:24.723929 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86dbcbdfc9-dl692" Nov 25 11:04:24 crc kubenswrapper[4776]: I1125 11:04:24.724453 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="58a77023-a881-4cb5-b38b-b7b7df6386e9" containerName="nova-api-api" containerID="cri-o://445bbbbf0f159b26fa7f1babf13213a21dec119fa98091968a8634be31a611cc" gracePeriod=30 Nov 25 11:04:24 crc kubenswrapper[4776]: I1125 11:04:24.746705 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86dbcbdfc9-dl692" podStartSLOduration=3.746682959 podStartE2EDuration="3.746682959s" podCreationTimestamp="2025-11-25 11:04:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:04:24.742824272 +0000 UTC m=+6009.783883825" watchObservedRunningTime="2025-11-25 11:04:24.746682959 +0000 UTC m=+6009.787742512" Nov 25 11:04:25 crc kubenswrapper[4776]: I1125 11:04:25.735869 4776 generic.go:334] "Generic (PLEG): container finished" podID="58a77023-a881-4cb5-b38b-b7b7df6386e9" containerID="236b32b9b55cd4563a526e3d631330c2d72821fd8bb7821ee92ac98d11e08f3c" exitCode=143 Nov 25 11:04:25 crc kubenswrapper[4776]: I1125 11:04:25.735945 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"58a77023-a881-4cb5-b38b-b7b7df6386e9","Type":"ContainerDied","Data":"236b32b9b55cd4563a526e3d631330c2d72821fd8bb7821ee92ac98d11e08f3c"} Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.311890 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.374237 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srpkz\" (UniqueName: \"kubernetes.io/projected/58a77023-a881-4cb5-b38b-b7b7df6386e9-kube-api-access-srpkz\") pod \"58a77023-a881-4cb5-b38b-b7b7df6386e9\" (UID: \"58a77023-a881-4cb5-b38b-b7b7df6386e9\") " Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.374856 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58a77023-a881-4cb5-b38b-b7b7df6386e9-config-data\") pod \"58a77023-a881-4cb5-b38b-b7b7df6386e9\" (UID: \"58a77023-a881-4cb5-b38b-b7b7df6386e9\") " Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.375282 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58a77023-a881-4cb5-b38b-b7b7df6386e9-combined-ca-bundle\") pod \"58a77023-a881-4cb5-b38b-b7b7df6386e9\" (UID: \"58a77023-a881-4cb5-b38b-b7b7df6386e9\") " Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.375551 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58a77023-a881-4cb5-b38b-b7b7df6386e9-logs\") pod \"58a77023-a881-4cb5-b38b-b7b7df6386e9\" (UID: \"58a77023-a881-4cb5-b38b-b7b7df6386e9\") " Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.376904 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58a77023-a881-4cb5-b38b-b7b7df6386e9-logs" (OuterVolumeSpecName: "logs") pod "58a77023-a881-4cb5-b38b-b7b7df6386e9" (UID: "58a77023-a881-4cb5-b38b-b7b7df6386e9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.379415 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58a77023-a881-4cb5-b38b-b7b7df6386e9-kube-api-access-srpkz" (OuterVolumeSpecName: "kube-api-access-srpkz") pod "58a77023-a881-4cb5-b38b-b7b7df6386e9" (UID: "58a77023-a881-4cb5-b38b-b7b7df6386e9"). InnerVolumeSpecName "kube-api-access-srpkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.406680 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58a77023-a881-4cb5-b38b-b7b7df6386e9-config-data" (OuterVolumeSpecName: "config-data") pod "58a77023-a881-4cb5-b38b-b7b7df6386e9" (UID: "58a77023-a881-4cb5-b38b-b7b7df6386e9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.408536 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58a77023-a881-4cb5-b38b-b7b7df6386e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58a77023-a881-4cb5-b38b-b7b7df6386e9" (UID: "58a77023-a881-4cb5-b38b-b7b7df6386e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.478380 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58a77023-a881-4cb5-b38b-b7b7df6386e9-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.478424 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58a77023-a881-4cb5-b38b-b7b7df6386e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.478438 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58a77023-a881-4cb5-b38b-b7b7df6386e9-logs\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.478448 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srpkz\" (UniqueName: \"kubernetes.io/projected/58a77023-a881-4cb5-b38b-b7b7df6386e9-kube-api-access-srpkz\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.782398 4776 generic.go:334] "Generic (PLEG): container finished" podID="58a77023-a881-4cb5-b38b-b7b7df6386e9" containerID="445bbbbf0f159b26fa7f1babf13213a21dec119fa98091968a8634be31a611cc" exitCode=0 Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.782455 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"58a77023-a881-4cb5-b38b-b7b7df6386e9","Type":"ContainerDied","Data":"445bbbbf0f159b26fa7f1babf13213a21dec119fa98091968a8634be31a611cc"} Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.782486 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"58a77023-a881-4cb5-b38b-b7b7df6386e9","Type":"ContainerDied","Data":"155ca79111445865d70046dc9605955be506ebc9abff695f04f085ce201996a0"} Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.782521 4776 scope.go:117] "RemoveContainer" containerID="445bbbbf0f159b26fa7f1babf13213a21dec119fa98091968a8634be31a611cc" Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.782711 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.830423 4776 scope.go:117] "RemoveContainer" containerID="236b32b9b55cd4563a526e3d631330c2d72821fd8bb7821ee92ac98d11e08f3c" Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.861619 4776 scope.go:117] "RemoveContainer" containerID="445bbbbf0f159b26fa7f1babf13213a21dec119fa98091968a8634be31a611cc" Nov 25 11:04:28 crc kubenswrapper[4776]: E1125 11:04:28.862616 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"445bbbbf0f159b26fa7f1babf13213a21dec119fa98091968a8634be31a611cc\": container with ID starting with 445bbbbf0f159b26fa7f1babf13213a21dec119fa98091968a8634be31a611cc not found: ID does not exist" containerID="445bbbbf0f159b26fa7f1babf13213a21dec119fa98091968a8634be31a611cc" Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.862656 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"445bbbbf0f159b26fa7f1babf13213a21dec119fa98091968a8634be31a611cc"} err="failed to get container status \"445bbbbf0f159b26fa7f1babf13213a21dec119fa98091968a8634be31a611cc\": rpc error: code = NotFound desc = could not find container \"445bbbbf0f159b26fa7f1babf13213a21dec119fa98091968a8634be31a611cc\": container with ID starting with 445bbbbf0f159b26fa7f1babf13213a21dec119fa98091968a8634be31a611cc not found: ID does not exist" Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.862687 4776 scope.go:117] "RemoveContainer" containerID="236b32b9b55cd4563a526e3d631330c2d72821fd8bb7821ee92ac98d11e08f3c" Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.878888 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 11:04:28 crc kubenswrapper[4776]: E1125 11:04:28.879015 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"236b32b9b55cd4563a526e3d631330c2d72821fd8bb7821ee92ac98d11e08f3c\": container with ID starting with 236b32b9b55cd4563a526e3d631330c2d72821fd8bb7821ee92ac98d11e08f3c not found: ID does not exist" containerID="236b32b9b55cd4563a526e3d631330c2d72821fd8bb7821ee92ac98d11e08f3c" Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.879042 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"236b32b9b55cd4563a526e3d631330c2d72821fd8bb7821ee92ac98d11e08f3c"} err="failed to get container status \"236b32b9b55cd4563a526e3d631330c2d72821fd8bb7821ee92ac98d11e08f3c\": rpc error: code = NotFound desc = could not find container \"236b32b9b55cd4563a526e3d631330c2d72821fd8bb7821ee92ac98d11e08f3c\": container with ID starting with 236b32b9b55cd4563a526e3d631330c2d72821fd8bb7821ee92ac98d11e08f3c not found: ID does not exist" Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.897161 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.917263 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 11:04:28 crc kubenswrapper[4776]: E1125 11:04:28.917701 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58a77023-a881-4cb5-b38b-b7b7df6386e9" containerName="nova-api-log" Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.917717 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="58a77023-a881-4cb5-b38b-b7b7df6386e9" containerName="nova-api-log" Nov 25 11:04:28 crc kubenswrapper[4776]: E1125 11:04:28.917754 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58a77023-a881-4cb5-b38b-b7b7df6386e9" containerName="nova-api-api" Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.917762 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="58a77023-a881-4cb5-b38b-b7b7df6386e9" containerName="nova-api-api" Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.917941 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="58a77023-a881-4cb5-b38b-b7b7df6386e9" containerName="nova-api-log" Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.917979 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="58a77023-a881-4cb5-b38b-b7b7df6386e9" containerName="nova-api-api" Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.919085 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.923485 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.929409 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.929554 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 11:04:28 crc kubenswrapper[4776]: I1125 11:04:28.935337 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 11:04:29 crc kubenswrapper[4776]: I1125 11:04:29.038547 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d854edbe-4354-4149-9d58-b503e44bd1c8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d854edbe-4354-4149-9d58-b503e44bd1c8\") " pod="openstack/nova-api-0" Nov 25 11:04:29 crc kubenswrapper[4776]: I1125 11:04:29.038622 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d854edbe-4354-4149-9d58-b503e44bd1c8-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d854edbe-4354-4149-9d58-b503e44bd1c8\") " pod="openstack/nova-api-0" Nov 25 11:04:29 crc kubenswrapper[4776]: I1125 11:04:29.038639 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d854edbe-4354-4149-9d58-b503e44bd1c8-public-tls-certs\") pod \"nova-api-0\" (UID: \"d854edbe-4354-4149-9d58-b503e44bd1c8\") " pod="openstack/nova-api-0" Nov 25 11:04:29 crc kubenswrapper[4776]: I1125 11:04:29.038683 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d854edbe-4354-4149-9d58-b503e44bd1c8-config-data\") pod \"nova-api-0\" (UID: \"d854edbe-4354-4149-9d58-b503e44bd1c8\") " pod="openstack/nova-api-0" Nov 25 11:04:29 crc kubenswrapper[4776]: I1125 11:04:29.038738 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmzcc\" (UniqueName: \"kubernetes.io/projected/d854edbe-4354-4149-9d58-b503e44bd1c8-kube-api-access-jmzcc\") pod \"nova-api-0\" (UID: \"d854edbe-4354-4149-9d58-b503e44bd1c8\") " pod="openstack/nova-api-0" Nov 25 11:04:29 crc kubenswrapper[4776]: I1125 11:04:29.038792 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d854edbe-4354-4149-9d58-b503e44bd1c8-logs\") pod \"nova-api-0\" (UID: \"d854edbe-4354-4149-9d58-b503e44bd1c8\") " pod="openstack/nova-api-0" Nov 25 11:04:29 crc kubenswrapper[4776]: I1125 11:04:29.140561 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d854edbe-4354-4149-9d58-b503e44bd1c8-logs\") pod \"nova-api-0\" (UID: \"d854edbe-4354-4149-9d58-b503e44bd1c8\") " pod="openstack/nova-api-0" Nov 25 11:04:29 crc kubenswrapper[4776]: I1125 11:04:29.140658 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d854edbe-4354-4149-9d58-b503e44bd1c8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d854edbe-4354-4149-9d58-b503e44bd1c8\") " pod="openstack/nova-api-0" Nov 25 11:04:29 crc kubenswrapper[4776]: I1125 11:04:29.140705 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d854edbe-4354-4149-9d58-b503e44bd1c8-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d854edbe-4354-4149-9d58-b503e44bd1c8\") " pod="openstack/nova-api-0" Nov 25 11:04:29 crc kubenswrapper[4776]: I1125 11:04:29.140725 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d854edbe-4354-4149-9d58-b503e44bd1c8-public-tls-certs\") pod \"nova-api-0\" (UID: \"d854edbe-4354-4149-9d58-b503e44bd1c8\") " pod="openstack/nova-api-0" Nov 25 11:04:29 crc kubenswrapper[4776]: I1125 11:04:29.140782 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d854edbe-4354-4149-9d58-b503e44bd1c8-config-data\") pod \"nova-api-0\" (UID: \"d854edbe-4354-4149-9d58-b503e44bd1c8\") " pod="openstack/nova-api-0" Nov 25 11:04:29 crc kubenswrapper[4776]: I1125 11:04:29.140847 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmzcc\" (UniqueName: \"kubernetes.io/projected/d854edbe-4354-4149-9d58-b503e44bd1c8-kube-api-access-jmzcc\") pod \"nova-api-0\" (UID: \"d854edbe-4354-4149-9d58-b503e44bd1c8\") " pod="openstack/nova-api-0" Nov 25 11:04:29 crc kubenswrapper[4776]: I1125 11:04:29.141642 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d854edbe-4354-4149-9d58-b503e44bd1c8-logs\") pod \"nova-api-0\" (UID: \"d854edbe-4354-4149-9d58-b503e44bd1c8\") " pod="openstack/nova-api-0" Nov 25 11:04:29 crc kubenswrapper[4776]: I1125 11:04:29.147587 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d854edbe-4354-4149-9d58-b503e44bd1c8-public-tls-certs\") pod \"nova-api-0\" (UID: \"d854edbe-4354-4149-9d58-b503e44bd1c8\") " pod="openstack/nova-api-0" Nov 25 11:04:29 crc kubenswrapper[4776]: I1125 11:04:29.147591 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d854edbe-4354-4149-9d58-b503e44bd1c8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d854edbe-4354-4149-9d58-b503e44bd1c8\") " pod="openstack/nova-api-0" Nov 25 11:04:29 crc kubenswrapper[4776]: I1125 11:04:29.149493 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d854edbe-4354-4149-9d58-b503e44bd1c8-internal-tls-certs\") pod \"nova-api-0\" (UID: \"d854edbe-4354-4149-9d58-b503e44bd1c8\") " pod="openstack/nova-api-0" Nov 25 11:04:29 crc kubenswrapper[4776]: I1125 11:04:29.154139 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d854edbe-4354-4149-9d58-b503e44bd1c8-config-data\") pod \"nova-api-0\" (UID: \"d854edbe-4354-4149-9d58-b503e44bd1c8\") " pod="openstack/nova-api-0" Nov 25 11:04:29 crc kubenswrapper[4776]: I1125 11:04:29.157375 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmzcc\" (UniqueName: \"kubernetes.io/projected/d854edbe-4354-4149-9d58-b503e44bd1c8-kube-api-access-jmzcc\") pod \"nova-api-0\" (UID: \"d854edbe-4354-4149-9d58-b503e44bd1c8\") " pod="openstack/nova-api-0" Nov 25 11:04:29 crc kubenswrapper[4776]: I1125 11:04:29.250593 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 11:04:29 crc kubenswrapper[4776]: I1125 11:04:29.676348 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58a77023-a881-4cb5-b38b-b7b7df6386e9" path="/var/lib/kubelet/pods/58a77023-a881-4cb5-b38b-b7b7df6386e9/volumes" Nov 25 11:04:29 crc kubenswrapper[4776]: I1125 11:04:29.715309 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 11:04:29 crc kubenswrapper[4776]: I1125 11:04:29.793152 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d854edbe-4354-4149-9d58-b503e44bd1c8","Type":"ContainerStarted","Data":"d0b260d7fdb4c737a5daa056f7515b10e6b2d43212d0556af09a948ed987e75b"} Nov 25 11:04:30 crc kubenswrapper[4776]: I1125 11:04:30.804090 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d854edbe-4354-4149-9d58-b503e44bd1c8","Type":"ContainerStarted","Data":"e2071c42555d92e072b179f72bc660dfec1e153919f6300461d43023f7c18e70"} Nov 25 11:04:30 crc kubenswrapper[4776]: I1125 11:04:30.804479 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d854edbe-4354-4149-9d58-b503e44bd1c8","Type":"ContainerStarted","Data":"d6e46cd49b2c18d00f7c5566b97c6a00bed98218a03cc5997bbd1e486a0d50d1"} Nov 25 11:04:30 crc kubenswrapper[4776]: I1125 11:04:30.826143 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.826116152 podStartE2EDuration="2.826116152s" podCreationTimestamp="2025-11-25 11:04:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:04:30.820764888 +0000 UTC m=+6015.861824461" watchObservedRunningTime="2025-11-25 11:04:30.826116152 +0000 UTC m=+6015.867175715" Nov 25 11:04:32 crc kubenswrapper[4776]: I1125 11:04:32.284336 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86dbcbdfc9-dl692" Nov 25 11:04:32 crc kubenswrapper[4776]: I1125 11:04:32.350841 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bdbcf4d67-qklzz"] Nov 25 11:04:32 crc kubenswrapper[4776]: I1125 11:04:32.351202 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5bdbcf4d67-qklzz" podUID="386eb4a1-747e-45d3-b665-fbb1a6c7b06b" containerName="dnsmasq-dns" containerID="cri-o://9b819bc49d6e3e198bb8825480a3e5ed1756ef83efdf29cd7e01996b5f25e992" gracePeriod=10 Nov 25 11:04:32 crc kubenswrapper[4776]: I1125 11:04:32.825025 4776 generic.go:334] "Generic (PLEG): container finished" podID="386eb4a1-747e-45d3-b665-fbb1a6c7b06b" containerID="9b819bc49d6e3e198bb8825480a3e5ed1756ef83efdf29cd7e01996b5f25e992" exitCode=0 Nov 25 11:04:32 crc kubenswrapper[4776]: I1125 11:04:32.825104 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bdbcf4d67-qklzz" event={"ID":"386eb4a1-747e-45d3-b665-fbb1a6c7b06b","Type":"ContainerDied","Data":"9b819bc49d6e3e198bb8825480a3e5ed1756ef83efdf29cd7e01996b5f25e992"} Nov 25 11:04:32 crc kubenswrapper[4776]: I1125 11:04:32.825378 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bdbcf4d67-qklzz" event={"ID":"386eb4a1-747e-45d3-b665-fbb1a6c7b06b","Type":"ContainerDied","Data":"4fb31ebee5814aaa27e8e537b779385a3308b339fe6d5fda34f50880de3e44a7"} Nov 25 11:04:32 crc kubenswrapper[4776]: I1125 11:04:32.825396 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4fb31ebee5814aaa27e8e537b779385a3308b339fe6d5fda34f50880de3e44a7" Nov 25 11:04:32 crc kubenswrapper[4776]: I1125 11:04:32.866453 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bdbcf4d67-qklzz" Nov 25 11:04:33 crc kubenswrapper[4776]: I1125 11:04:33.011607 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-config\") pod \"386eb4a1-747e-45d3-b665-fbb1a6c7b06b\" (UID: \"386eb4a1-747e-45d3-b665-fbb1a6c7b06b\") " Nov 25 11:04:33 crc kubenswrapper[4776]: I1125 11:04:33.011695 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-ovsdbserver-nb\") pod \"386eb4a1-747e-45d3-b665-fbb1a6c7b06b\" (UID: \"386eb4a1-747e-45d3-b665-fbb1a6c7b06b\") " Nov 25 11:04:33 crc kubenswrapper[4776]: I1125 11:04:33.011946 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-ovsdbserver-sb\") pod \"386eb4a1-747e-45d3-b665-fbb1a6c7b06b\" (UID: \"386eb4a1-747e-45d3-b665-fbb1a6c7b06b\") " Nov 25 11:04:33 crc kubenswrapper[4776]: I1125 11:04:33.012403 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5blf\" (UniqueName: \"kubernetes.io/projected/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-kube-api-access-b5blf\") pod \"386eb4a1-747e-45d3-b665-fbb1a6c7b06b\" (UID: \"386eb4a1-747e-45d3-b665-fbb1a6c7b06b\") " Nov 25 11:04:33 crc kubenswrapper[4776]: I1125 11:04:33.012472 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-dns-svc\") pod \"386eb4a1-747e-45d3-b665-fbb1a6c7b06b\" (UID: \"386eb4a1-747e-45d3-b665-fbb1a6c7b06b\") " Nov 25 11:04:33 crc kubenswrapper[4776]: I1125 11:04:33.016818 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-kube-api-access-b5blf" (OuterVolumeSpecName: "kube-api-access-b5blf") pod "386eb4a1-747e-45d3-b665-fbb1a6c7b06b" (UID: "386eb4a1-747e-45d3-b665-fbb1a6c7b06b"). InnerVolumeSpecName "kube-api-access-b5blf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:04:33 crc kubenswrapper[4776]: I1125 11:04:33.060587 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "386eb4a1-747e-45d3-b665-fbb1a6c7b06b" (UID: "386eb4a1-747e-45d3-b665-fbb1a6c7b06b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:04:33 crc kubenswrapper[4776]: I1125 11:04:33.069469 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "386eb4a1-747e-45d3-b665-fbb1a6c7b06b" (UID: "386eb4a1-747e-45d3-b665-fbb1a6c7b06b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:04:33 crc kubenswrapper[4776]: I1125 11:04:33.069726 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-config" (OuterVolumeSpecName: "config") pod "386eb4a1-747e-45d3-b665-fbb1a6c7b06b" (UID: "386eb4a1-747e-45d3-b665-fbb1a6c7b06b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:04:33 crc kubenswrapper[4776]: I1125 11:04:33.071926 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "386eb4a1-747e-45d3-b665-fbb1a6c7b06b" (UID: "386eb4a1-747e-45d3-b665-fbb1a6c7b06b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:04:33 crc kubenswrapper[4776]: I1125 11:04:33.115441 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:33 crc kubenswrapper[4776]: I1125 11:04:33.115469 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5blf\" (UniqueName: \"kubernetes.io/projected/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-kube-api-access-b5blf\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:33 crc kubenswrapper[4776]: I1125 11:04:33.115500 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:33 crc kubenswrapper[4776]: I1125 11:04:33.115519 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-config\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:33 crc kubenswrapper[4776]: I1125 11:04:33.115530 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/386eb4a1-747e-45d3-b665-fbb1a6c7b06b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:33 crc kubenswrapper[4776]: I1125 11:04:33.662436 4776 scope.go:117] "RemoveContainer" containerID="8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698" Nov 25 11:04:33 crc kubenswrapper[4776]: E1125 11:04:33.663077 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:04:33 crc kubenswrapper[4776]: I1125 11:04:33.834118 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bdbcf4d67-qklzz" Nov 25 11:04:33 crc kubenswrapper[4776]: I1125 11:04:33.866922 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bdbcf4d67-qklzz"] Nov 25 11:04:33 crc kubenswrapper[4776]: I1125 11:04:33.880625 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bdbcf4d67-qklzz"] Nov 25 11:04:35 crc kubenswrapper[4776]: I1125 11:04:35.678630 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="386eb4a1-747e-45d3-b665-fbb1a6c7b06b" path="/var/lib/kubelet/pods/386eb4a1-747e-45d3-b665-fbb1a6c7b06b/volumes" Nov 25 11:04:39 crc kubenswrapper[4776]: I1125 11:04:39.060234 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-3a30-account-create-rwgn5"] Nov 25 11:04:39 crc kubenswrapper[4776]: I1125 11:04:39.068341 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-9zjdz"] Nov 25 11:04:39 crc kubenswrapper[4776]: I1125 11:04:39.077975 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-9zjdz"] Nov 25 11:04:39 crc kubenswrapper[4776]: I1125 11:04:39.085247 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-3a30-account-create-rwgn5"] Nov 25 11:04:39 crc kubenswrapper[4776]: I1125 11:04:39.250780 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 11:04:39 crc kubenswrapper[4776]: I1125 11:04:39.250847 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 11:04:39 crc kubenswrapper[4776]: I1125 11:04:39.675457 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e9ca664-92bc-49cc-8bca-66d0fbf38343" path="/var/lib/kubelet/pods/0e9ca664-92bc-49cc-8bca-66d0fbf38343/volumes" Nov 25 11:04:39 crc kubenswrapper[4776]: I1125 11:04:39.676164 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3d99a94-8d26-4627-bd67-818fcca07854" path="/var/lib/kubelet/pods/b3d99a94-8d26-4627-bd67-818fcca07854/volumes" Nov 25 11:04:40 crc kubenswrapper[4776]: I1125 11:04:40.265218 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d854edbe-4354-4149-9d58-b503e44bd1c8" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.101:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 11:04:40 crc kubenswrapper[4776]: I1125 11:04:40.265490 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d854edbe-4354-4149-9d58-b503e44bd1c8" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.101:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.141956 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-7wxbq"] Nov 25 11:04:42 crc kubenswrapper[4776]: E1125 11:04:42.142704 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="386eb4a1-747e-45d3-b665-fbb1a6c7b06b" containerName="init" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.142720 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="386eb4a1-747e-45d3-b665-fbb1a6c7b06b" containerName="init" Nov 25 11:04:42 crc kubenswrapper[4776]: E1125 11:04:42.142741 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="386eb4a1-747e-45d3-b665-fbb1a6c7b06b" containerName="dnsmasq-dns" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.142747 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="386eb4a1-747e-45d3-b665-fbb1a6c7b06b" containerName="dnsmasq-dns" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.142922 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="386eb4a1-747e-45d3-b665-fbb1a6c7b06b" containerName="dnsmasq-dns" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.143567 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-7wxbq" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.146510 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.146771 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.153088 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-7wxbq"] Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.213332 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8df08eab-1375-42cd-b1dc-31a953e2bca6-scripts\") pod \"swift-ring-rebalance-7wxbq\" (UID: \"8df08eab-1375-42cd-b1dc-31a953e2bca6\") " pod="openstack/swift-ring-rebalance-7wxbq" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.213415 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8df08eab-1375-42cd-b1dc-31a953e2bca6-combined-ca-bundle\") pod \"swift-ring-rebalance-7wxbq\" (UID: \"8df08eab-1375-42cd-b1dc-31a953e2bca6\") " pod="openstack/swift-ring-rebalance-7wxbq" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.213448 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8df08eab-1375-42cd-b1dc-31a953e2bca6-swiftconf\") pod \"swift-ring-rebalance-7wxbq\" (UID: \"8df08eab-1375-42cd-b1dc-31a953e2bca6\") " pod="openstack/swift-ring-rebalance-7wxbq" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.213659 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8df08eab-1375-42cd-b1dc-31a953e2bca6-ring-data-devices\") pod \"swift-ring-rebalance-7wxbq\" (UID: \"8df08eab-1375-42cd-b1dc-31a953e2bca6\") " pod="openstack/swift-ring-rebalance-7wxbq" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.213833 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4587\" (UniqueName: \"kubernetes.io/projected/8df08eab-1375-42cd-b1dc-31a953e2bca6-kube-api-access-x4587\") pod \"swift-ring-rebalance-7wxbq\" (UID: \"8df08eab-1375-42cd-b1dc-31a953e2bca6\") " pod="openstack/swift-ring-rebalance-7wxbq" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.213895 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8df08eab-1375-42cd-b1dc-31a953e2bca6-etc-swift\") pod \"swift-ring-rebalance-7wxbq\" (UID: \"8df08eab-1375-42cd-b1dc-31a953e2bca6\") " pod="openstack/swift-ring-rebalance-7wxbq" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.214116 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8df08eab-1375-42cd-b1dc-31a953e2bca6-dispersionconf\") pod \"swift-ring-rebalance-7wxbq\" (UID: \"8df08eab-1375-42cd-b1dc-31a953e2bca6\") " pod="openstack/swift-ring-rebalance-7wxbq" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.315290 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8df08eab-1375-42cd-b1dc-31a953e2bca6-dispersionconf\") pod \"swift-ring-rebalance-7wxbq\" (UID: \"8df08eab-1375-42cd-b1dc-31a953e2bca6\") " pod="openstack/swift-ring-rebalance-7wxbq" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.315348 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8df08eab-1375-42cd-b1dc-31a953e2bca6-scripts\") pod \"swift-ring-rebalance-7wxbq\" (UID: \"8df08eab-1375-42cd-b1dc-31a953e2bca6\") " pod="openstack/swift-ring-rebalance-7wxbq" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.315389 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8df08eab-1375-42cd-b1dc-31a953e2bca6-combined-ca-bundle\") pod \"swift-ring-rebalance-7wxbq\" (UID: \"8df08eab-1375-42cd-b1dc-31a953e2bca6\") " pod="openstack/swift-ring-rebalance-7wxbq" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.315420 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8df08eab-1375-42cd-b1dc-31a953e2bca6-swiftconf\") pod \"swift-ring-rebalance-7wxbq\" (UID: \"8df08eab-1375-42cd-b1dc-31a953e2bca6\") " pod="openstack/swift-ring-rebalance-7wxbq" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.315503 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8df08eab-1375-42cd-b1dc-31a953e2bca6-ring-data-devices\") pod \"swift-ring-rebalance-7wxbq\" (UID: \"8df08eab-1375-42cd-b1dc-31a953e2bca6\") " pod="openstack/swift-ring-rebalance-7wxbq" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.315562 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4587\" (UniqueName: \"kubernetes.io/projected/8df08eab-1375-42cd-b1dc-31a953e2bca6-kube-api-access-x4587\") pod \"swift-ring-rebalance-7wxbq\" (UID: \"8df08eab-1375-42cd-b1dc-31a953e2bca6\") " pod="openstack/swift-ring-rebalance-7wxbq" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.315844 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8df08eab-1375-42cd-b1dc-31a953e2bca6-etc-swift\") pod \"swift-ring-rebalance-7wxbq\" (UID: \"8df08eab-1375-42cd-b1dc-31a953e2bca6\") " pod="openstack/swift-ring-rebalance-7wxbq" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.316323 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8df08eab-1375-42cd-b1dc-31a953e2bca6-etc-swift\") pod \"swift-ring-rebalance-7wxbq\" (UID: \"8df08eab-1375-42cd-b1dc-31a953e2bca6\") " pod="openstack/swift-ring-rebalance-7wxbq" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.316642 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8df08eab-1375-42cd-b1dc-31a953e2bca6-scripts\") pod \"swift-ring-rebalance-7wxbq\" (UID: \"8df08eab-1375-42cd-b1dc-31a953e2bca6\") " pod="openstack/swift-ring-rebalance-7wxbq" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.316742 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8df08eab-1375-42cd-b1dc-31a953e2bca6-ring-data-devices\") pod \"swift-ring-rebalance-7wxbq\" (UID: \"8df08eab-1375-42cd-b1dc-31a953e2bca6\") " pod="openstack/swift-ring-rebalance-7wxbq" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.321979 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8df08eab-1375-42cd-b1dc-31a953e2bca6-dispersionconf\") pod \"swift-ring-rebalance-7wxbq\" (UID: \"8df08eab-1375-42cd-b1dc-31a953e2bca6\") " pod="openstack/swift-ring-rebalance-7wxbq" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.322762 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8df08eab-1375-42cd-b1dc-31a953e2bca6-swiftconf\") pod \"swift-ring-rebalance-7wxbq\" (UID: \"8df08eab-1375-42cd-b1dc-31a953e2bca6\") " pod="openstack/swift-ring-rebalance-7wxbq" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.323187 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8df08eab-1375-42cd-b1dc-31a953e2bca6-combined-ca-bundle\") pod \"swift-ring-rebalance-7wxbq\" (UID: \"8df08eab-1375-42cd-b1dc-31a953e2bca6\") " pod="openstack/swift-ring-rebalance-7wxbq" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.333622 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4587\" (UniqueName: \"kubernetes.io/projected/8df08eab-1375-42cd-b1dc-31a953e2bca6-kube-api-access-x4587\") pod \"swift-ring-rebalance-7wxbq\" (UID: \"8df08eab-1375-42cd-b1dc-31a953e2bca6\") " pod="openstack/swift-ring-rebalance-7wxbq" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.472522 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-7wxbq" Nov 25 11:04:42 crc kubenswrapper[4776]: I1125 11:04:42.956616 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-7wxbq"] Nov 25 11:04:43 crc kubenswrapper[4776]: I1125 11:04:43.939138 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-7wxbq" event={"ID":"8df08eab-1375-42cd-b1dc-31a953e2bca6","Type":"ContainerStarted","Data":"217b34fc9df4ff20b62948a9319c19cf41a8a91e92f8b822729e7bc18ea823bd"} Nov 25 11:04:43 crc kubenswrapper[4776]: I1125 11:04:43.939217 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-7wxbq" event={"ID":"8df08eab-1375-42cd-b1dc-31a953e2bca6","Type":"ContainerStarted","Data":"a3fca680b356195aa5821fe1114d3dfcce0f3326997046a06c6307178429ab79"} Nov 25 11:04:43 crc kubenswrapper[4776]: I1125 11:04:43.964774 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-7wxbq" podStartSLOduration=1.9647483270000001 podStartE2EDuration="1.964748327s" podCreationTimestamp="2025-11-25 11:04:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:04:43.956778687 +0000 UTC m=+6028.997838280" watchObservedRunningTime="2025-11-25 11:04:43.964748327 +0000 UTC m=+6029.005807910" Nov 25 11:04:45 crc kubenswrapper[4776]: I1125 11:04:45.039950 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-jg4k7"] Nov 25 11:04:45 crc kubenswrapper[4776]: I1125 11:04:45.051798 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-jg4k7"] Nov 25 11:04:45 crc kubenswrapper[4776]: I1125 11:04:45.678976 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0da5ee40-85c5-4cf9-bf03-a1184d208541" path="/var/lib/kubelet/pods/0da5ee40-85c5-4cf9-bf03-a1184d208541/volumes" Nov 25 11:04:46 crc kubenswrapper[4776]: E1125 11:04:46.401686 4776 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.193:36072->38.102.83.193:36915: write tcp 38.102.83.193:36072->38.102.83.193:36915: write: broken pipe Nov 25 11:04:48 crc kubenswrapper[4776]: I1125 11:04:48.005828 4776 generic.go:334] "Generic (PLEG): container finished" podID="8df08eab-1375-42cd-b1dc-31a953e2bca6" containerID="217b34fc9df4ff20b62948a9319c19cf41a8a91e92f8b822729e7bc18ea823bd" exitCode=0 Nov 25 11:04:48 crc kubenswrapper[4776]: I1125 11:04:48.005921 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-7wxbq" event={"ID":"8df08eab-1375-42cd-b1dc-31a953e2bca6","Type":"ContainerDied","Data":"217b34fc9df4ff20b62948a9319c19cf41a8a91e92f8b822729e7bc18ea823bd"} Nov 25 11:04:48 crc kubenswrapper[4776]: I1125 11:04:48.662404 4776 scope.go:117] "RemoveContainer" containerID="8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698" Nov 25 11:04:48 crc kubenswrapper[4776]: E1125 11:04:48.662679 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:04:49 crc kubenswrapper[4776]: I1125 11:04:49.260544 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 11:04:49 crc kubenswrapper[4776]: I1125 11:04:49.262285 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 11:04:49 crc kubenswrapper[4776]: I1125 11:04:49.267220 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 11:04:49 crc kubenswrapper[4776]: I1125 11:04:49.272579 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 11:04:49 crc kubenswrapper[4776]: I1125 11:04:49.365637 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-7wxbq" Nov 25 11:04:49 crc kubenswrapper[4776]: I1125 11:04:49.459293 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8df08eab-1375-42cd-b1dc-31a953e2bca6-scripts\") pod \"8df08eab-1375-42cd-b1dc-31a953e2bca6\" (UID: \"8df08eab-1375-42cd-b1dc-31a953e2bca6\") " Nov 25 11:04:49 crc kubenswrapper[4776]: I1125 11:04:49.459418 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4587\" (UniqueName: \"kubernetes.io/projected/8df08eab-1375-42cd-b1dc-31a953e2bca6-kube-api-access-x4587\") pod \"8df08eab-1375-42cd-b1dc-31a953e2bca6\" (UID: \"8df08eab-1375-42cd-b1dc-31a953e2bca6\") " Nov 25 11:04:49 crc kubenswrapper[4776]: I1125 11:04:49.459601 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8df08eab-1375-42cd-b1dc-31a953e2bca6-etc-swift\") pod \"8df08eab-1375-42cd-b1dc-31a953e2bca6\" (UID: \"8df08eab-1375-42cd-b1dc-31a953e2bca6\") " Nov 25 11:04:49 crc kubenswrapper[4776]: I1125 11:04:49.459650 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8df08eab-1375-42cd-b1dc-31a953e2bca6-combined-ca-bundle\") pod \"8df08eab-1375-42cd-b1dc-31a953e2bca6\" (UID: \"8df08eab-1375-42cd-b1dc-31a953e2bca6\") " Nov 25 11:04:49 crc kubenswrapper[4776]: I1125 11:04:49.459677 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8df08eab-1375-42cd-b1dc-31a953e2bca6-swiftconf\") pod \"8df08eab-1375-42cd-b1dc-31a953e2bca6\" (UID: \"8df08eab-1375-42cd-b1dc-31a953e2bca6\") " Nov 25 11:04:49 crc kubenswrapper[4776]: I1125 11:04:49.459754 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8df08eab-1375-42cd-b1dc-31a953e2bca6-dispersionconf\") pod \"8df08eab-1375-42cd-b1dc-31a953e2bca6\" (UID: \"8df08eab-1375-42cd-b1dc-31a953e2bca6\") " Nov 25 11:04:49 crc kubenswrapper[4776]: I1125 11:04:49.459801 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8df08eab-1375-42cd-b1dc-31a953e2bca6-ring-data-devices\") pod \"8df08eab-1375-42cd-b1dc-31a953e2bca6\" (UID: \"8df08eab-1375-42cd-b1dc-31a953e2bca6\") " Nov 25 11:04:49 crc kubenswrapper[4776]: I1125 11:04:49.460657 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8df08eab-1375-42cd-b1dc-31a953e2bca6-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "8df08eab-1375-42cd-b1dc-31a953e2bca6" (UID: "8df08eab-1375-42cd-b1dc-31a953e2bca6"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:04:49 crc kubenswrapper[4776]: I1125 11:04:49.460960 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8df08eab-1375-42cd-b1dc-31a953e2bca6-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "8df08eab-1375-42cd-b1dc-31a953e2bca6" (UID: "8df08eab-1375-42cd-b1dc-31a953e2bca6"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:04:49 crc kubenswrapper[4776]: I1125 11:04:49.465252 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8df08eab-1375-42cd-b1dc-31a953e2bca6-kube-api-access-x4587" (OuterVolumeSpecName: "kube-api-access-x4587") pod "8df08eab-1375-42cd-b1dc-31a953e2bca6" (UID: "8df08eab-1375-42cd-b1dc-31a953e2bca6"). InnerVolumeSpecName "kube-api-access-x4587". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:04:49 crc kubenswrapper[4776]: I1125 11:04:49.484642 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8df08eab-1375-42cd-b1dc-31a953e2bca6-scripts" (OuterVolumeSpecName: "scripts") pod "8df08eab-1375-42cd-b1dc-31a953e2bca6" (UID: "8df08eab-1375-42cd-b1dc-31a953e2bca6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:04:49 crc kubenswrapper[4776]: I1125 11:04:49.486258 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8df08eab-1375-42cd-b1dc-31a953e2bca6-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "8df08eab-1375-42cd-b1dc-31a953e2bca6" (UID: "8df08eab-1375-42cd-b1dc-31a953e2bca6"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:04:49 crc kubenswrapper[4776]: I1125 11:04:49.487264 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8df08eab-1375-42cd-b1dc-31a953e2bca6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8df08eab-1375-42cd-b1dc-31a953e2bca6" (UID: "8df08eab-1375-42cd-b1dc-31a953e2bca6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:04:49 crc kubenswrapper[4776]: I1125 11:04:49.488148 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8df08eab-1375-42cd-b1dc-31a953e2bca6-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "8df08eab-1375-42cd-b1dc-31a953e2bca6" (UID: "8df08eab-1375-42cd-b1dc-31a953e2bca6"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:04:49 crc kubenswrapper[4776]: I1125 11:04:49.562014 4776 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8df08eab-1375-42cd-b1dc-31a953e2bca6-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:49 crc kubenswrapper[4776]: I1125 11:04:49.562105 4776 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8df08eab-1375-42cd-b1dc-31a953e2bca6-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:49 crc kubenswrapper[4776]: I1125 11:04:49.562119 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8df08eab-1375-42cd-b1dc-31a953e2bca6-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:49 crc kubenswrapper[4776]: I1125 11:04:49.562133 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4587\" (UniqueName: \"kubernetes.io/projected/8df08eab-1375-42cd-b1dc-31a953e2bca6-kube-api-access-x4587\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:49 crc kubenswrapper[4776]: I1125 11:04:49.562144 4776 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8df08eab-1375-42cd-b1dc-31a953e2bca6-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:49 crc kubenswrapper[4776]: I1125 11:04:49.562155 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8df08eab-1375-42cd-b1dc-31a953e2bca6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:49 crc kubenswrapper[4776]: I1125 11:04:49.562167 4776 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8df08eab-1375-42cd-b1dc-31a953e2bca6-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 25 11:04:50 crc kubenswrapper[4776]: I1125 11:04:50.026877 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-7wxbq" event={"ID":"8df08eab-1375-42cd-b1dc-31a953e2bca6","Type":"ContainerDied","Data":"a3fca680b356195aa5821fe1114d3dfcce0f3326997046a06c6307178429ab79"} Nov 25 11:04:50 crc kubenswrapper[4776]: I1125 11:04:50.026918 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3fca680b356195aa5821fe1114d3dfcce0f3326997046a06c6307178429ab79" Nov 25 11:04:50 crc kubenswrapper[4776]: I1125 11:04:50.027396 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-7wxbq" Nov 25 11:04:50 crc kubenswrapper[4776]: I1125 11:04:50.027426 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 11:04:50 crc kubenswrapper[4776]: I1125 11:04:50.034041 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 11:04:50 crc kubenswrapper[4776]: I1125 11:04:50.410719 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-r7nb6"] Nov 25 11:04:50 crc kubenswrapper[4776]: E1125 11:04:50.411257 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8df08eab-1375-42cd-b1dc-31a953e2bca6" containerName="swift-ring-rebalance" Nov 25 11:04:50 crc kubenswrapper[4776]: I1125 11:04:50.411274 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="8df08eab-1375-42cd-b1dc-31a953e2bca6" containerName="swift-ring-rebalance" Nov 25 11:04:50 crc kubenswrapper[4776]: I1125 11:04:50.411498 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="8df08eab-1375-42cd-b1dc-31a953e2bca6" containerName="swift-ring-rebalance" Nov 25 11:04:50 crc kubenswrapper[4776]: I1125 11:04:50.413131 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r7nb6" Nov 25 11:04:50 crc kubenswrapper[4776]: I1125 11:04:50.427814 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r7nb6"] Nov 25 11:04:50 crc kubenswrapper[4776]: I1125 11:04:50.487305 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwv2k\" (UniqueName: \"kubernetes.io/projected/59bd3761-a05d-41f5-95e6-8adc39cc936b-kube-api-access-bwv2k\") pod \"community-operators-r7nb6\" (UID: \"59bd3761-a05d-41f5-95e6-8adc39cc936b\") " pod="openshift-marketplace/community-operators-r7nb6" Nov 25 11:04:50 crc kubenswrapper[4776]: I1125 11:04:50.487395 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59bd3761-a05d-41f5-95e6-8adc39cc936b-utilities\") pod \"community-operators-r7nb6\" (UID: \"59bd3761-a05d-41f5-95e6-8adc39cc936b\") " pod="openshift-marketplace/community-operators-r7nb6" Nov 25 11:04:50 crc kubenswrapper[4776]: I1125 11:04:50.487685 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59bd3761-a05d-41f5-95e6-8adc39cc936b-catalog-content\") pod \"community-operators-r7nb6\" (UID: \"59bd3761-a05d-41f5-95e6-8adc39cc936b\") " pod="openshift-marketplace/community-operators-r7nb6" Nov 25 11:04:50 crc kubenswrapper[4776]: I1125 11:04:50.590020 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwv2k\" (UniqueName: \"kubernetes.io/projected/59bd3761-a05d-41f5-95e6-8adc39cc936b-kube-api-access-bwv2k\") pod \"community-operators-r7nb6\" (UID: \"59bd3761-a05d-41f5-95e6-8adc39cc936b\") " pod="openshift-marketplace/community-operators-r7nb6" Nov 25 11:04:50 crc kubenswrapper[4776]: I1125 11:04:50.590081 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59bd3761-a05d-41f5-95e6-8adc39cc936b-utilities\") pod \"community-operators-r7nb6\" (UID: \"59bd3761-a05d-41f5-95e6-8adc39cc936b\") " pod="openshift-marketplace/community-operators-r7nb6" Nov 25 11:04:50 crc kubenswrapper[4776]: I1125 11:04:50.590171 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59bd3761-a05d-41f5-95e6-8adc39cc936b-catalog-content\") pod \"community-operators-r7nb6\" (UID: \"59bd3761-a05d-41f5-95e6-8adc39cc936b\") " pod="openshift-marketplace/community-operators-r7nb6" Nov 25 11:04:50 crc kubenswrapper[4776]: I1125 11:04:50.590593 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59bd3761-a05d-41f5-95e6-8adc39cc936b-catalog-content\") pod \"community-operators-r7nb6\" (UID: \"59bd3761-a05d-41f5-95e6-8adc39cc936b\") " pod="openshift-marketplace/community-operators-r7nb6" Nov 25 11:04:50 crc kubenswrapper[4776]: I1125 11:04:50.590711 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59bd3761-a05d-41f5-95e6-8adc39cc936b-utilities\") pod \"community-operators-r7nb6\" (UID: \"59bd3761-a05d-41f5-95e6-8adc39cc936b\") " pod="openshift-marketplace/community-operators-r7nb6" Nov 25 11:04:50 crc kubenswrapper[4776]: I1125 11:04:50.615948 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwv2k\" (UniqueName: \"kubernetes.io/projected/59bd3761-a05d-41f5-95e6-8adc39cc936b-kube-api-access-bwv2k\") pod \"community-operators-r7nb6\" (UID: \"59bd3761-a05d-41f5-95e6-8adc39cc936b\") " pod="openshift-marketplace/community-operators-r7nb6" Nov 25 11:04:50 crc kubenswrapper[4776]: I1125 11:04:50.735743 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r7nb6" Nov 25 11:04:51 crc kubenswrapper[4776]: I1125 11:04:51.301716 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r7nb6"] Nov 25 11:04:52 crc kubenswrapper[4776]: I1125 11:04:52.048334 4776 generic.go:334] "Generic (PLEG): container finished" podID="59bd3761-a05d-41f5-95e6-8adc39cc936b" containerID="6d0d742f631b00c937940addc634ec8a8c45bf83772b9d578c46848f0c90e42e" exitCode=0 Nov 25 11:04:52 crc kubenswrapper[4776]: I1125 11:04:52.048395 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r7nb6" event={"ID":"59bd3761-a05d-41f5-95e6-8adc39cc936b","Type":"ContainerDied","Data":"6d0d742f631b00c937940addc634ec8a8c45bf83772b9d578c46848f0c90e42e"} Nov 25 11:04:52 crc kubenswrapper[4776]: I1125 11:04:52.048659 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r7nb6" event={"ID":"59bd3761-a05d-41f5-95e6-8adc39cc936b","Type":"ContainerStarted","Data":"a896bf8bb8046fdd4dba2266ed21c249ec16b627292465525640476e30064bf6"} Nov 25 11:04:52 crc kubenswrapper[4776]: I1125 11:04:52.051690 4776 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 11:04:54 crc kubenswrapper[4776]: I1125 11:04:54.068605 4776 generic.go:334] "Generic (PLEG): container finished" podID="59bd3761-a05d-41f5-95e6-8adc39cc936b" containerID="04b32211e0b3a697e78b717f3710b206c993d5976938176a8067e87f21db5051" exitCode=0 Nov 25 11:04:54 crc kubenswrapper[4776]: I1125 11:04:54.068848 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r7nb6" event={"ID":"59bd3761-a05d-41f5-95e6-8adc39cc936b","Type":"ContainerDied","Data":"04b32211e0b3a697e78b717f3710b206c993d5976938176a8067e87f21db5051"} Nov 25 11:04:55 crc kubenswrapper[4776]: I1125 11:04:55.082039 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r7nb6" event={"ID":"59bd3761-a05d-41f5-95e6-8adc39cc936b","Type":"ContainerStarted","Data":"f2505b03b7740082b21306bb9234c17bf471cd5742608e50ae098c10739fe4fc"} Nov 25 11:04:55 crc kubenswrapper[4776]: I1125 11:04:55.106799 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-r7nb6" podStartSLOduration=2.444593594 podStartE2EDuration="5.106780745s" podCreationTimestamp="2025-11-25 11:04:50 +0000 UTC" firstStartedPulling="2025-11-25 11:04:52.051490568 +0000 UTC m=+6037.092550121" lastFinishedPulling="2025-11-25 11:04:54.713677719 +0000 UTC m=+6039.754737272" observedRunningTime="2025-11-25 11:04:55.102923398 +0000 UTC m=+6040.143982951" watchObservedRunningTime="2025-11-25 11:04:55.106780745 +0000 UTC m=+6040.147840298" Nov 25 11:05:00 crc kubenswrapper[4776]: I1125 11:05:00.045873 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-wknmm"] Nov 25 11:05:00 crc kubenswrapper[4776]: I1125 11:05:00.054824 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-wknmm"] Nov 25 11:05:00 crc kubenswrapper[4776]: I1125 11:05:00.735958 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-r7nb6" Nov 25 11:05:00 crc kubenswrapper[4776]: I1125 11:05:00.736020 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-r7nb6" Nov 25 11:05:00 crc kubenswrapper[4776]: I1125 11:05:00.788317 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-r7nb6" Nov 25 11:05:01 crc kubenswrapper[4776]: I1125 11:05:01.182442 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-r7nb6" Nov 25 11:05:01 crc kubenswrapper[4776]: I1125 11:05:01.238924 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r7nb6"] Nov 25 11:05:01 crc kubenswrapper[4776]: I1125 11:05:01.671412 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a8513c8-6195-4ffe-8765-d3554db89257" path="/var/lib/kubelet/pods/6a8513c8-6195-4ffe-8765-d3554db89257/volumes" Nov 25 11:05:03 crc kubenswrapper[4776]: I1125 11:05:03.153669 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-r7nb6" podUID="59bd3761-a05d-41f5-95e6-8adc39cc936b" containerName="registry-server" containerID="cri-o://f2505b03b7740082b21306bb9234c17bf471cd5742608e50ae098c10739fe4fc" gracePeriod=2 Nov 25 11:05:03 crc kubenswrapper[4776]: I1125 11:05:03.663378 4776 scope.go:117] "RemoveContainer" containerID="8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698" Nov 25 11:05:03 crc kubenswrapper[4776]: E1125 11:05:03.663819 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:05:03 crc kubenswrapper[4776]: I1125 11:05:03.735022 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r7nb6" Nov 25 11:05:03 crc kubenswrapper[4776]: I1125 11:05:03.840556 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59bd3761-a05d-41f5-95e6-8adc39cc936b-utilities\") pod \"59bd3761-a05d-41f5-95e6-8adc39cc936b\" (UID: \"59bd3761-a05d-41f5-95e6-8adc39cc936b\") " Nov 25 11:05:03 crc kubenswrapper[4776]: I1125 11:05:03.840685 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwv2k\" (UniqueName: \"kubernetes.io/projected/59bd3761-a05d-41f5-95e6-8adc39cc936b-kube-api-access-bwv2k\") pod \"59bd3761-a05d-41f5-95e6-8adc39cc936b\" (UID: \"59bd3761-a05d-41f5-95e6-8adc39cc936b\") " Nov 25 11:05:03 crc kubenswrapper[4776]: I1125 11:05:03.840728 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59bd3761-a05d-41f5-95e6-8adc39cc936b-catalog-content\") pod \"59bd3761-a05d-41f5-95e6-8adc39cc936b\" (UID: \"59bd3761-a05d-41f5-95e6-8adc39cc936b\") " Nov 25 11:05:03 crc kubenswrapper[4776]: I1125 11:05:03.842699 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59bd3761-a05d-41f5-95e6-8adc39cc936b-utilities" (OuterVolumeSpecName: "utilities") pod "59bd3761-a05d-41f5-95e6-8adc39cc936b" (UID: "59bd3761-a05d-41f5-95e6-8adc39cc936b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:05:03 crc kubenswrapper[4776]: I1125 11:05:03.853189 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59bd3761-a05d-41f5-95e6-8adc39cc936b-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:03 crc kubenswrapper[4776]: I1125 11:05:03.867713 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59bd3761-a05d-41f5-95e6-8adc39cc936b-kube-api-access-bwv2k" (OuterVolumeSpecName: "kube-api-access-bwv2k") pod "59bd3761-a05d-41f5-95e6-8adc39cc936b" (UID: "59bd3761-a05d-41f5-95e6-8adc39cc936b"). InnerVolumeSpecName "kube-api-access-bwv2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:05:03 crc kubenswrapper[4776]: I1125 11:05:03.898738 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59bd3761-a05d-41f5-95e6-8adc39cc936b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "59bd3761-a05d-41f5-95e6-8adc39cc936b" (UID: "59bd3761-a05d-41f5-95e6-8adc39cc936b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:05:03 crc kubenswrapper[4776]: I1125 11:05:03.954306 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwv2k\" (UniqueName: \"kubernetes.io/projected/59bd3761-a05d-41f5-95e6-8adc39cc936b-kube-api-access-bwv2k\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:03 crc kubenswrapper[4776]: I1125 11:05:03.954348 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59bd3761-a05d-41f5-95e6-8adc39cc936b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:04 crc kubenswrapper[4776]: I1125 11:05:04.168958 4776 generic.go:334] "Generic (PLEG): container finished" podID="59bd3761-a05d-41f5-95e6-8adc39cc936b" containerID="f2505b03b7740082b21306bb9234c17bf471cd5742608e50ae098c10739fe4fc" exitCode=0 Nov 25 11:05:04 crc kubenswrapper[4776]: I1125 11:05:04.169002 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r7nb6" event={"ID":"59bd3761-a05d-41f5-95e6-8adc39cc936b","Type":"ContainerDied","Data":"f2505b03b7740082b21306bb9234c17bf471cd5742608e50ae098c10739fe4fc"} Nov 25 11:05:04 crc kubenswrapper[4776]: I1125 11:05:04.169044 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r7nb6" event={"ID":"59bd3761-a05d-41f5-95e6-8adc39cc936b","Type":"ContainerDied","Data":"a896bf8bb8046fdd4dba2266ed21c249ec16b627292465525640476e30064bf6"} Nov 25 11:05:04 crc kubenswrapper[4776]: I1125 11:05:04.169079 4776 scope.go:117] "RemoveContainer" containerID="f2505b03b7740082b21306bb9234c17bf471cd5742608e50ae098c10739fe4fc" Nov 25 11:05:04 crc kubenswrapper[4776]: I1125 11:05:04.169115 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r7nb6" Nov 25 11:05:04 crc kubenswrapper[4776]: I1125 11:05:04.192457 4776 scope.go:117] "RemoveContainer" containerID="04b32211e0b3a697e78b717f3710b206c993d5976938176a8067e87f21db5051" Nov 25 11:05:04 crc kubenswrapper[4776]: I1125 11:05:04.217759 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r7nb6"] Nov 25 11:05:04 crc kubenswrapper[4776]: I1125 11:05:04.224468 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-r7nb6"] Nov 25 11:05:04 crc kubenswrapper[4776]: I1125 11:05:04.241621 4776 scope.go:117] "RemoveContainer" containerID="6d0d742f631b00c937940addc634ec8a8c45bf83772b9d578c46848f0c90e42e" Nov 25 11:05:04 crc kubenswrapper[4776]: I1125 11:05:04.264512 4776 scope.go:117] "RemoveContainer" containerID="f2505b03b7740082b21306bb9234c17bf471cd5742608e50ae098c10739fe4fc" Nov 25 11:05:04 crc kubenswrapper[4776]: E1125 11:05:04.264997 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2505b03b7740082b21306bb9234c17bf471cd5742608e50ae098c10739fe4fc\": container with ID starting with f2505b03b7740082b21306bb9234c17bf471cd5742608e50ae098c10739fe4fc not found: ID does not exist" containerID="f2505b03b7740082b21306bb9234c17bf471cd5742608e50ae098c10739fe4fc" Nov 25 11:05:04 crc kubenswrapper[4776]: I1125 11:05:04.265049 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2505b03b7740082b21306bb9234c17bf471cd5742608e50ae098c10739fe4fc"} err="failed to get container status \"f2505b03b7740082b21306bb9234c17bf471cd5742608e50ae098c10739fe4fc\": rpc error: code = NotFound desc = could not find container \"f2505b03b7740082b21306bb9234c17bf471cd5742608e50ae098c10739fe4fc\": container with ID starting with f2505b03b7740082b21306bb9234c17bf471cd5742608e50ae098c10739fe4fc not found: ID does not exist" Nov 25 11:05:04 crc kubenswrapper[4776]: I1125 11:05:04.265100 4776 scope.go:117] "RemoveContainer" containerID="04b32211e0b3a697e78b717f3710b206c993d5976938176a8067e87f21db5051" Nov 25 11:05:04 crc kubenswrapper[4776]: E1125 11:05:04.265647 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04b32211e0b3a697e78b717f3710b206c993d5976938176a8067e87f21db5051\": container with ID starting with 04b32211e0b3a697e78b717f3710b206c993d5976938176a8067e87f21db5051 not found: ID does not exist" containerID="04b32211e0b3a697e78b717f3710b206c993d5976938176a8067e87f21db5051" Nov 25 11:05:04 crc kubenswrapper[4776]: I1125 11:05:04.265679 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04b32211e0b3a697e78b717f3710b206c993d5976938176a8067e87f21db5051"} err="failed to get container status \"04b32211e0b3a697e78b717f3710b206c993d5976938176a8067e87f21db5051\": rpc error: code = NotFound desc = could not find container \"04b32211e0b3a697e78b717f3710b206c993d5976938176a8067e87f21db5051\": container with ID starting with 04b32211e0b3a697e78b717f3710b206c993d5976938176a8067e87f21db5051 not found: ID does not exist" Nov 25 11:05:04 crc kubenswrapper[4776]: I1125 11:05:04.265698 4776 scope.go:117] "RemoveContainer" containerID="6d0d742f631b00c937940addc634ec8a8c45bf83772b9d578c46848f0c90e42e" Nov 25 11:05:04 crc kubenswrapper[4776]: E1125 11:05:04.266121 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d0d742f631b00c937940addc634ec8a8c45bf83772b9d578c46848f0c90e42e\": container with ID starting with 6d0d742f631b00c937940addc634ec8a8c45bf83772b9d578c46848f0c90e42e not found: ID does not exist" containerID="6d0d742f631b00c937940addc634ec8a8c45bf83772b9d578c46848f0c90e42e" Nov 25 11:05:04 crc kubenswrapper[4776]: I1125 11:05:04.266174 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d0d742f631b00c937940addc634ec8a8c45bf83772b9d578c46848f0c90e42e"} err="failed to get container status \"6d0d742f631b00c937940addc634ec8a8c45bf83772b9d578c46848f0c90e42e\": rpc error: code = NotFound desc = could not find container \"6d0d742f631b00c937940addc634ec8a8c45bf83772b9d578c46848f0c90e42e\": container with ID starting with 6d0d742f631b00c937940addc634ec8a8c45bf83772b9d578c46848f0c90e42e not found: ID does not exist" Nov 25 11:05:05 crc kubenswrapper[4776]: I1125 11:05:05.677992 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59bd3761-a05d-41f5-95e6-8adc39cc936b" path="/var/lib/kubelet/pods/59bd3761-a05d-41f5-95e6-8adc39cc936b/volumes" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.461736 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-85l79"] Nov 25 11:05:13 crc kubenswrapper[4776]: E1125 11:05:13.462753 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59bd3761-a05d-41f5-95e6-8adc39cc936b" containerName="registry-server" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.462775 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="59bd3761-a05d-41f5-95e6-8adc39cc936b" containerName="registry-server" Nov 25 11:05:13 crc kubenswrapper[4776]: E1125 11:05:13.462812 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59bd3761-a05d-41f5-95e6-8adc39cc936b" containerName="extract-utilities" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.462820 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="59bd3761-a05d-41f5-95e6-8adc39cc936b" containerName="extract-utilities" Nov 25 11:05:13 crc kubenswrapper[4776]: E1125 11:05:13.462847 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59bd3761-a05d-41f5-95e6-8adc39cc936b" containerName="extract-content" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.462855 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="59bd3761-a05d-41f5-95e6-8adc39cc936b" containerName="extract-content" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.463120 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="59bd3761-a05d-41f5-95e6-8adc39cc936b" containerName="registry-server" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.463964 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-85l79" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.473525 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-wlphl"] Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.475914 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-wlphl" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.482819 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.483107 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-bgj92" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.483447 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.494182 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-85l79"] Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.505715 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-wlphl"] Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.646674 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnssv\" (UniqueName: \"kubernetes.io/projected/a6e2d225-3db6-4189-a956-fd709c242387-kube-api-access-jnssv\") pod \"ovn-controller-85l79\" (UID: \"a6e2d225-3db6-4189-a956-fd709c242387\") " pod="openstack/ovn-controller-85l79" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.646755 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/845d70c5-c0ad-4143-8778-3bccb4724aa4-var-log\") pod \"ovn-controller-ovs-wlphl\" (UID: \"845d70c5-c0ad-4143-8778-3bccb4724aa4\") " pod="openstack/ovn-controller-ovs-wlphl" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.646785 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a6e2d225-3db6-4189-a956-fd709c242387-var-run\") pod \"ovn-controller-85l79\" (UID: \"a6e2d225-3db6-4189-a956-fd709c242387\") " pod="openstack/ovn-controller-85l79" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.646805 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rph2c\" (UniqueName: \"kubernetes.io/projected/845d70c5-c0ad-4143-8778-3bccb4724aa4-kube-api-access-rph2c\") pod \"ovn-controller-ovs-wlphl\" (UID: \"845d70c5-c0ad-4143-8778-3bccb4724aa4\") " pod="openstack/ovn-controller-ovs-wlphl" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.646868 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/845d70c5-c0ad-4143-8778-3bccb4724aa4-etc-ovs\") pod \"ovn-controller-ovs-wlphl\" (UID: \"845d70c5-c0ad-4143-8778-3bccb4724aa4\") " pod="openstack/ovn-controller-ovs-wlphl" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.646896 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/845d70c5-c0ad-4143-8778-3bccb4724aa4-var-run\") pod \"ovn-controller-ovs-wlphl\" (UID: \"845d70c5-c0ad-4143-8778-3bccb4724aa4\") " pod="openstack/ovn-controller-ovs-wlphl" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.646915 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6e2d225-3db6-4189-a956-fd709c242387-combined-ca-bundle\") pod \"ovn-controller-85l79\" (UID: \"a6e2d225-3db6-4189-a956-fd709c242387\") " pod="openstack/ovn-controller-85l79" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.646938 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6e2d225-3db6-4189-a956-fd709c242387-scripts\") pod \"ovn-controller-85l79\" (UID: \"a6e2d225-3db6-4189-a956-fd709c242387\") " pod="openstack/ovn-controller-85l79" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.646956 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a6e2d225-3db6-4189-a956-fd709c242387-var-log-ovn\") pod \"ovn-controller-85l79\" (UID: \"a6e2d225-3db6-4189-a956-fd709c242387\") " pod="openstack/ovn-controller-85l79" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.646990 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6e2d225-3db6-4189-a956-fd709c242387-ovn-controller-tls-certs\") pod \"ovn-controller-85l79\" (UID: \"a6e2d225-3db6-4189-a956-fd709c242387\") " pod="openstack/ovn-controller-85l79" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.647015 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/845d70c5-c0ad-4143-8778-3bccb4724aa4-var-lib\") pod \"ovn-controller-ovs-wlphl\" (UID: \"845d70c5-c0ad-4143-8778-3bccb4724aa4\") " pod="openstack/ovn-controller-ovs-wlphl" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.647041 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a6e2d225-3db6-4189-a956-fd709c242387-var-run-ovn\") pod \"ovn-controller-85l79\" (UID: \"a6e2d225-3db6-4189-a956-fd709c242387\") " pod="openstack/ovn-controller-85l79" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.647180 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/845d70c5-c0ad-4143-8778-3bccb4724aa4-scripts\") pod \"ovn-controller-ovs-wlphl\" (UID: \"845d70c5-c0ad-4143-8778-3bccb4724aa4\") " pod="openstack/ovn-controller-ovs-wlphl" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.749043 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/845d70c5-c0ad-4143-8778-3bccb4724aa4-var-log\") pod \"ovn-controller-ovs-wlphl\" (UID: \"845d70c5-c0ad-4143-8778-3bccb4724aa4\") " pod="openstack/ovn-controller-ovs-wlphl" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.749146 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a6e2d225-3db6-4189-a956-fd709c242387-var-run\") pod \"ovn-controller-85l79\" (UID: \"a6e2d225-3db6-4189-a956-fd709c242387\") " pod="openstack/ovn-controller-85l79" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.749182 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rph2c\" (UniqueName: \"kubernetes.io/projected/845d70c5-c0ad-4143-8778-3bccb4724aa4-kube-api-access-rph2c\") pod \"ovn-controller-ovs-wlphl\" (UID: \"845d70c5-c0ad-4143-8778-3bccb4724aa4\") " pod="openstack/ovn-controller-ovs-wlphl" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.749228 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/845d70c5-c0ad-4143-8778-3bccb4724aa4-etc-ovs\") pod \"ovn-controller-ovs-wlphl\" (UID: \"845d70c5-c0ad-4143-8778-3bccb4724aa4\") " pod="openstack/ovn-controller-ovs-wlphl" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.749269 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/845d70c5-c0ad-4143-8778-3bccb4724aa4-var-run\") pod \"ovn-controller-ovs-wlphl\" (UID: \"845d70c5-c0ad-4143-8778-3bccb4724aa4\") " pod="openstack/ovn-controller-ovs-wlphl" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.749293 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6e2d225-3db6-4189-a956-fd709c242387-combined-ca-bundle\") pod \"ovn-controller-85l79\" (UID: \"a6e2d225-3db6-4189-a956-fd709c242387\") " pod="openstack/ovn-controller-85l79" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.749323 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6e2d225-3db6-4189-a956-fd709c242387-scripts\") pod \"ovn-controller-85l79\" (UID: \"a6e2d225-3db6-4189-a956-fd709c242387\") " pod="openstack/ovn-controller-85l79" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.749343 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a6e2d225-3db6-4189-a956-fd709c242387-var-log-ovn\") pod \"ovn-controller-85l79\" (UID: \"a6e2d225-3db6-4189-a956-fd709c242387\") " pod="openstack/ovn-controller-85l79" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.749384 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6e2d225-3db6-4189-a956-fd709c242387-ovn-controller-tls-certs\") pod \"ovn-controller-85l79\" (UID: \"a6e2d225-3db6-4189-a956-fd709c242387\") " pod="openstack/ovn-controller-85l79" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.749407 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/845d70c5-c0ad-4143-8778-3bccb4724aa4-var-lib\") pod \"ovn-controller-ovs-wlphl\" (UID: \"845d70c5-c0ad-4143-8778-3bccb4724aa4\") " pod="openstack/ovn-controller-ovs-wlphl" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.749442 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a6e2d225-3db6-4189-a956-fd709c242387-var-run-ovn\") pod \"ovn-controller-85l79\" (UID: \"a6e2d225-3db6-4189-a956-fd709c242387\") " pod="openstack/ovn-controller-85l79" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.749459 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/845d70c5-c0ad-4143-8778-3bccb4724aa4-var-log\") pod \"ovn-controller-ovs-wlphl\" (UID: \"845d70c5-c0ad-4143-8778-3bccb4724aa4\") " pod="openstack/ovn-controller-ovs-wlphl" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.749478 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a6e2d225-3db6-4189-a956-fd709c242387-var-run\") pod \"ovn-controller-85l79\" (UID: \"a6e2d225-3db6-4189-a956-fd709c242387\") " pod="openstack/ovn-controller-85l79" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.749470 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/845d70c5-c0ad-4143-8778-3bccb4724aa4-scripts\") pod \"ovn-controller-ovs-wlphl\" (UID: \"845d70c5-c0ad-4143-8778-3bccb4724aa4\") " pod="openstack/ovn-controller-ovs-wlphl" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.749650 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnssv\" (UniqueName: \"kubernetes.io/projected/a6e2d225-3db6-4189-a956-fd709c242387-kube-api-access-jnssv\") pod \"ovn-controller-85l79\" (UID: \"a6e2d225-3db6-4189-a956-fd709c242387\") " pod="openstack/ovn-controller-85l79" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.750143 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/845d70c5-c0ad-4143-8778-3bccb4724aa4-etc-ovs\") pod \"ovn-controller-ovs-wlphl\" (UID: \"845d70c5-c0ad-4143-8778-3bccb4724aa4\") " pod="openstack/ovn-controller-ovs-wlphl" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.750327 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/845d70c5-c0ad-4143-8778-3bccb4724aa4-var-run\") pod \"ovn-controller-ovs-wlphl\" (UID: \"845d70c5-c0ad-4143-8778-3bccb4724aa4\") " pod="openstack/ovn-controller-ovs-wlphl" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.751164 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/845d70c5-c0ad-4143-8778-3bccb4724aa4-var-lib\") pod \"ovn-controller-ovs-wlphl\" (UID: \"845d70c5-c0ad-4143-8778-3bccb4724aa4\") " pod="openstack/ovn-controller-ovs-wlphl" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.751236 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a6e2d225-3db6-4189-a956-fd709c242387-var-run-ovn\") pod \"ovn-controller-85l79\" (UID: \"a6e2d225-3db6-4189-a956-fd709c242387\") " pod="openstack/ovn-controller-85l79" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.751375 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a6e2d225-3db6-4189-a956-fd709c242387-var-log-ovn\") pod \"ovn-controller-85l79\" (UID: \"a6e2d225-3db6-4189-a956-fd709c242387\") " pod="openstack/ovn-controller-85l79" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.752018 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6e2d225-3db6-4189-a956-fd709c242387-scripts\") pod \"ovn-controller-85l79\" (UID: \"a6e2d225-3db6-4189-a956-fd709c242387\") " pod="openstack/ovn-controller-85l79" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.752024 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/845d70c5-c0ad-4143-8778-3bccb4724aa4-scripts\") pod \"ovn-controller-ovs-wlphl\" (UID: \"845d70c5-c0ad-4143-8778-3bccb4724aa4\") " pod="openstack/ovn-controller-ovs-wlphl" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.756679 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6e2d225-3db6-4189-a956-fd709c242387-ovn-controller-tls-certs\") pod \"ovn-controller-85l79\" (UID: \"a6e2d225-3db6-4189-a956-fd709c242387\") " pod="openstack/ovn-controller-85l79" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.757166 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6e2d225-3db6-4189-a956-fd709c242387-combined-ca-bundle\") pod \"ovn-controller-85l79\" (UID: \"a6e2d225-3db6-4189-a956-fd709c242387\") " pod="openstack/ovn-controller-85l79" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.775600 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rph2c\" (UniqueName: \"kubernetes.io/projected/845d70c5-c0ad-4143-8778-3bccb4724aa4-kube-api-access-rph2c\") pod \"ovn-controller-ovs-wlphl\" (UID: \"845d70c5-c0ad-4143-8778-3bccb4724aa4\") " pod="openstack/ovn-controller-ovs-wlphl" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.776852 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnssv\" (UniqueName: \"kubernetes.io/projected/a6e2d225-3db6-4189-a956-fd709c242387-kube-api-access-jnssv\") pod \"ovn-controller-85l79\" (UID: \"a6e2d225-3db6-4189-a956-fd709c242387\") " pod="openstack/ovn-controller-85l79" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.792091 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-85l79" Nov 25 11:05:13 crc kubenswrapper[4776]: I1125 11:05:13.806676 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-wlphl" Nov 25 11:05:14 crc kubenswrapper[4776]: I1125 11:05:14.305630 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-85l79"] Nov 25 11:05:14 crc kubenswrapper[4776]: I1125 11:05:14.720597 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-wlphl"] Nov 25 11:05:14 crc kubenswrapper[4776]: W1125 11:05:14.730846 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod845d70c5_c0ad_4143_8778_3bccb4724aa4.slice/crio-a6b2c32acb8a94fc8c21743049bac59993866ed930c3e6fd22aa99962ec49449 WatchSource:0}: Error finding container a6b2c32acb8a94fc8c21743049bac59993866ed930c3e6fd22aa99962ec49449: Status 404 returned error can't find the container with id a6b2c32acb8a94fc8c21743049bac59993866ed930c3e6fd22aa99962ec49449 Nov 25 11:05:14 crc kubenswrapper[4776]: I1125 11:05:14.964529 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-jgqps"] Nov 25 11:05:14 crc kubenswrapper[4776]: I1125 11:05:14.966224 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-jgqps" Nov 25 11:05:14 crc kubenswrapper[4776]: I1125 11:05:14.968951 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 25 11:05:14 crc kubenswrapper[4776]: I1125 11:05:14.986302 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6-ovs-rundir\") pod \"ovn-controller-metrics-jgqps\" (UID: \"5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6\") " pod="openstack/ovn-controller-metrics-jgqps" Nov 25 11:05:14 crc kubenswrapper[4776]: I1125 11:05:14.986376 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-jgqps\" (UID: \"5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6\") " pod="openstack/ovn-controller-metrics-jgqps" Nov 25 11:05:14 crc kubenswrapper[4776]: I1125 11:05:14.986400 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6-ovn-rundir\") pod \"ovn-controller-metrics-jgqps\" (UID: \"5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6\") " pod="openstack/ovn-controller-metrics-jgqps" Nov 25 11:05:14 crc kubenswrapper[4776]: I1125 11:05:14.986478 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fz9md\" (UniqueName: \"kubernetes.io/projected/5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6-kube-api-access-fz9md\") pod \"ovn-controller-metrics-jgqps\" (UID: \"5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6\") " pod="openstack/ovn-controller-metrics-jgqps" Nov 25 11:05:14 crc kubenswrapper[4776]: I1125 11:05:14.986554 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6-combined-ca-bundle\") pod \"ovn-controller-metrics-jgqps\" (UID: \"5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6\") " pod="openstack/ovn-controller-metrics-jgqps" Nov 25 11:05:14 crc kubenswrapper[4776]: I1125 11:05:14.986575 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6-config\") pod \"ovn-controller-metrics-jgqps\" (UID: \"5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6\") " pod="openstack/ovn-controller-metrics-jgqps" Nov 25 11:05:14 crc kubenswrapper[4776]: I1125 11:05:14.993316 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-jgqps"] Nov 25 11:05:15 crc kubenswrapper[4776]: I1125 11:05:15.088195 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6-combined-ca-bundle\") pod \"ovn-controller-metrics-jgqps\" (UID: \"5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6\") " pod="openstack/ovn-controller-metrics-jgqps" Nov 25 11:05:15 crc kubenswrapper[4776]: I1125 11:05:15.088265 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6-config\") pod \"ovn-controller-metrics-jgqps\" (UID: \"5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6\") " pod="openstack/ovn-controller-metrics-jgqps" Nov 25 11:05:15 crc kubenswrapper[4776]: I1125 11:05:15.088379 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6-ovs-rundir\") pod \"ovn-controller-metrics-jgqps\" (UID: \"5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6\") " pod="openstack/ovn-controller-metrics-jgqps" Nov 25 11:05:15 crc kubenswrapper[4776]: I1125 11:05:15.088418 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-jgqps\" (UID: \"5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6\") " pod="openstack/ovn-controller-metrics-jgqps" Nov 25 11:05:15 crc kubenswrapper[4776]: I1125 11:05:15.088446 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6-ovn-rundir\") pod \"ovn-controller-metrics-jgqps\" (UID: \"5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6\") " pod="openstack/ovn-controller-metrics-jgqps" Nov 25 11:05:15 crc kubenswrapper[4776]: I1125 11:05:15.088535 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fz9md\" (UniqueName: \"kubernetes.io/projected/5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6-kube-api-access-fz9md\") pod \"ovn-controller-metrics-jgqps\" (UID: \"5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6\") " pod="openstack/ovn-controller-metrics-jgqps" Nov 25 11:05:15 crc kubenswrapper[4776]: I1125 11:05:15.089498 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6-ovs-rundir\") pod \"ovn-controller-metrics-jgqps\" (UID: \"5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6\") " pod="openstack/ovn-controller-metrics-jgqps" Nov 25 11:05:15 crc kubenswrapper[4776]: I1125 11:05:15.089491 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6-ovn-rundir\") pod \"ovn-controller-metrics-jgqps\" (UID: \"5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6\") " pod="openstack/ovn-controller-metrics-jgqps" Nov 25 11:05:15 crc kubenswrapper[4776]: I1125 11:05:15.089885 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6-config\") pod \"ovn-controller-metrics-jgqps\" (UID: \"5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6\") " pod="openstack/ovn-controller-metrics-jgqps" Nov 25 11:05:15 crc kubenswrapper[4776]: I1125 11:05:15.096250 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-jgqps\" (UID: \"5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6\") " pod="openstack/ovn-controller-metrics-jgqps" Nov 25 11:05:15 crc kubenswrapper[4776]: I1125 11:05:15.096990 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6-combined-ca-bundle\") pod \"ovn-controller-metrics-jgqps\" (UID: \"5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6\") " pod="openstack/ovn-controller-metrics-jgqps" Nov 25 11:05:15 crc kubenswrapper[4776]: I1125 11:05:15.121449 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fz9md\" (UniqueName: \"kubernetes.io/projected/5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6-kube-api-access-fz9md\") pod \"ovn-controller-metrics-jgqps\" (UID: \"5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6\") " pod="openstack/ovn-controller-metrics-jgqps" Nov 25 11:05:15 crc kubenswrapper[4776]: I1125 11:05:15.270086 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-85l79" event={"ID":"a6e2d225-3db6-4189-a956-fd709c242387","Type":"ContainerStarted","Data":"100c8baf1be56560e21f6a94f724d72088dc744647c303d14153c3dbf786d3aa"} Nov 25 11:05:15 crc kubenswrapper[4776]: I1125 11:05:15.270132 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-85l79" event={"ID":"a6e2d225-3db6-4189-a956-fd709c242387","Type":"ContainerStarted","Data":"ef65f8f93add88f7aa5e3d0625c6e1957140f16e1a8f30b30e70d115fdc5ad8e"} Nov 25 11:05:15 crc kubenswrapper[4776]: I1125 11:05:15.270477 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-85l79" Nov 25 11:05:15 crc kubenswrapper[4776]: I1125 11:05:15.274188 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-wlphl" event={"ID":"845d70c5-c0ad-4143-8778-3bccb4724aa4","Type":"ContainerStarted","Data":"b88a23cc7996228d50b5eb43124e279401023dece4ba70278741db418e95a198"} Nov 25 11:05:15 crc kubenswrapper[4776]: I1125 11:05:15.274256 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-wlphl" event={"ID":"845d70c5-c0ad-4143-8778-3bccb4724aa4","Type":"ContainerStarted","Data":"a6b2c32acb8a94fc8c21743049bac59993866ed930c3e6fd22aa99962ec49449"} Nov 25 11:05:15 crc kubenswrapper[4776]: I1125 11:05:15.300985 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-85l79" podStartSLOduration=2.300964011 podStartE2EDuration="2.300964011s" podCreationTimestamp="2025-11-25 11:05:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:05:15.2905726 +0000 UTC m=+6060.331632153" watchObservedRunningTime="2025-11-25 11:05:15.300964011 +0000 UTC m=+6060.342023564" Nov 25 11:05:15 crc kubenswrapper[4776]: I1125 11:05:15.317955 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-jgqps" Nov 25 11:05:15 crc kubenswrapper[4776]: I1125 11:05:15.798376 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-jgqps"] Nov 25 11:05:15 crc kubenswrapper[4776]: W1125 11:05:15.799576 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5dcfc2e2_07bd_4f08_b20f_604e1a6a49e6.slice/crio-2a391a47e978b7cd6d8ca4b4ce3e4442617226d26a1894ea1ba2a23f348e93b6 WatchSource:0}: Error finding container 2a391a47e978b7cd6d8ca4b4ce3e4442617226d26a1894ea1ba2a23f348e93b6: Status 404 returned error can't find the container with id 2a391a47e978b7cd6d8ca4b4ce3e4442617226d26a1894ea1ba2a23f348e93b6 Nov 25 11:05:15 crc kubenswrapper[4776]: I1125 11:05:15.987306 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-create-cgttx"] Nov 25 11:05:15 crc kubenswrapper[4776]: I1125 11:05:15.988863 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-cgttx" Nov 25 11:05:15 crc kubenswrapper[4776]: I1125 11:05:15.997377 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-cgttx"] Nov 25 11:05:16 crc kubenswrapper[4776]: I1125 11:05:16.008745 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r9h2\" (UniqueName: \"kubernetes.io/projected/62d9656d-fca0-4900-a375-764c2567734e-kube-api-access-8r9h2\") pod \"octavia-db-create-cgttx\" (UID: \"62d9656d-fca0-4900-a375-764c2567734e\") " pod="openstack/octavia-db-create-cgttx" Nov 25 11:05:16 crc kubenswrapper[4776]: I1125 11:05:16.008827 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62d9656d-fca0-4900-a375-764c2567734e-operator-scripts\") pod \"octavia-db-create-cgttx\" (UID: \"62d9656d-fca0-4900-a375-764c2567734e\") " pod="openstack/octavia-db-create-cgttx" Nov 25 11:05:16 crc kubenswrapper[4776]: I1125 11:05:16.111767 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8r9h2\" (UniqueName: \"kubernetes.io/projected/62d9656d-fca0-4900-a375-764c2567734e-kube-api-access-8r9h2\") pod \"octavia-db-create-cgttx\" (UID: \"62d9656d-fca0-4900-a375-764c2567734e\") " pod="openstack/octavia-db-create-cgttx" Nov 25 11:05:16 crc kubenswrapper[4776]: I1125 11:05:16.111892 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62d9656d-fca0-4900-a375-764c2567734e-operator-scripts\") pod \"octavia-db-create-cgttx\" (UID: \"62d9656d-fca0-4900-a375-764c2567734e\") " pod="openstack/octavia-db-create-cgttx" Nov 25 11:05:16 crc kubenswrapper[4776]: I1125 11:05:16.113134 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62d9656d-fca0-4900-a375-764c2567734e-operator-scripts\") pod \"octavia-db-create-cgttx\" (UID: \"62d9656d-fca0-4900-a375-764c2567734e\") " pod="openstack/octavia-db-create-cgttx" Nov 25 11:05:16 crc kubenswrapper[4776]: I1125 11:05:16.134659 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r9h2\" (UniqueName: \"kubernetes.io/projected/62d9656d-fca0-4900-a375-764c2567734e-kube-api-access-8r9h2\") pod \"octavia-db-create-cgttx\" (UID: \"62d9656d-fca0-4900-a375-764c2567734e\") " pod="openstack/octavia-db-create-cgttx" Nov 25 11:05:16 crc kubenswrapper[4776]: I1125 11:05:16.284901 4776 generic.go:334] "Generic (PLEG): container finished" podID="845d70c5-c0ad-4143-8778-3bccb4724aa4" containerID="b88a23cc7996228d50b5eb43124e279401023dece4ba70278741db418e95a198" exitCode=0 Nov 25 11:05:16 crc kubenswrapper[4776]: I1125 11:05:16.284973 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-wlphl" event={"ID":"845d70c5-c0ad-4143-8778-3bccb4724aa4","Type":"ContainerDied","Data":"b88a23cc7996228d50b5eb43124e279401023dece4ba70278741db418e95a198"} Nov 25 11:05:16 crc kubenswrapper[4776]: I1125 11:05:16.287051 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-jgqps" event={"ID":"5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6","Type":"ContainerStarted","Data":"83d7e451707a725f47bba6017e4a3b3b3c1c7aea1eb2cd5e4b8b3d6cc7ee07ec"} Nov 25 11:05:16 crc kubenswrapper[4776]: I1125 11:05:16.287126 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-jgqps" event={"ID":"5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6","Type":"ContainerStarted","Data":"2a391a47e978b7cd6d8ca4b4ce3e4442617226d26a1894ea1ba2a23f348e93b6"} Nov 25 11:05:16 crc kubenswrapper[4776]: I1125 11:05:16.354760 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-cgttx" Nov 25 11:05:16 crc kubenswrapper[4776]: I1125 11:05:16.871453 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-cgttx"] Nov 25 11:05:16 crc kubenswrapper[4776]: I1125 11:05:16.917493 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-ac46-account-create-9jb68"] Nov 25 11:05:16 crc kubenswrapper[4776]: I1125 11:05:16.918931 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-ac46-account-create-9jb68" Nov 25 11:05:16 crc kubenswrapper[4776]: I1125 11:05:16.921491 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-db-secret" Nov 25 11:05:16 crc kubenswrapper[4776]: I1125 11:05:16.928729 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc137cb9-0788-455d-9f5f-2d7530c0c20e-operator-scripts\") pod \"octavia-ac46-account-create-9jb68\" (UID: \"dc137cb9-0788-455d-9f5f-2d7530c0c20e\") " pod="openstack/octavia-ac46-account-create-9jb68" Nov 25 11:05:16 crc kubenswrapper[4776]: I1125 11:05:16.928846 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cm8wz\" (UniqueName: \"kubernetes.io/projected/dc137cb9-0788-455d-9f5f-2d7530c0c20e-kube-api-access-cm8wz\") pod \"octavia-ac46-account-create-9jb68\" (UID: \"dc137cb9-0788-455d-9f5f-2d7530c0c20e\") " pod="openstack/octavia-ac46-account-create-9jb68" Nov 25 11:05:16 crc kubenswrapper[4776]: I1125 11:05:16.935557 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-ac46-account-create-9jb68"] Nov 25 11:05:17 crc kubenswrapper[4776]: I1125 11:05:17.029828 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc137cb9-0788-455d-9f5f-2d7530c0c20e-operator-scripts\") pod \"octavia-ac46-account-create-9jb68\" (UID: \"dc137cb9-0788-455d-9f5f-2d7530c0c20e\") " pod="openstack/octavia-ac46-account-create-9jb68" Nov 25 11:05:17 crc kubenswrapper[4776]: I1125 11:05:17.030223 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cm8wz\" (UniqueName: \"kubernetes.io/projected/dc137cb9-0788-455d-9f5f-2d7530c0c20e-kube-api-access-cm8wz\") pod \"octavia-ac46-account-create-9jb68\" (UID: \"dc137cb9-0788-455d-9f5f-2d7530c0c20e\") " pod="openstack/octavia-ac46-account-create-9jb68" Nov 25 11:05:17 crc kubenswrapper[4776]: I1125 11:05:17.030716 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc137cb9-0788-455d-9f5f-2d7530c0c20e-operator-scripts\") pod \"octavia-ac46-account-create-9jb68\" (UID: \"dc137cb9-0788-455d-9f5f-2d7530c0c20e\") " pod="openstack/octavia-ac46-account-create-9jb68" Nov 25 11:05:17 crc kubenswrapper[4776]: I1125 11:05:17.050423 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cm8wz\" (UniqueName: \"kubernetes.io/projected/dc137cb9-0788-455d-9f5f-2d7530c0c20e-kube-api-access-cm8wz\") pod \"octavia-ac46-account-create-9jb68\" (UID: \"dc137cb9-0788-455d-9f5f-2d7530c0c20e\") " pod="openstack/octavia-ac46-account-create-9jb68" Nov 25 11:05:17 crc kubenswrapper[4776]: I1125 11:05:17.142445 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-ac46-account-create-9jb68" Nov 25 11:05:17 crc kubenswrapper[4776]: I1125 11:05:17.304694 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-wlphl" event={"ID":"845d70c5-c0ad-4143-8778-3bccb4724aa4","Type":"ContainerStarted","Data":"2c42fbc0b8ea892f0707aa7823e4a2cc61b70159c3daca150f2d5eef9c36f78a"} Nov 25 11:05:17 crc kubenswrapper[4776]: I1125 11:05:17.304740 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-wlphl" event={"ID":"845d70c5-c0ad-4143-8778-3bccb4724aa4","Type":"ContainerStarted","Data":"c687d01235421a462f47c4efd62ad476a7408f55aed455d2cdb961f530602c09"} Nov 25 11:05:17 crc kubenswrapper[4776]: I1125 11:05:17.307495 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-cgttx" event={"ID":"62d9656d-fca0-4900-a375-764c2567734e","Type":"ContainerStarted","Data":"eff1494da1033d724e3b72a4150f8e8a07e6d08e2897283feee65477acd13f62"} Nov 25 11:05:17 crc kubenswrapper[4776]: I1125 11:05:17.348496 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-jgqps" podStartSLOduration=3.348475746 podStartE2EDuration="3.348475746s" podCreationTimestamp="2025-11-25 11:05:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:05:17.348452196 +0000 UTC m=+6062.389511749" watchObservedRunningTime="2025-11-25 11:05:17.348475746 +0000 UTC m=+6062.389535299" Nov 25 11:05:17 crc kubenswrapper[4776]: I1125 11:05:17.627090 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-ac46-account-create-9jb68"] Nov 25 11:05:18 crc kubenswrapper[4776]: I1125 11:05:18.316868 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-ac46-account-create-9jb68" event={"ID":"dc137cb9-0788-455d-9f5f-2d7530c0c20e","Type":"ContainerStarted","Data":"f025724c8ab81ee0a83b81d6cbcb981381614ada45b3b21ef81fbf4082a243d3"} Nov 25 11:05:18 crc kubenswrapper[4776]: I1125 11:05:18.317364 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-ac46-account-create-9jb68" event={"ID":"dc137cb9-0788-455d-9f5f-2d7530c0c20e","Type":"ContainerStarted","Data":"abb6cb03364257e1241e3258533ec0d28e2d4265f4176e281a2a8ed23f3ea16c"} Nov 25 11:05:18 crc kubenswrapper[4776]: I1125 11:05:18.319604 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-cgttx" event={"ID":"62d9656d-fca0-4900-a375-764c2567734e","Type":"ContainerStarted","Data":"15701fba63ade2b2868713c0edcf9c8ebd367f8392edeb5225004c085debdaf5"} Nov 25 11:05:18 crc kubenswrapper[4776]: I1125 11:05:18.319698 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-wlphl" Nov 25 11:05:18 crc kubenswrapper[4776]: I1125 11:05:18.319727 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-wlphl" Nov 25 11:05:18 crc kubenswrapper[4776]: I1125 11:05:18.345296 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-wlphl" podStartSLOduration=5.345269473 podStartE2EDuration="5.345269473s" podCreationTimestamp="2025-11-25 11:05:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:05:18.340982155 +0000 UTC m=+6063.382041728" watchObservedRunningTime="2025-11-25 11:05:18.345269473 +0000 UTC m=+6063.386329046" Nov 25 11:05:18 crc kubenswrapper[4776]: I1125 11:05:18.359732 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-db-create-cgttx" podStartSLOduration=3.359710125 podStartE2EDuration="3.359710125s" podCreationTimestamp="2025-11-25 11:05:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:05:18.356286669 +0000 UTC m=+6063.397346222" watchObservedRunningTime="2025-11-25 11:05:18.359710125 +0000 UTC m=+6063.400769688" Nov 25 11:05:18 crc kubenswrapper[4776]: I1125 11:05:18.663554 4776 scope.go:117] "RemoveContainer" containerID="8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698" Nov 25 11:05:18 crc kubenswrapper[4776]: E1125 11:05:18.664109 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:05:19 crc kubenswrapper[4776]: I1125 11:05:19.333299 4776 generic.go:334] "Generic (PLEG): container finished" podID="62d9656d-fca0-4900-a375-764c2567734e" containerID="15701fba63ade2b2868713c0edcf9c8ebd367f8392edeb5225004c085debdaf5" exitCode=0 Nov 25 11:05:19 crc kubenswrapper[4776]: I1125 11:05:19.333372 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-cgttx" event={"ID":"62d9656d-fca0-4900-a375-764c2567734e","Type":"ContainerDied","Data":"15701fba63ade2b2868713c0edcf9c8ebd367f8392edeb5225004c085debdaf5"} Nov 25 11:05:19 crc kubenswrapper[4776]: I1125 11:05:19.336625 4776 generic.go:334] "Generic (PLEG): container finished" podID="dc137cb9-0788-455d-9f5f-2d7530c0c20e" containerID="f025724c8ab81ee0a83b81d6cbcb981381614ada45b3b21ef81fbf4082a243d3" exitCode=0 Nov 25 11:05:19 crc kubenswrapper[4776]: I1125 11:05:19.336815 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-ac46-account-create-9jb68" event={"ID":"dc137cb9-0788-455d-9f5f-2d7530c0c20e","Type":"ContainerDied","Data":"f025724c8ab81ee0a83b81d6cbcb981381614ada45b3b21ef81fbf4082a243d3"} Nov 25 11:05:20 crc kubenswrapper[4776]: I1125 11:05:20.876922 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-cgttx" Nov 25 11:05:20 crc kubenswrapper[4776]: I1125 11:05:20.883975 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-ac46-account-create-9jb68" Nov 25 11:05:21 crc kubenswrapper[4776]: I1125 11:05:21.005826 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc137cb9-0788-455d-9f5f-2d7530c0c20e-operator-scripts\") pod \"dc137cb9-0788-455d-9f5f-2d7530c0c20e\" (UID: \"dc137cb9-0788-455d-9f5f-2d7530c0c20e\") " Nov 25 11:05:21 crc kubenswrapper[4776]: I1125 11:05:21.006011 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8r9h2\" (UniqueName: \"kubernetes.io/projected/62d9656d-fca0-4900-a375-764c2567734e-kube-api-access-8r9h2\") pod \"62d9656d-fca0-4900-a375-764c2567734e\" (UID: \"62d9656d-fca0-4900-a375-764c2567734e\") " Nov 25 11:05:21 crc kubenswrapper[4776]: I1125 11:05:21.006140 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62d9656d-fca0-4900-a375-764c2567734e-operator-scripts\") pod \"62d9656d-fca0-4900-a375-764c2567734e\" (UID: \"62d9656d-fca0-4900-a375-764c2567734e\") " Nov 25 11:05:21 crc kubenswrapper[4776]: I1125 11:05:21.006184 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cm8wz\" (UniqueName: \"kubernetes.io/projected/dc137cb9-0788-455d-9f5f-2d7530c0c20e-kube-api-access-cm8wz\") pod \"dc137cb9-0788-455d-9f5f-2d7530c0c20e\" (UID: \"dc137cb9-0788-455d-9f5f-2d7530c0c20e\") " Nov 25 11:05:21 crc kubenswrapper[4776]: I1125 11:05:21.007372 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc137cb9-0788-455d-9f5f-2d7530c0c20e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dc137cb9-0788-455d-9f5f-2d7530c0c20e" (UID: "dc137cb9-0788-455d-9f5f-2d7530c0c20e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:05:21 crc kubenswrapper[4776]: I1125 11:05:21.007512 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62d9656d-fca0-4900-a375-764c2567734e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "62d9656d-fca0-4900-a375-764c2567734e" (UID: "62d9656d-fca0-4900-a375-764c2567734e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:05:21 crc kubenswrapper[4776]: I1125 11:05:21.012819 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc137cb9-0788-455d-9f5f-2d7530c0c20e-kube-api-access-cm8wz" (OuterVolumeSpecName: "kube-api-access-cm8wz") pod "dc137cb9-0788-455d-9f5f-2d7530c0c20e" (UID: "dc137cb9-0788-455d-9f5f-2d7530c0c20e"). InnerVolumeSpecName "kube-api-access-cm8wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:05:21 crc kubenswrapper[4776]: I1125 11:05:21.012878 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62d9656d-fca0-4900-a375-764c2567734e-kube-api-access-8r9h2" (OuterVolumeSpecName: "kube-api-access-8r9h2") pod "62d9656d-fca0-4900-a375-764c2567734e" (UID: "62d9656d-fca0-4900-a375-764c2567734e"). InnerVolumeSpecName "kube-api-access-8r9h2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:05:21 crc kubenswrapper[4776]: I1125 11:05:21.108093 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/62d9656d-fca0-4900-a375-764c2567734e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:21 crc kubenswrapper[4776]: I1125 11:05:21.108128 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cm8wz\" (UniqueName: \"kubernetes.io/projected/dc137cb9-0788-455d-9f5f-2d7530c0c20e-kube-api-access-cm8wz\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:21 crc kubenswrapper[4776]: I1125 11:05:21.108140 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc137cb9-0788-455d-9f5f-2d7530c0c20e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:21 crc kubenswrapper[4776]: I1125 11:05:21.108148 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8r9h2\" (UniqueName: \"kubernetes.io/projected/62d9656d-fca0-4900-a375-764c2567734e-kube-api-access-8r9h2\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:21 crc kubenswrapper[4776]: I1125 11:05:21.364559 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-cgttx" event={"ID":"62d9656d-fca0-4900-a375-764c2567734e","Type":"ContainerDied","Data":"eff1494da1033d724e3b72a4150f8e8a07e6d08e2897283feee65477acd13f62"} Nov 25 11:05:21 crc kubenswrapper[4776]: I1125 11:05:21.364897 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eff1494da1033d724e3b72a4150f8e8a07e6d08e2897283feee65477acd13f62" Nov 25 11:05:21 crc kubenswrapper[4776]: I1125 11:05:21.364598 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-cgttx" Nov 25 11:05:21 crc kubenswrapper[4776]: I1125 11:05:21.366103 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-ac46-account-create-9jb68" event={"ID":"dc137cb9-0788-455d-9f5f-2d7530c0c20e","Type":"ContainerDied","Data":"abb6cb03364257e1241e3258533ec0d28e2d4265f4176e281a2a8ed23f3ea16c"} Nov 25 11:05:21 crc kubenswrapper[4776]: I1125 11:05:21.366138 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="abb6cb03364257e1241e3258533ec0d28e2d4265f4176e281a2a8ed23f3ea16c" Nov 25 11:05:21 crc kubenswrapper[4776]: I1125 11:05:21.366201 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-ac46-account-create-9jb68" Nov 25 11:05:27 crc kubenswrapper[4776]: I1125 11:05:27.521771 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-persistence-db-create-gjqsq"] Nov 25 11:05:27 crc kubenswrapper[4776]: E1125 11:05:27.522685 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc137cb9-0788-455d-9f5f-2d7530c0c20e" containerName="mariadb-account-create" Nov 25 11:05:27 crc kubenswrapper[4776]: I1125 11:05:27.522698 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc137cb9-0788-455d-9f5f-2d7530c0c20e" containerName="mariadb-account-create" Nov 25 11:05:27 crc kubenswrapper[4776]: E1125 11:05:27.522719 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62d9656d-fca0-4900-a375-764c2567734e" containerName="mariadb-database-create" Nov 25 11:05:27 crc kubenswrapper[4776]: I1125 11:05:27.522726 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="62d9656d-fca0-4900-a375-764c2567734e" containerName="mariadb-database-create" Nov 25 11:05:27 crc kubenswrapper[4776]: I1125 11:05:27.522883 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc137cb9-0788-455d-9f5f-2d7530c0c20e" containerName="mariadb-account-create" Nov 25 11:05:27 crc kubenswrapper[4776]: I1125 11:05:27.522908 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="62d9656d-fca0-4900-a375-764c2567734e" containerName="mariadb-database-create" Nov 25 11:05:27 crc kubenswrapper[4776]: I1125 11:05:27.523539 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-gjqsq" Nov 25 11:05:27 crc kubenswrapper[4776]: I1125 11:05:27.547838 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-gjqsq"] Nov 25 11:05:27 crc kubenswrapper[4776]: I1125 11:05:27.643274 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/703a295e-547f-4e35-a67a-dd2d77361161-operator-scripts\") pod \"octavia-persistence-db-create-gjqsq\" (UID: \"703a295e-547f-4e35-a67a-dd2d77361161\") " pod="openstack/octavia-persistence-db-create-gjqsq" Nov 25 11:05:27 crc kubenswrapper[4776]: I1125 11:05:27.644028 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzg8t\" (UniqueName: \"kubernetes.io/projected/703a295e-547f-4e35-a67a-dd2d77361161-kube-api-access-bzg8t\") pod \"octavia-persistence-db-create-gjqsq\" (UID: \"703a295e-547f-4e35-a67a-dd2d77361161\") " pod="openstack/octavia-persistence-db-create-gjqsq" Nov 25 11:05:27 crc kubenswrapper[4776]: I1125 11:05:27.746980 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzg8t\" (UniqueName: \"kubernetes.io/projected/703a295e-547f-4e35-a67a-dd2d77361161-kube-api-access-bzg8t\") pod \"octavia-persistence-db-create-gjqsq\" (UID: \"703a295e-547f-4e35-a67a-dd2d77361161\") " pod="openstack/octavia-persistence-db-create-gjqsq" Nov 25 11:05:27 crc kubenswrapper[4776]: I1125 11:05:27.747112 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/703a295e-547f-4e35-a67a-dd2d77361161-operator-scripts\") pod \"octavia-persistence-db-create-gjqsq\" (UID: \"703a295e-547f-4e35-a67a-dd2d77361161\") " pod="openstack/octavia-persistence-db-create-gjqsq" Nov 25 11:05:27 crc kubenswrapper[4776]: I1125 11:05:27.748157 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/703a295e-547f-4e35-a67a-dd2d77361161-operator-scripts\") pod \"octavia-persistence-db-create-gjqsq\" (UID: \"703a295e-547f-4e35-a67a-dd2d77361161\") " pod="openstack/octavia-persistence-db-create-gjqsq" Nov 25 11:05:27 crc kubenswrapper[4776]: I1125 11:05:27.769417 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzg8t\" (UniqueName: \"kubernetes.io/projected/703a295e-547f-4e35-a67a-dd2d77361161-kube-api-access-bzg8t\") pod \"octavia-persistence-db-create-gjqsq\" (UID: \"703a295e-547f-4e35-a67a-dd2d77361161\") " pod="openstack/octavia-persistence-db-create-gjqsq" Nov 25 11:05:27 crc kubenswrapper[4776]: I1125 11:05:27.845029 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-gjqsq" Nov 25 11:05:28 crc kubenswrapper[4776]: I1125 11:05:28.345398 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-gjqsq"] Nov 25 11:05:28 crc kubenswrapper[4776]: I1125 11:05:28.435056 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-gjqsq" event={"ID":"703a295e-547f-4e35-a67a-dd2d77361161","Type":"ContainerStarted","Data":"1cb4756e3cc77436bf7a806f0c66ba262223fa889def46c7ce8774c88cfaf36f"} Nov 25 11:05:28 crc kubenswrapper[4776]: I1125 11:05:28.973544 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-116d-account-create-jf6wx"] Nov 25 11:05:28 crc kubenswrapper[4776]: I1125 11:05:28.975287 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-116d-account-create-jf6wx" Nov 25 11:05:28 crc kubenswrapper[4776]: I1125 11:05:28.977113 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-persistence-db-secret" Nov 25 11:05:28 crc kubenswrapper[4776]: I1125 11:05:28.989668 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-116d-account-create-jf6wx"] Nov 25 11:05:29 crc kubenswrapper[4776]: I1125 11:05:29.178273 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kx8xv\" (UniqueName: \"kubernetes.io/projected/53c3e53a-7c78-4607-923c-604eed89a9df-kube-api-access-kx8xv\") pod \"octavia-116d-account-create-jf6wx\" (UID: \"53c3e53a-7c78-4607-923c-604eed89a9df\") " pod="openstack/octavia-116d-account-create-jf6wx" Nov 25 11:05:29 crc kubenswrapper[4776]: I1125 11:05:29.178753 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53c3e53a-7c78-4607-923c-604eed89a9df-operator-scripts\") pod \"octavia-116d-account-create-jf6wx\" (UID: \"53c3e53a-7c78-4607-923c-604eed89a9df\") " pod="openstack/octavia-116d-account-create-jf6wx" Nov 25 11:05:29 crc kubenswrapper[4776]: I1125 11:05:29.281560 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kx8xv\" (UniqueName: \"kubernetes.io/projected/53c3e53a-7c78-4607-923c-604eed89a9df-kube-api-access-kx8xv\") pod \"octavia-116d-account-create-jf6wx\" (UID: \"53c3e53a-7c78-4607-923c-604eed89a9df\") " pod="openstack/octavia-116d-account-create-jf6wx" Nov 25 11:05:29 crc kubenswrapper[4776]: I1125 11:05:29.281695 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53c3e53a-7c78-4607-923c-604eed89a9df-operator-scripts\") pod \"octavia-116d-account-create-jf6wx\" (UID: \"53c3e53a-7c78-4607-923c-604eed89a9df\") " pod="openstack/octavia-116d-account-create-jf6wx" Nov 25 11:05:29 crc kubenswrapper[4776]: I1125 11:05:29.282687 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53c3e53a-7c78-4607-923c-604eed89a9df-operator-scripts\") pod \"octavia-116d-account-create-jf6wx\" (UID: \"53c3e53a-7c78-4607-923c-604eed89a9df\") " pod="openstack/octavia-116d-account-create-jf6wx" Nov 25 11:05:29 crc kubenswrapper[4776]: I1125 11:05:29.304319 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kx8xv\" (UniqueName: \"kubernetes.io/projected/53c3e53a-7c78-4607-923c-604eed89a9df-kube-api-access-kx8xv\") pod \"octavia-116d-account-create-jf6wx\" (UID: \"53c3e53a-7c78-4607-923c-604eed89a9df\") " pod="openstack/octavia-116d-account-create-jf6wx" Nov 25 11:05:29 crc kubenswrapper[4776]: I1125 11:05:29.323735 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-116d-account-create-jf6wx" Nov 25 11:05:29 crc kubenswrapper[4776]: I1125 11:05:29.444401 4776 generic.go:334] "Generic (PLEG): container finished" podID="703a295e-547f-4e35-a67a-dd2d77361161" containerID="abbfd694167843fa80feaa11dbb30b5a0931e128c7fec98484adaddf3c487af7" exitCode=0 Nov 25 11:05:29 crc kubenswrapper[4776]: I1125 11:05:29.444473 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-gjqsq" event={"ID":"703a295e-547f-4e35-a67a-dd2d77361161","Type":"ContainerDied","Data":"abbfd694167843fa80feaa11dbb30b5a0931e128c7fec98484adaddf3c487af7"} Nov 25 11:05:29 crc kubenswrapper[4776]: I1125 11:05:29.782141 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-116d-account-create-jf6wx"] Nov 25 11:05:29 crc kubenswrapper[4776]: W1125 11:05:29.789314 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53c3e53a_7c78_4607_923c_604eed89a9df.slice/crio-1813c7b40b4cac99f99b304d8facef1ecd259f24bd25f4ba31cc216dcdc0c159 WatchSource:0}: Error finding container 1813c7b40b4cac99f99b304d8facef1ecd259f24bd25f4ba31cc216dcdc0c159: Status 404 returned error can't find the container with id 1813c7b40b4cac99f99b304d8facef1ecd259f24bd25f4ba31cc216dcdc0c159 Nov 25 11:05:30 crc kubenswrapper[4776]: I1125 11:05:30.453889 4776 generic.go:334] "Generic (PLEG): container finished" podID="53c3e53a-7c78-4607-923c-604eed89a9df" containerID="7968e159f463c8259bd2c25b1735de14db0d97b3eacbf0f1fca4fa992873748d" exitCode=0 Nov 25 11:05:30 crc kubenswrapper[4776]: I1125 11:05:30.453995 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-116d-account-create-jf6wx" event={"ID":"53c3e53a-7c78-4607-923c-604eed89a9df","Type":"ContainerDied","Data":"7968e159f463c8259bd2c25b1735de14db0d97b3eacbf0f1fca4fa992873748d"} Nov 25 11:05:30 crc kubenswrapper[4776]: I1125 11:05:30.454422 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-116d-account-create-jf6wx" event={"ID":"53c3e53a-7c78-4607-923c-604eed89a9df","Type":"ContainerStarted","Data":"1813c7b40b4cac99f99b304d8facef1ecd259f24bd25f4ba31cc216dcdc0c159"} Nov 25 11:05:30 crc kubenswrapper[4776]: I1125 11:05:30.846572 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-gjqsq" Nov 25 11:05:31 crc kubenswrapper[4776]: I1125 11:05:31.016484 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/703a295e-547f-4e35-a67a-dd2d77361161-operator-scripts\") pod \"703a295e-547f-4e35-a67a-dd2d77361161\" (UID: \"703a295e-547f-4e35-a67a-dd2d77361161\") " Nov 25 11:05:31 crc kubenswrapper[4776]: I1125 11:05:31.016805 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzg8t\" (UniqueName: \"kubernetes.io/projected/703a295e-547f-4e35-a67a-dd2d77361161-kube-api-access-bzg8t\") pod \"703a295e-547f-4e35-a67a-dd2d77361161\" (UID: \"703a295e-547f-4e35-a67a-dd2d77361161\") " Nov 25 11:05:31 crc kubenswrapper[4776]: I1125 11:05:31.017334 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/703a295e-547f-4e35-a67a-dd2d77361161-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "703a295e-547f-4e35-a67a-dd2d77361161" (UID: "703a295e-547f-4e35-a67a-dd2d77361161"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:05:31 crc kubenswrapper[4776]: I1125 11:05:31.022368 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/703a295e-547f-4e35-a67a-dd2d77361161-kube-api-access-bzg8t" (OuterVolumeSpecName: "kube-api-access-bzg8t") pod "703a295e-547f-4e35-a67a-dd2d77361161" (UID: "703a295e-547f-4e35-a67a-dd2d77361161"). InnerVolumeSpecName "kube-api-access-bzg8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:05:31 crc kubenswrapper[4776]: I1125 11:05:31.119199 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzg8t\" (UniqueName: \"kubernetes.io/projected/703a295e-547f-4e35-a67a-dd2d77361161-kube-api-access-bzg8t\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:31 crc kubenswrapper[4776]: I1125 11:05:31.119255 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/703a295e-547f-4e35-a67a-dd2d77361161-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:31 crc kubenswrapper[4776]: I1125 11:05:31.469738 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-gjqsq" Nov 25 11:05:31 crc kubenswrapper[4776]: I1125 11:05:31.469837 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-gjqsq" event={"ID":"703a295e-547f-4e35-a67a-dd2d77361161","Type":"ContainerDied","Data":"1cb4756e3cc77436bf7a806f0c66ba262223fa889def46c7ce8774c88cfaf36f"} Nov 25 11:05:31 crc kubenswrapper[4776]: I1125 11:05:31.469885 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cb4756e3cc77436bf7a806f0c66ba262223fa889def46c7ce8774c88cfaf36f" Nov 25 11:05:31 crc kubenswrapper[4776]: I1125 11:05:31.663370 4776 scope.go:117] "RemoveContainer" containerID="8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698" Nov 25 11:05:31 crc kubenswrapper[4776]: E1125 11:05:31.663779 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:05:31 crc kubenswrapper[4776]: I1125 11:05:31.800743 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-116d-account-create-jf6wx" Nov 25 11:05:31 crc kubenswrapper[4776]: I1125 11:05:31.935455 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53c3e53a-7c78-4607-923c-604eed89a9df-operator-scripts\") pod \"53c3e53a-7c78-4607-923c-604eed89a9df\" (UID: \"53c3e53a-7c78-4607-923c-604eed89a9df\") " Nov 25 11:05:31 crc kubenswrapper[4776]: I1125 11:05:31.935650 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kx8xv\" (UniqueName: \"kubernetes.io/projected/53c3e53a-7c78-4607-923c-604eed89a9df-kube-api-access-kx8xv\") pod \"53c3e53a-7c78-4607-923c-604eed89a9df\" (UID: \"53c3e53a-7c78-4607-923c-604eed89a9df\") " Nov 25 11:05:31 crc kubenswrapper[4776]: I1125 11:05:31.936135 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53c3e53a-7c78-4607-923c-604eed89a9df-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "53c3e53a-7c78-4607-923c-604eed89a9df" (UID: "53c3e53a-7c78-4607-923c-604eed89a9df"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:05:31 crc kubenswrapper[4776]: I1125 11:05:31.936330 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/53c3e53a-7c78-4607-923c-604eed89a9df-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:31 crc kubenswrapper[4776]: I1125 11:05:31.940518 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53c3e53a-7c78-4607-923c-604eed89a9df-kube-api-access-kx8xv" (OuterVolumeSpecName: "kube-api-access-kx8xv") pod "53c3e53a-7c78-4607-923c-604eed89a9df" (UID: "53c3e53a-7c78-4607-923c-604eed89a9df"). InnerVolumeSpecName "kube-api-access-kx8xv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:05:32 crc kubenswrapper[4776]: I1125 11:05:32.037525 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kx8xv\" (UniqueName: \"kubernetes.io/projected/53c3e53a-7c78-4607-923c-604eed89a9df-kube-api-access-kx8xv\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:32 crc kubenswrapper[4776]: I1125 11:05:32.479193 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-116d-account-create-jf6wx" event={"ID":"53c3e53a-7c78-4607-923c-604eed89a9df","Type":"ContainerDied","Data":"1813c7b40b4cac99f99b304d8facef1ecd259f24bd25f4ba31cc216dcdc0c159"} Nov 25 11:05:32 crc kubenswrapper[4776]: I1125 11:05:32.479235 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1813c7b40b4cac99f99b304d8facef1ecd259f24bd25f4ba31cc216dcdc0c159" Nov 25 11:05:32 crc kubenswrapper[4776]: I1125 11:05:32.479291 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-116d-account-create-jf6wx" Nov 25 11:05:33 crc kubenswrapper[4776]: I1125 11:05:33.190594 4776 scope.go:117] "RemoveContainer" containerID="99918cbec2ca01810476b8024781dfaddbdd25714aa55a102117b6fa681ba895" Nov 25 11:05:33 crc kubenswrapper[4776]: I1125 11:05:33.227595 4776 scope.go:117] "RemoveContainer" containerID="b281721f6ad7427a1e3f77d7eeaedc45a3a0ae1b688aa850dc6f343ee2a59668" Nov 25 11:05:33 crc kubenswrapper[4776]: I1125 11:05:33.397652 4776 scope.go:117] "RemoveContainer" containerID="2d918aa82f7a519fd0d43d5392ee5dfc86d0b00a061e1b3f1ddfa9b88d81c779" Nov 25 11:05:33 crc kubenswrapper[4776]: I1125 11:05:33.428653 4776 scope.go:117] "RemoveContainer" containerID="54822ea4c536a47287cc033d08783b1b390937d0ee1f7d8174f2653ca047f4df" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.130571 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-api-dccc45cc4-75rpv"] Nov 25 11:05:35 crc kubenswrapper[4776]: E1125 11:05:35.131346 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53c3e53a-7c78-4607-923c-604eed89a9df" containerName="mariadb-account-create" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.131361 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="53c3e53a-7c78-4607-923c-604eed89a9df" containerName="mariadb-account-create" Nov 25 11:05:35 crc kubenswrapper[4776]: E1125 11:05:35.131378 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="703a295e-547f-4e35-a67a-dd2d77361161" containerName="mariadb-database-create" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.131385 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="703a295e-547f-4e35-a67a-dd2d77361161" containerName="mariadb-database-create" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.131552 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="53c3e53a-7c78-4607-923c-604eed89a9df" containerName="mariadb-account-create" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.131569 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="703a295e-547f-4e35-a67a-dd2d77361161" containerName="mariadb-database-create" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.133784 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-dccc45cc4-75rpv" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.135181 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-scripts" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.135932 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-octavia-dockercfg-xkbkk" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.136148 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-config-data" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.136387 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-octavia-ovndbs" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.157515 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-dccc45cc4-75rpv"] Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.297759 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-config-data-merged\") pod \"octavia-api-dccc45cc4-75rpv\" (UID: \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\") " pod="openstack/octavia-api-dccc45cc4-75rpv" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.298018 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-config-data\") pod \"octavia-api-dccc45cc4-75rpv\" (UID: \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\") " pod="openstack/octavia-api-dccc45cc4-75rpv" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.298173 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-octavia-run\") pod \"octavia-api-dccc45cc4-75rpv\" (UID: \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\") " pod="openstack/octavia-api-dccc45cc4-75rpv" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.298260 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-ovndb-tls-certs\") pod \"octavia-api-dccc45cc4-75rpv\" (UID: \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\") " pod="openstack/octavia-api-dccc45cc4-75rpv" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.298388 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-combined-ca-bundle\") pod \"octavia-api-dccc45cc4-75rpv\" (UID: \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\") " pod="openstack/octavia-api-dccc45cc4-75rpv" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.299216 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-scripts\") pod \"octavia-api-dccc45cc4-75rpv\" (UID: \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\") " pod="openstack/octavia-api-dccc45cc4-75rpv" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.400556 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-combined-ca-bundle\") pod \"octavia-api-dccc45cc4-75rpv\" (UID: \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\") " pod="openstack/octavia-api-dccc45cc4-75rpv" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.400626 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-scripts\") pod \"octavia-api-dccc45cc4-75rpv\" (UID: \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\") " pod="openstack/octavia-api-dccc45cc4-75rpv" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.400668 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-config-data-merged\") pod \"octavia-api-dccc45cc4-75rpv\" (UID: \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\") " pod="openstack/octavia-api-dccc45cc4-75rpv" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.400685 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-config-data\") pod \"octavia-api-dccc45cc4-75rpv\" (UID: \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\") " pod="openstack/octavia-api-dccc45cc4-75rpv" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.400757 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-octavia-run\") pod \"octavia-api-dccc45cc4-75rpv\" (UID: \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\") " pod="openstack/octavia-api-dccc45cc4-75rpv" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.400796 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-ovndb-tls-certs\") pod \"octavia-api-dccc45cc4-75rpv\" (UID: \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\") " pod="openstack/octavia-api-dccc45cc4-75rpv" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.401283 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-config-data-merged\") pod \"octavia-api-dccc45cc4-75rpv\" (UID: \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\") " pod="openstack/octavia-api-dccc45cc4-75rpv" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.401322 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-octavia-run\") pod \"octavia-api-dccc45cc4-75rpv\" (UID: \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\") " pod="openstack/octavia-api-dccc45cc4-75rpv" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.406278 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-config-data\") pod \"octavia-api-dccc45cc4-75rpv\" (UID: \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\") " pod="openstack/octavia-api-dccc45cc4-75rpv" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.406340 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-combined-ca-bundle\") pod \"octavia-api-dccc45cc4-75rpv\" (UID: \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\") " pod="openstack/octavia-api-dccc45cc4-75rpv" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.406608 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-ovndb-tls-certs\") pod \"octavia-api-dccc45cc4-75rpv\" (UID: \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\") " pod="openstack/octavia-api-dccc45cc4-75rpv" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.420869 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-scripts\") pod \"octavia-api-dccc45cc4-75rpv\" (UID: \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\") " pod="openstack/octavia-api-dccc45cc4-75rpv" Nov 25 11:05:35 crc kubenswrapper[4776]: I1125 11:05:35.467278 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-dccc45cc4-75rpv" Nov 25 11:05:36 crc kubenswrapper[4776]: I1125 11:05:36.025488 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-dccc45cc4-75rpv"] Nov 25 11:05:36 crc kubenswrapper[4776]: W1125 11:05:36.033544 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cee8213_d6c2_4b71_a4f5_ee4baa239c4c.slice/crio-ee567d3eb7e86e1bbf570e74d3875d83fc2640958f10c0e23e28be5c1cfdbde2 WatchSource:0}: Error finding container ee567d3eb7e86e1bbf570e74d3875d83fc2640958f10c0e23e28be5c1cfdbde2: Status 404 returned error can't find the container with id ee567d3eb7e86e1bbf570e74d3875d83fc2640958f10c0e23e28be5c1cfdbde2 Nov 25 11:05:36 crc kubenswrapper[4776]: I1125 11:05:36.532437 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-dccc45cc4-75rpv" event={"ID":"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c","Type":"ContainerStarted","Data":"ee567d3eb7e86e1bbf570e74d3875d83fc2640958f10c0e23e28be5c1cfdbde2"} Nov 25 11:05:42 crc kubenswrapper[4776]: I1125 11:05:42.663507 4776 scope.go:117] "RemoveContainer" containerID="8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698" Nov 25 11:05:42 crc kubenswrapper[4776]: E1125 11:05:42.664454 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:05:48 crc kubenswrapper[4776]: I1125 11:05:48.838734 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-85l79" Nov 25 11:05:48 crc kubenswrapper[4776]: I1125 11:05:48.863369 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-wlphl" Nov 25 11:05:48 crc kubenswrapper[4776]: I1125 11:05:48.863422 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-wlphl" Nov 25 11:05:48 crc kubenswrapper[4776]: I1125 11:05:48.992605 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-85l79-config-nd5rg"] Nov 25 11:05:48 crc kubenswrapper[4776]: I1125 11:05:48.993798 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-85l79-config-nd5rg" Nov 25 11:05:48 crc kubenswrapper[4776]: I1125 11:05:48.996266 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 25 11:05:49 crc kubenswrapper[4776]: I1125 11:05:49.010419 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-85l79-config-nd5rg"] Nov 25 11:05:49 crc kubenswrapper[4776]: I1125 11:05:49.096823 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/76880f11-d657-4560-9205-3e86b2845cfc-additional-scripts\") pod \"ovn-controller-85l79-config-nd5rg\" (UID: \"76880f11-d657-4560-9205-3e86b2845cfc\") " pod="openstack/ovn-controller-85l79-config-nd5rg" Nov 25 11:05:49 crc kubenswrapper[4776]: I1125 11:05:49.097029 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76880f11-d657-4560-9205-3e86b2845cfc-scripts\") pod \"ovn-controller-85l79-config-nd5rg\" (UID: \"76880f11-d657-4560-9205-3e86b2845cfc\") " pod="openstack/ovn-controller-85l79-config-nd5rg" Nov 25 11:05:49 crc kubenswrapper[4776]: I1125 11:05:49.097056 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hr29\" (UniqueName: \"kubernetes.io/projected/76880f11-d657-4560-9205-3e86b2845cfc-kube-api-access-6hr29\") pod \"ovn-controller-85l79-config-nd5rg\" (UID: \"76880f11-d657-4560-9205-3e86b2845cfc\") " pod="openstack/ovn-controller-85l79-config-nd5rg" Nov 25 11:05:49 crc kubenswrapper[4776]: I1125 11:05:49.097123 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/76880f11-d657-4560-9205-3e86b2845cfc-var-run-ovn\") pod \"ovn-controller-85l79-config-nd5rg\" (UID: \"76880f11-d657-4560-9205-3e86b2845cfc\") " pod="openstack/ovn-controller-85l79-config-nd5rg" Nov 25 11:05:49 crc kubenswrapper[4776]: I1125 11:05:49.097163 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/76880f11-d657-4560-9205-3e86b2845cfc-var-run\") pod \"ovn-controller-85l79-config-nd5rg\" (UID: \"76880f11-d657-4560-9205-3e86b2845cfc\") " pod="openstack/ovn-controller-85l79-config-nd5rg" Nov 25 11:05:49 crc kubenswrapper[4776]: I1125 11:05:49.097445 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/76880f11-d657-4560-9205-3e86b2845cfc-var-log-ovn\") pod \"ovn-controller-85l79-config-nd5rg\" (UID: \"76880f11-d657-4560-9205-3e86b2845cfc\") " pod="openstack/ovn-controller-85l79-config-nd5rg" Nov 25 11:05:49 crc kubenswrapper[4776]: I1125 11:05:49.200020 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76880f11-d657-4560-9205-3e86b2845cfc-scripts\") pod \"ovn-controller-85l79-config-nd5rg\" (UID: \"76880f11-d657-4560-9205-3e86b2845cfc\") " pod="openstack/ovn-controller-85l79-config-nd5rg" Nov 25 11:05:49 crc kubenswrapper[4776]: I1125 11:05:49.200100 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hr29\" (UniqueName: \"kubernetes.io/projected/76880f11-d657-4560-9205-3e86b2845cfc-kube-api-access-6hr29\") pod \"ovn-controller-85l79-config-nd5rg\" (UID: \"76880f11-d657-4560-9205-3e86b2845cfc\") " pod="openstack/ovn-controller-85l79-config-nd5rg" Nov 25 11:05:49 crc kubenswrapper[4776]: I1125 11:05:49.200154 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/76880f11-d657-4560-9205-3e86b2845cfc-var-run-ovn\") pod \"ovn-controller-85l79-config-nd5rg\" (UID: \"76880f11-d657-4560-9205-3e86b2845cfc\") " pod="openstack/ovn-controller-85l79-config-nd5rg" Nov 25 11:05:49 crc kubenswrapper[4776]: I1125 11:05:49.200202 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/76880f11-d657-4560-9205-3e86b2845cfc-var-run\") pod \"ovn-controller-85l79-config-nd5rg\" (UID: \"76880f11-d657-4560-9205-3e86b2845cfc\") " pod="openstack/ovn-controller-85l79-config-nd5rg" Nov 25 11:05:49 crc kubenswrapper[4776]: I1125 11:05:49.200296 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/76880f11-d657-4560-9205-3e86b2845cfc-var-log-ovn\") pod \"ovn-controller-85l79-config-nd5rg\" (UID: \"76880f11-d657-4560-9205-3e86b2845cfc\") " pod="openstack/ovn-controller-85l79-config-nd5rg" Nov 25 11:05:49 crc kubenswrapper[4776]: I1125 11:05:49.200342 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/76880f11-d657-4560-9205-3e86b2845cfc-additional-scripts\") pod \"ovn-controller-85l79-config-nd5rg\" (UID: \"76880f11-d657-4560-9205-3e86b2845cfc\") " pod="openstack/ovn-controller-85l79-config-nd5rg" Nov 25 11:05:49 crc kubenswrapper[4776]: I1125 11:05:49.200504 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/76880f11-d657-4560-9205-3e86b2845cfc-var-run-ovn\") pod \"ovn-controller-85l79-config-nd5rg\" (UID: \"76880f11-d657-4560-9205-3e86b2845cfc\") " pod="openstack/ovn-controller-85l79-config-nd5rg" Nov 25 11:05:49 crc kubenswrapper[4776]: I1125 11:05:49.200562 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/76880f11-d657-4560-9205-3e86b2845cfc-var-log-ovn\") pod \"ovn-controller-85l79-config-nd5rg\" (UID: \"76880f11-d657-4560-9205-3e86b2845cfc\") " pod="openstack/ovn-controller-85l79-config-nd5rg" Nov 25 11:05:49 crc kubenswrapper[4776]: I1125 11:05:49.201384 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/76880f11-d657-4560-9205-3e86b2845cfc-additional-scripts\") pod \"ovn-controller-85l79-config-nd5rg\" (UID: \"76880f11-d657-4560-9205-3e86b2845cfc\") " pod="openstack/ovn-controller-85l79-config-nd5rg" Nov 25 11:05:49 crc kubenswrapper[4776]: I1125 11:05:49.201619 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/76880f11-d657-4560-9205-3e86b2845cfc-var-run\") pod \"ovn-controller-85l79-config-nd5rg\" (UID: \"76880f11-d657-4560-9205-3e86b2845cfc\") " pod="openstack/ovn-controller-85l79-config-nd5rg" Nov 25 11:05:49 crc kubenswrapper[4776]: I1125 11:05:49.203000 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76880f11-d657-4560-9205-3e86b2845cfc-scripts\") pod \"ovn-controller-85l79-config-nd5rg\" (UID: \"76880f11-d657-4560-9205-3e86b2845cfc\") " pod="openstack/ovn-controller-85l79-config-nd5rg" Nov 25 11:05:49 crc kubenswrapper[4776]: I1125 11:05:49.218555 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hr29\" (UniqueName: \"kubernetes.io/projected/76880f11-d657-4560-9205-3e86b2845cfc-kube-api-access-6hr29\") pod \"ovn-controller-85l79-config-nd5rg\" (UID: \"76880f11-d657-4560-9205-3e86b2845cfc\") " pod="openstack/ovn-controller-85l79-config-nd5rg" Nov 25 11:05:49 crc kubenswrapper[4776]: I1125 11:05:49.320004 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-85l79-config-nd5rg" Nov 25 11:05:50 crc kubenswrapper[4776]: I1125 11:05:50.648411 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-85l79-config-nd5rg"] Nov 25 11:05:50 crc kubenswrapper[4776]: I1125 11:05:50.680835 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-85l79-config-nd5rg" event={"ID":"76880f11-d657-4560-9205-3e86b2845cfc","Type":"ContainerStarted","Data":"23abe12e3116f8ecdbdf6fa1851c34e346ba351c96fce48b3a563d76f8a9a2a9"} Nov 25 11:05:51 crc kubenswrapper[4776]: I1125 11:05:51.690191 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-85l79-config-nd5rg" event={"ID":"76880f11-d657-4560-9205-3e86b2845cfc","Type":"ContainerStarted","Data":"6478752a998200191ccc75ba8bbac28e78c153aa2bac4f02c428d64aa8be44fe"} Nov 25 11:05:51 crc kubenswrapper[4776]: I1125 11:05:51.713535 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-85l79-config-nd5rg" podStartSLOduration=3.713514105 podStartE2EDuration="3.713514105s" podCreationTimestamp="2025-11-25 11:05:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:05:51.709772602 +0000 UTC m=+6096.750832155" watchObservedRunningTime="2025-11-25 11:05:51.713514105 +0000 UTC m=+6096.754573658" Nov 25 11:05:52 crc kubenswrapper[4776]: I1125 11:05:52.703441 4776 generic.go:334] "Generic (PLEG): container finished" podID="5cee8213-d6c2-4b71-a4f5-ee4baa239c4c" containerID="6fb82a453b0f536e3d172688f301b919bdc97263ee56a595fa3d5bf2f1a847df" exitCode=0 Nov 25 11:05:52 crc kubenswrapper[4776]: I1125 11:05:52.703527 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-dccc45cc4-75rpv" event={"ID":"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c","Type":"ContainerDied","Data":"6fb82a453b0f536e3d172688f301b919bdc97263ee56a595fa3d5bf2f1a847df"} Nov 25 11:05:52 crc kubenswrapper[4776]: I1125 11:05:52.707109 4776 generic.go:334] "Generic (PLEG): container finished" podID="76880f11-d657-4560-9205-3e86b2845cfc" containerID="6478752a998200191ccc75ba8bbac28e78c153aa2bac4f02c428d64aa8be44fe" exitCode=0 Nov 25 11:05:52 crc kubenswrapper[4776]: I1125 11:05:52.707174 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-85l79-config-nd5rg" event={"ID":"76880f11-d657-4560-9205-3e86b2845cfc","Type":"ContainerDied","Data":"6478752a998200191ccc75ba8bbac28e78c153aa2bac4f02c428d64aa8be44fe"} Nov 25 11:05:53 crc kubenswrapper[4776]: I1125 11:05:53.719373 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-dccc45cc4-75rpv" event={"ID":"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c","Type":"ContainerStarted","Data":"45a105694c088eee3d549f74c1f7545507c3146be8cbc5bca8fba2cd43af9c58"} Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.116087 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-85l79-config-nd5rg" Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.130183 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/76880f11-d657-4560-9205-3e86b2845cfc-additional-scripts\") pod \"76880f11-d657-4560-9205-3e86b2845cfc\" (UID: \"76880f11-d657-4560-9205-3e86b2845cfc\") " Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.130267 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/76880f11-d657-4560-9205-3e86b2845cfc-var-run\") pod \"76880f11-d657-4560-9205-3e86b2845cfc\" (UID: \"76880f11-d657-4560-9205-3e86b2845cfc\") " Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.130307 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76880f11-d657-4560-9205-3e86b2845cfc-scripts\") pod \"76880f11-d657-4560-9205-3e86b2845cfc\" (UID: \"76880f11-d657-4560-9205-3e86b2845cfc\") " Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.130524 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hr29\" (UniqueName: \"kubernetes.io/projected/76880f11-d657-4560-9205-3e86b2845cfc-kube-api-access-6hr29\") pod \"76880f11-d657-4560-9205-3e86b2845cfc\" (UID: \"76880f11-d657-4560-9205-3e86b2845cfc\") " Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.130589 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/76880f11-d657-4560-9205-3e86b2845cfc-var-run-ovn\") pod \"76880f11-d657-4560-9205-3e86b2845cfc\" (UID: \"76880f11-d657-4560-9205-3e86b2845cfc\") " Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.130633 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/76880f11-d657-4560-9205-3e86b2845cfc-var-log-ovn\") pod \"76880f11-d657-4560-9205-3e86b2845cfc\" (UID: \"76880f11-d657-4560-9205-3e86b2845cfc\") " Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.131312 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/76880f11-d657-4560-9205-3e86b2845cfc-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "76880f11-d657-4560-9205-3e86b2845cfc" (UID: "76880f11-d657-4560-9205-3e86b2845cfc"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.131359 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/76880f11-d657-4560-9205-3e86b2845cfc-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "76880f11-d657-4560-9205-3e86b2845cfc" (UID: "76880f11-d657-4560-9205-3e86b2845cfc"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.131376 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/76880f11-d657-4560-9205-3e86b2845cfc-var-run" (OuterVolumeSpecName: "var-run") pod "76880f11-d657-4560-9205-3e86b2845cfc" (UID: "76880f11-d657-4560-9205-3e86b2845cfc"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.131581 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76880f11-d657-4560-9205-3e86b2845cfc-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "76880f11-d657-4560-9205-3e86b2845cfc" (UID: "76880f11-d657-4560-9205-3e86b2845cfc"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.132412 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76880f11-d657-4560-9205-3e86b2845cfc-scripts" (OuterVolumeSpecName: "scripts") pod "76880f11-d657-4560-9205-3e86b2845cfc" (UID: "76880f11-d657-4560-9205-3e86b2845cfc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.135990 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76880f11-d657-4560-9205-3e86b2845cfc-kube-api-access-6hr29" (OuterVolumeSpecName: "kube-api-access-6hr29") pod "76880f11-d657-4560-9205-3e86b2845cfc" (UID: "76880f11-d657-4560-9205-3e86b2845cfc"). InnerVolumeSpecName "kube-api-access-6hr29". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.232243 4776 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/76880f11-d657-4560-9205-3e86b2845cfc-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.232513 4776 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/76880f11-d657-4560-9205-3e86b2845cfc-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.232593 4776 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/76880f11-d657-4560-9205-3e86b2845cfc-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.232663 4776 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/76880f11-d657-4560-9205-3e86b2845cfc-var-run\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.232723 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76880f11-d657-4560-9205-3e86b2845cfc-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.232778 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hr29\" (UniqueName: \"kubernetes.io/projected/76880f11-d657-4560-9205-3e86b2845cfc-kube-api-access-6hr29\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.730460 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-dccc45cc4-75rpv" event={"ID":"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c","Type":"ContainerStarted","Data":"6e06d480bd5fa8bdf071006149cf24c73dcf92f43827208ae442911ea39c79e2"} Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.731636 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-dccc45cc4-75rpv" Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.731730 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-dccc45cc4-75rpv" Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.736395 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-85l79-config-nd5rg" event={"ID":"76880f11-d657-4560-9205-3e86b2845cfc","Type":"ContainerDied","Data":"23abe12e3116f8ecdbdf6fa1851c34e346ba351c96fce48b3a563d76f8a9a2a9"} Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.736490 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-85l79-config-nd5rg" Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.736640 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23abe12e3116f8ecdbdf6fa1851c34e346ba351c96fce48b3a563d76f8a9a2a9" Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.776314 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-api-dccc45cc4-75rpv" podStartSLOduration=4.322788109 podStartE2EDuration="19.776292291s" podCreationTimestamp="2025-11-25 11:05:35 +0000 UTC" firstStartedPulling="2025-11-25 11:05:36.036938965 +0000 UTC m=+6081.077998518" lastFinishedPulling="2025-11-25 11:05:51.490443147 +0000 UTC m=+6096.531502700" observedRunningTime="2025-11-25 11:05:54.761957841 +0000 UTC m=+6099.803017394" watchObservedRunningTime="2025-11-25 11:05:54.776292291 +0000 UTC m=+6099.817351844" Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.809256 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-85l79-config-nd5rg"] Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.823572 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-85l79-config-nd5rg"] Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.907424 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-85l79-config-2bddm"] Nov 25 11:05:54 crc kubenswrapper[4776]: E1125 11:05:54.907787 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76880f11-d657-4560-9205-3e86b2845cfc" containerName="ovn-config" Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.907803 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="76880f11-d657-4560-9205-3e86b2845cfc" containerName="ovn-config" Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.908000 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="76880f11-d657-4560-9205-3e86b2845cfc" containerName="ovn-config" Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.908679 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-85l79-config-2bddm" Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.910956 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 25 11:05:54 crc kubenswrapper[4776]: I1125 11:05:54.920888 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-85l79-config-2bddm"] Nov 25 11:05:55 crc kubenswrapper[4776]: I1125 11:05:55.046654 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/067ad60b-a932-46e6-b247-5c5b700fdf7a-var-run\") pod \"ovn-controller-85l79-config-2bddm\" (UID: \"067ad60b-a932-46e6-b247-5c5b700fdf7a\") " pod="openstack/ovn-controller-85l79-config-2bddm" Nov 25 11:05:55 crc kubenswrapper[4776]: I1125 11:05:55.046964 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/067ad60b-a932-46e6-b247-5c5b700fdf7a-additional-scripts\") pod \"ovn-controller-85l79-config-2bddm\" (UID: \"067ad60b-a932-46e6-b247-5c5b700fdf7a\") " pod="openstack/ovn-controller-85l79-config-2bddm" Nov 25 11:05:55 crc kubenswrapper[4776]: I1125 11:05:55.047197 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/067ad60b-a932-46e6-b247-5c5b700fdf7a-scripts\") pod \"ovn-controller-85l79-config-2bddm\" (UID: \"067ad60b-a932-46e6-b247-5c5b700fdf7a\") " pod="openstack/ovn-controller-85l79-config-2bddm" Nov 25 11:05:55 crc kubenswrapper[4776]: I1125 11:05:55.047414 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/067ad60b-a932-46e6-b247-5c5b700fdf7a-var-log-ovn\") pod \"ovn-controller-85l79-config-2bddm\" (UID: \"067ad60b-a932-46e6-b247-5c5b700fdf7a\") " pod="openstack/ovn-controller-85l79-config-2bddm" Nov 25 11:05:55 crc kubenswrapper[4776]: I1125 11:05:55.047485 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/067ad60b-a932-46e6-b247-5c5b700fdf7a-var-run-ovn\") pod \"ovn-controller-85l79-config-2bddm\" (UID: \"067ad60b-a932-46e6-b247-5c5b700fdf7a\") " pod="openstack/ovn-controller-85l79-config-2bddm" Nov 25 11:05:55 crc kubenswrapper[4776]: I1125 11:05:55.047591 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhlxs\" (UniqueName: \"kubernetes.io/projected/067ad60b-a932-46e6-b247-5c5b700fdf7a-kube-api-access-xhlxs\") pod \"ovn-controller-85l79-config-2bddm\" (UID: \"067ad60b-a932-46e6-b247-5c5b700fdf7a\") " pod="openstack/ovn-controller-85l79-config-2bddm" Nov 25 11:05:55 crc kubenswrapper[4776]: I1125 11:05:55.149269 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/067ad60b-a932-46e6-b247-5c5b700fdf7a-var-run\") pod \"ovn-controller-85l79-config-2bddm\" (UID: \"067ad60b-a932-46e6-b247-5c5b700fdf7a\") " pod="openstack/ovn-controller-85l79-config-2bddm" Nov 25 11:05:55 crc kubenswrapper[4776]: I1125 11:05:55.149546 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/067ad60b-a932-46e6-b247-5c5b700fdf7a-additional-scripts\") pod \"ovn-controller-85l79-config-2bddm\" (UID: \"067ad60b-a932-46e6-b247-5c5b700fdf7a\") " pod="openstack/ovn-controller-85l79-config-2bddm" Nov 25 11:05:55 crc kubenswrapper[4776]: I1125 11:05:55.149560 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/067ad60b-a932-46e6-b247-5c5b700fdf7a-var-run\") pod \"ovn-controller-85l79-config-2bddm\" (UID: \"067ad60b-a932-46e6-b247-5c5b700fdf7a\") " pod="openstack/ovn-controller-85l79-config-2bddm" Nov 25 11:05:55 crc kubenswrapper[4776]: I1125 11:05:55.149572 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/067ad60b-a932-46e6-b247-5c5b700fdf7a-scripts\") pod \"ovn-controller-85l79-config-2bddm\" (UID: \"067ad60b-a932-46e6-b247-5c5b700fdf7a\") " pod="openstack/ovn-controller-85l79-config-2bddm" Nov 25 11:05:55 crc kubenswrapper[4776]: I1125 11:05:55.149760 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/067ad60b-a932-46e6-b247-5c5b700fdf7a-var-log-ovn\") pod \"ovn-controller-85l79-config-2bddm\" (UID: \"067ad60b-a932-46e6-b247-5c5b700fdf7a\") " pod="openstack/ovn-controller-85l79-config-2bddm" Nov 25 11:05:55 crc kubenswrapper[4776]: I1125 11:05:55.149811 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/067ad60b-a932-46e6-b247-5c5b700fdf7a-var-run-ovn\") pod \"ovn-controller-85l79-config-2bddm\" (UID: \"067ad60b-a932-46e6-b247-5c5b700fdf7a\") " pod="openstack/ovn-controller-85l79-config-2bddm" Nov 25 11:05:55 crc kubenswrapper[4776]: I1125 11:05:55.149893 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhlxs\" (UniqueName: \"kubernetes.io/projected/067ad60b-a932-46e6-b247-5c5b700fdf7a-kube-api-access-xhlxs\") pod \"ovn-controller-85l79-config-2bddm\" (UID: \"067ad60b-a932-46e6-b247-5c5b700fdf7a\") " pod="openstack/ovn-controller-85l79-config-2bddm" Nov 25 11:05:55 crc kubenswrapper[4776]: I1125 11:05:55.150190 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/067ad60b-a932-46e6-b247-5c5b700fdf7a-var-run-ovn\") pod \"ovn-controller-85l79-config-2bddm\" (UID: \"067ad60b-a932-46e6-b247-5c5b700fdf7a\") " pod="openstack/ovn-controller-85l79-config-2bddm" Nov 25 11:05:55 crc kubenswrapper[4776]: I1125 11:05:55.150276 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/067ad60b-a932-46e6-b247-5c5b700fdf7a-additional-scripts\") pod \"ovn-controller-85l79-config-2bddm\" (UID: \"067ad60b-a932-46e6-b247-5c5b700fdf7a\") " pod="openstack/ovn-controller-85l79-config-2bddm" Nov 25 11:05:55 crc kubenswrapper[4776]: I1125 11:05:55.150385 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/067ad60b-a932-46e6-b247-5c5b700fdf7a-var-log-ovn\") pod \"ovn-controller-85l79-config-2bddm\" (UID: \"067ad60b-a932-46e6-b247-5c5b700fdf7a\") " pod="openstack/ovn-controller-85l79-config-2bddm" Nov 25 11:05:55 crc kubenswrapper[4776]: I1125 11:05:55.151447 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/067ad60b-a932-46e6-b247-5c5b700fdf7a-scripts\") pod \"ovn-controller-85l79-config-2bddm\" (UID: \"067ad60b-a932-46e6-b247-5c5b700fdf7a\") " pod="openstack/ovn-controller-85l79-config-2bddm" Nov 25 11:05:55 crc kubenswrapper[4776]: I1125 11:05:55.169012 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhlxs\" (UniqueName: \"kubernetes.io/projected/067ad60b-a932-46e6-b247-5c5b700fdf7a-kube-api-access-xhlxs\") pod \"ovn-controller-85l79-config-2bddm\" (UID: \"067ad60b-a932-46e6-b247-5c5b700fdf7a\") " pod="openstack/ovn-controller-85l79-config-2bddm" Nov 25 11:05:55 crc kubenswrapper[4776]: I1125 11:05:55.225163 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-85l79-config-2bddm" Nov 25 11:05:55 crc kubenswrapper[4776]: I1125 11:05:55.681671 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76880f11-d657-4560-9205-3e86b2845cfc" path="/var/lib/kubelet/pods/76880f11-d657-4560-9205-3e86b2845cfc/volumes" Nov 25 11:05:55 crc kubenswrapper[4776]: I1125 11:05:55.687457 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-85l79-config-2bddm"] Nov 25 11:05:55 crc kubenswrapper[4776]: I1125 11:05:55.749459 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-85l79-config-2bddm" event={"ID":"067ad60b-a932-46e6-b247-5c5b700fdf7a","Type":"ContainerStarted","Data":"177b8969104a79b38938ea6d69130183403fd02fe41bc1768d0808cb659d3bfc"} Nov 25 11:05:56 crc kubenswrapper[4776]: I1125 11:05:56.758010 4776 generic.go:334] "Generic (PLEG): container finished" podID="067ad60b-a932-46e6-b247-5c5b700fdf7a" containerID="0bead501d42c1f9defd4db9ce9dbd117c8813e94be8fe3f158d2f400855361cd" exitCode=0 Nov 25 11:05:56 crc kubenswrapper[4776]: I1125 11:05:56.758264 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-85l79-config-2bddm" event={"ID":"067ad60b-a932-46e6-b247-5c5b700fdf7a","Type":"ContainerDied","Data":"0bead501d42c1f9defd4db9ce9dbd117c8813e94be8fe3f158d2f400855361cd"} Nov 25 11:05:57 crc kubenswrapper[4776]: I1125 11:05:57.662652 4776 scope.go:117] "RemoveContainer" containerID="8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698" Nov 25 11:05:57 crc kubenswrapper[4776]: E1125 11:05:57.663465 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:05:58 crc kubenswrapper[4776]: I1125 11:05:58.150961 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-85l79-config-2bddm" Nov 25 11:05:58 crc kubenswrapper[4776]: I1125 11:05:58.315734 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/067ad60b-a932-46e6-b247-5c5b700fdf7a-additional-scripts\") pod \"067ad60b-a932-46e6-b247-5c5b700fdf7a\" (UID: \"067ad60b-a932-46e6-b247-5c5b700fdf7a\") " Nov 25 11:05:58 crc kubenswrapper[4776]: I1125 11:05:58.315968 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhlxs\" (UniqueName: \"kubernetes.io/projected/067ad60b-a932-46e6-b247-5c5b700fdf7a-kube-api-access-xhlxs\") pod \"067ad60b-a932-46e6-b247-5c5b700fdf7a\" (UID: \"067ad60b-a932-46e6-b247-5c5b700fdf7a\") " Nov 25 11:05:58 crc kubenswrapper[4776]: I1125 11:05:58.316170 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/067ad60b-a932-46e6-b247-5c5b700fdf7a-var-run\") pod \"067ad60b-a932-46e6-b247-5c5b700fdf7a\" (UID: \"067ad60b-a932-46e6-b247-5c5b700fdf7a\") " Nov 25 11:05:58 crc kubenswrapper[4776]: I1125 11:05:58.316204 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/067ad60b-a932-46e6-b247-5c5b700fdf7a-var-run-ovn\") pod \"067ad60b-a932-46e6-b247-5c5b700fdf7a\" (UID: \"067ad60b-a932-46e6-b247-5c5b700fdf7a\") " Nov 25 11:05:58 crc kubenswrapper[4776]: I1125 11:05:58.316280 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/067ad60b-a932-46e6-b247-5c5b700fdf7a-var-log-ovn\") pod \"067ad60b-a932-46e6-b247-5c5b700fdf7a\" (UID: \"067ad60b-a932-46e6-b247-5c5b700fdf7a\") " Nov 25 11:05:58 crc kubenswrapper[4776]: I1125 11:05:58.316323 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/067ad60b-a932-46e6-b247-5c5b700fdf7a-scripts\") pod \"067ad60b-a932-46e6-b247-5c5b700fdf7a\" (UID: \"067ad60b-a932-46e6-b247-5c5b700fdf7a\") " Nov 25 11:05:58 crc kubenswrapper[4776]: I1125 11:05:58.316312 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/067ad60b-a932-46e6-b247-5c5b700fdf7a-var-run" (OuterVolumeSpecName: "var-run") pod "067ad60b-a932-46e6-b247-5c5b700fdf7a" (UID: "067ad60b-a932-46e6-b247-5c5b700fdf7a"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 11:05:58 crc kubenswrapper[4776]: I1125 11:05:58.316390 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/067ad60b-a932-46e6-b247-5c5b700fdf7a-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "067ad60b-a932-46e6-b247-5c5b700fdf7a" (UID: "067ad60b-a932-46e6-b247-5c5b700fdf7a"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 11:05:58 crc kubenswrapper[4776]: I1125 11:05:58.316384 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/067ad60b-a932-46e6-b247-5c5b700fdf7a-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "067ad60b-a932-46e6-b247-5c5b700fdf7a" (UID: "067ad60b-a932-46e6-b247-5c5b700fdf7a"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 11:05:58 crc kubenswrapper[4776]: I1125 11:05:58.316740 4776 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/067ad60b-a932-46e6-b247-5c5b700fdf7a-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:58 crc kubenswrapper[4776]: I1125 11:05:58.316756 4776 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/067ad60b-a932-46e6-b247-5c5b700fdf7a-var-run\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:58 crc kubenswrapper[4776]: I1125 11:05:58.316765 4776 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/067ad60b-a932-46e6-b247-5c5b700fdf7a-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:58 crc kubenswrapper[4776]: I1125 11:05:58.317197 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/067ad60b-a932-46e6-b247-5c5b700fdf7a-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "067ad60b-a932-46e6-b247-5c5b700fdf7a" (UID: "067ad60b-a932-46e6-b247-5c5b700fdf7a"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:05:58 crc kubenswrapper[4776]: I1125 11:05:58.317418 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/067ad60b-a932-46e6-b247-5c5b700fdf7a-scripts" (OuterVolumeSpecName: "scripts") pod "067ad60b-a932-46e6-b247-5c5b700fdf7a" (UID: "067ad60b-a932-46e6-b247-5c5b700fdf7a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:05:58 crc kubenswrapper[4776]: I1125 11:05:58.418537 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/067ad60b-a932-46e6-b247-5c5b700fdf7a-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:58 crc kubenswrapper[4776]: I1125 11:05:58.418904 4776 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/067ad60b-a932-46e6-b247-5c5b700fdf7a-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:58 crc kubenswrapper[4776]: I1125 11:05:58.544491 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/067ad60b-a932-46e6-b247-5c5b700fdf7a-kube-api-access-xhlxs" (OuterVolumeSpecName: "kube-api-access-xhlxs") pod "067ad60b-a932-46e6-b247-5c5b700fdf7a" (UID: "067ad60b-a932-46e6-b247-5c5b700fdf7a"). InnerVolumeSpecName "kube-api-access-xhlxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:05:58 crc kubenswrapper[4776]: I1125 11:05:58.621764 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhlxs\" (UniqueName: \"kubernetes.io/projected/067ad60b-a932-46e6-b247-5c5b700fdf7a-kube-api-access-xhlxs\") on node \"crc\" DevicePath \"\"" Nov 25 11:05:58 crc kubenswrapper[4776]: I1125 11:05:58.775452 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-85l79-config-2bddm" event={"ID":"067ad60b-a932-46e6-b247-5c5b700fdf7a","Type":"ContainerDied","Data":"177b8969104a79b38938ea6d69130183403fd02fe41bc1768d0808cb659d3bfc"} Nov 25 11:05:58 crc kubenswrapper[4776]: I1125 11:05:58.775505 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="177b8969104a79b38938ea6d69130183403fd02fe41bc1768d0808cb659d3bfc" Nov 25 11:05:58 crc kubenswrapper[4776]: I1125 11:05:58.775540 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-85l79-config-2bddm" Nov 25 11:05:59 crc kubenswrapper[4776]: I1125 11:05:59.239707 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-85l79-config-2bddm"] Nov 25 11:05:59 crc kubenswrapper[4776]: I1125 11:05:59.250509 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-85l79-config-2bddm"] Nov 25 11:05:59 crc kubenswrapper[4776]: I1125 11:05:59.674004 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="067ad60b-a932-46e6-b247-5c5b700fdf7a" path="/var/lib/kubelet/pods/067ad60b-a932-46e6-b247-5c5b700fdf7a/volumes" Nov 25 11:06:09 crc kubenswrapper[4776]: I1125 11:06:09.757876 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-dccc45cc4-75rpv" Nov 25 11:06:09 crc kubenswrapper[4776]: I1125 11:06:09.788014 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-dccc45cc4-75rpv" Nov 25 11:06:11 crc kubenswrapper[4776]: I1125 11:06:11.670346 4776 scope.go:117] "RemoveContainer" containerID="8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698" Nov 25 11:06:11 crc kubenswrapper[4776]: E1125 11:06:11.671642 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.042271 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-rsyslog-nbkx7"] Nov 25 11:06:22 crc kubenswrapper[4776]: E1125 11:06:22.043331 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="067ad60b-a932-46e6-b247-5c5b700fdf7a" containerName="ovn-config" Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.043352 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="067ad60b-a932-46e6-b247-5c5b700fdf7a" containerName="ovn-config" Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.043657 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="067ad60b-a932-46e6-b247-5c5b700fdf7a" containerName="ovn-config" Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.044965 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-nbkx7" Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.047374 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-config-data" Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.047430 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-scripts" Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.047606 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"octavia-hmport-map" Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.082836 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-nbkx7"] Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.178674 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e7c13e5-cbf7-408b-8a3c-edfc1d208c12-config-data\") pod \"octavia-rsyslog-nbkx7\" (UID: \"2e7c13e5-cbf7-408b-8a3c-edfc1d208c12\") " pod="openstack/octavia-rsyslog-nbkx7" Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.178829 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/2e7c13e5-cbf7-408b-8a3c-edfc1d208c12-hm-ports\") pod \"octavia-rsyslog-nbkx7\" (UID: \"2e7c13e5-cbf7-408b-8a3c-edfc1d208c12\") " pod="openstack/octavia-rsyslog-nbkx7" Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.178874 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/2e7c13e5-cbf7-408b-8a3c-edfc1d208c12-config-data-merged\") pod \"octavia-rsyslog-nbkx7\" (UID: \"2e7c13e5-cbf7-408b-8a3c-edfc1d208c12\") " pod="openstack/octavia-rsyslog-nbkx7" Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.178909 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e7c13e5-cbf7-408b-8a3c-edfc1d208c12-scripts\") pod \"octavia-rsyslog-nbkx7\" (UID: \"2e7c13e5-cbf7-408b-8a3c-edfc1d208c12\") " pod="openstack/octavia-rsyslog-nbkx7" Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.280190 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e7c13e5-cbf7-408b-8a3c-edfc1d208c12-scripts\") pod \"octavia-rsyslog-nbkx7\" (UID: \"2e7c13e5-cbf7-408b-8a3c-edfc1d208c12\") " pod="openstack/octavia-rsyslog-nbkx7" Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.280343 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e7c13e5-cbf7-408b-8a3c-edfc1d208c12-config-data\") pod \"octavia-rsyslog-nbkx7\" (UID: \"2e7c13e5-cbf7-408b-8a3c-edfc1d208c12\") " pod="openstack/octavia-rsyslog-nbkx7" Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.280405 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/2e7c13e5-cbf7-408b-8a3c-edfc1d208c12-hm-ports\") pod \"octavia-rsyslog-nbkx7\" (UID: \"2e7c13e5-cbf7-408b-8a3c-edfc1d208c12\") " pod="openstack/octavia-rsyslog-nbkx7" Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.280451 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/2e7c13e5-cbf7-408b-8a3c-edfc1d208c12-config-data-merged\") pod \"octavia-rsyslog-nbkx7\" (UID: \"2e7c13e5-cbf7-408b-8a3c-edfc1d208c12\") " pod="openstack/octavia-rsyslog-nbkx7" Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.281315 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/2e7c13e5-cbf7-408b-8a3c-edfc1d208c12-config-data-merged\") pod \"octavia-rsyslog-nbkx7\" (UID: \"2e7c13e5-cbf7-408b-8a3c-edfc1d208c12\") " pod="openstack/octavia-rsyslog-nbkx7" Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.281765 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/2e7c13e5-cbf7-408b-8a3c-edfc1d208c12-hm-ports\") pod \"octavia-rsyslog-nbkx7\" (UID: \"2e7c13e5-cbf7-408b-8a3c-edfc1d208c12\") " pod="openstack/octavia-rsyslog-nbkx7" Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.286059 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e7c13e5-cbf7-408b-8a3c-edfc1d208c12-config-data\") pod \"octavia-rsyslog-nbkx7\" (UID: \"2e7c13e5-cbf7-408b-8a3c-edfc1d208c12\") " pod="openstack/octavia-rsyslog-nbkx7" Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.287476 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e7c13e5-cbf7-408b-8a3c-edfc1d208c12-scripts\") pod \"octavia-rsyslog-nbkx7\" (UID: \"2e7c13e5-cbf7-408b-8a3c-edfc1d208c12\") " pod="openstack/octavia-rsyslog-nbkx7" Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.381170 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-nbkx7" Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.673405 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-5955f5554b-ggjdr"] Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.680439 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-5955f5554b-ggjdr" Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.685203 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.688528 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-5955f5554b-ggjdr"] Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.799461 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/83f8c52f-018a-481f-baec-1a4a3df56aec-amphora-image\") pod \"octavia-image-upload-5955f5554b-ggjdr\" (UID: \"83f8c52f-018a-481f-baec-1a4a3df56aec\") " pod="openstack/octavia-image-upload-5955f5554b-ggjdr" Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.799519 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/83f8c52f-018a-481f-baec-1a4a3df56aec-httpd-config\") pod \"octavia-image-upload-5955f5554b-ggjdr\" (UID: \"83f8c52f-018a-481f-baec-1a4a3df56aec\") " pod="openstack/octavia-image-upload-5955f5554b-ggjdr" Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.901728 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/83f8c52f-018a-481f-baec-1a4a3df56aec-amphora-image\") pod \"octavia-image-upload-5955f5554b-ggjdr\" (UID: \"83f8c52f-018a-481f-baec-1a4a3df56aec\") " pod="openstack/octavia-image-upload-5955f5554b-ggjdr" Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.902157 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/83f8c52f-018a-481f-baec-1a4a3df56aec-httpd-config\") pod \"octavia-image-upload-5955f5554b-ggjdr\" (UID: \"83f8c52f-018a-481f-baec-1a4a3df56aec\") " pod="openstack/octavia-image-upload-5955f5554b-ggjdr" Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.902607 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/83f8c52f-018a-481f-baec-1a4a3df56aec-amphora-image\") pod \"octavia-image-upload-5955f5554b-ggjdr\" (UID: \"83f8c52f-018a-481f-baec-1a4a3df56aec\") " pod="openstack/octavia-image-upload-5955f5554b-ggjdr" Nov 25 11:06:22 crc kubenswrapper[4776]: I1125 11:06:22.908296 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/83f8c52f-018a-481f-baec-1a4a3df56aec-httpd-config\") pod \"octavia-image-upload-5955f5554b-ggjdr\" (UID: \"83f8c52f-018a-481f-baec-1a4a3df56aec\") " pod="openstack/octavia-image-upload-5955f5554b-ggjdr" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.015968 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-5955f5554b-ggjdr" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.539421 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-5955f5554b-ggjdr"] Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.587363 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-api-c694cd6fb-994xl"] Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.590743 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.596827 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-octavia-public-svc" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.597963 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-octavia-internal-svc" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.606933 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-c694cd6fb-994xl"] Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.649655 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-nbkx7"] Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.718559 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21522877-e085-4ff3-857d-edb4989664fd-internal-tls-certs\") pod \"octavia-api-c694cd6fb-994xl\" (UID: \"21522877-e085-4ff3-857d-edb4989664fd\") " pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.718886 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21522877-e085-4ff3-857d-edb4989664fd-config-data\") pod \"octavia-api-c694cd6fb-994xl\" (UID: \"21522877-e085-4ff3-857d-edb4989664fd\") " pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.718950 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21522877-e085-4ff3-857d-edb4989664fd-public-tls-certs\") pod \"octavia-api-c694cd6fb-994xl\" (UID: \"21522877-e085-4ff3-857d-edb4989664fd\") " pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.718997 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21522877-e085-4ff3-857d-edb4989664fd-combined-ca-bundle\") pod \"octavia-api-c694cd6fb-994xl\" (UID: \"21522877-e085-4ff3-857d-edb4989664fd\") " pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.719026 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/21522877-e085-4ff3-857d-edb4989664fd-octavia-run\") pod \"octavia-api-c694cd6fb-994xl\" (UID: \"21522877-e085-4ff3-857d-edb4989664fd\") " pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.719219 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/21522877-e085-4ff3-857d-edb4989664fd-ovndb-tls-certs\") pod \"octavia-api-c694cd6fb-994xl\" (UID: \"21522877-e085-4ff3-857d-edb4989664fd\") " pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.719389 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21522877-e085-4ff3-857d-edb4989664fd-scripts\") pod \"octavia-api-c694cd6fb-994xl\" (UID: \"21522877-e085-4ff3-857d-edb4989664fd\") " pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.719631 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/21522877-e085-4ff3-857d-edb4989664fd-config-data-merged\") pod \"octavia-api-c694cd6fb-994xl\" (UID: \"21522877-e085-4ff3-857d-edb4989664fd\") " pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.821608 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/21522877-e085-4ff3-857d-edb4989664fd-octavia-run\") pod \"octavia-api-c694cd6fb-994xl\" (UID: \"21522877-e085-4ff3-857d-edb4989664fd\") " pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.821685 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/21522877-e085-4ff3-857d-edb4989664fd-ovndb-tls-certs\") pod \"octavia-api-c694cd6fb-994xl\" (UID: \"21522877-e085-4ff3-857d-edb4989664fd\") " pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.821723 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21522877-e085-4ff3-857d-edb4989664fd-scripts\") pod \"octavia-api-c694cd6fb-994xl\" (UID: \"21522877-e085-4ff3-857d-edb4989664fd\") " pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.821791 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/21522877-e085-4ff3-857d-edb4989664fd-config-data-merged\") pod \"octavia-api-c694cd6fb-994xl\" (UID: \"21522877-e085-4ff3-857d-edb4989664fd\") " pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.821858 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21522877-e085-4ff3-857d-edb4989664fd-internal-tls-certs\") pod \"octavia-api-c694cd6fb-994xl\" (UID: \"21522877-e085-4ff3-857d-edb4989664fd\") " pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.822375 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21522877-e085-4ff3-857d-edb4989664fd-config-data\") pod \"octavia-api-c694cd6fb-994xl\" (UID: \"21522877-e085-4ff3-857d-edb4989664fd\") " pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.822582 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21522877-e085-4ff3-857d-edb4989664fd-public-tls-certs\") pod \"octavia-api-c694cd6fb-994xl\" (UID: \"21522877-e085-4ff3-857d-edb4989664fd\") " pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.822747 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21522877-e085-4ff3-857d-edb4989664fd-combined-ca-bundle\") pod \"octavia-api-c694cd6fb-994xl\" (UID: \"21522877-e085-4ff3-857d-edb4989664fd\") " pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.822750 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/21522877-e085-4ff3-857d-edb4989664fd-config-data-merged\") pod \"octavia-api-c694cd6fb-994xl\" (UID: \"21522877-e085-4ff3-857d-edb4989664fd\") " pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.823251 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/21522877-e085-4ff3-857d-edb4989664fd-octavia-run\") pod \"octavia-api-c694cd6fb-994xl\" (UID: \"21522877-e085-4ff3-857d-edb4989664fd\") " pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.829038 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21522877-e085-4ff3-857d-edb4989664fd-internal-tls-certs\") pod \"octavia-api-c694cd6fb-994xl\" (UID: \"21522877-e085-4ff3-857d-edb4989664fd\") " pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.829896 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21522877-e085-4ff3-857d-edb4989664fd-combined-ca-bundle\") pod \"octavia-api-c694cd6fb-994xl\" (UID: \"21522877-e085-4ff3-857d-edb4989664fd\") " pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.829928 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21522877-e085-4ff3-857d-edb4989664fd-public-tls-certs\") pod \"octavia-api-c694cd6fb-994xl\" (UID: \"21522877-e085-4ff3-857d-edb4989664fd\") " pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.831671 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/21522877-e085-4ff3-857d-edb4989664fd-ovndb-tls-certs\") pod \"octavia-api-c694cd6fb-994xl\" (UID: \"21522877-e085-4ff3-857d-edb4989664fd\") " pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.832108 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21522877-e085-4ff3-857d-edb4989664fd-config-data\") pod \"octavia-api-c694cd6fb-994xl\" (UID: \"21522877-e085-4ff3-857d-edb4989664fd\") " pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.835341 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21522877-e085-4ff3-857d-edb4989664fd-scripts\") pod \"octavia-api-c694cd6fb-994xl\" (UID: \"21522877-e085-4ff3-857d-edb4989664fd\") " pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:23 crc kubenswrapper[4776]: I1125 11:06:23.982058 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:24 crc kubenswrapper[4776]: I1125 11:06:24.104342 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-5955f5554b-ggjdr" event={"ID":"83f8c52f-018a-481f-baec-1a4a3df56aec","Type":"ContainerStarted","Data":"10b19c718e7232e0268afa29a2784ac8bdbad0651d17b815aeca9aaf3932d9a1"} Nov 25 11:06:24 crc kubenswrapper[4776]: I1125 11:06:24.106733 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-nbkx7" event={"ID":"2e7c13e5-cbf7-408b-8a3c-edfc1d208c12","Type":"ContainerStarted","Data":"9e6307184d2b5343779725d83ac6bb871e2e87215e04485847e83e7830768a5b"} Nov 25 11:06:24 crc kubenswrapper[4776]: I1125 11:06:24.472927 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-c694cd6fb-994xl"] Nov 25 11:06:24 crc kubenswrapper[4776]: W1125 11:06:24.494286 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21522877_e085_4ff3_857d_edb4989664fd.slice/crio-0fd10d8f9c0e52f358ec7a38bb8fb0dd508af3f45430889b07a47a8ea0ceb280 WatchSource:0}: Error finding container 0fd10d8f9c0e52f358ec7a38bb8fb0dd508af3f45430889b07a47a8ea0ceb280: Status 404 returned error can't find the container with id 0fd10d8f9c0e52f358ec7a38bb8fb0dd508af3f45430889b07a47a8ea0ceb280 Nov 25 11:06:24 crc kubenswrapper[4776]: I1125 11:06:24.663185 4776 scope.go:117] "RemoveContainer" containerID="8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698" Nov 25 11:06:24 crc kubenswrapper[4776]: E1125 11:06:24.663472 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:06:25 crc kubenswrapper[4776]: I1125 11:06:25.141320 4776 generic.go:334] "Generic (PLEG): container finished" podID="21522877-e085-4ff3-857d-edb4989664fd" containerID="55897df930b5d59a6c95bf3b6dcfa90a2d537223e89e802d083c3792c526a3bb" exitCode=0 Nov 25 11:06:25 crc kubenswrapper[4776]: I1125 11:06:25.141389 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-c694cd6fb-994xl" event={"ID":"21522877-e085-4ff3-857d-edb4989664fd","Type":"ContainerDied","Data":"55897df930b5d59a6c95bf3b6dcfa90a2d537223e89e802d083c3792c526a3bb"} Nov 25 11:06:25 crc kubenswrapper[4776]: I1125 11:06:25.141716 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-c694cd6fb-994xl" event={"ID":"21522877-e085-4ff3-857d-edb4989664fd","Type":"ContainerStarted","Data":"0fd10d8f9c0e52f358ec7a38bb8fb0dd508af3f45430889b07a47a8ea0ceb280"} Nov 25 11:06:27 crc kubenswrapper[4776]: I1125 11:06:27.535152 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-healthmanager-z8qlp"] Nov 25 11:06:27 crc kubenswrapper[4776]: I1125 11:06:27.538262 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-z8qlp" Nov 25 11:06:27 crc kubenswrapper[4776]: I1125 11:06:27.544759 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-z8qlp"] Nov 25 11:06:27 crc kubenswrapper[4776]: I1125 11:06:27.545688 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-config-data" Nov 25 11:06:27 crc kubenswrapper[4776]: I1125 11:06:27.546027 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-scripts" Nov 25 11:06:27 crc kubenswrapper[4776]: I1125 11:06:27.546211 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-certs-secret" Nov 25 11:06:27 crc kubenswrapper[4776]: I1125 11:06:27.707149 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d017f1f2-3baf-46a7-9dcd-d361a86c51f6-scripts\") pod \"octavia-healthmanager-z8qlp\" (UID: \"d017f1f2-3baf-46a7-9dcd-d361a86c51f6\") " pod="openstack/octavia-healthmanager-z8qlp" Nov 25 11:06:27 crc kubenswrapper[4776]: I1125 11:06:27.707574 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/d017f1f2-3baf-46a7-9dcd-d361a86c51f6-amphora-certs\") pod \"octavia-healthmanager-z8qlp\" (UID: \"d017f1f2-3baf-46a7-9dcd-d361a86c51f6\") " pod="openstack/octavia-healthmanager-z8qlp" Nov 25 11:06:27 crc kubenswrapper[4776]: I1125 11:06:27.707612 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/d017f1f2-3baf-46a7-9dcd-d361a86c51f6-hm-ports\") pod \"octavia-healthmanager-z8qlp\" (UID: \"d017f1f2-3baf-46a7-9dcd-d361a86c51f6\") " pod="openstack/octavia-healthmanager-z8qlp" Nov 25 11:06:27 crc kubenswrapper[4776]: I1125 11:06:27.707637 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/d017f1f2-3baf-46a7-9dcd-d361a86c51f6-config-data-merged\") pod \"octavia-healthmanager-z8qlp\" (UID: \"d017f1f2-3baf-46a7-9dcd-d361a86c51f6\") " pod="openstack/octavia-healthmanager-z8qlp" Nov 25 11:06:27 crc kubenswrapper[4776]: I1125 11:06:27.707683 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d017f1f2-3baf-46a7-9dcd-d361a86c51f6-combined-ca-bundle\") pod \"octavia-healthmanager-z8qlp\" (UID: \"d017f1f2-3baf-46a7-9dcd-d361a86c51f6\") " pod="openstack/octavia-healthmanager-z8qlp" Nov 25 11:06:27 crc kubenswrapper[4776]: I1125 11:06:27.707715 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d017f1f2-3baf-46a7-9dcd-d361a86c51f6-config-data\") pod \"octavia-healthmanager-z8qlp\" (UID: \"d017f1f2-3baf-46a7-9dcd-d361a86c51f6\") " pod="openstack/octavia-healthmanager-z8qlp" Nov 25 11:06:27 crc kubenswrapper[4776]: I1125 11:06:27.810151 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d017f1f2-3baf-46a7-9dcd-d361a86c51f6-scripts\") pod \"octavia-healthmanager-z8qlp\" (UID: \"d017f1f2-3baf-46a7-9dcd-d361a86c51f6\") " pod="openstack/octavia-healthmanager-z8qlp" Nov 25 11:06:27 crc kubenswrapper[4776]: I1125 11:06:27.810240 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/d017f1f2-3baf-46a7-9dcd-d361a86c51f6-amphora-certs\") pod \"octavia-healthmanager-z8qlp\" (UID: \"d017f1f2-3baf-46a7-9dcd-d361a86c51f6\") " pod="openstack/octavia-healthmanager-z8qlp" Nov 25 11:06:27 crc kubenswrapper[4776]: I1125 11:06:27.810322 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/d017f1f2-3baf-46a7-9dcd-d361a86c51f6-hm-ports\") pod \"octavia-healthmanager-z8qlp\" (UID: \"d017f1f2-3baf-46a7-9dcd-d361a86c51f6\") " pod="openstack/octavia-healthmanager-z8qlp" Nov 25 11:06:27 crc kubenswrapper[4776]: I1125 11:06:27.810354 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/d017f1f2-3baf-46a7-9dcd-d361a86c51f6-config-data-merged\") pod \"octavia-healthmanager-z8qlp\" (UID: \"d017f1f2-3baf-46a7-9dcd-d361a86c51f6\") " pod="openstack/octavia-healthmanager-z8qlp" Nov 25 11:06:27 crc kubenswrapper[4776]: I1125 11:06:27.810390 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d017f1f2-3baf-46a7-9dcd-d361a86c51f6-combined-ca-bundle\") pod \"octavia-healthmanager-z8qlp\" (UID: \"d017f1f2-3baf-46a7-9dcd-d361a86c51f6\") " pod="openstack/octavia-healthmanager-z8qlp" Nov 25 11:06:27 crc kubenswrapper[4776]: I1125 11:06:27.810429 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d017f1f2-3baf-46a7-9dcd-d361a86c51f6-config-data\") pod \"octavia-healthmanager-z8qlp\" (UID: \"d017f1f2-3baf-46a7-9dcd-d361a86c51f6\") " pod="openstack/octavia-healthmanager-z8qlp" Nov 25 11:06:27 crc kubenswrapper[4776]: I1125 11:06:27.811404 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/d017f1f2-3baf-46a7-9dcd-d361a86c51f6-config-data-merged\") pod \"octavia-healthmanager-z8qlp\" (UID: \"d017f1f2-3baf-46a7-9dcd-d361a86c51f6\") " pod="openstack/octavia-healthmanager-z8qlp" Nov 25 11:06:27 crc kubenswrapper[4776]: I1125 11:06:27.811705 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/d017f1f2-3baf-46a7-9dcd-d361a86c51f6-hm-ports\") pod \"octavia-healthmanager-z8qlp\" (UID: \"d017f1f2-3baf-46a7-9dcd-d361a86c51f6\") " pod="openstack/octavia-healthmanager-z8qlp" Nov 25 11:06:27 crc kubenswrapper[4776]: I1125 11:06:27.817510 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/d017f1f2-3baf-46a7-9dcd-d361a86c51f6-amphora-certs\") pod \"octavia-healthmanager-z8qlp\" (UID: \"d017f1f2-3baf-46a7-9dcd-d361a86c51f6\") " pod="openstack/octavia-healthmanager-z8qlp" Nov 25 11:06:27 crc kubenswrapper[4776]: I1125 11:06:27.818279 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d017f1f2-3baf-46a7-9dcd-d361a86c51f6-scripts\") pod \"octavia-healthmanager-z8qlp\" (UID: \"d017f1f2-3baf-46a7-9dcd-d361a86c51f6\") " pod="openstack/octavia-healthmanager-z8qlp" Nov 25 11:06:27 crc kubenswrapper[4776]: I1125 11:06:27.818470 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d017f1f2-3baf-46a7-9dcd-d361a86c51f6-config-data\") pod \"octavia-healthmanager-z8qlp\" (UID: \"d017f1f2-3baf-46a7-9dcd-d361a86c51f6\") " pod="openstack/octavia-healthmanager-z8qlp" Nov 25 11:06:27 crc kubenswrapper[4776]: I1125 11:06:27.818940 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d017f1f2-3baf-46a7-9dcd-d361a86c51f6-combined-ca-bundle\") pod \"octavia-healthmanager-z8qlp\" (UID: \"d017f1f2-3baf-46a7-9dcd-d361a86c51f6\") " pod="openstack/octavia-healthmanager-z8qlp" Nov 25 11:06:28 crc kubenswrapper[4776]: I1125 11:06:28.025818 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-z8qlp" Nov 25 11:06:28 crc kubenswrapper[4776]: I1125 11:06:28.176442 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-c694cd6fb-994xl" event={"ID":"21522877-e085-4ff3-857d-edb4989664fd","Type":"ContainerStarted","Data":"40dc010018bc90fb9d4a6ba6a08d9226bfb49a2040482fa4f187de9882d00c1d"} Nov 25 11:06:29 crc kubenswrapper[4776]: I1125 11:06:29.186171 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-c694cd6fb-994xl" event={"ID":"21522877-e085-4ff3-857d-edb4989664fd","Type":"ContainerStarted","Data":"f0117c5842bea7f9387e6c9e91c456c43bfa9b4caf1cd8d2295dcc244b895efb"} Nov 25 11:06:29 crc kubenswrapper[4776]: I1125 11:06:29.186617 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:29 crc kubenswrapper[4776]: I1125 11:06:29.186645 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:29 crc kubenswrapper[4776]: I1125 11:06:29.212684 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-api-c694cd6fb-994xl" podStartSLOduration=6.21266538 podStartE2EDuration="6.21266538s" podCreationTimestamp="2025-11-25 11:06:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:06:29.205814668 +0000 UTC m=+6134.246874221" watchObservedRunningTime="2025-11-25 11:06:29.21266538 +0000 UTC m=+6134.253724933" Nov 25 11:06:29 crc kubenswrapper[4776]: I1125 11:06:29.357081 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-z8qlp"] Nov 25 11:06:30 crc kubenswrapper[4776]: I1125 11:06:30.209981 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-z8qlp" event={"ID":"d017f1f2-3baf-46a7-9dcd-d361a86c51f6","Type":"ContainerStarted","Data":"8ef6e9881ed7ac3db7e07204396cdb0e664e76db1f037786d45f9776d0a0c1c5"} Nov 25 11:06:31 crc kubenswrapper[4776]: I1125 11:06:31.222052 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-z8qlp" event={"ID":"d017f1f2-3baf-46a7-9dcd-d361a86c51f6","Type":"ContainerStarted","Data":"f07a8b3b51f0da4f9160494adfe06a629c86c0d18881eb30b72837241a7ef37b"} Nov 25 11:06:31 crc kubenswrapper[4776]: I1125 11:06:31.225046 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-nbkx7" event={"ID":"2e7c13e5-cbf7-408b-8a3c-edfc1d208c12","Type":"ContainerStarted","Data":"152a34859add32433e7d4dcb24b390570741113ea0058f264ebe8cde28a1e8c6"} Nov 25 11:06:31 crc kubenswrapper[4776]: I1125 11:06:31.918172 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-sync-gwv2g"] Nov 25 11:06:31 crc kubenswrapper[4776]: I1125 11:06:31.920580 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-gwv2g" Nov 25 11:06:31 crc kubenswrapper[4776]: I1125 11:06:31.926475 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-scripts" Nov 25 11:06:31 crc kubenswrapper[4776]: I1125 11:06:31.942529 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-gwv2g"] Nov 25 11:06:32 crc kubenswrapper[4776]: I1125 11:06:32.105004 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e-combined-ca-bundle\") pod \"octavia-db-sync-gwv2g\" (UID: \"73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e\") " pod="openstack/octavia-db-sync-gwv2g" Nov 25 11:06:32 crc kubenswrapper[4776]: I1125 11:06:32.105113 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e-config-data-merged\") pod \"octavia-db-sync-gwv2g\" (UID: \"73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e\") " pod="openstack/octavia-db-sync-gwv2g" Nov 25 11:06:32 crc kubenswrapper[4776]: I1125 11:06:32.105145 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e-config-data\") pod \"octavia-db-sync-gwv2g\" (UID: \"73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e\") " pod="openstack/octavia-db-sync-gwv2g" Nov 25 11:06:32 crc kubenswrapper[4776]: I1125 11:06:32.105168 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e-scripts\") pod \"octavia-db-sync-gwv2g\" (UID: \"73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e\") " pod="openstack/octavia-db-sync-gwv2g" Nov 25 11:06:32 crc kubenswrapper[4776]: I1125 11:06:32.206857 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e-combined-ca-bundle\") pod \"octavia-db-sync-gwv2g\" (UID: \"73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e\") " pod="openstack/octavia-db-sync-gwv2g" Nov 25 11:06:32 crc kubenswrapper[4776]: I1125 11:06:32.206961 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e-config-data-merged\") pod \"octavia-db-sync-gwv2g\" (UID: \"73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e\") " pod="openstack/octavia-db-sync-gwv2g" Nov 25 11:06:32 crc kubenswrapper[4776]: I1125 11:06:32.207007 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e-config-data\") pod \"octavia-db-sync-gwv2g\" (UID: \"73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e\") " pod="openstack/octavia-db-sync-gwv2g" Nov 25 11:06:32 crc kubenswrapper[4776]: I1125 11:06:32.207041 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e-scripts\") pod \"octavia-db-sync-gwv2g\" (UID: \"73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e\") " pod="openstack/octavia-db-sync-gwv2g" Nov 25 11:06:32 crc kubenswrapper[4776]: I1125 11:06:32.208162 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e-config-data-merged\") pod \"octavia-db-sync-gwv2g\" (UID: \"73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e\") " pod="openstack/octavia-db-sync-gwv2g" Nov 25 11:06:32 crc kubenswrapper[4776]: I1125 11:06:32.213393 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e-combined-ca-bundle\") pod \"octavia-db-sync-gwv2g\" (UID: \"73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e\") " pod="openstack/octavia-db-sync-gwv2g" Nov 25 11:06:32 crc kubenswrapper[4776]: I1125 11:06:32.214389 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e-scripts\") pod \"octavia-db-sync-gwv2g\" (UID: \"73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e\") " pod="openstack/octavia-db-sync-gwv2g" Nov 25 11:06:32 crc kubenswrapper[4776]: I1125 11:06:32.215222 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e-config-data\") pod \"octavia-db-sync-gwv2g\" (UID: \"73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e\") " pod="openstack/octavia-db-sync-gwv2g" Nov 25 11:06:32 crc kubenswrapper[4776]: I1125 11:06:32.254461 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-gwv2g" Nov 25 11:06:33 crc kubenswrapper[4776]: I1125 11:06:33.285872 4776 generic.go:334] "Generic (PLEG): container finished" podID="d017f1f2-3baf-46a7-9dcd-d361a86c51f6" containerID="f07a8b3b51f0da4f9160494adfe06a629c86c0d18881eb30b72837241a7ef37b" exitCode=0 Nov 25 11:06:33 crc kubenswrapper[4776]: I1125 11:06:33.286187 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-z8qlp" event={"ID":"d017f1f2-3baf-46a7-9dcd-d361a86c51f6","Type":"ContainerDied","Data":"f07a8b3b51f0da4f9160494adfe06a629c86c0d18881eb30b72837241a7ef37b"} Nov 25 11:06:33 crc kubenswrapper[4776]: I1125 11:06:33.290457 4776 generic.go:334] "Generic (PLEG): container finished" podID="2e7c13e5-cbf7-408b-8a3c-edfc1d208c12" containerID="152a34859add32433e7d4dcb24b390570741113ea0058f264ebe8cde28a1e8c6" exitCode=0 Nov 25 11:06:33 crc kubenswrapper[4776]: I1125 11:06:33.290593 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-nbkx7" event={"ID":"2e7c13e5-cbf7-408b-8a3c-edfc1d208c12","Type":"ContainerDied","Data":"152a34859add32433e7d4dcb24b390570741113ea0058f264ebe8cde28a1e8c6"} Nov 25 11:06:38 crc kubenswrapper[4776]: I1125 11:06:38.662164 4776 scope.go:117] "RemoveContainer" containerID="8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698" Nov 25 11:06:38 crc kubenswrapper[4776]: E1125 11:06:38.663115 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:06:41 crc kubenswrapper[4776]: I1125 11:06:41.318244 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-gwv2g"] Nov 25 11:06:41 crc kubenswrapper[4776]: W1125 11:06:41.478182 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73f7c1dc_9edd_40f4_b8f4_16b99ea7ac7e.slice/crio-189015d6bbe5ba24b36abc11ddcb2eecfac641bf820a01c940c8f4e61d2fd6a5 WatchSource:0}: Error finding container 189015d6bbe5ba24b36abc11ddcb2eecfac641bf820a01c940c8f4e61d2fd6a5: Status 404 returned error can't find the container with id 189015d6bbe5ba24b36abc11ddcb2eecfac641bf820a01c940c8f4e61d2fd6a5 Nov 25 11:06:42 crc kubenswrapper[4776]: I1125 11:06:42.402244 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-z8qlp" event={"ID":"d017f1f2-3baf-46a7-9dcd-d361a86c51f6","Type":"ContainerStarted","Data":"1c88dc6c577f5a267168eb20ceefdf4c7c965d5e5de026476299f604fc1de145"} Nov 25 11:06:42 crc kubenswrapper[4776]: I1125 11:06:42.402924 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-healthmanager-z8qlp" Nov 25 11:06:42 crc kubenswrapper[4776]: I1125 11:06:42.405766 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-5955f5554b-ggjdr" event={"ID":"83f8c52f-018a-481f-baec-1a4a3df56aec","Type":"ContainerStarted","Data":"e2fafd645d0c7a698d4d0ea45ee33fe28c2310f8df722e9a8f90d816881bea03"} Nov 25 11:06:42 crc kubenswrapper[4776]: I1125 11:06:42.408716 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-gwv2g" event={"ID":"73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e","Type":"ContainerStarted","Data":"184170080ae095a5b221c8b587a3a4dfa3ecfbae323b183d5873a8c506cbc423"} Nov 25 11:06:42 crc kubenswrapper[4776]: I1125 11:06:42.408771 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-gwv2g" event={"ID":"73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e","Type":"ContainerStarted","Data":"189015d6bbe5ba24b36abc11ddcb2eecfac641bf820a01c940c8f4e61d2fd6a5"} Nov 25 11:06:42 crc kubenswrapper[4776]: I1125 11:06:42.411216 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-nbkx7" event={"ID":"2e7c13e5-cbf7-408b-8a3c-edfc1d208c12","Type":"ContainerStarted","Data":"8568c235fb52cb1b51adebf4506c27bc18b358f556a0c88ee6e142a7fd0b8104"} Nov 25 11:06:42 crc kubenswrapper[4776]: I1125 11:06:42.411815 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-rsyslog-nbkx7" Nov 25 11:06:42 crc kubenswrapper[4776]: I1125 11:06:42.433141 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-healthmanager-z8qlp" podStartSLOduration=15.433122168 podStartE2EDuration="15.433122168s" podCreationTimestamp="2025-11-25 11:06:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:06:42.423489847 +0000 UTC m=+6147.464549400" watchObservedRunningTime="2025-11-25 11:06:42.433122168 +0000 UTC m=+6147.474181721" Nov 25 11:06:42 crc kubenswrapper[4776]: I1125 11:06:42.466283 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-rsyslog-nbkx7" podStartSLOduration=2.5674450699999998 podStartE2EDuration="20.4662668s" podCreationTimestamp="2025-11-25 11:06:22 +0000 UTC" firstStartedPulling="2025-11-25 11:06:23.651827141 +0000 UTC m=+6128.692886694" lastFinishedPulling="2025-11-25 11:06:41.550648851 +0000 UTC m=+6146.591708424" observedRunningTime="2025-11-25 11:06:42.464855245 +0000 UTC m=+6147.505914818" watchObservedRunningTime="2025-11-25 11:06:42.4662668 +0000 UTC m=+6147.507326353" Nov 25 11:06:43 crc kubenswrapper[4776]: I1125 11:06:43.199703 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:43 crc kubenswrapper[4776]: I1125 11:06:43.343250 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-c694cd6fb-994xl" Nov 25 11:06:43 crc kubenswrapper[4776]: I1125 11:06:43.432824 4776 generic.go:334] "Generic (PLEG): container finished" podID="83f8c52f-018a-481f-baec-1a4a3df56aec" containerID="e2fafd645d0c7a698d4d0ea45ee33fe28c2310f8df722e9a8f90d816881bea03" exitCode=0 Nov 25 11:06:43 crc kubenswrapper[4776]: I1125 11:06:43.433258 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-5955f5554b-ggjdr" event={"ID":"83f8c52f-018a-481f-baec-1a4a3df56aec","Type":"ContainerDied","Data":"e2fafd645d0c7a698d4d0ea45ee33fe28c2310f8df722e9a8f90d816881bea03"} Nov 25 11:06:43 crc kubenswrapper[4776]: I1125 11:06:43.435338 4776 generic.go:334] "Generic (PLEG): container finished" podID="73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e" containerID="184170080ae095a5b221c8b587a3a4dfa3ecfbae323b183d5873a8c506cbc423" exitCode=0 Nov 25 11:06:43 crc kubenswrapper[4776]: I1125 11:06:43.435574 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-gwv2g" event={"ID":"73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e","Type":"ContainerDied","Data":"184170080ae095a5b221c8b587a3a4dfa3ecfbae323b183d5873a8c506cbc423"} Nov 25 11:06:43 crc kubenswrapper[4776]: I1125 11:06:43.505639 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-api-dccc45cc4-75rpv"] Nov 25 11:06:43 crc kubenswrapper[4776]: I1125 11:06:43.505948 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/octavia-api-dccc45cc4-75rpv" podUID="5cee8213-d6c2-4b71-a4f5-ee4baa239c4c" containerName="octavia-api" containerID="cri-o://45a105694c088eee3d549f74c1f7545507c3146be8cbc5bca8fba2cd43af9c58" gracePeriod=30 Nov 25 11:06:43 crc kubenswrapper[4776]: I1125 11:06:43.506482 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/octavia-api-dccc45cc4-75rpv" podUID="5cee8213-d6c2-4b71-a4f5-ee4baa239c4c" containerName="octavia-api-provider-agent" containerID="cri-o://6e06d480bd5fa8bdf071006149cf24c73dcf92f43827208ae442911ea39c79e2" gracePeriod=30 Nov 25 11:06:44 crc kubenswrapper[4776]: I1125 11:06:44.444595 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-5955f5554b-ggjdr" event={"ID":"83f8c52f-018a-481f-baec-1a4a3df56aec","Type":"ContainerStarted","Data":"c725e457a4142652cd427f4db06eefa741ae225608abc2eaa5bc74ee3e5d061a"} Nov 25 11:06:44 crc kubenswrapper[4776]: I1125 11:06:44.447451 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-gwv2g" event={"ID":"73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e","Type":"ContainerStarted","Data":"0241f0320df1e6c1e5c6ca440d91ab0ec49adce12766987f286c37e32eb369e7"} Nov 25 11:06:44 crc kubenswrapper[4776]: I1125 11:06:44.451490 4776 generic.go:334] "Generic (PLEG): container finished" podID="5cee8213-d6c2-4b71-a4f5-ee4baa239c4c" containerID="6e06d480bd5fa8bdf071006149cf24c73dcf92f43827208ae442911ea39c79e2" exitCode=0 Nov 25 11:06:44 crc kubenswrapper[4776]: I1125 11:06:44.451851 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-dccc45cc4-75rpv" event={"ID":"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c","Type":"ContainerDied","Data":"6e06d480bd5fa8bdf071006149cf24c73dcf92f43827208ae442911ea39c79e2"} Nov 25 11:06:44 crc kubenswrapper[4776]: I1125 11:06:44.484484 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-5955f5554b-ggjdr" podStartSLOduration=4.564807098 podStartE2EDuration="22.484463921s" podCreationTimestamp="2025-11-25 11:06:22 +0000 UTC" firstStartedPulling="2025-11-25 11:06:23.557254927 +0000 UTC m=+6128.598314480" lastFinishedPulling="2025-11-25 11:06:41.47691175 +0000 UTC m=+6146.517971303" observedRunningTime="2025-11-25 11:06:44.464586682 +0000 UTC m=+6149.505646235" watchObservedRunningTime="2025-11-25 11:06:44.484463921 +0000 UTC m=+6149.525523474" Nov 25 11:06:44 crc kubenswrapper[4776]: I1125 11:06:44.500583 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-db-sync-gwv2g" podStartSLOduration=13.500568605 podStartE2EDuration="13.500568605s" podCreationTimestamp="2025-11-25 11:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:06:44.499483148 +0000 UTC m=+6149.540542701" watchObservedRunningTime="2025-11-25 11:06:44.500568605 +0000 UTC m=+6149.541628158" Nov 25 11:06:47 crc kubenswrapper[4776]: I1125 11:06:47.483964 4776 generic.go:334] "Generic (PLEG): container finished" podID="5cee8213-d6c2-4b71-a4f5-ee4baa239c4c" containerID="45a105694c088eee3d549f74c1f7545507c3146be8cbc5bca8fba2cd43af9c58" exitCode=0 Nov 25 11:06:47 crc kubenswrapper[4776]: I1125 11:06:47.484443 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-dccc45cc4-75rpv" event={"ID":"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c","Type":"ContainerDied","Data":"45a105694c088eee3d549f74c1f7545507c3146be8cbc5bca8fba2cd43af9c58"} Nov 25 11:06:48 crc kubenswrapper[4776]: I1125 11:06:48.281341 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-dccc45cc4-75rpv" Nov 25 11:06:48 crc kubenswrapper[4776]: I1125 11:06:48.418963 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-config-data-merged\") pod \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\" (UID: \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\") " Nov 25 11:06:48 crc kubenswrapper[4776]: I1125 11:06:48.419060 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-scripts\") pod \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\" (UID: \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\") " Nov 25 11:06:48 crc kubenswrapper[4776]: I1125 11:06:48.419197 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-config-data\") pod \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\" (UID: \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\") " Nov 25 11:06:48 crc kubenswrapper[4776]: I1125 11:06:48.419227 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-ovndb-tls-certs\") pod \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\" (UID: \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\") " Nov 25 11:06:48 crc kubenswrapper[4776]: I1125 11:06:48.419318 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-octavia-run\") pod \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\" (UID: \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\") " Nov 25 11:06:48 crc kubenswrapper[4776]: I1125 11:06:48.419445 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-combined-ca-bundle\") pod \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\" (UID: \"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c\") " Nov 25 11:06:48 crc kubenswrapper[4776]: I1125 11:06:48.420595 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-octavia-run" (OuterVolumeSpecName: "octavia-run") pod "5cee8213-d6c2-4b71-a4f5-ee4baa239c4c" (UID: "5cee8213-d6c2-4b71-a4f5-ee4baa239c4c"). InnerVolumeSpecName "octavia-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:06:48 crc kubenswrapper[4776]: I1125 11:06:48.425259 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-scripts" (OuterVolumeSpecName: "scripts") pod "5cee8213-d6c2-4b71-a4f5-ee4baa239c4c" (UID: "5cee8213-d6c2-4b71-a4f5-ee4baa239c4c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:06:48 crc kubenswrapper[4776]: I1125 11:06:48.442431 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-config-data" (OuterVolumeSpecName: "config-data") pod "5cee8213-d6c2-4b71-a4f5-ee4baa239c4c" (UID: "5cee8213-d6c2-4b71-a4f5-ee4baa239c4c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:06:48 crc kubenswrapper[4776]: I1125 11:06:48.474069 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "5cee8213-d6c2-4b71-a4f5-ee4baa239c4c" (UID: "5cee8213-d6c2-4b71-a4f5-ee4baa239c4c"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:06:48 crc kubenswrapper[4776]: I1125 11:06:48.476898 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5cee8213-d6c2-4b71-a4f5-ee4baa239c4c" (UID: "5cee8213-d6c2-4b71-a4f5-ee4baa239c4c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:06:48 crc kubenswrapper[4776]: I1125 11:06:48.497679 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-dccc45cc4-75rpv" event={"ID":"5cee8213-d6c2-4b71-a4f5-ee4baa239c4c","Type":"ContainerDied","Data":"ee567d3eb7e86e1bbf570e74d3875d83fc2640958f10c0e23e28be5c1cfdbde2"} Nov 25 11:06:48 crc kubenswrapper[4776]: I1125 11:06:48.497747 4776 scope.go:117] "RemoveContainer" containerID="6e06d480bd5fa8bdf071006149cf24c73dcf92f43827208ae442911ea39c79e2" Nov 25 11:06:48 crc kubenswrapper[4776]: I1125 11:06:48.497749 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-dccc45cc4-75rpv" Nov 25 11:06:48 crc kubenswrapper[4776]: I1125 11:06:48.521330 4776 reconciler_common.go:293] "Volume detached for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-octavia-run\") on node \"crc\" DevicePath \"\"" Nov 25 11:06:48 crc kubenswrapper[4776]: I1125 11:06:48.521642 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:06:48 crc kubenswrapper[4776]: I1125 11:06:48.521655 4776 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-config-data-merged\") on node \"crc\" DevicePath \"\"" Nov 25 11:06:48 crc kubenswrapper[4776]: I1125 11:06:48.521664 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:06:48 crc kubenswrapper[4776]: I1125 11:06:48.521672 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:06:48 crc kubenswrapper[4776]: I1125 11:06:48.571819 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "5cee8213-d6c2-4b71-a4f5-ee4baa239c4c" (UID: "5cee8213-d6c2-4b71-a4f5-ee4baa239c4c"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:06:48 crc kubenswrapper[4776]: I1125 11:06:48.623391 4776 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 11:06:48 crc kubenswrapper[4776]: I1125 11:06:48.844828 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-api-dccc45cc4-75rpv"] Nov 25 11:06:48 crc kubenswrapper[4776]: I1125 11:06:48.854788 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-api-dccc45cc4-75rpv"] Nov 25 11:06:49 crc kubenswrapper[4776]: I1125 11:06:49.186674 4776 scope.go:117] "RemoveContainer" containerID="45a105694c088eee3d549f74c1f7545507c3146be8cbc5bca8fba2cd43af9c58" Nov 25 11:06:49 crc kubenswrapper[4776]: I1125 11:06:49.277165 4776 scope.go:117] "RemoveContainer" containerID="6fb82a453b0f536e3d172688f301b919bdc97263ee56a595fa3d5bf2f1a847df" Nov 25 11:06:49 crc kubenswrapper[4776]: I1125 11:06:49.675651 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cee8213-d6c2-4b71-a4f5-ee4baa239c4c" path="/var/lib/kubelet/pods/5cee8213-d6c2-4b71-a4f5-ee4baa239c4c/volumes" Nov 25 11:06:52 crc kubenswrapper[4776]: I1125 11:06:52.411293 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-rsyslog-nbkx7" Nov 25 11:06:52 crc kubenswrapper[4776]: I1125 11:06:52.662281 4776 scope.go:117] "RemoveContainer" containerID="8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698" Nov 25 11:06:52 crc kubenswrapper[4776]: E1125 11:06:52.662884 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:06:58 crc kubenswrapper[4776]: I1125 11:06:58.056018 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-healthmanager-z8qlp" Nov 25 11:06:59 crc kubenswrapper[4776]: I1125 11:06:59.621276 4776 generic.go:334] "Generic (PLEG): container finished" podID="73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e" containerID="0241f0320df1e6c1e5c6ca440d91ab0ec49adce12766987f286c37e32eb369e7" exitCode=0 Nov 25 11:06:59 crc kubenswrapper[4776]: I1125 11:06:59.621464 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-gwv2g" event={"ID":"73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e","Type":"ContainerDied","Data":"0241f0320df1e6c1e5c6ca440d91ab0ec49adce12766987f286c37e32eb369e7"} Nov 25 11:07:01 crc kubenswrapper[4776]: I1125 11:07:01.039358 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-gwv2g" Nov 25 11:07:01 crc kubenswrapper[4776]: I1125 11:07:01.158972 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e-scripts\") pod \"73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e\" (UID: \"73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e\") " Nov 25 11:07:01 crc kubenswrapper[4776]: I1125 11:07:01.159022 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e-config-data\") pod \"73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e\" (UID: \"73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e\") " Nov 25 11:07:01 crc kubenswrapper[4776]: I1125 11:07:01.159894 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e-combined-ca-bundle\") pod \"73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e\" (UID: \"73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e\") " Nov 25 11:07:01 crc kubenswrapper[4776]: I1125 11:07:01.159992 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e-config-data-merged\") pod \"73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e\" (UID: \"73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e\") " Nov 25 11:07:01 crc kubenswrapper[4776]: I1125 11:07:01.164668 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e-config-data" (OuterVolumeSpecName: "config-data") pod "73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e" (UID: "73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:07:01 crc kubenswrapper[4776]: I1125 11:07:01.164714 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e-scripts" (OuterVolumeSpecName: "scripts") pod "73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e" (UID: "73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:07:01 crc kubenswrapper[4776]: I1125 11:07:01.187648 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e" (UID: "73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:07:01 crc kubenswrapper[4776]: I1125 11:07:01.190647 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e" (UID: "73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:07:01 crc kubenswrapper[4776]: I1125 11:07:01.262675 4776 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e-config-data-merged\") on node \"crc\" DevicePath \"\"" Nov 25 11:07:01 crc kubenswrapper[4776]: I1125 11:07:01.262707 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:07:01 crc kubenswrapper[4776]: I1125 11:07:01.262718 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:07:01 crc kubenswrapper[4776]: I1125 11:07:01.262729 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:07:01 crc kubenswrapper[4776]: I1125 11:07:01.643926 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-gwv2g" event={"ID":"73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e","Type":"ContainerDied","Data":"189015d6bbe5ba24b36abc11ddcb2eecfac641bf820a01c940c8f4e61d2fd6a5"} Nov 25 11:07:01 crc kubenswrapper[4776]: I1125 11:07:01.643962 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-gwv2g" Nov 25 11:07:01 crc kubenswrapper[4776]: I1125 11:07:01.643971 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="189015d6bbe5ba24b36abc11ddcb2eecfac641bf820a01c940c8f4e61d2fd6a5" Nov 25 11:07:07 crc kubenswrapper[4776]: I1125 11:07:07.662699 4776 scope.go:117] "RemoveContainer" containerID="8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698" Nov 25 11:07:07 crc kubenswrapper[4776]: E1125 11:07:07.663855 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:07:17 crc kubenswrapper[4776]: I1125 11:07:17.041739 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-x7zns"] Nov 25 11:07:17 crc kubenswrapper[4776]: I1125 11:07:17.058385 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-60f3-account-create-2ffkm"] Nov 25 11:07:17 crc kubenswrapper[4776]: I1125 11:07:17.066699 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-x7zns"] Nov 25 11:07:17 crc kubenswrapper[4776]: I1125 11:07:17.076022 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-60f3-account-create-2ffkm"] Nov 25 11:07:17 crc kubenswrapper[4776]: I1125 11:07:17.696846 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37973950-513b-427d-b640-cbd21bc1fd65" path="/var/lib/kubelet/pods/37973950-513b-427d-b640-cbd21bc1fd65/volumes" Nov 25 11:07:17 crc kubenswrapper[4776]: I1125 11:07:17.697607 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d2684fe-a46c-4bcc-9f5b-8d06e5966073" path="/var/lib/kubelet/pods/3d2684fe-a46c-4bcc-9f5b-8d06e5966073/volumes" Nov 25 11:07:18 crc kubenswrapper[4776]: I1125 11:07:18.663338 4776 scope.go:117] "RemoveContainer" containerID="8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698" Nov 25 11:07:18 crc kubenswrapper[4776]: E1125 11:07:18.663599 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:07:23 crc kubenswrapper[4776]: I1125 11:07:23.170581 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8k6jm"] Nov 25 11:07:23 crc kubenswrapper[4776]: E1125 11:07:23.173977 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cee8213-d6c2-4b71-a4f5-ee4baa239c4c" containerName="init" Nov 25 11:07:23 crc kubenswrapper[4776]: I1125 11:07:23.173993 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cee8213-d6c2-4b71-a4f5-ee4baa239c4c" containerName="init" Nov 25 11:07:23 crc kubenswrapper[4776]: E1125 11:07:23.174007 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cee8213-d6c2-4b71-a4f5-ee4baa239c4c" containerName="octavia-api" Nov 25 11:07:23 crc kubenswrapper[4776]: I1125 11:07:23.174014 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cee8213-d6c2-4b71-a4f5-ee4baa239c4c" containerName="octavia-api" Nov 25 11:07:23 crc kubenswrapper[4776]: E1125 11:07:23.174031 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e" containerName="octavia-db-sync" Nov 25 11:07:23 crc kubenswrapper[4776]: I1125 11:07:23.174039 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e" containerName="octavia-db-sync" Nov 25 11:07:23 crc kubenswrapper[4776]: E1125 11:07:23.174061 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e" containerName="init" Nov 25 11:07:23 crc kubenswrapper[4776]: I1125 11:07:23.174067 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e" containerName="init" Nov 25 11:07:23 crc kubenswrapper[4776]: E1125 11:07:23.174091 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cee8213-d6c2-4b71-a4f5-ee4baa239c4c" containerName="octavia-api-provider-agent" Nov 25 11:07:23 crc kubenswrapper[4776]: I1125 11:07:23.174097 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cee8213-d6c2-4b71-a4f5-ee4baa239c4c" containerName="octavia-api-provider-agent" Nov 25 11:07:23 crc kubenswrapper[4776]: I1125 11:07:23.174261 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e" containerName="octavia-db-sync" Nov 25 11:07:23 crc kubenswrapper[4776]: I1125 11:07:23.174270 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cee8213-d6c2-4b71-a4f5-ee4baa239c4c" containerName="octavia-api-provider-agent" Nov 25 11:07:23 crc kubenswrapper[4776]: I1125 11:07:23.174292 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cee8213-d6c2-4b71-a4f5-ee4baa239c4c" containerName="octavia-api" Nov 25 11:07:23 crc kubenswrapper[4776]: I1125 11:07:23.175619 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8k6jm" Nov 25 11:07:23 crc kubenswrapper[4776]: I1125 11:07:23.179884 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8k6jm"] Nov 25 11:07:23 crc kubenswrapper[4776]: I1125 11:07:23.223589 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24wmc\" (UniqueName: \"kubernetes.io/projected/44844357-d27a-454b-adbc-be91bed3a5a0-kube-api-access-24wmc\") pod \"redhat-operators-8k6jm\" (UID: \"44844357-d27a-454b-adbc-be91bed3a5a0\") " pod="openshift-marketplace/redhat-operators-8k6jm" Nov 25 11:07:23 crc kubenswrapper[4776]: I1125 11:07:23.223685 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44844357-d27a-454b-adbc-be91bed3a5a0-utilities\") pod \"redhat-operators-8k6jm\" (UID: \"44844357-d27a-454b-adbc-be91bed3a5a0\") " pod="openshift-marketplace/redhat-operators-8k6jm" Nov 25 11:07:23 crc kubenswrapper[4776]: I1125 11:07:23.223729 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44844357-d27a-454b-adbc-be91bed3a5a0-catalog-content\") pod \"redhat-operators-8k6jm\" (UID: \"44844357-d27a-454b-adbc-be91bed3a5a0\") " pod="openshift-marketplace/redhat-operators-8k6jm" Nov 25 11:07:23 crc kubenswrapper[4776]: I1125 11:07:23.325832 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44844357-d27a-454b-adbc-be91bed3a5a0-utilities\") pod \"redhat-operators-8k6jm\" (UID: \"44844357-d27a-454b-adbc-be91bed3a5a0\") " pod="openshift-marketplace/redhat-operators-8k6jm" Nov 25 11:07:23 crc kubenswrapper[4776]: I1125 11:07:23.325943 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44844357-d27a-454b-adbc-be91bed3a5a0-catalog-content\") pod \"redhat-operators-8k6jm\" (UID: \"44844357-d27a-454b-adbc-be91bed3a5a0\") " pod="openshift-marketplace/redhat-operators-8k6jm" Nov 25 11:07:23 crc kubenswrapper[4776]: I1125 11:07:23.326383 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24wmc\" (UniqueName: \"kubernetes.io/projected/44844357-d27a-454b-adbc-be91bed3a5a0-kube-api-access-24wmc\") pod \"redhat-operators-8k6jm\" (UID: \"44844357-d27a-454b-adbc-be91bed3a5a0\") " pod="openshift-marketplace/redhat-operators-8k6jm" Nov 25 11:07:23 crc kubenswrapper[4776]: I1125 11:07:23.326490 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44844357-d27a-454b-adbc-be91bed3a5a0-utilities\") pod \"redhat-operators-8k6jm\" (UID: \"44844357-d27a-454b-adbc-be91bed3a5a0\") " pod="openshift-marketplace/redhat-operators-8k6jm" Nov 25 11:07:23 crc kubenswrapper[4776]: I1125 11:07:23.326505 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44844357-d27a-454b-adbc-be91bed3a5a0-catalog-content\") pod \"redhat-operators-8k6jm\" (UID: \"44844357-d27a-454b-adbc-be91bed3a5a0\") " pod="openshift-marketplace/redhat-operators-8k6jm" Nov 25 11:07:23 crc kubenswrapper[4776]: I1125 11:07:23.345542 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24wmc\" (UniqueName: \"kubernetes.io/projected/44844357-d27a-454b-adbc-be91bed3a5a0-kube-api-access-24wmc\") pod \"redhat-operators-8k6jm\" (UID: \"44844357-d27a-454b-adbc-be91bed3a5a0\") " pod="openshift-marketplace/redhat-operators-8k6jm" Nov 25 11:07:23 crc kubenswrapper[4776]: I1125 11:07:23.502794 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8k6jm" Nov 25 11:07:23 crc kubenswrapper[4776]: I1125 11:07:23.937628 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8k6jm"] Nov 25 11:07:24 crc kubenswrapper[4776]: I1125 11:07:24.911118 4776 generic.go:334] "Generic (PLEG): container finished" podID="44844357-d27a-454b-adbc-be91bed3a5a0" containerID="fd7f93b212affaa250086af9ce571010f33bd3f5c0388a0a7199f70d0dcbbbff" exitCode=0 Nov 25 11:07:24 crc kubenswrapper[4776]: I1125 11:07:24.911810 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8k6jm" event={"ID":"44844357-d27a-454b-adbc-be91bed3a5a0","Type":"ContainerDied","Data":"fd7f93b212affaa250086af9ce571010f33bd3f5c0388a0a7199f70d0dcbbbff"} Nov 25 11:07:24 crc kubenswrapper[4776]: I1125 11:07:24.911869 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8k6jm" event={"ID":"44844357-d27a-454b-adbc-be91bed3a5a0","Type":"ContainerStarted","Data":"c32496cff092b861908d282f267eb4c61341f3fb38d90f5e8dcba309394eff4c"} Nov 25 11:07:26 crc kubenswrapper[4776]: I1125 11:07:26.029089 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-5ppz8"] Nov 25 11:07:26 crc kubenswrapper[4776]: I1125 11:07:26.038284 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-5ppz8"] Nov 25 11:07:26 crc kubenswrapper[4776]: I1125 11:07:26.932235 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8k6jm" event={"ID":"44844357-d27a-454b-adbc-be91bed3a5a0","Type":"ContainerStarted","Data":"f96ad2eff5a5b75c94badf0988d86a8b7e56febec7ed2df11d2e7dba80fb9bbc"} Nov 25 11:07:27 crc kubenswrapper[4776]: I1125 11:07:27.676208 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68963dc0-5607-450a-8bf3-b2a92abdefc9" path="/var/lib/kubelet/pods/68963dc0-5607-450a-8bf3-b2a92abdefc9/volumes" Nov 25 11:07:28 crc kubenswrapper[4776]: I1125 11:07:28.629526 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-5955f5554b-ggjdr"] Nov 25 11:07:28 crc kubenswrapper[4776]: I1125 11:07:28.629791 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/octavia-image-upload-5955f5554b-ggjdr" podUID="83f8c52f-018a-481f-baec-1a4a3df56aec" containerName="octavia-amphora-httpd" containerID="cri-o://c725e457a4142652cd427f4db06eefa741ae225608abc2eaa5bc74ee3e5d061a" gracePeriod=30 Nov 25 11:07:28 crc kubenswrapper[4776]: I1125 11:07:28.967016 4776 generic.go:334] "Generic (PLEG): container finished" podID="83f8c52f-018a-481f-baec-1a4a3df56aec" containerID="c725e457a4142652cd427f4db06eefa741ae225608abc2eaa5bc74ee3e5d061a" exitCode=0 Nov 25 11:07:28 crc kubenswrapper[4776]: I1125 11:07:28.967206 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-5955f5554b-ggjdr" event={"ID":"83f8c52f-018a-481f-baec-1a4a3df56aec","Type":"ContainerDied","Data":"c725e457a4142652cd427f4db06eefa741ae225608abc2eaa5bc74ee3e5d061a"} Nov 25 11:07:28 crc kubenswrapper[4776]: I1125 11:07:28.976259 4776 generic.go:334] "Generic (PLEG): container finished" podID="44844357-d27a-454b-adbc-be91bed3a5a0" containerID="f96ad2eff5a5b75c94badf0988d86a8b7e56febec7ed2df11d2e7dba80fb9bbc" exitCode=0 Nov 25 11:07:28 crc kubenswrapper[4776]: I1125 11:07:28.976304 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8k6jm" event={"ID":"44844357-d27a-454b-adbc-be91bed3a5a0","Type":"ContainerDied","Data":"f96ad2eff5a5b75c94badf0988d86a8b7e56febec7ed2df11d2e7dba80fb9bbc"} Nov 25 11:07:29 crc kubenswrapper[4776]: I1125 11:07:29.171165 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-5955f5554b-ggjdr" Nov 25 11:07:29 crc kubenswrapper[4776]: I1125 11:07:29.351236 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/83f8c52f-018a-481f-baec-1a4a3df56aec-httpd-config\") pod \"83f8c52f-018a-481f-baec-1a4a3df56aec\" (UID: \"83f8c52f-018a-481f-baec-1a4a3df56aec\") " Nov 25 11:07:29 crc kubenswrapper[4776]: I1125 11:07:29.351621 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/83f8c52f-018a-481f-baec-1a4a3df56aec-amphora-image\") pod \"83f8c52f-018a-481f-baec-1a4a3df56aec\" (UID: \"83f8c52f-018a-481f-baec-1a4a3df56aec\") " Nov 25 11:07:29 crc kubenswrapper[4776]: I1125 11:07:29.388485 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83f8c52f-018a-481f-baec-1a4a3df56aec-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "83f8c52f-018a-481f-baec-1a4a3df56aec" (UID: "83f8c52f-018a-481f-baec-1a4a3df56aec"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:07:29 crc kubenswrapper[4776]: I1125 11:07:29.457399 4776 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/83f8c52f-018a-481f-baec-1a4a3df56aec-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 25 11:07:29 crc kubenswrapper[4776]: I1125 11:07:29.470430 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83f8c52f-018a-481f-baec-1a4a3df56aec-amphora-image" (OuterVolumeSpecName: "amphora-image") pod "83f8c52f-018a-481f-baec-1a4a3df56aec" (UID: "83f8c52f-018a-481f-baec-1a4a3df56aec"). InnerVolumeSpecName "amphora-image". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:07:29 crc kubenswrapper[4776]: I1125 11:07:29.559900 4776 reconciler_common.go:293] "Volume detached for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/83f8c52f-018a-481f-baec-1a4a3df56aec-amphora-image\") on node \"crc\" DevicePath \"\"" Nov 25 11:07:29 crc kubenswrapper[4776]: I1125 11:07:29.990692 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-5955f5554b-ggjdr" event={"ID":"83f8c52f-018a-481f-baec-1a4a3df56aec","Type":"ContainerDied","Data":"10b19c718e7232e0268afa29a2784ac8bdbad0651d17b815aeca9aaf3932d9a1"} Nov 25 11:07:29 crc kubenswrapper[4776]: I1125 11:07:29.990816 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-5955f5554b-ggjdr" Nov 25 11:07:29 crc kubenswrapper[4776]: I1125 11:07:29.991917 4776 scope.go:117] "RemoveContainer" containerID="c725e457a4142652cd427f4db06eefa741ae225608abc2eaa5bc74ee3e5d061a" Nov 25 11:07:29 crc kubenswrapper[4776]: I1125 11:07:29.998720 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8k6jm" event={"ID":"44844357-d27a-454b-adbc-be91bed3a5a0","Type":"ContainerStarted","Data":"cca0d85f30f5a0aa861d35d468ba1704f94763a74cdd9fcadb0dbfd3b1d044a6"} Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.016619 4776 scope.go:117] "RemoveContainer" containerID="e2fafd645d0c7a698d4d0ea45ee33fe28c2310f8df722e9a8f90d816881bea03" Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.024911 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-5955f5554b-ggjdr"] Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.034366 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-image-upload-5955f5554b-ggjdr"] Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.046170 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8k6jm" podStartSLOduration=2.5390850240000002 podStartE2EDuration="7.046154397s" podCreationTimestamp="2025-11-25 11:07:23 +0000 UTC" firstStartedPulling="2025-11-25 11:07:24.913183096 +0000 UTC m=+6189.954242649" lastFinishedPulling="2025-11-25 11:07:29.420252479 +0000 UTC m=+6194.461312022" observedRunningTime="2025-11-25 11:07:30.042531036 +0000 UTC m=+6195.083590599" watchObservedRunningTime="2025-11-25 11:07:30.046154397 +0000 UTC m=+6195.087213940" Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.259824 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-housekeeping-bd8xp"] Nov 25 11:07:30 crc kubenswrapper[4776]: E1125 11:07:30.260277 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83f8c52f-018a-481f-baec-1a4a3df56aec" containerName="init" Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.260296 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="83f8c52f-018a-481f-baec-1a4a3df56aec" containerName="init" Nov 25 11:07:30 crc kubenswrapper[4776]: E1125 11:07:30.260320 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83f8c52f-018a-481f-baec-1a4a3df56aec" containerName="octavia-amphora-httpd" Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.260329 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="83f8c52f-018a-481f-baec-1a4a3df56aec" containerName="octavia-amphora-httpd" Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.260579 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="83f8c52f-018a-481f-baec-1a4a3df56aec" containerName="octavia-amphora-httpd" Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.262452 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-bd8xp" Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.270138 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-config-data" Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.270834 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-scripts" Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.317103 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-bd8xp"] Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.391468 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de8204cc-520a-41de-8079-2b25831ae0ae-combined-ca-bundle\") pod \"octavia-housekeeping-bd8xp\" (UID: \"de8204cc-520a-41de-8079-2b25831ae0ae\") " pod="openstack/octavia-housekeeping-bd8xp" Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.391532 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/de8204cc-520a-41de-8079-2b25831ae0ae-amphora-certs\") pod \"octavia-housekeeping-bd8xp\" (UID: \"de8204cc-520a-41de-8079-2b25831ae0ae\") " pod="openstack/octavia-housekeeping-bd8xp" Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.391621 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/de8204cc-520a-41de-8079-2b25831ae0ae-config-data-merged\") pod \"octavia-housekeeping-bd8xp\" (UID: \"de8204cc-520a-41de-8079-2b25831ae0ae\") " pod="openstack/octavia-housekeeping-bd8xp" Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.391668 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de8204cc-520a-41de-8079-2b25831ae0ae-scripts\") pod \"octavia-housekeeping-bd8xp\" (UID: \"de8204cc-520a-41de-8079-2b25831ae0ae\") " pod="openstack/octavia-housekeeping-bd8xp" Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.391698 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/de8204cc-520a-41de-8079-2b25831ae0ae-hm-ports\") pod \"octavia-housekeeping-bd8xp\" (UID: \"de8204cc-520a-41de-8079-2b25831ae0ae\") " pod="openstack/octavia-housekeeping-bd8xp" Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.391898 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de8204cc-520a-41de-8079-2b25831ae0ae-config-data\") pod \"octavia-housekeeping-bd8xp\" (UID: \"de8204cc-520a-41de-8079-2b25831ae0ae\") " pod="openstack/octavia-housekeeping-bd8xp" Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.493673 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/de8204cc-520a-41de-8079-2b25831ae0ae-config-data-merged\") pod \"octavia-housekeeping-bd8xp\" (UID: \"de8204cc-520a-41de-8079-2b25831ae0ae\") " pod="openstack/octavia-housekeeping-bd8xp" Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.493817 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de8204cc-520a-41de-8079-2b25831ae0ae-scripts\") pod \"octavia-housekeeping-bd8xp\" (UID: \"de8204cc-520a-41de-8079-2b25831ae0ae\") " pod="openstack/octavia-housekeeping-bd8xp" Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.493863 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/de8204cc-520a-41de-8079-2b25831ae0ae-hm-ports\") pod \"octavia-housekeeping-bd8xp\" (UID: \"de8204cc-520a-41de-8079-2b25831ae0ae\") " pod="openstack/octavia-housekeeping-bd8xp" Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.493915 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de8204cc-520a-41de-8079-2b25831ae0ae-config-data\") pod \"octavia-housekeeping-bd8xp\" (UID: \"de8204cc-520a-41de-8079-2b25831ae0ae\") " pod="openstack/octavia-housekeeping-bd8xp" Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.494119 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de8204cc-520a-41de-8079-2b25831ae0ae-combined-ca-bundle\") pod \"octavia-housekeeping-bd8xp\" (UID: \"de8204cc-520a-41de-8079-2b25831ae0ae\") " pod="openstack/octavia-housekeeping-bd8xp" Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.494149 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/de8204cc-520a-41de-8079-2b25831ae0ae-amphora-certs\") pod \"octavia-housekeeping-bd8xp\" (UID: \"de8204cc-520a-41de-8079-2b25831ae0ae\") " pod="openstack/octavia-housekeeping-bd8xp" Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.494681 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/de8204cc-520a-41de-8079-2b25831ae0ae-config-data-merged\") pod \"octavia-housekeeping-bd8xp\" (UID: \"de8204cc-520a-41de-8079-2b25831ae0ae\") " pod="openstack/octavia-housekeeping-bd8xp" Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.495416 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/de8204cc-520a-41de-8079-2b25831ae0ae-hm-ports\") pod \"octavia-housekeeping-bd8xp\" (UID: \"de8204cc-520a-41de-8079-2b25831ae0ae\") " pod="openstack/octavia-housekeeping-bd8xp" Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.500080 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de8204cc-520a-41de-8079-2b25831ae0ae-combined-ca-bundle\") pod \"octavia-housekeeping-bd8xp\" (UID: \"de8204cc-520a-41de-8079-2b25831ae0ae\") " pod="openstack/octavia-housekeeping-bd8xp" Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.500091 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de8204cc-520a-41de-8079-2b25831ae0ae-scripts\") pod \"octavia-housekeeping-bd8xp\" (UID: \"de8204cc-520a-41de-8079-2b25831ae0ae\") " pod="openstack/octavia-housekeeping-bd8xp" Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.500185 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/de8204cc-520a-41de-8079-2b25831ae0ae-amphora-certs\") pod \"octavia-housekeeping-bd8xp\" (UID: \"de8204cc-520a-41de-8079-2b25831ae0ae\") " pod="openstack/octavia-housekeeping-bd8xp" Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.501000 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de8204cc-520a-41de-8079-2b25831ae0ae-config-data\") pod \"octavia-housekeeping-bd8xp\" (UID: \"de8204cc-520a-41de-8079-2b25831ae0ae\") " pod="openstack/octavia-housekeeping-bd8xp" Nov 25 11:07:30 crc kubenswrapper[4776]: I1125 11:07:30.592567 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-bd8xp" Nov 25 11:07:31 crc kubenswrapper[4776]: I1125 11:07:31.158603 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-bd8xp"] Nov 25 11:07:31 crc kubenswrapper[4776]: I1125 11:07:31.675243 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83f8c52f-018a-481f-baec-1a4a3df56aec" path="/var/lib/kubelet/pods/83f8c52f-018a-481f-baec-1a4a3df56aec/volumes" Nov 25 11:07:32 crc kubenswrapper[4776]: I1125 11:07:32.026521 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-bd8xp" event={"ID":"de8204cc-520a-41de-8079-2b25831ae0ae","Type":"ContainerStarted","Data":"eb6df9757b0aaa9857697bdda00c465c04372713013b91937e543ae5025c3bba"} Nov 25 11:07:32 crc kubenswrapper[4776]: I1125 11:07:32.663714 4776 scope.go:117] "RemoveContainer" containerID="8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698" Nov 25 11:07:32 crc kubenswrapper[4776]: E1125 11:07:32.664314 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:07:32 crc kubenswrapper[4776]: I1125 11:07:32.699240 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-worker-jzw7v"] Nov 25 11:07:32 crc kubenswrapper[4776]: I1125 11:07:32.701059 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-jzw7v" Nov 25 11:07:32 crc kubenswrapper[4776]: I1125 11:07:32.705479 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-scripts" Nov 25 11:07:32 crc kubenswrapper[4776]: I1125 11:07:32.705834 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-config-data" Nov 25 11:07:32 crc kubenswrapper[4776]: I1125 11:07:32.706125 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-jzw7v"] Nov 25 11:07:32 crc kubenswrapper[4776]: I1125 11:07:32.742809 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb21722f-848b-44ff-8658-a5a890bf0855-config-data\") pod \"octavia-worker-jzw7v\" (UID: \"eb21722f-848b-44ff-8658-a5a890bf0855\") " pod="openstack/octavia-worker-jzw7v" Nov 25 11:07:32 crc kubenswrapper[4776]: I1125 11:07:32.743056 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb21722f-848b-44ff-8658-a5a890bf0855-scripts\") pod \"octavia-worker-jzw7v\" (UID: \"eb21722f-848b-44ff-8658-a5a890bf0855\") " pod="openstack/octavia-worker-jzw7v" Nov 25 11:07:32 crc kubenswrapper[4776]: I1125 11:07:32.743708 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/eb21722f-848b-44ff-8658-a5a890bf0855-config-data-merged\") pod \"octavia-worker-jzw7v\" (UID: \"eb21722f-848b-44ff-8658-a5a890bf0855\") " pod="openstack/octavia-worker-jzw7v" Nov 25 11:07:32 crc kubenswrapper[4776]: I1125 11:07:32.743918 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/eb21722f-848b-44ff-8658-a5a890bf0855-amphora-certs\") pod \"octavia-worker-jzw7v\" (UID: \"eb21722f-848b-44ff-8658-a5a890bf0855\") " pod="openstack/octavia-worker-jzw7v" Nov 25 11:07:32 crc kubenswrapper[4776]: I1125 11:07:32.744006 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb21722f-848b-44ff-8658-a5a890bf0855-combined-ca-bundle\") pod \"octavia-worker-jzw7v\" (UID: \"eb21722f-848b-44ff-8658-a5a890bf0855\") " pod="openstack/octavia-worker-jzw7v" Nov 25 11:07:32 crc kubenswrapper[4776]: I1125 11:07:32.744099 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/eb21722f-848b-44ff-8658-a5a890bf0855-hm-ports\") pod \"octavia-worker-jzw7v\" (UID: \"eb21722f-848b-44ff-8658-a5a890bf0855\") " pod="openstack/octavia-worker-jzw7v" Nov 25 11:07:32 crc kubenswrapper[4776]: I1125 11:07:32.846094 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb21722f-848b-44ff-8658-a5a890bf0855-scripts\") pod \"octavia-worker-jzw7v\" (UID: \"eb21722f-848b-44ff-8658-a5a890bf0855\") " pod="openstack/octavia-worker-jzw7v" Nov 25 11:07:32 crc kubenswrapper[4776]: I1125 11:07:32.846145 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/eb21722f-848b-44ff-8658-a5a890bf0855-config-data-merged\") pod \"octavia-worker-jzw7v\" (UID: \"eb21722f-848b-44ff-8658-a5a890bf0855\") " pod="openstack/octavia-worker-jzw7v" Nov 25 11:07:32 crc kubenswrapper[4776]: I1125 11:07:32.846232 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/eb21722f-848b-44ff-8658-a5a890bf0855-amphora-certs\") pod \"octavia-worker-jzw7v\" (UID: \"eb21722f-848b-44ff-8658-a5a890bf0855\") " pod="openstack/octavia-worker-jzw7v" Nov 25 11:07:32 crc kubenswrapper[4776]: I1125 11:07:32.846282 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb21722f-848b-44ff-8658-a5a890bf0855-combined-ca-bundle\") pod \"octavia-worker-jzw7v\" (UID: \"eb21722f-848b-44ff-8658-a5a890bf0855\") " pod="openstack/octavia-worker-jzw7v" Nov 25 11:07:32 crc kubenswrapper[4776]: I1125 11:07:32.846313 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/eb21722f-848b-44ff-8658-a5a890bf0855-hm-ports\") pod \"octavia-worker-jzw7v\" (UID: \"eb21722f-848b-44ff-8658-a5a890bf0855\") " pod="openstack/octavia-worker-jzw7v" Nov 25 11:07:32 crc kubenswrapper[4776]: I1125 11:07:32.846339 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb21722f-848b-44ff-8658-a5a890bf0855-config-data\") pod \"octavia-worker-jzw7v\" (UID: \"eb21722f-848b-44ff-8658-a5a890bf0855\") " pod="openstack/octavia-worker-jzw7v" Nov 25 11:07:32 crc kubenswrapper[4776]: I1125 11:07:32.848786 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/eb21722f-848b-44ff-8658-a5a890bf0855-config-data-merged\") pod \"octavia-worker-jzw7v\" (UID: \"eb21722f-848b-44ff-8658-a5a890bf0855\") " pod="openstack/octavia-worker-jzw7v" Nov 25 11:07:32 crc kubenswrapper[4776]: I1125 11:07:32.849584 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/eb21722f-848b-44ff-8658-a5a890bf0855-hm-ports\") pod \"octavia-worker-jzw7v\" (UID: \"eb21722f-848b-44ff-8658-a5a890bf0855\") " pod="openstack/octavia-worker-jzw7v" Nov 25 11:07:32 crc kubenswrapper[4776]: I1125 11:07:32.856617 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/eb21722f-848b-44ff-8658-a5a890bf0855-amphora-certs\") pod \"octavia-worker-jzw7v\" (UID: \"eb21722f-848b-44ff-8658-a5a890bf0855\") " pod="openstack/octavia-worker-jzw7v" Nov 25 11:07:32 crc kubenswrapper[4776]: I1125 11:07:32.856678 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb21722f-848b-44ff-8658-a5a890bf0855-combined-ca-bundle\") pod \"octavia-worker-jzw7v\" (UID: \"eb21722f-848b-44ff-8658-a5a890bf0855\") " pod="openstack/octavia-worker-jzw7v" Nov 25 11:07:32 crc kubenswrapper[4776]: I1125 11:07:32.857463 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb21722f-848b-44ff-8658-a5a890bf0855-config-data\") pod \"octavia-worker-jzw7v\" (UID: \"eb21722f-848b-44ff-8658-a5a890bf0855\") " pod="openstack/octavia-worker-jzw7v" Nov 25 11:07:32 crc kubenswrapper[4776]: I1125 11:07:32.858566 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb21722f-848b-44ff-8658-a5a890bf0855-scripts\") pod \"octavia-worker-jzw7v\" (UID: \"eb21722f-848b-44ff-8658-a5a890bf0855\") " pod="openstack/octavia-worker-jzw7v" Nov 25 11:07:33 crc kubenswrapper[4776]: I1125 11:07:33.028591 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-jzw7v" Nov 25 11:07:33 crc kubenswrapper[4776]: I1125 11:07:33.503201 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8k6jm" Nov 25 11:07:33 crc kubenswrapper[4776]: I1125 11:07:33.503530 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8k6jm" Nov 25 11:07:33 crc kubenswrapper[4776]: I1125 11:07:33.594124 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-jzw7v"] Nov 25 11:07:33 crc kubenswrapper[4776]: W1125 11:07:33.600282 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb21722f_848b_44ff_8658_a5a890bf0855.slice/crio-4acabe0e741830315e9e99dd28f796efb8608a045cd1de350f755edba58b8949 WatchSource:0}: Error finding container 4acabe0e741830315e9e99dd28f796efb8608a045cd1de350f755edba58b8949: Status 404 returned error can't find the container with id 4acabe0e741830315e9e99dd28f796efb8608a045cd1de350f755edba58b8949 Nov 25 11:07:33 crc kubenswrapper[4776]: I1125 11:07:33.639741 4776 scope.go:117] "RemoveContainer" containerID="f6e9f3a5d505b4c545dbbead2d55fa069dbc2b11537a386b96f16848ee212cfd" Nov 25 11:07:33 crc kubenswrapper[4776]: I1125 11:07:33.696587 4776 scope.go:117] "RemoveContainer" containerID="fd039443c5df559bc8d525197566de07fa763f9c155c919c5dfb3352f577160a" Nov 25 11:07:33 crc kubenswrapper[4776]: I1125 11:07:33.743742 4776 scope.go:117] "RemoveContainer" containerID="d94e4d0d2d389030273930dc6b70c474012aa26f933c8c8c8ede8945ce5e2d05" Nov 25 11:07:34 crc kubenswrapper[4776]: I1125 11:07:34.046260 4776 generic.go:334] "Generic (PLEG): container finished" podID="de8204cc-520a-41de-8079-2b25831ae0ae" containerID="9b9de572d553bfba25af339ab86c291870425d99f0d6a6f4352e923679cd12f8" exitCode=0 Nov 25 11:07:34 crc kubenswrapper[4776]: I1125 11:07:34.046308 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-bd8xp" event={"ID":"de8204cc-520a-41de-8079-2b25831ae0ae","Type":"ContainerDied","Data":"9b9de572d553bfba25af339ab86c291870425d99f0d6a6f4352e923679cd12f8"} Nov 25 11:07:34 crc kubenswrapper[4776]: I1125 11:07:34.048703 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-jzw7v" event={"ID":"eb21722f-848b-44ff-8658-a5a890bf0855","Type":"ContainerStarted","Data":"4acabe0e741830315e9e99dd28f796efb8608a045cd1de350f755edba58b8949"} Nov 25 11:07:34 crc kubenswrapper[4776]: I1125 11:07:34.564875 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8k6jm" podUID="44844357-d27a-454b-adbc-be91bed3a5a0" containerName="registry-server" probeResult="failure" output=< Nov 25 11:07:34 crc kubenswrapper[4776]: timeout: failed to connect service ":50051" within 1s Nov 25 11:07:34 crc kubenswrapper[4776]: > Nov 25 11:07:36 crc kubenswrapper[4776]: I1125 11:07:36.072667 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-bd8xp" event={"ID":"de8204cc-520a-41de-8079-2b25831ae0ae","Type":"ContainerStarted","Data":"66e063a3dc2c1ecebbb9f6cd1626a4ba08fb849e71752de2727400c99ef0a41b"} Nov 25 11:07:36 crc kubenswrapper[4776]: I1125 11:07:36.073215 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-housekeeping-bd8xp" Nov 25 11:07:36 crc kubenswrapper[4776]: I1125 11:07:36.093670 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-housekeeping-bd8xp" podStartSLOduration=4.800424773 podStartE2EDuration="6.093647098s" podCreationTimestamp="2025-11-25 11:07:30 +0000 UTC" firstStartedPulling="2025-11-25 11:07:31.167424477 +0000 UTC m=+6196.208484030" lastFinishedPulling="2025-11-25 11:07:32.460646812 +0000 UTC m=+6197.501706355" observedRunningTime="2025-11-25 11:07:36.089158556 +0000 UTC m=+6201.130218109" watchObservedRunningTime="2025-11-25 11:07:36.093647098 +0000 UTC m=+6201.134706651" Nov 25 11:07:39 crc kubenswrapper[4776]: I1125 11:07:39.104156 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-jzw7v" event={"ID":"eb21722f-848b-44ff-8658-a5a890bf0855","Type":"ContainerStarted","Data":"fdeb6e9676cc1896e63631019d70d090a9b9d5fc34abe5348a907946cfb5fd5a"} Nov 25 11:07:41 crc kubenswrapper[4776]: I1125 11:07:41.122968 4776 generic.go:334] "Generic (PLEG): container finished" podID="eb21722f-848b-44ff-8658-a5a890bf0855" containerID="fdeb6e9676cc1896e63631019d70d090a9b9d5fc34abe5348a907946cfb5fd5a" exitCode=0 Nov 25 11:07:41 crc kubenswrapper[4776]: I1125 11:07:41.123174 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-jzw7v" event={"ID":"eb21722f-848b-44ff-8658-a5a890bf0855","Type":"ContainerDied","Data":"fdeb6e9676cc1896e63631019d70d090a9b9d5fc34abe5348a907946cfb5fd5a"} Nov 25 11:07:42 crc kubenswrapper[4776]: I1125 11:07:42.136209 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-jzw7v" event={"ID":"eb21722f-848b-44ff-8658-a5a890bf0855","Type":"ContainerStarted","Data":"8ccca6725beff2fbdf345db466aeba06ff3f51d5d90e756fd4825074df74ea4e"} Nov 25 11:07:42 crc kubenswrapper[4776]: I1125 11:07:42.136809 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-worker-jzw7v" Nov 25 11:07:42 crc kubenswrapper[4776]: I1125 11:07:42.157820 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-worker-jzw7v" podStartSLOduration=5.705734065 podStartE2EDuration="10.157796338s" podCreationTimestamp="2025-11-25 11:07:32 +0000 UTC" firstStartedPulling="2025-11-25 11:07:33.603905884 +0000 UTC m=+6198.644965437" lastFinishedPulling="2025-11-25 11:07:38.055968157 +0000 UTC m=+6203.097027710" observedRunningTime="2025-11-25 11:07:42.156576968 +0000 UTC m=+6207.197636521" watchObservedRunningTime="2025-11-25 11:07:42.157796338 +0000 UTC m=+6207.198855891" Nov 25 11:07:43 crc kubenswrapper[4776]: I1125 11:07:43.552176 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8k6jm" Nov 25 11:07:43 crc kubenswrapper[4776]: I1125 11:07:43.605268 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8k6jm" Nov 25 11:07:43 crc kubenswrapper[4776]: I1125 11:07:43.788085 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8k6jm"] Nov 25 11:07:44 crc kubenswrapper[4776]: I1125 11:07:44.662188 4776 scope.go:117] "RemoveContainer" containerID="8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698" Nov 25 11:07:44 crc kubenswrapper[4776]: E1125 11:07:44.662805 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:07:45 crc kubenswrapper[4776]: I1125 11:07:45.162251 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8k6jm" podUID="44844357-d27a-454b-adbc-be91bed3a5a0" containerName="registry-server" containerID="cri-o://cca0d85f30f5a0aa861d35d468ba1704f94763a74cdd9fcadb0dbfd3b1d044a6" gracePeriod=2 Nov 25 11:07:45 crc kubenswrapper[4776]: I1125 11:07:45.642421 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-housekeeping-bd8xp" Nov 25 11:07:45 crc kubenswrapper[4776]: I1125 11:07:45.779265 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8k6jm" Nov 25 11:07:45 crc kubenswrapper[4776]: I1125 11:07:45.828369 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44844357-d27a-454b-adbc-be91bed3a5a0-utilities\") pod \"44844357-d27a-454b-adbc-be91bed3a5a0\" (UID: \"44844357-d27a-454b-adbc-be91bed3a5a0\") " Nov 25 11:07:45 crc kubenswrapper[4776]: I1125 11:07:45.829095 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44844357-d27a-454b-adbc-be91bed3a5a0-utilities" (OuterVolumeSpecName: "utilities") pod "44844357-d27a-454b-adbc-be91bed3a5a0" (UID: "44844357-d27a-454b-adbc-be91bed3a5a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:07:45 crc kubenswrapper[4776]: I1125 11:07:45.929613 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24wmc\" (UniqueName: \"kubernetes.io/projected/44844357-d27a-454b-adbc-be91bed3a5a0-kube-api-access-24wmc\") pod \"44844357-d27a-454b-adbc-be91bed3a5a0\" (UID: \"44844357-d27a-454b-adbc-be91bed3a5a0\") " Nov 25 11:07:45 crc kubenswrapper[4776]: I1125 11:07:45.929742 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44844357-d27a-454b-adbc-be91bed3a5a0-catalog-content\") pod \"44844357-d27a-454b-adbc-be91bed3a5a0\" (UID: \"44844357-d27a-454b-adbc-be91bed3a5a0\") " Nov 25 11:07:45 crc kubenswrapper[4776]: I1125 11:07:45.930230 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44844357-d27a-454b-adbc-be91bed3a5a0-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:07:45 crc kubenswrapper[4776]: I1125 11:07:45.943891 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44844357-d27a-454b-adbc-be91bed3a5a0-kube-api-access-24wmc" (OuterVolumeSpecName: "kube-api-access-24wmc") pod "44844357-d27a-454b-adbc-be91bed3a5a0" (UID: "44844357-d27a-454b-adbc-be91bed3a5a0"). InnerVolumeSpecName "kube-api-access-24wmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:07:46 crc kubenswrapper[4776]: I1125 11:07:46.030775 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44844357-d27a-454b-adbc-be91bed3a5a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "44844357-d27a-454b-adbc-be91bed3a5a0" (UID: "44844357-d27a-454b-adbc-be91bed3a5a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:07:46 crc kubenswrapper[4776]: I1125 11:07:46.033463 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24wmc\" (UniqueName: \"kubernetes.io/projected/44844357-d27a-454b-adbc-be91bed3a5a0-kube-api-access-24wmc\") on node \"crc\" DevicePath \"\"" Nov 25 11:07:46 crc kubenswrapper[4776]: I1125 11:07:46.033680 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44844357-d27a-454b-adbc-be91bed3a5a0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:07:46 crc kubenswrapper[4776]: I1125 11:07:46.175323 4776 generic.go:334] "Generic (PLEG): container finished" podID="44844357-d27a-454b-adbc-be91bed3a5a0" containerID="cca0d85f30f5a0aa861d35d468ba1704f94763a74cdd9fcadb0dbfd3b1d044a6" exitCode=0 Nov 25 11:07:46 crc kubenswrapper[4776]: I1125 11:07:46.175419 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8k6jm" Nov 25 11:07:46 crc kubenswrapper[4776]: I1125 11:07:46.175410 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8k6jm" event={"ID":"44844357-d27a-454b-adbc-be91bed3a5a0","Type":"ContainerDied","Data":"cca0d85f30f5a0aa861d35d468ba1704f94763a74cdd9fcadb0dbfd3b1d044a6"} Nov 25 11:07:46 crc kubenswrapper[4776]: I1125 11:07:46.176026 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8k6jm" event={"ID":"44844357-d27a-454b-adbc-be91bed3a5a0","Type":"ContainerDied","Data":"c32496cff092b861908d282f267eb4c61341f3fb38d90f5e8dcba309394eff4c"} Nov 25 11:07:46 crc kubenswrapper[4776]: I1125 11:07:46.176108 4776 scope.go:117] "RemoveContainer" containerID="cca0d85f30f5a0aa861d35d468ba1704f94763a74cdd9fcadb0dbfd3b1d044a6" Nov 25 11:07:46 crc kubenswrapper[4776]: I1125 11:07:46.227107 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8k6jm"] Nov 25 11:07:46 crc kubenswrapper[4776]: I1125 11:07:46.227792 4776 scope.go:117] "RemoveContainer" containerID="f96ad2eff5a5b75c94badf0988d86a8b7e56febec7ed2df11d2e7dba80fb9bbc" Nov 25 11:07:46 crc kubenswrapper[4776]: I1125 11:07:46.236780 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8k6jm"] Nov 25 11:07:46 crc kubenswrapper[4776]: I1125 11:07:46.260297 4776 scope.go:117] "RemoveContainer" containerID="fd7f93b212affaa250086af9ce571010f33bd3f5c0388a0a7199f70d0dcbbbff" Nov 25 11:07:46 crc kubenswrapper[4776]: I1125 11:07:46.313684 4776 scope.go:117] "RemoveContainer" containerID="cca0d85f30f5a0aa861d35d468ba1704f94763a74cdd9fcadb0dbfd3b1d044a6" Nov 25 11:07:46 crc kubenswrapper[4776]: E1125 11:07:46.314254 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cca0d85f30f5a0aa861d35d468ba1704f94763a74cdd9fcadb0dbfd3b1d044a6\": container with ID starting with cca0d85f30f5a0aa861d35d468ba1704f94763a74cdd9fcadb0dbfd3b1d044a6 not found: ID does not exist" containerID="cca0d85f30f5a0aa861d35d468ba1704f94763a74cdd9fcadb0dbfd3b1d044a6" Nov 25 11:07:46 crc kubenswrapper[4776]: I1125 11:07:46.314364 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cca0d85f30f5a0aa861d35d468ba1704f94763a74cdd9fcadb0dbfd3b1d044a6"} err="failed to get container status \"cca0d85f30f5a0aa861d35d468ba1704f94763a74cdd9fcadb0dbfd3b1d044a6\": rpc error: code = NotFound desc = could not find container \"cca0d85f30f5a0aa861d35d468ba1704f94763a74cdd9fcadb0dbfd3b1d044a6\": container with ID starting with cca0d85f30f5a0aa861d35d468ba1704f94763a74cdd9fcadb0dbfd3b1d044a6 not found: ID does not exist" Nov 25 11:07:46 crc kubenswrapper[4776]: I1125 11:07:46.314459 4776 scope.go:117] "RemoveContainer" containerID="f96ad2eff5a5b75c94badf0988d86a8b7e56febec7ed2df11d2e7dba80fb9bbc" Nov 25 11:07:46 crc kubenswrapper[4776]: E1125 11:07:46.315051 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f96ad2eff5a5b75c94badf0988d86a8b7e56febec7ed2df11d2e7dba80fb9bbc\": container with ID starting with f96ad2eff5a5b75c94badf0988d86a8b7e56febec7ed2df11d2e7dba80fb9bbc not found: ID does not exist" containerID="f96ad2eff5a5b75c94badf0988d86a8b7e56febec7ed2df11d2e7dba80fb9bbc" Nov 25 11:07:46 crc kubenswrapper[4776]: I1125 11:07:46.315198 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f96ad2eff5a5b75c94badf0988d86a8b7e56febec7ed2df11d2e7dba80fb9bbc"} err="failed to get container status \"f96ad2eff5a5b75c94badf0988d86a8b7e56febec7ed2df11d2e7dba80fb9bbc\": rpc error: code = NotFound desc = could not find container \"f96ad2eff5a5b75c94badf0988d86a8b7e56febec7ed2df11d2e7dba80fb9bbc\": container with ID starting with f96ad2eff5a5b75c94badf0988d86a8b7e56febec7ed2df11d2e7dba80fb9bbc not found: ID does not exist" Nov 25 11:07:46 crc kubenswrapper[4776]: I1125 11:07:46.315278 4776 scope.go:117] "RemoveContainer" containerID="fd7f93b212affaa250086af9ce571010f33bd3f5c0388a0a7199f70d0dcbbbff" Nov 25 11:07:46 crc kubenswrapper[4776]: E1125 11:07:46.315568 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd7f93b212affaa250086af9ce571010f33bd3f5c0388a0a7199f70d0dcbbbff\": container with ID starting with fd7f93b212affaa250086af9ce571010f33bd3f5c0388a0a7199f70d0dcbbbff not found: ID does not exist" containerID="fd7f93b212affaa250086af9ce571010f33bd3f5c0388a0a7199f70d0dcbbbff" Nov 25 11:07:46 crc kubenswrapper[4776]: I1125 11:07:46.315670 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd7f93b212affaa250086af9ce571010f33bd3f5c0388a0a7199f70d0dcbbbff"} err="failed to get container status \"fd7f93b212affaa250086af9ce571010f33bd3f5c0388a0a7199f70d0dcbbbff\": rpc error: code = NotFound desc = could not find container \"fd7f93b212affaa250086af9ce571010f33bd3f5c0388a0a7199f70d0dcbbbff\": container with ID starting with fd7f93b212affaa250086af9ce571010f33bd3f5c0388a0a7199f70d0dcbbbff not found: ID does not exist" Nov 25 11:07:47 crc kubenswrapper[4776]: I1125 11:07:47.674329 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44844357-d27a-454b-adbc-be91bed3a5a0" path="/var/lib/kubelet/pods/44844357-d27a-454b-adbc-be91bed3a5a0/volumes" Nov 25 11:07:48 crc kubenswrapper[4776]: I1125 11:07:48.117303 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-worker-jzw7v" Nov 25 11:07:52 crc kubenswrapper[4776]: I1125 11:07:52.041585 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-6t7lg"] Nov 25 11:07:52 crc kubenswrapper[4776]: I1125 11:07:52.055727 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-2c0f-account-create-pn6jc"] Nov 25 11:07:52 crc kubenswrapper[4776]: I1125 11:07:52.064408 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-2c0f-account-create-pn6jc"] Nov 25 11:07:52 crc kubenswrapper[4776]: I1125 11:07:52.072689 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-6t7lg"] Nov 25 11:07:53 crc kubenswrapper[4776]: I1125 11:07:53.695425 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="410928cc-80b1-415a-a02c-83ca27897dca" path="/var/lib/kubelet/pods/410928cc-80b1-415a-a02c-83ca27897dca/volumes" Nov 25 11:07:53 crc kubenswrapper[4776]: I1125 11:07:53.697619 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7eac8e5-4d91-4316-981c-14c58d3a261a" path="/var/lib/kubelet/pods/b7eac8e5-4d91-4316-981c-14c58d3a261a/volumes" Nov 25 11:07:55 crc kubenswrapper[4776]: I1125 11:07:55.670589 4776 scope.go:117] "RemoveContainer" containerID="8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698" Nov 25 11:07:55 crc kubenswrapper[4776]: E1125 11:07:55.671164 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:08:02 crc kubenswrapper[4776]: I1125 11:08:02.037619 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-gwjl8"] Nov 25 11:08:02 crc kubenswrapper[4776]: I1125 11:08:02.046764 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-gwjl8"] Nov 25 11:08:03 crc kubenswrapper[4776]: I1125 11:08:03.674164 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c027d749-e028-4f8e-9b29-f70d6608c04f" path="/var/lib/kubelet/pods/c027d749-e028-4f8e-9b29-f70d6608c04f/volumes" Nov 25 11:08:08 crc kubenswrapper[4776]: I1125 11:08:08.662930 4776 scope.go:117] "RemoveContainer" containerID="8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698" Nov 25 11:08:08 crc kubenswrapper[4776]: E1125 11:08:08.663944 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:08:23 crc kubenswrapper[4776]: I1125 11:08:23.662917 4776 scope.go:117] "RemoveContainer" containerID="8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698" Nov 25 11:08:23 crc kubenswrapper[4776]: E1125 11:08:23.663771 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:08:33 crc kubenswrapper[4776]: I1125 11:08:33.938097 4776 scope.go:117] "RemoveContainer" containerID="5be9efcdeacb408f3fd872d2ba344432f9c962a46ad4a8ab56fa7d01586c114c" Nov 25 11:08:33 crc kubenswrapper[4776]: I1125 11:08:33.972831 4776 scope.go:117] "RemoveContainer" containerID="7fa743ab4a5cc8af2b03d5467b695a058f40063b22924ac92defa7c75c4ad680" Nov 25 11:08:34 crc kubenswrapper[4776]: I1125 11:08:34.054799 4776 scope.go:117] "RemoveContainer" containerID="82259d19d7e1f4c13b2c491ddc71c89165b933f1c30a6993ba43a9831cd7b6a6" Nov 25 11:08:35 crc kubenswrapper[4776]: I1125 11:08:35.669989 4776 scope.go:117] "RemoveContainer" containerID="8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698" Nov 25 11:08:35 crc kubenswrapper[4776]: E1125 11:08:35.670678 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.330551 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7b6b74d589-7lsxv"] Nov 25 11:08:37 crc kubenswrapper[4776]: E1125 11:08:37.331314 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44844357-d27a-454b-adbc-be91bed3a5a0" containerName="extract-content" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.331329 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="44844357-d27a-454b-adbc-be91bed3a5a0" containerName="extract-content" Nov 25 11:08:37 crc kubenswrapper[4776]: E1125 11:08:37.331344 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44844357-d27a-454b-adbc-be91bed3a5a0" containerName="registry-server" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.331351 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="44844357-d27a-454b-adbc-be91bed3a5a0" containerName="registry-server" Nov 25 11:08:37 crc kubenswrapper[4776]: E1125 11:08:37.331371 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44844357-d27a-454b-adbc-be91bed3a5a0" containerName="extract-utilities" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.331378 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="44844357-d27a-454b-adbc-be91bed3a5a0" containerName="extract-utilities" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.331614 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="44844357-d27a-454b-adbc-be91bed3a5a0" containerName="registry-server" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.340940 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b6b74d589-7lsxv" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.345400 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.345635 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-d8vr2" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.345829 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.346163 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7b6b74d589-7lsxv"] Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.346303 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.372834 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.373094 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="afe1ae15-e85d-4d25-81ef-445a574d8222" containerName="glance-log" containerID="cri-o://4b90d93d9326d7485e4ab371e06f94e76c0b64f3ce6adcbb172886a6fa5348d8" gracePeriod=30 Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.373144 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="afe1ae15-e85d-4d25-81ef-445a574d8222" containerName="glance-httpd" containerID="cri-o://614824f19f67266106422717756035ef58d54d132e6357b434ced6ff82605737" gracePeriod=30 Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.445581 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-599bd8d6b9-hz9cl"] Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.450729 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-599bd8d6b9-hz9cl" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.467456 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.467764 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a" containerName="glance-log" containerID="cri-o://5d69cfffdca6866fa1479b16f66e7d4638a5e2afa489a7050cef817c49652ad8" gracePeriod=30 Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.467953 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a" containerName="glance-httpd" containerID="cri-o://d67251a37a2930b6bc7b71c9e7bce4cd60e86e67780a1487e294e57b5a93ca21" gracePeriod=30 Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.473873 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c465b70c-b76d-430c-954e-b8408390cd99-config-data\") pod \"horizon-7b6b74d589-7lsxv\" (UID: \"c465b70c-b76d-430c-954e-b8408390cd99\") " pod="openstack/horizon-7b6b74d589-7lsxv" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.473951 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c465b70c-b76d-430c-954e-b8408390cd99-horizon-secret-key\") pod \"horizon-7b6b74d589-7lsxv\" (UID: \"c465b70c-b76d-430c-954e-b8408390cd99\") " pod="openstack/horizon-7b6b74d589-7lsxv" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.474051 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c465b70c-b76d-430c-954e-b8408390cd99-logs\") pod \"horizon-7b6b74d589-7lsxv\" (UID: \"c465b70c-b76d-430c-954e-b8408390cd99\") " pod="openstack/horizon-7b6b74d589-7lsxv" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.474201 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c465b70c-b76d-430c-954e-b8408390cd99-scripts\") pod \"horizon-7b6b74d589-7lsxv\" (UID: \"c465b70c-b76d-430c-954e-b8408390cd99\") " pod="openstack/horizon-7b6b74d589-7lsxv" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.474336 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w26nr\" (UniqueName: \"kubernetes.io/projected/c465b70c-b76d-430c-954e-b8408390cd99-kube-api-access-w26nr\") pod \"horizon-7b6b74d589-7lsxv\" (UID: \"c465b70c-b76d-430c-954e-b8408390cd99\") " pod="openstack/horizon-7b6b74d589-7lsxv" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.488591 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-599bd8d6b9-hz9cl"] Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.576531 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkx9l\" (UniqueName: \"kubernetes.io/projected/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-kube-api-access-vkx9l\") pod \"horizon-599bd8d6b9-hz9cl\" (UID: \"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c\") " pod="openstack/horizon-599bd8d6b9-hz9cl" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.576604 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c465b70c-b76d-430c-954e-b8408390cd99-scripts\") pod \"horizon-7b6b74d589-7lsxv\" (UID: \"c465b70c-b76d-430c-954e-b8408390cd99\") " pod="openstack/horizon-7b6b74d589-7lsxv" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.576689 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w26nr\" (UniqueName: \"kubernetes.io/projected/c465b70c-b76d-430c-954e-b8408390cd99-kube-api-access-w26nr\") pod \"horizon-7b6b74d589-7lsxv\" (UID: \"c465b70c-b76d-430c-954e-b8408390cd99\") " pod="openstack/horizon-7b6b74d589-7lsxv" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.576754 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c465b70c-b76d-430c-954e-b8408390cd99-config-data\") pod \"horizon-7b6b74d589-7lsxv\" (UID: \"c465b70c-b76d-430c-954e-b8408390cd99\") " pod="openstack/horizon-7b6b74d589-7lsxv" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.576780 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c465b70c-b76d-430c-954e-b8408390cd99-horizon-secret-key\") pod \"horizon-7b6b74d589-7lsxv\" (UID: \"c465b70c-b76d-430c-954e-b8408390cd99\") " pod="openstack/horizon-7b6b74d589-7lsxv" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.576801 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-logs\") pod \"horizon-599bd8d6b9-hz9cl\" (UID: \"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c\") " pod="openstack/horizon-599bd8d6b9-hz9cl" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.576837 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-config-data\") pod \"horizon-599bd8d6b9-hz9cl\" (UID: \"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c\") " pod="openstack/horizon-599bd8d6b9-hz9cl" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.576854 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c465b70c-b76d-430c-954e-b8408390cd99-logs\") pod \"horizon-7b6b74d589-7lsxv\" (UID: \"c465b70c-b76d-430c-954e-b8408390cd99\") " pod="openstack/horizon-7b6b74d589-7lsxv" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.576888 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-horizon-secret-key\") pod \"horizon-599bd8d6b9-hz9cl\" (UID: \"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c\") " pod="openstack/horizon-599bd8d6b9-hz9cl" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.576917 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-scripts\") pod \"horizon-599bd8d6b9-hz9cl\" (UID: \"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c\") " pod="openstack/horizon-599bd8d6b9-hz9cl" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.577762 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c465b70c-b76d-430c-954e-b8408390cd99-scripts\") pod \"horizon-7b6b74d589-7lsxv\" (UID: \"c465b70c-b76d-430c-954e-b8408390cd99\") " pod="openstack/horizon-7b6b74d589-7lsxv" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.578693 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c465b70c-b76d-430c-954e-b8408390cd99-logs\") pod \"horizon-7b6b74d589-7lsxv\" (UID: \"c465b70c-b76d-430c-954e-b8408390cd99\") " pod="openstack/horizon-7b6b74d589-7lsxv" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.579798 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c465b70c-b76d-430c-954e-b8408390cd99-config-data\") pod \"horizon-7b6b74d589-7lsxv\" (UID: \"c465b70c-b76d-430c-954e-b8408390cd99\") " pod="openstack/horizon-7b6b74d589-7lsxv" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.584589 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c465b70c-b76d-430c-954e-b8408390cd99-horizon-secret-key\") pod \"horizon-7b6b74d589-7lsxv\" (UID: \"c465b70c-b76d-430c-954e-b8408390cd99\") " pod="openstack/horizon-7b6b74d589-7lsxv" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.602620 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w26nr\" (UniqueName: \"kubernetes.io/projected/c465b70c-b76d-430c-954e-b8408390cd99-kube-api-access-w26nr\") pod \"horizon-7b6b74d589-7lsxv\" (UID: \"c465b70c-b76d-430c-954e-b8408390cd99\") " pod="openstack/horizon-7b6b74d589-7lsxv" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.639133 4776 generic.go:334] "Generic (PLEG): container finished" podID="2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a" containerID="5d69cfffdca6866fa1479b16f66e7d4638a5e2afa489a7050cef817c49652ad8" exitCode=143 Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.639209 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a","Type":"ContainerDied","Data":"5d69cfffdca6866fa1479b16f66e7d4638a5e2afa489a7050cef817c49652ad8"} Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.640884 4776 generic.go:334] "Generic (PLEG): container finished" podID="afe1ae15-e85d-4d25-81ef-445a574d8222" containerID="4b90d93d9326d7485e4ab371e06f94e76c0b64f3ce6adcbb172886a6fa5348d8" exitCode=143 Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.642055 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"afe1ae15-e85d-4d25-81ef-445a574d8222","Type":"ContainerDied","Data":"4b90d93d9326d7485e4ab371e06f94e76c0b64f3ce6adcbb172886a6fa5348d8"} Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.661482 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b6b74d589-7lsxv" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.686215 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-logs\") pod \"horizon-599bd8d6b9-hz9cl\" (UID: \"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c\") " pod="openstack/horizon-599bd8d6b9-hz9cl" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.686677 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-config-data\") pod \"horizon-599bd8d6b9-hz9cl\" (UID: \"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c\") " pod="openstack/horizon-599bd8d6b9-hz9cl" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.686875 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-horizon-secret-key\") pod \"horizon-599bd8d6b9-hz9cl\" (UID: \"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c\") " pod="openstack/horizon-599bd8d6b9-hz9cl" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.687025 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-scripts\") pod \"horizon-599bd8d6b9-hz9cl\" (UID: \"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c\") " pod="openstack/horizon-599bd8d6b9-hz9cl" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.687203 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkx9l\" (UniqueName: \"kubernetes.io/projected/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-kube-api-access-vkx9l\") pod \"horizon-599bd8d6b9-hz9cl\" (UID: \"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c\") " pod="openstack/horizon-599bd8d6b9-hz9cl" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.687666 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-config-data\") pod \"horizon-599bd8d6b9-hz9cl\" (UID: \"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c\") " pod="openstack/horizon-599bd8d6b9-hz9cl" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.688663 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-scripts\") pod \"horizon-599bd8d6b9-hz9cl\" (UID: \"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c\") " pod="openstack/horizon-599bd8d6b9-hz9cl" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.686683 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-logs\") pod \"horizon-599bd8d6b9-hz9cl\" (UID: \"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c\") " pod="openstack/horizon-599bd8d6b9-hz9cl" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.694356 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-horizon-secret-key\") pod \"horizon-599bd8d6b9-hz9cl\" (UID: \"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c\") " pod="openstack/horizon-599bd8d6b9-hz9cl" Nov 25 11:08:37 crc kubenswrapper[4776]: I1125 11:08:37.708721 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkx9l\" (UniqueName: \"kubernetes.io/projected/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-kube-api-access-vkx9l\") pod \"horizon-599bd8d6b9-hz9cl\" (UID: \"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c\") " pod="openstack/horizon-599bd8d6b9-hz9cl" Nov 25 11:08:38 crc kubenswrapper[4776]: I1125 11:08:37.780977 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-599bd8d6b9-hz9cl" Nov 25 11:08:38 crc kubenswrapper[4776]: I1125 11:08:38.924859 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7b6b74d589-7lsxv"] Nov 25 11:08:38 crc kubenswrapper[4776]: I1125 11:08:38.941301 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-599bd8d6b9-hz9cl"] Nov 25 11:08:38 crc kubenswrapper[4776]: W1125 11:08:38.945739 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4ceac98_94a9_4cc2_a405_d57a1a15cf9c.slice/crio-d3a620cc9e70afefea5e7c11ed49253141433709c1cfbc61a3eee63384412e88 WatchSource:0}: Error finding container d3a620cc9e70afefea5e7c11ed49253141433709c1cfbc61a3eee63384412e88: Status 404 returned error can't find the container with id d3a620cc9e70afefea5e7c11ed49253141433709c1cfbc61a3eee63384412e88 Nov 25 11:08:39 crc kubenswrapper[4776]: I1125 11:08:39.681495 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b6b74d589-7lsxv" event={"ID":"c465b70c-b76d-430c-954e-b8408390cd99","Type":"ContainerStarted","Data":"adf9775a0780deb6241108dfcb0371e0702b8e4cd74cf9d350386abcd8cf9cc6"} Nov 25 11:08:39 crc kubenswrapper[4776]: I1125 11:08:39.685518 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-599bd8d6b9-hz9cl" event={"ID":"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c","Type":"ContainerStarted","Data":"d3a620cc9e70afefea5e7c11ed49253141433709c1cfbc61a3eee63384412e88"} Nov 25 11:08:39 crc kubenswrapper[4776]: I1125 11:08:39.827835 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-599bd8d6b9-hz9cl"] Nov 25 11:08:39 crc kubenswrapper[4776]: I1125 11:08:39.865165 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7c9cb59558-gpdwk"] Nov 25 11:08:39 crc kubenswrapper[4776]: I1125 11:08:39.866838 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:08:39 crc kubenswrapper[4776]: I1125 11:08:39.874887 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Nov 25 11:08:39 crc kubenswrapper[4776]: I1125 11:08:39.881118 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c9cb59558-gpdwk"] Nov 25 11:08:39 crc kubenswrapper[4776]: I1125 11:08:39.949880 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/238717c2-18fa-47d1-b39e-4d2f0b3619de-horizon-tls-certs\") pod \"horizon-7c9cb59558-gpdwk\" (UID: \"238717c2-18fa-47d1-b39e-4d2f0b3619de\") " pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:08:39 crc kubenswrapper[4776]: I1125 11:08:39.949936 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/238717c2-18fa-47d1-b39e-4d2f0b3619de-horizon-secret-key\") pod \"horizon-7c9cb59558-gpdwk\" (UID: \"238717c2-18fa-47d1-b39e-4d2f0b3619de\") " pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:08:39 crc kubenswrapper[4776]: I1125 11:08:39.949966 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/238717c2-18fa-47d1-b39e-4d2f0b3619de-logs\") pod \"horizon-7c9cb59558-gpdwk\" (UID: \"238717c2-18fa-47d1-b39e-4d2f0b3619de\") " pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:08:39 crc kubenswrapper[4776]: I1125 11:08:39.949993 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-659v5\" (UniqueName: \"kubernetes.io/projected/238717c2-18fa-47d1-b39e-4d2f0b3619de-kube-api-access-659v5\") pod \"horizon-7c9cb59558-gpdwk\" (UID: \"238717c2-18fa-47d1-b39e-4d2f0b3619de\") " pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:08:39 crc kubenswrapper[4776]: I1125 11:08:39.950047 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/238717c2-18fa-47d1-b39e-4d2f0b3619de-scripts\") pod \"horizon-7c9cb59558-gpdwk\" (UID: \"238717c2-18fa-47d1-b39e-4d2f0b3619de\") " pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:08:39 crc kubenswrapper[4776]: I1125 11:08:39.950119 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/238717c2-18fa-47d1-b39e-4d2f0b3619de-combined-ca-bundle\") pod \"horizon-7c9cb59558-gpdwk\" (UID: \"238717c2-18fa-47d1-b39e-4d2f0b3619de\") " pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:08:39 crc kubenswrapper[4776]: I1125 11:08:39.950159 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/238717c2-18fa-47d1-b39e-4d2f0b3619de-config-data\") pod \"horizon-7c9cb59558-gpdwk\" (UID: \"238717c2-18fa-47d1-b39e-4d2f0b3619de\") " pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:08:39 crc kubenswrapper[4776]: I1125 11:08:39.952819 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7b6b74d589-7lsxv"] Nov 25 11:08:39 crc kubenswrapper[4776]: I1125 11:08:39.973239 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-656cc574-dstj4"] Nov 25 11:08:39 crc kubenswrapper[4776]: I1125 11:08:39.975401 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-656cc574-dstj4" Nov 25 11:08:39 crc kubenswrapper[4776]: I1125 11:08:39.983255 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-656cc574-dstj4"] Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.051997 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/238717c2-18fa-47d1-b39e-4d2f0b3619de-horizon-tls-certs\") pod \"horizon-7c9cb59558-gpdwk\" (UID: \"238717c2-18fa-47d1-b39e-4d2f0b3619de\") " pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.052051 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/238717c2-18fa-47d1-b39e-4d2f0b3619de-horizon-secret-key\") pod \"horizon-7c9cb59558-gpdwk\" (UID: \"238717c2-18fa-47d1-b39e-4d2f0b3619de\") " pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.052096 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/219a7347-5102-4e6c-9b38-deea6275f715-scripts\") pod \"horizon-656cc574-dstj4\" (UID: \"219a7347-5102-4e6c-9b38-deea6275f715\") " pod="openstack/horizon-656cc574-dstj4" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.052121 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/238717c2-18fa-47d1-b39e-4d2f0b3619de-logs\") pod \"horizon-7c9cb59558-gpdwk\" (UID: \"238717c2-18fa-47d1-b39e-4d2f0b3619de\") " pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.052142 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-659v5\" (UniqueName: \"kubernetes.io/projected/238717c2-18fa-47d1-b39e-4d2f0b3619de-kube-api-access-659v5\") pod \"horizon-7c9cb59558-gpdwk\" (UID: \"238717c2-18fa-47d1-b39e-4d2f0b3619de\") " pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.052162 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkvvq\" (UniqueName: \"kubernetes.io/projected/219a7347-5102-4e6c-9b38-deea6275f715-kube-api-access-kkvvq\") pod \"horizon-656cc574-dstj4\" (UID: \"219a7347-5102-4e6c-9b38-deea6275f715\") " pod="openstack/horizon-656cc574-dstj4" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.052186 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/219a7347-5102-4e6c-9b38-deea6275f715-horizon-secret-key\") pod \"horizon-656cc574-dstj4\" (UID: \"219a7347-5102-4e6c-9b38-deea6275f715\") " pod="openstack/horizon-656cc574-dstj4" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.052214 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/219a7347-5102-4e6c-9b38-deea6275f715-config-data\") pod \"horizon-656cc574-dstj4\" (UID: \"219a7347-5102-4e6c-9b38-deea6275f715\") " pod="openstack/horizon-656cc574-dstj4" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.052231 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/238717c2-18fa-47d1-b39e-4d2f0b3619de-scripts\") pod \"horizon-7c9cb59558-gpdwk\" (UID: \"238717c2-18fa-47d1-b39e-4d2f0b3619de\") " pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.052266 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/219a7347-5102-4e6c-9b38-deea6275f715-combined-ca-bundle\") pod \"horizon-656cc574-dstj4\" (UID: \"219a7347-5102-4e6c-9b38-deea6275f715\") " pod="openstack/horizon-656cc574-dstj4" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.052289 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/238717c2-18fa-47d1-b39e-4d2f0b3619de-combined-ca-bundle\") pod \"horizon-7c9cb59558-gpdwk\" (UID: \"238717c2-18fa-47d1-b39e-4d2f0b3619de\") " pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.052308 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/219a7347-5102-4e6c-9b38-deea6275f715-logs\") pod \"horizon-656cc574-dstj4\" (UID: \"219a7347-5102-4e6c-9b38-deea6275f715\") " pod="openstack/horizon-656cc574-dstj4" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.052334 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/238717c2-18fa-47d1-b39e-4d2f0b3619de-config-data\") pod \"horizon-7c9cb59558-gpdwk\" (UID: \"238717c2-18fa-47d1-b39e-4d2f0b3619de\") " pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.052371 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/219a7347-5102-4e6c-9b38-deea6275f715-horizon-tls-certs\") pod \"horizon-656cc574-dstj4\" (UID: \"219a7347-5102-4e6c-9b38-deea6275f715\") " pod="openstack/horizon-656cc574-dstj4" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.053827 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/238717c2-18fa-47d1-b39e-4d2f0b3619de-logs\") pod \"horizon-7c9cb59558-gpdwk\" (UID: \"238717c2-18fa-47d1-b39e-4d2f0b3619de\") " pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.054773 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/238717c2-18fa-47d1-b39e-4d2f0b3619de-scripts\") pod \"horizon-7c9cb59558-gpdwk\" (UID: \"238717c2-18fa-47d1-b39e-4d2f0b3619de\") " pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.055664 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/238717c2-18fa-47d1-b39e-4d2f0b3619de-config-data\") pod \"horizon-7c9cb59558-gpdwk\" (UID: \"238717c2-18fa-47d1-b39e-4d2f0b3619de\") " pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.060509 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/238717c2-18fa-47d1-b39e-4d2f0b3619de-horizon-secret-key\") pod \"horizon-7c9cb59558-gpdwk\" (UID: \"238717c2-18fa-47d1-b39e-4d2f0b3619de\") " pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.060766 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/238717c2-18fa-47d1-b39e-4d2f0b3619de-combined-ca-bundle\") pod \"horizon-7c9cb59558-gpdwk\" (UID: \"238717c2-18fa-47d1-b39e-4d2f0b3619de\") " pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.073529 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/238717c2-18fa-47d1-b39e-4d2f0b3619de-horizon-tls-certs\") pod \"horizon-7c9cb59558-gpdwk\" (UID: \"238717c2-18fa-47d1-b39e-4d2f0b3619de\") " pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.076349 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-659v5\" (UniqueName: \"kubernetes.io/projected/238717c2-18fa-47d1-b39e-4d2f0b3619de-kube-api-access-659v5\") pod \"horizon-7c9cb59558-gpdwk\" (UID: \"238717c2-18fa-47d1-b39e-4d2f0b3619de\") " pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.153762 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/219a7347-5102-4e6c-9b38-deea6275f715-config-data\") pod \"horizon-656cc574-dstj4\" (UID: \"219a7347-5102-4e6c-9b38-deea6275f715\") " pod="openstack/horizon-656cc574-dstj4" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.153855 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/219a7347-5102-4e6c-9b38-deea6275f715-combined-ca-bundle\") pod \"horizon-656cc574-dstj4\" (UID: \"219a7347-5102-4e6c-9b38-deea6275f715\") " pod="openstack/horizon-656cc574-dstj4" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.153899 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/219a7347-5102-4e6c-9b38-deea6275f715-logs\") pod \"horizon-656cc574-dstj4\" (UID: \"219a7347-5102-4e6c-9b38-deea6275f715\") " pod="openstack/horizon-656cc574-dstj4" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.153955 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/219a7347-5102-4e6c-9b38-deea6275f715-horizon-tls-certs\") pod \"horizon-656cc574-dstj4\" (UID: \"219a7347-5102-4e6c-9b38-deea6275f715\") " pod="openstack/horizon-656cc574-dstj4" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.154084 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/219a7347-5102-4e6c-9b38-deea6275f715-scripts\") pod \"horizon-656cc574-dstj4\" (UID: \"219a7347-5102-4e6c-9b38-deea6275f715\") " pod="openstack/horizon-656cc574-dstj4" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.154134 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkvvq\" (UniqueName: \"kubernetes.io/projected/219a7347-5102-4e6c-9b38-deea6275f715-kube-api-access-kkvvq\") pod \"horizon-656cc574-dstj4\" (UID: \"219a7347-5102-4e6c-9b38-deea6275f715\") " pod="openstack/horizon-656cc574-dstj4" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.154165 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/219a7347-5102-4e6c-9b38-deea6275f715-horizon-secret-key\") pod \"horizon-656cc574-dstj4\" (UID: \"219a7347-5102-4e6c-9b38-deea6275f715\") " pod="openstack/horizon-656cc574-dstj4" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.155254 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/219a7347-5102-4e6c-9b38-deea6275f715-config-data\") pod \"horizon-656cc574-dstj4\" (UID: \"219a7347-5102-4e6c-9b38-deea6275f715\") " pod="openstack/horizon-656cc574-dstj4" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.155737 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/219a7347-5102-4e6c-9b38-deea6275f715-logs\") pod \"horizon-656cc574-dstj4\" (UID: \"219a7347-5102-4e6c-9b38-deea6275f715\") " pod="openstack/horizon-656cc574-dstj4" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.155827 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/219a7347-5102-4e6c-9b38-deea6275f715-scripts\") pod \"horizon-656cc574-dstj4\" (UID: \"219a7347-5102-4e6c-9b38-deea6275f715\") " pod="openstack/horizon-656cc574-dstj4" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.162723 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/219a7347-5102-4e6c-9b38-deea6275f715-horizon-tls-certs\") pod \"horizon-656cc574-dstj4\" (UID: \"219a7347-5102-4e6c-9b38-deea6275f715\") " pod="openstack/horizon-656cc574-dstj4" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.164012 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/219a7347-5102-4e6c-9b38-deea6275f715-horizon-secret-key\") pod \"horizon-656cc574-dstj4\" (UID: \"219a7347-5102-4e6c-9b38-deea6275f715\") " pod="openstack/horizon-656cc574-dstj4" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.167902 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/219a7347-5102-4e6c-9b38-deea6275f715-combined-ca-bundle\") pod \"horizon-656cc574-dstj4\" (UID: \"219a7347-5102-4e6c-9b38-deea6275f715\") " pod="openstack/horizon-656cc574-dstj4" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.173911 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkvvq\" (UniqueName: \"kubernetes.io/projected/219a7347-5102-4e6c-9b38-deea6275f715-kube-api-access-kkvvq\") pod \"horizon-656cc574-dstj4\" (UID: \"219a7347-5102-4e6c-9b38-deea6275f715\") " pod="openstack/horizon-656cc574-dstj4" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.217348 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.305251 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-656cc574-dstj4" Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.699106 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c9cb59558-gpdwk"] Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.705527 4776 generic.go:334] "Generic (PLEG): container finished" podID="afe1ae15-e85d-4d25-81ef-445a574d8222" containerID="614824f19f67266106422717756035ef58d54d132e6357b434ced6ff82605737" exitCode=0 Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.705601 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"afe1ae15-e85d-4d25-81ef-445a574d8222","Type":"ContainerDied","Data":"614824f19f67266106422717756035ef58d54d132e6357b434ced6ff82605737"} Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.716697 4776 generic.go:334] "Generic (PLEG): container finished" podID="2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a" containerID="d67251a37a2930b6bc7b71c9e7bce4cd60e86e67780a1487e294e57b5a93ca21" exitCode=0 Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.716761 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a","Type":"ContainerDied","Data":"d67251a37a2930b6bc7b71c9e7bce4cd60e86e67780a1487e294e57b5a93ca21"} Nov 25 11:08:40 crc kubenswrapper[4776]: W1125 11:08:40.742411 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod238717c2_18fa_47d1_b39e_4d2f0b3619de.slice/crio-d4cdf4801299a49009f51e63142359764d6d309af9ebd9974481e2eed74bc82e WatchSource:0}: Error finding container d4cdf4801299a49009f51e63142359764d6d309af9ebd9974481e2eed74bc82e: Status 404 returned error can't find the container with id d4cdf4801299a49009f51e63142359764d6d309af9ebd9974481e2eed74bc82e Nov 25 11:08:40 crc kubenswrapper[4776]: I1125 11:08:40.848362 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-656cc574-dstj4"] Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.344280 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.348969 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.490269 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-scripts\") pod \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\" (UID: \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\") " Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.490538 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afe1ae15-e85d-4d25-81ef-445a574d8222-scripts\") pod \"afe1ae15-e85d-4d25-81ef-445a574d8222\" (UID: \"afe1ae15-e85d-4d25-81ef-445a574d8222\") " Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.490743 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afe1ae15-e85d-4d25-81ef-445a574d8222-config-data\") pod \"afe1ae15-e85d-4d25-81ef-445a574d8222\" (UID: \"afe1ae15-e85d-4d25-81ef-445a574d8222\") " Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.490815 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-combined-ca-bundle\") pod \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\" (UID: \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\") " Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.490903 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbpzr\" (UniqueName: \"kubernetes.io/projected/afe1ae15-e85d-4d25-81ef-445a574d8222-kube-api-access-kbpzr\") pod \"afe1ae15-e85d-4d25-81ef-445a574d8222\" (UID: \"afe1ae15-e85d-4d25-81ef-445a574d8222\") " Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.490973 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-internal-tls-certs\") pod \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\" (UID: \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\") " Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.491036 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afe1ae15-e85d-4d25-81ef-445a574d8222-logs\") pod \"afe1ae15-e85d-4d25-81ef-445a574d8222\" (UID: \"afe1ae15-e85d-4d25-81ef-445a574d8222\") " Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.491119 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-httpd-run\") pod \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\" (UID: \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\") " Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.491241 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-config-data\") pod \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\" (UID: \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\") " Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.491325 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afe1ae15-e85d-4d25-81ef-445a574d8222-combined-ca-bundle\") pod \"afe1ae15-e85d-4d25-81ef-445a574d8222\" (UID: \"afe1ae15-e85d-4d25-81ef-445a574d8222\") " Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.491405 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/afe1ae15-e85d-4d25-81ef-445a574d8222-public-tls-certs\") pod \"afe1ae15-e85d-4d25-81ef-445a574d8222\" (UID: \"afe1ae15-e85d-4d25-81ef-445a574d8222\") " Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.491521 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/afe1ae15-e85d-4d25-81ef-445a574d8222-httpd-run\") pod \"afe1ae15-e85d-4d25-81ef-445a574d8222\" (UID: \"afe1ae15-e85d-4d25-81ef-445a574d8222\") " Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.491591 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-logs\") pod \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\" (UID: \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\") " Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.491655 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cvhn\" (UniqueName: \"kubernetes.io/projected/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-kube-api-access-6cvhn\") pod \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\" (UID: \"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a\") " Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.494472 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afe1ae15-e85d-4d25-81ef-445a574d8222-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "afe1ae15-e85d-4d25-81ef-445a574d8222" (UID: "afe1ae15-e85d-4d25-81ef-445a574d8222"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.494744 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-logs" (OuterVolumeSpecName: "logs") pod "2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a" (UID: "2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.495183 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afe1ae15-e85d-4d25-81ef-445a574d8222-logs" (OuterVolumeSpecName: "logs") pod "afe1ae15-e85d-4d25-81ef-445a574d8222" (UID: "afe1ae15-e85d-4d25-81ef-445a574d8222"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.496255 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a" (UID: "2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.500300 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afe1ae15-e85d-4d25-81ef-445a574d8222-scripts" (OuterVolumeSpecName: "scripts") pod "afe1ae15-e85d-4d25-81ef-445a574d8222" (UID: "afe1ae15-e85d-4d25-81ef-445a574d8222"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.500980 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-kube-api-access-6cvhn" (OuterVolumeSpecName: "kube-api-access-6cvhn") pod "2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a" (UID: "2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a"). InnerVolumeSpecName "kube-api-access-6cvhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.509468 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afe1ae15-e85d-4d25-81ef-445a574d8222-kube-api-access-kbpzr" (OuterVolumeSpecName: "kube-api-access-kbpzr") pod "afe1ae15-e85d-4d25-81ef-445a574d8222" (UID: "afe1ae15-e85d-4d25-81ef-445a574d8222"). InnerVolumeSpecName "kube-api-access-kbpzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.514297 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-scripts" (OuterVolumeSpecName: "scripts") pod "2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a" (UID: "2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.542272 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a" (UID: "2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.552017 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afe1ae15-e85d-4d25-81ef-445a574d8222-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "afe1ae15-e85d-4d25-81ef-445a574d8222" (UID: "afe1ae15-e85d-4d25-81ef-445a574d8222"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.572760 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a" (UID: "2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.577880 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afe1ae15-e85d-4d25-81ef-445a574d8222-config-data" (OuterVolumeSpecName: "config-data") pod "afe1ae15-e85d-4d25-81ef-445a574d8222" (UID: "afe1ae15-e85d-4d25-81ef-445a574d8222"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.584986 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-config-data" (OuterVolumeSpecName: "config-data") pod "2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a" (UID: "2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.586844 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afe1ae15-e85d-4d25-81ef-445a574d8222-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "afe1ae15-e85d-4d25-81ef-445a574d8222" (UID: "afe1ae15-e85d-4d25-81ef-445a574d8222"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.594705 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.594757 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afe1ae15-e85d-4d25-81ef-445a574d8222-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.594770 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afe1ae15-e85d-4d25-81ef-445a574d8222-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.594784 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.594799 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbpzr\" (UniqueName: \"kubernetes.io/projected/afe1ae15-e85d-4d25-81ef-445a574d8222-kube-api-access-kbpzr\") on node \"crc\" DevicePath \"\"" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.594853 4776 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.594875 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afe1ae15-e85d-4d25-81ef-445a574d8222-logs\") on node \"crc\" DevicePath \"\"" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.594896 4776 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.594906 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.594916 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afe1ae15-e85d-4d25-81ef-445a574d8222-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.594927 4776 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/afe1ae15-e85d-4d25-81ef-445a574d8222-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.594937 4776 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/afe1ae15-e85d-4d25-81ef-445a574d8222-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.594947 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-logs\") on node \"crc\" DevicePath \"\"" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.594957 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cvhn\" (UniqueName: \"kubernetes.io/projected/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a-kube-api-access-6cvhn\") on node \"crc\" DevicePath \"\"" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.741633 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.742449 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"afe1ae15-e85d-4d25-81ef-445a574d8222","Type":"ContainerDied","Data":"5fe6069ad61b78b91b5e7db1494a23f0e4776a679e6abbd5081e8bf56edefc26"} Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.742483 4776 scope.go:117] "RemoveContainer" containerID="614824f19f67266106422717756035ef58d54d132e6357b434ced6ff82605737" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.758763 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c9cb59558-gpdwk" event={"ID":"238717c2-18fa-47d1-b39e-4d2f0b3619de","Type":"ContainerStarted","Data":"d4cdf4801299a49009f51e63142359764d6d309af9ebd9974481e2eed74bc82e"} Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.762713 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-656cc574-dstj4" event={"ID":"219a7347-5102-4e6c-9b38-deea6275f715","Type":"ContainerStarted","Data":"accf9af77c7a917385e318edd9989c36af54f20883603f01cbe8c6df33d9d891"} Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.767007 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a","Type":"ContainerDied","Data":"92594241f6fc33b4adbb62148ccdf2e313b680db092c248427a6a0eb82e8c4ea"} Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.767329 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.776206 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.782755 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.804657 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 11:08:41 crc kubenswrapper[4776]: E1125 11:08:41.805390 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a" containerName="glance-log" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.805433 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a" containerName="glance-log" Nov 25 11:08:41 crc kubenswrapper[4776]: E1125 11:08:41.805452 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a" containerName="glance-httpd" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.805461 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a" containerName="glance-httpd" Nov 25 11:08:41 crc kubenswrapper[4776]: E1125 11:08:41.805501 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afe1ae15-e85d-4d25-81ef-445a574d8222" containerName="glance-log" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.805510 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="afe1ae15-e85d-4d25-81ef-445a574d8222" containerName="glance-log" Nov 25 11:08:41 crc kubenswrapper[4776]: E1125 11:08:41.805527 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afe1ae15-e85d-4d25-81ef-445a574d8222" containerName="glance-httpd" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.805534 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="afe1ae15-e85d-4d25-81ef-445a574d8222" containerName="glance-httpd" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.805871 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="afe1ae15-e85d-4d25-81ef-445a574d8222" containerName="glance-httpd" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.805919 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a" containerName="glance-log" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.805940 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a" containerName="glance-httpd" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.805957 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="afe1ae15-e85d-4d25-81ef-445a574d8222" containerName="glance-log" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.809111 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.820993 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.821347 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.821489 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-xlswr" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.826624 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.837442 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.844192 4776 scope.go:117] "RemoveContainer" containerID="4b90d93d9326d7485e4ab371e06f94e76c0b64f3ce6adcbb172886a6fa5348d8" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.874119 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.891137 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.911148 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54\") " pod="openstack/glance-default-external-api-0" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.911292 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54\") " pod="openstack/glance-default-external-api-0" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.911359 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54-logs\") pod \"glance-default-external-api-0\" (UID: \"6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54\") " pod="openstack/glance-default-external-api-0" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.911428 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54-scripts\") pod \"glance-default-external-api-0\" (UID: \"6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54\") " pod="openstack/glance-default-external-api-0" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.911453 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54\") " pod="openstack/glance-default-external-api-0" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.911484 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62d2w\" (UniqueName: \"kubernetes.io/projected/6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54-kube-api-access-62d2w\") pod \"glance-default-external-api-0\" (UID: \"6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54\") " pod="openstack/glance-default-external-api-0" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.911529 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54-config-data\") pod \"glance-default-external-api-0\" (UID: \"6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54\") " pod="openstack/glance-default-external-api-0" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.930177 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.933138 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.951495 4776 scope.go:117] "RemoveContainer" containerID="d67251a37a2930b6bc7b71c9e7bce4cd60e86e67780a1487e294e57b5a93ca21" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.952706 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.953821 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.958553 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 11:08:41 crc kubenswrapper[4776]: I1125 11:08:41.992381 4776 scope.go:117] "RemoveContainer" containerID="5d69cfffdca6866fa1479b16f66e7d4638a5e2afa489a7050cef817c49652ad8" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.013944 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54\") " pod="openstack/glance-default-external-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.014013 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54\") " pod="openstack/glance-default-external-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.015372 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54-logs\") pod \"glance-default-external-api-0\" (UID: \"6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54\") " pod="openstack/glance-default-external-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.015447 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54-scripts\") pod \"glance-default-external-api-0\" (UID: \"6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54\") " pod="openstack/glance-default-external-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.015475 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54\") " pod="openstack/glance-default-external-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.015495 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62d2w\" (UniqueName: \"kubernetes.io/projected/6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54-kube-api-access-62d2w\") pod \"glance-default-external-api-0\" (UID: \"6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54\") " pod="openstack/glance-default-external-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.015532 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54-config-data\") pod \"glance-default-external-api-0\" (UID: \"6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54\") " pod="openstack/glance-default-external-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.016601 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54\") " pod="openstack/glance-default-external-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.016873 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54-logs\") pod \"glance-default-external-api-0\" (UID: \"6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54\") " pod="openstack/glance-default-external-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.019887 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54\") " pod="openstack/glance-default-external-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.020445 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54-scripts\") pod \"glance-default-external-api-0\" (UID: \"6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54\") " pod="openstack/glance-default-external-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.033641 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54-config-data\") pod \"glance-default-external-api-0\" (UID: \"6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54\") " pod="openstack/glance-default-external-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.036849 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62d2w\" (UniqueName: \"kubernetes.io/projected/6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54-kube-api-access-62d2w\") pod \"glance-default-external-api-0\" (UID: \"6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54\") " pod="openstack/glance-default-external-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.040835 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54\") " pod="openstack/glance-default-external-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.118854 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea2660a8-255a-45db-a684-d089438652e8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ea2660a8-255a-45db-a684-d089438652e8\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.119653 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ea2660a8-255a-45db-a684-d089438652e8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ea2660a8-255a-45db-a684-d089438652e8\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.119724 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbb74\" (UniqueName: \"kubernetes.io/projected/ea2660a8-255a-45db-a684-d089438652e8-kube-api-access-gbb74\") pod \"glance-default-internal-api-0\" (UID: \"ea2660a8-255a-45db-a684-d089438652e8\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.119782 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea2660a8-255a-45db-a684-d089438652e8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ea2660a8-255a-45db-a684-d089438652e8\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.119835 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea2660a8-255a-45db-a684-d089438652e8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ea2660a8-255a-45db-a684-d089438652e8\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.119858 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea2660a8-255a-45db-a684-d089438652e8-logs\") pod \"glance-default-internal-api-0\" (UID: \"ea2660a8-255a-45db-a684-d089438652e8\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.120224 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea2660a8-255a-45db-a684-d089438652e8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ea2660a8-255a-45db-a684-d089438652e8\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.157782 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.221984 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea2660a8-255a-45db-a684-d089438652e8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ea2660a8-255a-45db-a684-d089438652e8\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.222232 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ea2660a8-255a-45db-a684-d089438652e8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ea2660a8-255a-45db-a684-d089438652e8\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.222280 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbb74\" (UniqueName: \"kubernetes.io/projected/ea2660a8-255a-45db-a684-d089438652e8-kube-api-access-gbb74\") pod \"glance-default-internal-api-0\" (UID: \"ea2660a8-255a-45db-a684-d089438652e8\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.222358 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea2660a8-255a-45db-a684-d089438652e8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ea2660a8-255a-45db-a684-d089438652e8\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.222457 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea2660a8-255a-45db-a684-d089438652e8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ea2660a8-255a-45db-a684-d089438652e8\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.222478 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea2660a8-255a-45db-a684-d089438652e8-logs\") pod \"glance-default-internal-api-0\" (UID: \"ea2660a8-255a-45db-a684-d089438652e8\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.222625 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea2660a8-255a-45db-a684-d089438652e8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ea2660a8-255a-45db-a684-d089438652e8\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.223088 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ea2660a8-255a-45db-a684-d089438652e8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ea2660a8-255a-45db-a684-d089438652e8\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.224151 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea2660a8-255a-45db-a684-d089438652e8-logs\") pod \"glance-default-internal-api-0\" (UID: \"ea2660a8-255a-45db-a684-d089438652e8\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.228284 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea2660a8-255a-45db-a684-d089438652e8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ea2660a8-255a-45db-a684-d089438652e8\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.228307 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea2660a8-255a-45db-a684-d089438652e8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ea2660a8-255a-45db-a684-d089438652e8\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.229241 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea2660a8-255a-45db-a684-d089438652e8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ea2660a8-255a-45db-a684-d089438652e8\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.231166 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea2660a8-255a-45db-a684-d089438652e8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ea2660a8-255a-45db-a684-d089438652e8\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.251912 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbb74\" (UniqueName: \"kubernetes.io/projected/ea2660a8-255a-45db-a684-d089438652e8-kube-api-access-gbb74\") pod \"glance-default-internal-api-0\" (UID: \"ea2660a8-255a-45db-a684-d089438652e8\") " pod="openstack/glance-default-internal-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.302727 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.743437 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 11:08:42 crc kubenswrapper[4776]: W1125 11:08:42.757166 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a43c9d6_0d1b_48a5_8fe6_07abcf9c2b54.slice/crio-7e49d116558fb60961cc4ffbc24238868c16e204a1f9ecda2394aa077aac65f4 WatchSource:0}: Error finding container 7e49d116558fb60961cc4ffbc24238868c16e204a1f9ecda2394aa077aac65f4: Status 404 returned error can't find the container with id 7e49d116558fb60961cc4ffbc24238868c16e204a1f9ecda2394aa077aac65f4 Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.780741 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54","Type":"ContainerStarted","Data":"7e49d116558fb60961cc4ffbc24238868c16e204a1f9ecda2394aa077aac65f4"} Nov 25 11:08:42 crc kubenswrapper[4776]: I1125 11:08:42.905898 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 11:08:42 crc kubenswrapper[4776]: W1125 11:08:42.919195 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea2660a8_255a_45db_a684_d089438652e8.slice/crio-0ebe76e43cdadc65f6bdd449b2c7e1d7dc3fed165f46153e211f48a0398ff01a WatchSource:0}: Error finding container 0ebe76e43cdadc65f6bdd449b2c7e1d7dc3fed165f46153e211f48a0398ff01a: Status 404 returned error can't find the container with id 0ebe76e43cdadc65f6bdd449b2c7e1d7dc3fed165f46153e211f48a0398ff01a Nov 25 11:08:43 crc kubenswrapper[4776]: I1125 11:08:43.677817 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a" path="/var/lib/kubelet/pods/2bdf5cf7-3d1f-43b3-8d5a-b7b5253a502a/volumes" Nov 25 11:08:43 crc kubenswrapper[4776]: I1125 11:08:43.679113 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afe1ae15-e85d-4d25-81ef-445a574d8222" path="/var/lib/kubelet/pods/afe1ae15-e85d-4d25-81ef-445a574d8222/volumes" Nov 25 11:08:43 crc kubenswrapper[4776]: I1125 11:08:43.793749 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ea2660a8-255a-45db-a684-d089438652e8","Type":"ContainerStarted","Data":"30036a3b3e4bfcd3a7e1a42929f13500dccd4c68c6fa35478b684bed47f2ab93"} Nov 25 11:08:43 crc kubenswrapper[4776]: I1125 11:08:43.793793 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ea2660a8-255a-45db-a684-d089438652e8","Type":"ContainerStarted","Data":"0ebe76e43cdadc65f6bdd449b2c7e1d7dc3fed165f46153e211f48a0398ff01a"} Nov 25 11:08:43 crc kubenswrapper[4776]: I1125 11:08:43.796075 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54","Type":"ContainerStarted","Data":"136f1c96d077a648378dd5f89ccdf249d11bb8af46f10ec7ef4ef97886abdecd"} Nov 25 11:08:49 crc kubenswrapper[4776]: I1125 11:08:49.859612 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b6b74d589-7lsxv" event={"ID":"c465b70c-b76d-430c-954e-b8408390cd99","Type":"ContainerStarted","Data":"56d410c33ee33045ef294b6712c131e27295599e88c30964cdf01877e975e5d2"} Nov 25 11:08:49 crc kubenswrapper[4776]: I1125 11:08:49.860334 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b6b74d589-7lsxv" event={"ID":"c465b70c-b76d-430c-954e-b8408390cd99","Type":"ContainerStarted","Data":"7d3bb0850a55fbd25b5001e9f538ee89de9452f36bc4762d93883d48fc5c517d"} Nov 25 11:08:49 crc kubenswrapper[4776]: I1125 11:08:49.862429 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-656cc574-dstj4" event={"ID":"219a7347-5102-4e6c-9b38-deea6275f715","Type":"ContainerStarted","Data":"f5dd0b1f4ea70853679a58ed439e71548cd36135ef8443036178352353139632"} Nov 25 11:08:49 crc kubenswrapper[4776]: I1125 11:08:49.862468 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-656cc574-dstj4" event={"ID":"219a7347-5102-4e6c-9b38-deea6275f715","Type":"ContainerStarted","Data":"75624a8110711b2f1d37d5065d9a5ab2035226b6e3ac4acf037118c35cb7d28c"} Nov 25 11:08:49 crc kubenswrapper[4776]: I1125 11:08:49.864777 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-599bd8d6b9-hz9cl" event={"ID":"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c","Type":"ContainerStarted","Data":"5a471f50de7f68f576aa9a22499ed07079aa755d8fac9e3a4b4e59dac08f9a08"} Nov 25 11:08:49 crc kubenswrapper[4776]: I1125 11:08:49.864815 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-599bd8d6b9-hz9cl" event={"ID":"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c","Type":"ContainerStarted","Data":"4b0b5e26d059853bd5e7dc65017bff796f62d2a3bc79f08a51fd586e15e82a1f"} Nov 25 11:08:49 crc kubenswrapper[4776]: I1125 11:08:49.864884 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-599bd8d6b9-hz9cl" podUID="c4ceac98-94a9-4cc2-a405-d57a1a15cf9c" containerName="horizon-log" containerID="cri-o://4b0b5e26d059853bd5e7dc65017bff796f62d2a3bc79f08a51fd586e15e82a1f" gracePeriod=30 Nov 25 11:08:49 crc kubenswrapper[4776]: I1125 11:08:49.864893 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-599bd8d6b9-hz9cl" podUID="c4ceac98-94a9-4cc2-a405-d57a1a15cf9c" containerName="horizon" containerID="cri-o://5a471f50de7f68f576aa9a22499ed07079aa755d8fac9e3a4b4e59dac08f9a08" gracePeriod=30 Nov 25 11:08:49 crc kubenswrapper[4776]: I1125 11:08:49.871533 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54","Type":"ContainerStarted","Data":"dd7a6df1498e35915aebfe1dda5fed1d4c4dbd415c295040c48a21c39251d8b9"} Nov 25 11:08:49 crc kubenswrapper[4776]: I1125 11:08:49.874607 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c9cb59558-gpdwk" event={"ID":"238717c2-18fa-47d1-b39e-4d2f0b3619de","Type":"ContainerStarted","Data":"89f14821d2aaeda2739c1d10adaa4b87fb5a83414dc465a734f6b375da24fdee"} Nov 25 11:08:49 crc kubenswrapper[4776]: I1125 11:08:49.874644 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c9cb59558-gpdwk" event={"ID":"238717c2-18fa-47d1-b39e-4d2f0b3619de","Type":"ContainerStarted","Data":"048841f4110bcd3f5595becc35292ea8885ba21d1bde0f0d3e40f649c52ef3a8"} Nov 25 11:08:49 crc kubenswrapper[4776]: I1125 11:08:49.876943 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ea2660a8-255a-45db-a684-d089438652e8","Type":"ContainerStarted","Data":"5e1ffbdeb442b020f908239432d469a4a8d5ea58ec0d64d5cac6bcb42d92bac2"} Nov 25 11:08:49 crc kubenswrapper[4776]: I1125 11:08:49.897842 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-656cc574-dstj4" podStartSLOduration=2.971056214 podStartE2EDuration="10.897819069s" podCreationTimestamp="2025-11-25 11:08:39 +0000 UTC" firstStartedPulling="2025-11-25 11:08:40.875725474 +0000 UTC m=+6265.916785027" lastFinishedPulling="2025-11-25 11:08:48.802488319 +0000 UTC m=+6273.843547882" observedRunningTime="2025-11-25 11:08:49.883486439 +0000 UTC m=+6274.924545992" watchObservedRunningTime="2025-11-25 11:08:49.897819069 +0000 UTC m=+6274.938878622" Nov 25 11:08:49 crc kubenswrapper[4776]: I1125 11:08:49.917662 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=8.917639126 podStartE2EDuration="8.917639126s" podCreationTimestamp="2025-11-25 11:08:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:08:49.908613749 +0000 UTC m=+6274.949673322" watchObservedRunningTime="2025-11-25 11:08:49.917639126 +0000 UTC m=+6274.958698679" Nov 25 11:08:49 crc kubenswrapper[4776]: I1125 11:08:49.946749 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-599bd8d6b9-hz9cl" podStartSLOduration=3.129339082 podStartE2EDuration="12.946727236s" podCreationTimestamp="2025-11-25 11:08:37 +0000 UTC" firstStartedPulling="2025-11-25 11:08:38.948268871 +0000 UTC m=+6263.989328424" lastFinishedPulling="2025-11-25 11:08:48.765657015 +0000 UTC m=+6273.806716578" observedRunningTime="2025-11-25 11:08:49.943020023 +0000 UTC m=+6274.984079576" watchObservedRunningTime="2025-11-25 11:08:49.946727236 +0000 UTC m=+6274.987786789" Nov 25 11:08:49 crc kubenswrapper[4776]: I1125 11:08:49.980137 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=8.980114274 podStartE2EDuration="8.980114274s" podCreationTimestamp="2025-11-25 11:08:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:08:49.970358909 +0000 UTC m=+6275.011418462" watchObservedRunningTime="2025-11-25 11:08:49.980114274 +0000 UTC m=+6275.021173827" Nov 25 11:08:49 crc kubenswrapper[4776]: I1125 11:08:49.992238 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7c9cb59558-gpdwk" podStartSLOduration=2.952677592 podStartE2EDuration="10.992219518s" podCreationTimestamp="2025-11-25 11:08:39 +0000 UTC" firstStartedPulling="2025-11-25 11:08:40.75085109 +0000 UTC m=+6265.791910643" lastFinishedPulling="2025-11-25 11:08:48.790393016 +0000 UTC m=+6273.831452569" observedRunningTime="2025-11-25 11:08:49.988383651 +0000 UTC m=+6275.029443204" watchObservedRunningTime="2025-11-25 11:08:49.992219518 +0000 UTC m=+6275.033279071" Nov 25 11:08:50 crc kubenswrapper[4776]: I1125 11:08:50.218650 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:08:50 crc kubenswrapper[4776]: I1125 11:08:50.218875 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:08:50 crc kubenswrapper[4776]: I1125 11:08:50.306193 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-656cc574-dstj4" Nov 25 11:08:50 crc kubenswrapper[4776]: I1125 11:08:50.306262 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-656cc574-dstj4" Nov 25 11:08:50 crc kubenswrapper[4776]: I1125 11:08:50.664212 4776 scope.go:117] "RemoveContainer" containerID="8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698" Nov 25 11:08:50 crc kubenswrapper[4776]: E1125 11:08:50.665039 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:08:50 crc kubenswrapper[4776]: I1125 11:08:50.887848 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7b6b74d589-7lsxv" podUID="c465b70c-b76d-430c-954e-b8408390cd99" containerName="horizon-log" containerID="cri-o://7d3bb0850a55fbd25b5001e9f538ee89de9452f36bc4762d93883d48fc5c517d" gracePeriod=30 Nov 25 11:08:50 crc kubenswrapper[4776]: I1125 11:08:50.888231 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7b6b74d589-7lsxv" podUID="c465b70c-b76d-430c-954e-b8408390cd99" containerName="horizon" containerID="cri-o://56d410c33ee33045ef294b6712c131e27295599e88c30964cdf01877e975e5d2" gracePeriod=30 Nov 25 11:08:50 crc kubenswrapper[4776]: I1125 11:08:50.917368 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7b6b74d589-7lsxv" podStartSLOduration=3.877243052 podStartE2EDuration="13.917346705s" podCreationTimestamp="2025-11-25 11:08:37 +0000 UTC" firstStartedPulling="2025-11-25 11:08:38.932437094 +0000 UTC m=+6263.973496647" lastFinishedPulling="2025-11-25 11:08:48.972540747 +0000 UTC m=+6274.013600300" observedRunningTime="2025-11-25 11:08:50.909038626 +0000 UTC m=+6275.950098179" watchObservedRunningTime="2025-11-25 11:08:50.917346705 +0000 UTC m=+6275.958406258" Nov 25 11:08:52 crc kubenswrapper[4776]: I1125 11:08:52.159833 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 25 11:08:52 crc kubenswrapper[4776]: I1125 11:08:52.160255 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 25 11:08:52 crc kubenswrapper[4776]: I1125 11:08:52.206108 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 25 11:08:52 crc kubenswrapper[4776]: I1125 11:08:52.213813 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 25 11:08:52 crc kubenswrapper[4776]: I1125 11:08:52.303981 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 25 11:08:52 crc kubenswrapper[4776]: I1125 11:08:52.304047 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 25 11:08:52 crc kubenswrapper[4776]: I1125 11:08:52.341447 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 25 11:08:52 crc kubenswrapper[4776]: I1125 11:08:52.349766 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 25 11:08:52 crc kubenswrapper[4776]: I1125 11:08:52.902711 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 25 11:08:52 crc kubenswrapper[4776]: I1125 11:08:52.902777 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 25 11:08:52 crc kubenswrapper[4776]: I1125 11:08:52.902796 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 25 11:08:52 crc kubenswrapper[4776]: I1125 11:08:52.904259 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 25 11:08:54 crc kubenswrapper[4776]: I1125 11:08:54.917188 4776 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 11:08:55 crc kubenswrapper[4776]: I1125 11:08:55.162665 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 25 11:08:55 crc kubenswrapper[4776]: I1125 11:08:55.164248 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 25 11:08:55 crc kubenswrapper[4776]: I1125 11:08:55.286216 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 25 11:08:55 crc kubenswrapper[4776]: I1125 11:08:55.300481 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 25 11:08:57 crc kubenswrapper[4776]: I1125 11:08:57.672771 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7b6b74d589-7lsxv" Nov 25 11:08:57 crc kubenswrapper[4776]: I1125 11:08:57.781277 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-599bd8d6b9-hz9cl" Nov 25 11:09:00 crc kubenswrapper[4776]: I1125 11:09:00.221544 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7c9cb59558-gpdwk" podUID="238717c2-18fa-47d1-b39e-4d2f0b3619de" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.124:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.124:8443: connect: connection refused" Nov 25 11:09:00 crc kubenswrapper[4776]: I1125 11:09:00.307092 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-656cc574-dstj4" podUID="219a7347-5102-4e6c-9b38-deea6275f715" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.125:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.125:8443: connect: connection refused" Nov 25 11:09:03 crc kubenswrapper[4776]: I1125 11:09:03.666266 4776 scope.go:117] "RemoveContainer" containerID="8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698" Nov 25 11:09:03 crc kubenswrapper[4776]: E1125 11:09:03.666716 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:09:12 crc kubenswrapper[4776]: I1125 11:09:12.117741 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-656cc574-dstj4" Nov 25 11:09:12 crc kubenswrapper[4776]: I1125 11:09:12.122491 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:09:13 crc kubenswrapper[4776]: I1125 11:09:13.963607 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-656cc574-dstj4" Nov 25 11:09:14 crc kubenswrapper[4776]: I1125 11:09:14.035737 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7c9cb59558-gpdwk"] Nov 25 11:09:14 crc kubenswrapper[4776]: I1125 11:09:14.035948 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7c9cb59558-gpdwk" podUID="238717c2-18fa-47d1-b39e-4d2f0b3619de" containerName="horizon-log" containerID="cri-o://048841f4110bcd3f5595becc35292ea8885ba21d1bde0f0d3e40f649c52ef3a8" gracePeriod=30 Nov 25 11:09:14 crc kubenswrapper[4776]: I1125 11:09:14.036597 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7c9cb59558-gpdwk" podUID="238717c2-18fa-47d1-b39e-4d2f0b3619de" containerName="horizon" containerID="cri-o://89f14821d2aaeda2739c1d10adaa4b87fb5a83414dc465a734f6b375da24fdee" gracePeriod=30 Nov 25 11:09:14 crc kubenswrapper[4776]: I1125 11:09:14.052579 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7c9cb59558-gpdwk" podUID="238717c2-18fa-47d1-b39e-4d2f0b3619de" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.124:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Nov 25 11:09:15 crc kubenswrapper[4776]: I1125 11:09:15.883980 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:09:18 crc kubenswrapper[4776]: I1125 11:09:18.663302 4776 scope.go:117] "RemoveContainer" containerID="8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698" Nov 25 11:09:18 crc kubenswrapper[4776]: E1125 11:09:18.664259 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:09:20 crc kubenswrapper[4776]: I1125 11:09:20.218369 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7c9cb59558-gpdwk" podUID="238717c2-18fa-47d1-b39e-4d2f0b3619de" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.124:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.124:8443: connect: connection refused" Nov 25 11:09:20 crc kubenswrapper[4776]: I1125 11:09:20.229379 4776 generic.go:334] "Generic (PLEG): container finished" podID="c4ceac98-94a9-4cc2-a405-d57a1a15cf9c" containerID="5a471f50de7f68f576aa9a22499ed07079aa755d8fac9e3a4b4e59dac08f9a08" exitCode=137 Nov 25 11:09:20 crc kubenswrapper[4776]: I1125 11:09:20.229419 4776 generic.go:334] "Generic (PLEG): container finished" podID="c4ceac98-94a9-4cc2-a405-d57a1a15cf9c" containerID="4b0b5e26d059853bd5e7dc65017bff796f62d2a3bc79f08a51fd586e15e82a1f" exitCode=137 Nov 25 11:09:20 crc kubenswrapper[4776]: I1125 11:09:20.229464 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-599bd8d6b9-hz9cl" event={"ID":"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c","Type":"ContainerDied","Data":"5a471f50de7f68f576aa9a22499ed07079aa755d8fac9e3a4b4e59dac08f9a08"} Nov 25 11:09:20 crc kubenswrapper[4776]: I1125 11:09:20.229496 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-599bd8d6b9-hz9cl" event={"ID":"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c","Type":"ContainerDied","Data":"4b0b5e26d059853bd5e7dc65017bff796f62d2a3bc79f08a51fd586e15e82a1f"} Nov 25 11:09:20 crc kubenswrapper[4776]: I1125 11:09:20.231003 4776 generic.go:334] "Generic (PLEG): container finished" podID="238717c2-18fa-47d1-b39e-4d2f0b3619de" containerID="89f14821d2aaeda2739c1d10adaa4b87fb5a83414dc465a734f6b375da24fdee" exitCode=0 Nov 25 11:09:20 crc kubenswrapper[4776]: I1125 11:09:20.231040 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c9cb59558-gpdwk" event={"ID":"238717c2-18fa-47d1-b39e-4d2f0b3619de","Type":"ContainerDied","Data":"89f14821d2aaeda2739c1d10adaa4b87fb5a83414dc465a734f6b375da24fdee"} Nov 25 11:09:20 crc kubenswrapper[4776]: I1125 11:09:20.334473 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-599bd8d6b9-hz9cl" Nov 25 11:09:20 crc kubenswrapper[4776]: I1125 11:09:20.457849 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-horizon-secret-key\") pod \"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c\" (UID: \"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c\") " Nov 25 11:09:20 crc kubenswrapper[4776]: I1125 11:09:20.457955 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkx9l\" (UniqueName: \"kubernetes.io/projected/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-kube-api-access-vkx9l\") pod \"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c\" (UID: \"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c\") " Nov 25 11:09:20 crc kubenswrapper[4776]: I1125 11:09:20.458008 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-logs\") pod \"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c\" (UID: \"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c\") " Nov 25 11:09:20 crc kubenswrapper[4776]: I1125 11:09:20.458259 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-config-data\") pod \"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c\" (UID: \"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c\") " Nov 25 11:09:20 crc kubenswrapper[4776]: I1125 11:09:20.458317 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-scripts\") pod \"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c\" (UID: \"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c\") " Nov 25 11:09:20 crc kubenswrapper[4776]: I1125 11:09:20.458676 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-logs" (OuterVolumeSpecName: "logs") pod "c4ceac98-94a9-4cc2-a405-d57a1a15cf9c" (UID: "c4ceac98-94a9-4cc2-a405-d57a1a15cf9c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:09:20 crc kubenswrapper[4776]: I1125 11:09:20.459757 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-logs\") on node \"crc\" DevicePath \"\"" Nov 25 11:09:20 crc kubenswrapper[4776]: I1125 11:09:20.463713 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-kube-api-access-vkx9l" (OuterVolumeSpecName: "kube-api-access-vkx9l") pod "c4ceac98-94a9-4cc2-a405-d57a1a15cf9c" (UID: "c4ceac98-94a9-4cc2-a405-d57a1a15cf9c"). InnerVolumeSpecName "kube-api-access-vkx9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:09:20 crc kubenswrapper[4776]: I1125 11:09:20.464486 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "c4ceac98-94a9-4cc2-a405-d57a1a15cf9c" (UID: "c4ceac98-94a9-4cc2-a405-d57a1a15cf9c"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:09:20 crc kubenswrapper[4776]: I1125 11:09:20.485568 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-scripts" (OuterVolumeSpecName: "scripts") pod "c4ceac98-94a9-4cc2-a405-d57a1a15cf9c" (UID: "c4ceac98-94a9-4cc2-a405-d57a1a15cf9c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:09:20 crc kubenswrapper[4776]: I1125 11:09:20.501281 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-config-data" (OuterVolumeSpecName: "config-data") pod "c4ceac98-94a9-4cc2-a405-d57a1a15cf9c" (UID: "c4ceac98-94a9-4cc2-a405-d57a1a15cf9c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:09:20 crc kubenswrapper[4776]: I1125 11:09:20.562353 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:09:20 crc kubenswrapper[4776]: I1125 11:09:20.562403 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:09:20 crc kubenswrapper[4776]: I1125 11:09:20.562413 4776 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:09:20 crc kubenswrapper[4776]: I1125 11:09:20.562424 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkx9l\" (UniqueName: \"kubernetes.io/projected/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c-kube-api-access-vkx9l\") on node \"crc\" DevicePath \"\"" Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.249137 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b6b74d589-7lsxv" Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.250865 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-599bd8d6b9-hz9cl" event={"ID":"c4ceac98-94a9-4cc2-a405-d57a1a15cf9c","Type":"ContainerDied","Data":"d3a620cc9e70afefea5e7c11ed49253141433709c1cfbc61a3eee63384412e88"} Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.250951 4776 scope.go:117] "RemoveContainer" containerID="5a471f50de7f68f576aa9a22499ed07079aa755d8fac9e3a4b4e59dac08f9a08" Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.251023 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-599bd8d6b9-hz9cl" Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.254989 4776 generic.go:334] "Generic (PLEG): container finished" podID="c465b70c-b76d-430c-954e-b8408390cd99" containerID="56d410c33ee33045ef294b6712c131e27295599e88c30964cdf01877e975e5d2" exitCode=137 Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.255012 4776 generic.go:334] "Generic (PLEG): container finished" podID="c465b70c-b76d-430c-954e-b8408390cd99" containerID="7d3bb0850a55fbd25b5001e9f538ee89de9452f36bc4762d93883d48fc5c517d" exitCode=137 Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.255035 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b6b74d589-7lsxv" event={"ID":"c465b70c-b76d-430c-954e-b8408390cd99","Type":"ContainerDied","Data":"56d410c33ee33045ef294b6712c131e27295599e88c30964cdf01877e975e5d2"} Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.255054 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b6b74d589-7lsxv" event={"ID":"c465b70c-b76d-430c-954e-b8408390cd99","Type":"ContainerDied","Data":"7d3bb0850a55fbd25b5001e9f538ee89de9452f36bc4762d93883d48fc5c517d"} Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.255054 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b6b74d589-7lsxv" Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.255082 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b6b74d589-7lsxv" event={"ID":"c465b70c-b76d-430c-954e-b8408390cd99","Type":"ContainerDied","Data":"adf9775a0780deb6241108dfcb0371e0702b8e4cd74cf9d350386abcd8cf9cc6"} Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.321910 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-599bd8d6b9-hz9cl"] Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.333790 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-599bd8d6b9-hz9cl"] Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.380304 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c465b70c-b76d-430c-954e-b8408390cd99-logs\") pod \"c465b70c-b76d-430c-954e-b8408390cd99\" (UID: \"c465b70c-b76d-430c-954e-b8408390cd99\") " Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.380640 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c465b70c-b76d-430c-954e-b8408390cd99-config-data\") pod \"c465b70c-b76d-430c-954e-b8408390cd99\" (UID: \"c465b70c-b76d-430c-954e-b8408390cd99\") " Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.380720 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w26nr\" (UniqueName: \"kubernetes.io/projected/c465b70c-b76d-430c-954e-b8408390cd99-kube-api-access-w26nr\") pod \"c465b70c-b76d-430c-954e-b8408390cd99\" (UID: \"c465b70c-b76d-430c-954e-b8408390cd99\") " Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.381005 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c465b70c-b76d-430c-954e-b8408390cd99-logs" (OuterVolumeSpecName: "logs") pod "c465b70c-b76d-430c-954e-b8408390cd99" (UID: "c465b70c-b76d-430c-954e-b8408390cd99"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.381237 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c465b70c-b76d-430c-954e-b8408390cd99-horizon-secret-key\") pod \"c465b70c-b76d-430c-954e-b8408390cd99\" (UID: \"c465b70c-b76d-430c-954e-b8408390cd99\") " Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.381354 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c465b70c-b76d-430c-954e-b8408390cd99-scripts\") pod \"c465b70c-b76d-430c-954e-b8408390cd99\" (UID: \"c465b70c-b76d-430c-954e-b8408390cd99\") " Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.381997 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c465b70c-b76d-430c-954e-b8408390cd99-logs\") on node \"crc\" DevicePath \"\"" Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.385654 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c465b70c-b76d-430c-954e-b8408390cd99-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "c465b70c-b76d-430c-954e-b8408390cd99" (UID: "c465b70c-b76d-430c-954e-b8408390cd99"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.385860 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c465b70c-b76d-430c-954e-b8408390cd99-kube-api-access-w26nr" (OuterVolumeSpecName: "kube-api-access-w26nr") pod "c465b70c-b76d-430c-954e-b8408390cd99" (UID: "c465b70c-b76d-430c-954e-b8408390cd99"). InnerVolumeSpecName "kube-api-access-w26nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.406877 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c465b70c-b76d-430c-954e-b8408390cd99-config-data" (OuterVolumeSpecName: "config-data") pod "c465b70c-b76d-430c-954e-b8408390cd99" (UID: "c465b70c-b76d-430c-954e-b8408390cd99"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.408642 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c465b70c-b76d-430c-954e-b8408390cd99-scripts" (OuterVolumeSpecName: "scripts") pod "c465b70c-b76d-430c-954e-b8408390cd99" (UID: "c465b70c-b76d-430c-954e-b8408390cd99"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.421659 4776 scope.go:117] "RemoveContainer" containerID="4b0b5e26d059853bd5e7dc65017bff796f62d2a3bc79f08a51fd586e15e82a1f" Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.445452 4776 scope.go:117] "RemoveContainer" containerID="56d410c33ee33045ef294b6712c131e27295599e88c30964cdf01877e975e5d2" Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.483801 4776 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c465b70c-b76d-430c-954e-b8408390cd99-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.483836 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c465b70c-b76d-430c-954e-b8408390cd99-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.483846 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c465b70c-b76d-430c-954e-b8408390cd99-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.483854 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w26nr\" (UniqueName: \"kubernetes.io/projected/c465b70c-b76d-430c-954e-b8408390cd99-kube-api-access-w26nr\") on node \"crc\" DevicePath \"\"" Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.601558 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7b6b74d589-7lsxv"] Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.610227 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7b6b74d589-7lsxv"] Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.624275 4776 scope.go:117] "RemoveContainer" containerID="7d3bb0850a55fbd25b5001e9f538ee89de9452f36bc4762d93883d48fc5c517d" Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.646316 4776 scope.go:117] "RemoveContainer" containerID="56d410c33ee33045ef294b6712c131e27295599e88c30964cdf01877e975e5d2" Nov 25 11:09:21 crc kubenswrapper[4776]: E1125 11:09:21.646786 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56d410c33ee33045ef294b6712c131e27295599e88c30964cdf01877e975e5d2\": container with ID starting with 56d410c33ee33045ef294b6712c131e27295599e88c30964cdf01877e975e5d2 not found: ID does not exist" containerID="56d410c33ee33045ef294b6712c131e27295599e88c30964cdf01877e975e5d2" Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.646832 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56d410c33ee33045ef294b6712c131e27295599e88c30964cdf01877e975e5d2"} err="failed to get container status \"56d410c33ee33045ef294b6712c131e27295599e88c30964cdf01877e975e5d2\": rpc error: code = NotFound desc = could not find container \"56d410c33ee33045ef294b6712c131e27295599e88c30964cdf01877e975e5d2\": container with ID starting with 56d410c33ee33045ef294b6712c131e27295599e88c30964cdf01877e975e5d2 not found: ID does not exist" Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.646878 4776 scope.go:117] "RemoveContainer" containerID="7d3bb0850a55fbd25b5001e9f538ee89de9452f36bc4762d93883d48fc5c517d" Nov 25 11:09:21 crc kubenswrapper[4776]: E1125 11:09:21.647498 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d3bb0850a55fbd25b5001e9f538ee89de9452f36bc4762d93883d48fc5c517d\": container with ID starting with 7d3bb0850a55fbd25b5001e9f538ee89de9452f36bc4762d93883d48fc5c517d not found: ID does not exist" containerID="7d3bb0850a55fbd25b5001e9f538ee89de9452f36bc4762d93883d48fc5c517d" Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.647552 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d3bb0850a55fbd25b5001e9f538ee89de9452f36bc4762d93883d48fc5c517d"} err="failed to get container status \"7d3bb0850a55fbd25b5001e9f538ee89de9452f36bc4762d93883d48fc5c517d\": rpc error: code = NotFound desc = could not find container \"7d3bb0850a55fbd25b5001e9f538ee89de9452f36bc4762d93883d48fc5c517d\": container with ID starting with 7d3bb0850a55fbd25b5001e9f538ee89de9452f36bc4762d93883d48fc5c517d not found: ID does not exist" Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.647570 4776 scope.go:117] "RemoveContainer" containerID="56d410c33ee33045ef294b6712c131e27295599e88c30964cdf01877e975e5d2" Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.648114 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56d410c33ee33045ef294b6712c131e27295599e88c30964cdf01877e975e5d2"} err="failed to get container status \"56d410c33ee33045ef294b6712c131e27295599e88c30964cdf01877e975e5d2\": rpc error: code = NotFound desc = could not find container \"56d410c33ee33045ef294b6712c131e27295599e88c30964cdf01877e975e5d2\": container with ID starting with 56d410c33ee33045ef294b6712c131e27295599e88c30964cdf01877e975e5d2 not found: ID does not exist" Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.648156 4776 scope.go:117] "RemoveContainer" containerID="7d3bb0850a55fbd25b5001e9f538ee89de9452f36bc4762d93883d48fc5c517d" Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.648458 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d3bb0850a55fbd25b5001e9f538ee89de9452f36bc4762d93883d48fc5c517d"} err="failed to get container status \"7d3bb0850a55fbd25b5001e9f538ee89de9452f36bc4762d93883d48fc5c517d\": rpc error: code = NotFound desc = could not find container \"7d3bb0850a55fbd25b5001e9f538ee89de9452f36bc4762d93883d48fc5c517d\": container with ID starting with 7d3bb0850a55fbd25b5001e9f538ee89de9452f36bc4762d93883d48fc5c517d not found: ID does not exist" Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.682879 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c465b70c-b76d-430c-954e-b8408390cd99" path="/var/lib/kubelet/pods/c465b70c-b76d-430c-954e-b8408390cd99/volumes" Nov 25 11:09:21 crc kubenswrapper[4776]: I1125 11:09:21.684313 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4ceac98-94a9-4cc2-a405-d57a1a15cf9c" path="/var/lib/kubelet/pods/c4ceac98-94a9-4cc2-a405-d57a1a15cf9c/volumes" Nov 25 11:09:28 crc kubenswrapper[4776]: I1125 11:09:28.056161 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-99m2h"] Nov 25 11:09:28 crc kubenswrapper[4776]: I1125 11:09:28.067766 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-ea9d-account-create-gtwm6"] Nov 25 11:09:28 crc kubenswrapper[4776]: I1125 11:09:28.079265 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-99m2h"] Nov 25 11:09:28 crc kubenswrapper[4776]: I1125 11:09:28.089528 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-ea9d-account-create-gtwm6"] Nov 25 11:09:29 crc kubenswrapper[4776]: I1125 11:09:29.680967 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54887439-da3f-418c-85bd-3754d009e466" path="/var/lib/kubelet/pods/54887439-da3f-418c-85bd-3754d009e466/volumes" Nov 25 11:09:29 crc kubenswrapper[4776]: I1125 11:09:29.682466 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55d65b36-5e84-4842-8184-2bf6ed987f45" path="/var/lib/kubelet/pods/55d65b36-5e84-4842-8184-2bf6ed987f45/volumes" Nov 25 11:09:30 crc kubenswrapper[4776]: I1125 11:09:30.218947 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7c9cb59558-gpdwk" podUID="238717c2-18fa-47d1-b39e-4d2f0b3619de" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.124:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.124:8443: connect: connection refused" Nov 25 11:09:32 crc kubenswrapper[4776]: I1125 11:09:32.663434 4776 scope.go:117] "RemoveContainer" containerID="8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698" Nov 25 11:09:33 crc kubenswrapper[4776]: I1125 11:09:33.372942 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"d9823ef55a496477edc68cef40859ee8ddcd3f8a6fd6685be1f1eda0a305897e"} Nov 25 11:09:34 crc kubenswrapper[4776]: I1125 11:09:34.197176 4776 scope.go:117] "RemoveContainer" containerID="bc26d14a9a86b5d052663b9b6e0df405a57c8fd4a7e1a22acae4fa55c7eccd22" Nov 25 11:09:34 crc kubenswrapper[4776]: I1125 11:09:34.228564 4776 scope.go:117] "RemoveContainer" containerID="4dade33a1326ebc318b0c703187ab1716a56503c32445a3856d0ad68bf7c7564" Nov 25 11:09:34 crc kubenswrapper[4776]: I1125 11:09:34.310215 4776 scope.go:117] "RemoveContainer" containerID="ceadafb6f7dd12bf36e87137b5b7f7a26ddf7fac48ce172f88cbba9e25611f51" Nov 25 11:09:34 crc kubenswrapper[4776]: I1125 11:09:34.360248 4776 scope.go:117] "RemoveContainer" containerID="9b819bc49d6e3e198bb8825480a3e5ed1756ef83efdf29cd7e01996b5f25e992" Nov 25 11:09:38 crc kubenswrapper[4776]: E1125 11:09:38.932862 4776 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/ef2e022a71fe372ca04399153ddedf5ba9036a462cf84fde4dcafe59daae2ccb/diff" to get inode usage: stat /var/lib/containers/storage/overlay/ef2e022a71fe372ca04399153ddedf5ba9036a462cf84fde4dcafe59daae2ccb/diff: no such file or directory, extraDiskErr: Nov 25 11:09:39 crc kubenswrapper[4776]: I1125 11:09:39.033897 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-4f94t"] Nov 25 11:09:39 crc kubenswrapper[4776]: I1125 11:09:39.048346 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-4f94t"] Nov 25 11:09:39 crc kubenswrapper[4776]: I1125 11:09:39.672883 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90693935-44c2-4e50-a8ad-5e2967ac582f" path="/var/lib/kubelet/pods/90693935-44c2-4e50-a8ad-5e2967ac582f/volumes" Nov 25 11:09:40 crc kubenswrapper[4776]: I1125 11:09:40.218499 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7c9cb59558-gpdwk" podUID="238717c2-18fa-47d1-b39e-4d2f0b3619de" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.124:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.124:8443: connect: connection refused" Nov 25 11:09:40 crc kubenswrapper[4776]: I1125 11:09:40.219857 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:09:44 crc kubenswrapper[4776]: E1125 11:09:44.086777 4776 manager.go:1116] Failed to create existing container: /kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4ceac98_94a9_4cc2_a405_d57a1a15cf9c.slice/crio-d3a620cc9e70afefea5e7c11ed49253141433709c1cfbc61a3eee63384412e88: Error finding container d3a620cc9e70afefea5e7c11ed49253141433709c1cfbc61a3eee63384412e88: Status 404 returned error can't find the container with id d3a620cc9e70afefea5e7c11ed49253141433709c1cfbc61a3eee63384412e88 Nov 25 11:09:44 crc kubenswrapper[4776]: I1125 11:09:44.496139 4776 generic.go:334] "Generic (PLEG): container finished" podID="238717c2-18fa-47d1-b39e-4d2f0b3619de" containerID="048841f4110bcd3f5595becc35292ea8885ba21d1bde0f0d3e40f649c52ef3a8" exitCode=137 Nov 25 11:09:44 crc kubenswrapper[4776]: I1125 11:09:44.496178 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c9cb59558-gpdwk" event={"ID":"238717c2-18fa-47d1-b39e-4d2f0b3619de","Type":"ContainerDied","Data":"048841f4110bcd3f5595becc35292ea8885ba21d1bde0f0d3e40f649c52ef3a8"} Nov 25 11:09:44 crc kubenswrapper[4776]: I1125 11:09:44.496732 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c9cb59558-gpdwk" event={"ID":"238717c2-18fa-47d1-b39e-4d2f0b3619de","Type":"ContainerDied","Data":"d4cdf4801299a49009f51e63142359764d6d309af9ebd9974481e2eed74bc82e"} Nov 25 11:09:44 crc kubenswrapper[4776]: I1125 11:09:44.496745 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4cdf4801299a49009f51e63142359764d6d309af9ebd9974481e2eed74bc82e" Nov 25 11:09:44 crc kubenswrapper[4776]: I1125 11:09:44.514950 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:09:44 crc kubenswrapper[4776]: I1125 11:09:44.662020 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/238717c2-18fa-47d1-b39e-4d2f0b3619de-config-data\") pod \"238717c2-18fa-47d1-b39e-4d2f0b3619de\" (UID: \"238717c2-18fa-47d1-b39e-4d2f0b3619de\") " Nov 25 11:09:44 crc kubenswrapper[4776]: I1125 11:09:44.662076 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/238717c2-18fa-47d1-b39e-4d2f0b3619de-horizon-secret-key\") pod \"238717c2-18fa-47d1-b39e-4d2f0b3619de\" (UID: \"238717c2-18fa-47d1-b39e-4d2f0b3619de\") " Nov 25 11:09:44 crc kubenswrapper[4776]: I1125 11:09:44.662103 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/238717c2-18fa-47d1-b39e-4d2f0b3619de-logs\") pod \"238717c2-18fa-47d1-b39e-4d2f0b3619de\" (UID: \"238717c2-18fa-47d1-b39e-4d2f0b3619de\") " Nov 25 11:09:44 crc kubenswrapper[4776]: I1125 11:09:44.662163 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/238717c2-18fa-47d1-b39e-4d2f0b3619de-scripts\") pod \"238717c2-18fa-47d1-b39e-4d2f0b3619de\" (UID: \"238717c2-18fa-47d1-b39e-4d2f0b3619de\") " Nov 25 11:09:44 crc kubenswrapper[4776]: I1125 11:09:44.662198 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/238717c2-18fa-47d1-b39e-4d2f0b3619de-horizon-tls-certs\") pod \"238717c2-18fa-47d1-b39e-4d2f0b3619de\" (UID: \"238717c2-18fa-47d1-b39e-4d2f0b3619de\") " Nov 25 11:09:44 crc kubenswrapper[4776]: I1125 11:09:44.662224 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-659v5\" (UniqueName: \"kubernetes.io/projected/238717c2-18fa-47d1-b39e-4d2f0b3619de-kube-api-access-659v5\") pod \"238717c2-18fa-47d1-b39e-4d2f0b3619de\" (UID: \"238717c2-18fa-47d1-b39e-4d2f0b3619de\") " Nov 25 11:09:44 crc kubenswrapper[4776]: I1125 11:09:44.662275 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/238717c2-18fa-47d1-b39e-4d2f0b3619de-combined-ca-bundle\") pod \"238717c2-18fa-47d1-b39e-4d2f0b3619de\" (UID: \"238717c2-18fa-47d1-b39e-4d2f0b3619de\") " Nov 25 11:09:44 crc kubenswrapper[4776]: I1125 11:09:44.663056 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/238717c2-18fa-47d1-b39e-4d2f0b3619de-logs" (OuterVolumeSpecName: "logs") pod "238717c2-18fa-47d1-b39e-4d2f0b3619de" (UID: "238717c2-18fa-47d1-b39e-4d2f0b3619de"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:09:44 crc kubenswrapper[4776]: I1125 11:09:44.664555 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/238717c2-18fa-47d1-b39e-4d2f0b3619de-logs\") on node \"crc\" DevicePath \"\"" Nov 25 11:09:44 crc kubenswrapper[4776]: I1125 11:09:44.668434 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/238717c2-18fa-47d1-b39e-4d2f0b3619de-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "238717c2-18fa-47d1-b39e-4d2f0b3619de" (UID: "238717c2-18fa-47d1-b39e-4d2f0b3619de"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:09:44 crc kubenswrapper[4776]: I1125 11:09:44.680371 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/238717c2-18fa-47d1-b39e-4d2f0b3619de-kube-api-access-659v5" (OuterVolumeSpecName: "kube-api-access-659v5") pod "238717c2-18fa-47d1-b39e-4d2f0b3619de" (UID: "238717c2-18fa-47d1-b39e-4d2f0b3619de"). InnerVolumeSpecName "kube-api-access-659v5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:09:44 crc kubenswrapper[4776]: I1125 11:09:44.694573 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/238717c2-18fa-47d1-b39e-4d2f0b3619de-scripts" (OuterVolumeSpecName: "scripts") pod "238717c2-18fa-47d1-b39e-4d2f0b3619de" (UID: "238717c2-18fa-47d1-b39e-4d2f0b3619de"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:09:44 crc kubenswrapper[4776]: I1125 11:09:44.694878 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/238717c2-18fa-47d1-b39e-4d2f0b3619de-config-data" (OuterVolumeSpecName: "config-data") pod "238717c2-18fa-47d1-b39e-4d2f0b3619de" (UID: "238717c2-18fa-47d1-b39e-4d2f0b3619de"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:09:44 crc kubenswrapper[4776]: I1125 11:09:44.697231 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/238717c2-18fa-47d1-b39e-4d2f0b3619de-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "238717c2-18fa-47d1-b39e-4d2f0b3619de" (UID: "238717c2-18fa-47d1-b39e-4d2f0b3619de"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:09:44 crc kubenswrapper[4776]: I1125 11:09:44.726600 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/238717c2-18fa-47d1-b39e-4d2f0b3619de-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "238717c2-18fa-47d1-b39e-4d2f0b3619de" (UID: "238717c2-18fa-47d1-b39e-4d2f0b3619de"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:09:44 crc kubenswrapper[4776]: I1125 11:09:44.766238 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/238717c2-18fa-47d1-b39e-4d2f0b3619de-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:09:44 crc kubenswrapper[4776]: I1125 11:09:44.766270 4776 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/238717c2-18fa-47d1-b39e-4d2f0b3619de-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:09:44 crc kubenswrapper[4776]: I1125 11:09:44.766281 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/238717c2-18fa-47d1-b39e-4d2f0b3619de-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:09:44 crc kubenswrapper[4776]: I1125 11:09:44.766290 4776 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/238717c2-18fa-47d1-b39e-4d2f0b3619de-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 11:09:44 crc kubenswrapper[4776]: I1125 11:09:44.766300 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-659v5\" (UniqueName: \"kubernetes.io/projected/238717c2-18fa-47d1-b39e-4d2f0b3619de-kube-api-access-659v5\") on node \"crc\" DevicePath \"\"" Nov 25 11:09:44 crc kubenswrapper[4776]: I1125 11:09:44.766364 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/238717c2-18fa-47d1-b39e-4d2f0b3619de-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:09:45 crc kubenswrapper[4776]: I1125 11:09:45.504908 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c9cb59558-gpdwk" Nov 25 11:09:45 crc kubenswrapper[4776]: I1125 11:09:45.537428 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7c9cb59558-gpdwk"] Nov 25 11:09:45 crc kubenswrapper[4776]: I1125 11:09:45.545553 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7c9cb59558-gpdwk"] Nov 25 11:09:45 crc kubenswrapper[4776]: I1125 11:09:45.689664 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="238717c2-18fa-47d1-b39e-4d2f0b3619de" path="/var/lib/kubelet/pods/238717c2-18fa-47d1-b39e-4d2f0b3619de/volumes" Nov 25 11:09:54 crc kubenswrapper[4776]: I1125 11:09:54.859490 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-f4f475bb-4hkzw"] Nov 25 11:09:54 crc kubenswrapper[4776]: E1125 11:09:54.860861 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c465b70c-b76d-430c-954e-b8408390cd99" containerName="horizon-log" Nov 25 11:09:54 crc kubenswrapper[4776]: I1125 11:09:54.860880 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="c465b70c-b76d-430c-954e-b8408390cd99" containerName="horizon-log" Nov 25 11:09:54 crc kubenswrapper[4776]: E1125 11:09:54.860897 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="238717c2-18fa-47d1-b39e-4d2f0b3619de" containerName="horizon" Nov 25 11:09:54 crc kubenswrapper[4776]: I1125 11:09:54.860904 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="238717c2-18fa-47d1-b39e-4d2f0b3619de" containerName="horizon" Nov 25 11:09:54 crc kubenswrapper[4776]: E1125 11:09:54.860930 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="238717c2-18fa-47d1-b39e-4d2f0b3619de" containerName="horizon-log" Nov 25 11:09:54 crc kubenswrapper[4776]: I1125 11:09:54.860939 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="238717c2-18fa-47d1-b39e-4d2f0b3619de" containerName="horizon-log" Nov 25 11:09:54 crc kubenswrapper[4776]: E1125 11:09:54.860950 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4ceac98-94a9-4cc2-a405-d57a1a15cf9c" containerName="horizon-log" Nov 25 11:09:54 crc kubenswrapper[4776]: I1125 11:09:54.860957 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4ceac98-94a9-4cc2-a405-d57a1a15cf9c" containerName="horizon-log" Nov 25 11:09:54 crc kubenswrapper[4776]: E1125 11:09:54.860970 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c465b70c-b76d-430c-954e-b8408390cd99" containerName="horizon" Nov 25 11:09:54 crc kubenswrapper[4776]: I1125 11:09:54.860977 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="c465b70c-b76d-430c-954e-b8408390cd99" containerName="horizon" Nov 25 11:09:54 crc kubenswrapper[4776]: E1125 11:09:54.860988 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4ceac98-94a9-4cc2-a405-d57a1a15cf9c" containerName="horizon" Nov 25 11:09:54 crc kubenswrapper[4776]: I1125 11:09:54.860995 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4ceac98-94a9-4cc2-a405-d57a1a15cf9c" containerName="horizon" Nov 25 11:09:54 crc kubenswrapper[4776]: I1125 11:09:54.861307 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="c465b70c-b76d-430c-954e-b8408390cd99" containerName="horizon-log" Nov 25 11:09:54 crc kubenswrapper[4776]: I1125 11:09:54.861330 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4ceac98-94a9-4cc2-a405-d57a1a15cf9c" containerName="horizon-log" Nov 25 11:09:54 crc kubenswrapper[4776]: I1125 11:09:54.861341 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="238717c2-18fa-47d1-b39e-4d2f0b3619de" containerName="horizon" Nov 25 11:09:54 crc kubenswrapper[4776]: I1125 11:09:54.861359 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="c465b70c-b76d-430c-954e-b8408390cd99" containerName="horizon" Nov 25 11:09:54 crc kubenswrapper[4776]: I1125 11:09:54.861375 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="238717c2-18fa-47d1-b39e-4d2f0b3619de" containerName="horizon-log" Nov 25 11:09:54 crc kubenswrapper[4776]: I1125 11:09:54.861389 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4ceac98-94a9-4cc2-a405-d57a1a15cf9c" containerName="horizon" Nov 25 11:09:54 crc kubenswrapper[4776]: I1125 11:09:54.897008 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f4f475bb-4hkzw" Nov 25 11:09:54 crc kubenswrapper[4776]: I1125 11:09:54.899458 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-f4f475bb-4hkzw"] Nov 25 11:09:54 crc kubenswrapper[4776]: I1125 11:09:54.903124 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/029b9a31-76a8-446d-be2c-53760cb3980a-combined-ca-bundle\") pod \"horizon-f4f475bb-4hkzw\" (UID: \"029b9a31-76a8-446d-be2c-53760cb3980a\") " pod="openstack/horizon-f4f475bb-4hkzw" Nov 25 11:09:54 crc kubenswrapper[4776]: I1125 11:09:54.903225 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/029b9a31-76a8-446d-be2c-53760cb3980a-logs\") pod \"horizon-f4f475bb-4hkzw\" (UID: \"029b9a31-76a8-446d-be2c-53760cb3980a\") " pod="openstack/horizon-f4f475bb-4hkzw" Nov 25 11:09:54 crc kubenswrapper[4776]: I1125 11:09:54.903251 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/029b9a31-76a8-446d-be2c-53760cb3980a-scripts\") pod \"horizon-f4f475bb-4hkzw\" (UID: \"029b9a31-76a8-446d-be2c-53760cb3980a\") " pod="openstack/horizon-f4f475bb-4hkzw" Nov 25 11:09:54 crc kubenswrapper[4776]: I1125 11:09:54.903279 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/029b9a31-76a8-446d-be2c-53760cb3980a-horizon-secret-key\") pod \"horizon-f4f475bb-4hkzw\" (UID: \"029b9a31-76a8-446d-be2c-53760cb3980a\") " pod="openstack/horizon-f4f475bb-4hkzw" Nov 25 11:09:54 crc kubenswrapper[4776]: I1125 11:09:54.903300 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/029b9a31-76a8-446d-be2c-53760cb3980a-horizon-tls-certs\") pod \"horizon-f4f475bb-4hkzw\" (UID: \"029b9a31-76a8-446d-be2c-53760cb3980a\") " pod="openstack/horizon-f4f475bb-4hkzw" Nov 25 11:09:54 crc kubenswrapper[4776]: I1125 11:09:54.903318 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/029b9a31-76a8-446d-be2c-53760cb3980a-config-data\") pod \"horizon-f4f475bb-4hkzw\" (UID: \"029b9a31-76a8-446d-be2c-53760cb3980a\") " pod="openstack/horizon-f4f475bb-4hkzw" Nov 25 11:09:54 crc kubenswrapper[4776]: I1125 11:09:54.903393 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlxkm\" (UniqueName: \"kubernetes.io/projected/029b9a31-76a8-446d-be2c-53760cb3980a-kube-api-access-wlxkm\") pod \"horizon-f4f475bb-4hkzw\" (UID: \"029b9a31-76a8-446d-be2c-53760cb3980a\") " pod="openstack/horizon-f4f475bb-4hkzw" Nov 25 11:09:55 crc kubenswrapper[4776]: I1125 11:09:55.004692 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/029b9a31-76a8-446d-be2c-53760cb3980a-logs\") pod \"horizon-f4f475bb-4hkzw\" (UID: \"029b9a31-76a8-446d-be2c-53760cb3980a\") " pod="openstack/horizon-f4f475bb-4hkzw" Nov 25 11:09:55 crc kubenswrapper[4776]: I1125 11:09:55.004755 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/029b9a31-76a8-446d-be2c-53760cb3980a-scripts\") pod \"horizon-f4f475bb-4hkzw\" (UID: \"029b9a31-76a8-446d-be2c-53760cb3980a\") " pod="openstack/horizon-f4f475bb-4hkzw" Nov 25 11:09:55 crc kubenswrapper[4776]: I1125 11:09:55.004788 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/029b9a31-76a8-446d-be2c-53760cb3980a-config-data\") pod \"horizon-f4f475bb-4hkzw\" (UID: \"029b9a31-76a8-446d-be2c-53760cb3980a\") " pod="openstack/horizon-f4f475bb-4hkzw" Nov 25 11:09:55 crc kubenswrapper[4776]: I1125 11:09:55.004810 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/029b9a31-76a8-446d-be2c-53760cb3980a-horizon-secret-key\") pod \"horizon-f4f475bb-4hkzw\" (UID: \"029b9a31-76a8-446d-be2c-53760cb3980a\") " pod="openstack/horizon-f4f475bb-4hkzw" Nov 25 11:09:55 crc kubenswrapper[4776]: I1125 11:09:55.004834 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/029b9a31-76a8-446d-be2c-53760cb3980a-horizon-tls-certs\") pod \"horizon-f4f475bb-4hkzw\" (UID: \"029b9a31-76a8-446d-be2c-53760cb3980a\") " pod="openstack/horizon-f4f475bb-4hkzw" Nov 25 11:09:55 crc kubenswrapper[4776]: I1125 11:09:55.004925 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlxkm\" (UniqueName: \"kubernetes.io/projected/029b9a31-76a8-446d-be2c-53760cb3980a-kube-api-access-wlxkm\") pod \"horizon-f4f475bb-4hkzw\" (UID: \"029b9a31-76a8-446d-be2c-53760cb3980a\") " pod="openstack/horizon-f4f475bb-4hkzw" Nov 25 11:09:55 crc kubenswrapper[4776]: I1125 11:09:55.005002 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/029b9a31-76a8-446d-be2c-53760cb3980a-combined-ca-bundle\") pod \"horizon-f4f475bb-4hkzw\" (UID: \"029b9a31-76a8-446d-be2c-53760cb3980a\") " pod="openstack/horizon-f4f475bb-4hkzw" Nov 25 11:09:55 crc kubenswrapper[4776]: I1125 11:09:55.005370 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/029b9a31-76a8-446d-be2c-53760cb3980a-logs\") pod \"horizon-f4f475bb-4hkzw\" (UID: \"029b9a31-76a8-446d-be2c-53760cb3980a\") " pod="openstack/horizon-f4f475bb-4hkzw" Nov 25 11:09:55 crc kubenswrapper[4776]: I1125 11:09:55.005715 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/029b9a31-76a8-446d-be2c-53760cb3980a-scripts\") pod \"horizon-f4f475bb-4hkzw\" (UID: \"029b9a31-76a8-446d-be2c-53760cb3980a\") " pod="openstack/horizon-f4f475bb-4hkzw" Nov 25 11:09:55 crc kubenswrapper[4776]: I1125 11:09:55.006527 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/029b9a31-76a8-446d-be2c-53760cb3980a-config-data\") pod \"horizon-f4f475bb-4hkzw\" (UID: \"029b9a31-76a8-446d-be2c-53760cb3980a\") " pod="openstack/horizon-f4f475bb-4hkzw" Nov 25 11:09:55 crc kubenswrapper[4776]: I1125 11:09:55.012414 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/029b9a31-76a8-446d-be2c-53760cb3980a-horizon-secret-key\") pod \"horizon-f4f475bb-4hkzw\" (UID: \"029b9a31-76a8-446d-be2c-53760cb3980a\") " pod="openstack/horizon-f4f475bb-4hkzw" Nov 25 11:09:55 crc kubenswrapper[4776]: I1125 11:09:55.012485 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/029b9a31-76a8-446d-be2c-53760cb3980a-horizon-tls-certs\") pod \"horizon-f4f475bb-4hkzw\" (UID: \"029b9a31-76a8-446d-be2c-53760cb3980a\") " pod="openstack/horizon-f4f475bb-4hkzw" Nov 25 11:09:55 crc kubenswrapper[4776]: I1125 11:09:55.013899 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/029b9a31-76a8-446d-be2c-53760cb3980a-combined-ca-bundle\") pod \"horizon-f4f475bb-4hkzw\" (UID: \"029b9a31-76a8-446d-be2c-53760cb3980a\") " pod="openstack/horizon-f4f475bb-4hkzw" Nov 25 11:09:55 crc kubenswrapper[4776]: I1125 11:09:55.024437 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlxkm\" (UniqueName: \"kubernetes.io/projected/029b9a31-76a8-446d-be2c-53760cb3980a-kube-api-access-wlxkm\") pod \"horizon-f4f475bb-4hkzw\" (UID: \"029b9a31-76a8-446d-be2c-53760cb3980a\") " pod="openstack/horizon-f4f475bb-4hkzw" Nov 25 11:09:55 crc kubenswrapper[4776]: I1125 11:09:55.233573 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f4f475bb-4hkzw" Nov 25 11:09:55 crc kubenswrapper[4776]: I1125 11:09:55.917736 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-f4f475bb-4hkzw"] Nov 25 11:09:56 crc kubenswrapper[4776]: I1125 11:09:56.344536 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-2jqfn"] Nov 25 11:09:56 crc kubenswrapper[4776]: I1125 11:09:56.346639 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-2jqfn" Nov 25 11:09:56 crc kubenswrapper[4776]: I1125 11:09:56.354772 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-2jqfn"] Nov 25 11:09:56 crc kubenswrapper[4776]: I1125 11:09:56.439411 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/559c0e1c-d5ad-46f7-b9c5-3597858cc244-operator-scripts\") pod \"heat-db-create-2jqfn\" (UID: \"559c0e1c-d5ad-46f7-b9c5-3597858cc244\") " pod="openstack/heat-db-create-2jqfn" Nov 25 11:09:56 crc kubenswrapper[4776]: I1125 11:09:56.439729 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbq99\" (UniqueName: \"kubernetes.io/projected/559c0e1c-d5ad-46f7-b9c5-3597858cc244-kube-api-access-pbq99\") pod \"heat-db-create-2jqfn\" (UID: \"559c0e1c-d5ad-46f7-b9c5-3597858cc244\") " pod="openstack/heat-db-create-2jqfn" Nov 25 11:09:56 crc kubenswrapper[4776]: I1125 11:09:56.444685 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-695d-account-create-fq4l2"] Nov 25 11:09:56 crc kubenswrapper[4776]: I1125 11:09:56.445946 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-695d-account-create-fq4l2" Nov 25 11:09:56 crc kubenswrapper[4776]: I1125 11:09:56.448767 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Nov 25 11:09:56 crc kubenswrapper[4776]: I1125 11:09:56.467608 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-695d-account-create-fq4l2"] Nov 25 11:09:56 crc kubenswrapper[4776]: I1125 11:09:56.541917 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzlkq\" (UniqueName: \"kubernetes.io/projected/b650f074-0767-497f-8a4d-d7e477b759f6-kube-api-access-mzlkq\") pod \"heat-695d-account-create-fq4l2\" (UID: \"b650f074-0767-497f-8a4d-d7e477b759f6\") " pod="openstack/heat-695d-account-create-fq4l2" Nov 25 11:09:56 crc kubenswrapper[4776]: I1125 11:09:56.542061 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b650f074-0767-497f-8a4d-d7e477b759f6-operator-scripts\") pod \"heat-695d-account-create-fq4l2\" (UID: \"b650f074-0767-497f-8a4d-d7e477b759f6\") " pod="openstack/heat-695d-account-create-fq4l2" Nov 25 11:09:56 crc kubenswrapper[4776]: I1125 11:09:56.542147 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/559c0e1c-d5ad-46f7-b9c5-3597858cc244-operator-scripts\") pod \"heat-db-create-2jqfn\" (UID: \"559c0e1c-d5ad-46f7-b9c5-3597858cc244\") " pod="openstack/heat-db-create-2jqfn" Nov 25 11:09:56 crc kubenswrapper[4776]: I1125 11:09:56.542340 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbq99\" (UniqueName: \"kubernetes.io/projected/559c0e1c-d5ad-46f7-b9c5-3597858cc244-kube-api-access-pbq99\") pod \"heat-db-create-2jqfn\" (UID: \"559c0e1c-d5ad-46f7-b9c5-3597858cc244\") " pod="openstack/heat-db-create-2jqfn" Nov 25 11:09:56 crc kubenswrapper[4776]: I1125 11:09:56.543027 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/559c0e1c-d5ad-46f7-b9c5-3597858cc244-operator-scripts\") pod \"heat-db-create-2jqfn\" (UID: \"559c0e1c-d5ad-46f7-b9c5-3597858cc244\") " pod="openstack/heat-db-create-2jqfn" Nov 25 11:09:56 crc kubenswrapper[4776]: I1125 11:09:56.562603 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbq99\" (UniqueName: \"kubernetes.io/projected/559c0e1c-d5ad-46f7-b9c5-3597858cc244-kube-api-access-pbq99\") pod \"heat-db-create-2jqfn\" (UID: \"559c0e1c-d5ad-46f7-b9c5-3597858cc244\") " pod="openstack/heat-db-create-2jqfn" Nov 25 11:09:56 crc kubenswrapper[4776]: I1125 11:09:56.635988 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f4f475bb-4hkzw" event={"ID":"029b9a31-76a8-446d-be2c-53760cb3980a","Type":"ContainerStarted","Data":"c64a57030dd474ffd2c8e200bf7e9c85970ab5f44bb4c50bb4f6354926779d60"} Nov 25 11:09:56 crc kubenswrapper[4776]: I1125 11:09:56.636034 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f4f475bb-4hkzw" event={"ID":"029b9a31-76a8-446d-be2c-53760cb3980a","Type":"ContainerStarted","Data":"052f7e98f69081f6d54f2bd7d4c68bd3cec71e3cdaa41403ac4cc11238d709a2"} Nov 25 11:09:56 crc kubenswrapper[4776]: I1125 11:09:56.643407 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b650f074-0767-497f-8a4d-d7e477b759f6-operator-scripts\") pod \"heat-695d-account-create-fq4l2\" (UID: \"b650f074-0767-497f-8a4d-d7e477b759f6\") " pod="openstack/heat-695d-account-create-fq4l2" Nov 25 11:09:56 crc kubenswrapper[4776]: I1125 11:09:56.643780 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzlkq\" (UniqueName: \"kubernetes.io/projected/b650f074-0767-497f-8a4d-d7e477b759f6-kube-api-access-mzlkq\") pod \"heat-695d-account-create-fq4l2\" (UID: \"b650f074-0767-497f-8a4d-d7e477b759f6\") " pod="openstack/heat-695d-account-create-fq4l2" Nov 25 11:09:56 crc kubenswrapper[4776]: I1125 11:09:56.651712 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b650f074-0767-497f-8a4d-d7e477b759f6-operator-scripts\") pod \"heat-695d-account-create-fq4l2\" (UID: \"b650f074-0767-497f-8a4d-d7e477b759f6\") " pod="openstack/heat-695d-account-create-fq4l2" Nov 25 11:09:56 crc kubenswrapper[4776]: I1125 11:09:56.663578 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzlkq\" (UniqueName: \"kubernetes.io/projected/b650f074-0767-497f-8a4d-d7e477b759f6-kube-api-access-mzlkq\") pod \"heat-695d-account-create-fq4l2\" (UID: \"b650f074-0767-497f-8a4d-d7e477b759f6\") " pod="openstack/heat-695d-account-create-fq4l2" Nov 25 11:09:56 crc kubenswrapper[4776]: I1125 11:09:56.666146 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-2jqfn" Nov 25 11:09:56 crc kubenswrapper[4776]: I1125 11:09:56.953710 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-695d-account-create-fq4l2" Nov 25 11:09:57 crc kubenswrapper[4776]: I1125 11:09:57.121770 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-2jqfn"] Nov 25 11:09:57 crc kubenswrapper[4776]: I1125 11:09:57.418239 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-695d-account-create-fq4l2"] Nov 25 11:09:57 crc kubenswrapper[4776]: W1125 11:09:57.423297 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb650f074_0767_497f_8a4d_d7e477b759f6.slice/crio-d8207908f348b47a43ead3f4754e9f10a85997fa38f6f5e0752e9d472815b0e2 WatchSource:0}: Error finding container d8207908f348b47a43ead3f4754e9f10a85997fa38f6f5e0752e9d472815b0e2: Status 404 returned error can't find the container with id d8207908f348b47a43ead3f4754e9f10a85997fa38f6f5e0752e9d472815b0e2 Nov 25 11:09:57 crc kubenswrapper[4776]: I1125 11:09:57.649904 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-695d-account-create-fq4l2" event={"ID":"b650f074-0767-497f-8a4d-d7e477b759f6","Type":"ContainerStarted","Data":"d8207908f348b47a43ead3f4754e9f10a85997fa38f6f5e0752e9d472815b0e2"} Nov 25 11:09:57 crc kubenswrapper[4776]: I1125 11:09:57.652021 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f4f475bb-4hkzw" event={"ID":"029b9a31-76a8-446d-be2c-53760cb3980a","Type":"ContainerStarted","Data":"00bda29ea7dc5b6e7ee3e165b712bd88ee3956a6d868f889fecc297f1fcad9b8"} Nov 25 11:09:57 crc kubenswrapper[4776]: I1125 11:09:57.657132 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-2jqfn" event={"ID":"559c0e1c-d5ad-46f7-b9c5-3597858cc244","Type":"ContainerStarted","Data":"ece40a72eb7c52a50691108104116e718253051f03f3be7484c2f5aed9864bfb"} Nov 25 11:09:57 crc kubenswrapper[4776]: I1125 11:09:57.657168 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-2jqfn" event={"ID":"559c0e1c-d5ad-46f7-b9c5-3597858cc244","Type":"ContainerStarted","Data":"9e2197045c32b925122ce979e7e9342c107fcb3684bf0deb833b6100d196ece4"} Nov 25 11:09:57 crc kubenswrapper[4776]: I1125 11:09:57.685326 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-f4f475bb-4hkzw" podStartSLOduration=3.685307171 podStartE2EDuration="3.685307171s" podCreationTimestamp="2025-11-25 11:09:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:09:57.675446304 +0000 UTC m=+6342.716505857" watchObservedRunningTime="2025-11-25 11:09:57.685307171 +0000 UTC m=+6342.726366724" Nov 25 11:09:57 crc kubenswrapper[4776]: I1125 11:09:57.700373 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-create-2jqfn" podStartSLOduration=1.7003548990000001 podStartE2EDuration="1.700354899s" podCreationTimestamp="2025-11-25 11:09:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:09:57.693987009 +0000 UTC m=+6342.735046562" watchObservedRunningTime="2025-11-25 11:09:57.700354899 +0000 UTC m=+6342.741414482" Nov 25 11:09:58 crc kubenswrapper[4776]: I1125 11:09:58.668283 4776 generic.go:334] "Generic (PLEG): container finished" podID="559c0e1c-d5ad-46f7-b9c5-3597858cc244" containerID="ece40a72eb7c52a50691108104116e718253051f03f3be7484c2f5aed9864bfb" exitCode=0 Nov 25 11:09:58 crc kubenswrapper[4776]: I1125 11:09:58.668442 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-2jqfn" event={"ID":"559c0e1c-d5ad-46f7-b9c5-3597858cc244","Type":"ContainerDied","Data":"ece40a72eb7c52a50691108104116e718253051f03f3be7484c2f5aed9864bfb"} Nov 25 11:09:58 crc kubenswrapper[4776]: I1125 11:09:58.670486 4776 generic.go:334] "Generic (PLEG): container finished" podID="b650f074-0767-497f-8a4d-d7e477b759f6" containerID="c7a14088844565f332ea7726f4c85eed0320545074691122dd0f6f2040e3ed6d" exitCode=0 Nov 25 11:09:58 crc kubenswrapper[4776]: I1125 11:09:58.670721 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-695d-account-create-fq4l2" event={"ID":"b650f074-0767-497f-8a4d-d7e477b759f6","Type":"ContainerDied","Data":"c7a14088844565f332ea7726f4c85eed0320545074691122dd0f6f2040e3ed6d"} Nov 25 11:10:00 crc kubenswrapper[4776]: I1125 11:10:00.124229 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-2jqfn" Nov 25 11:10:00 crc kubenswrapper[4776]: I1125 11:10:00.132763 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-695d-account-create-fq4l2" Nov 25 11:10:00 crc kubenswrapper[4776]: I1125 11:10:00.316955 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/559c0e1c-d5ad-46f7-b9c5-3597858cc244-operator-scripts\") pod \"559c0e1c-d5ad-46f7-b9c5-3597858cc244\" (UID: \"559c0e1c-d5ad-46f7-b9c5-3597858cc244\") " Nov 25 11:10:00 crc kubenswrapper[4776]: I1125 11:10:00.317034 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b650f074-0767-497f-8a4d-d7e477b759f6-operator-scripts\") pod \"b650f074-0767-497f-8a4d-d7e477b759f6\" (UID: \"b650f074-0767-497f-8a4d-d7e477b759f6\") " Nov 25 11:10:00 crc kubenswrapper[4776]: I1125 11:10:00.317142 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbq99\" (UniqueName: \"kubernetes.io/projected/559c0e1c-d5ad-46f7-b9c5-3597858cc244-kube-api-access-pbq99\") pod \"559c0e1c-d5ad-46f7-b9c5-3597858cc244\" (UID: \"559c0e1c-d5ad-46f7-b9c5-3597858cc244\") " Nov 25 11:10:00 crc kubenswrapper[4776]: I1125 11:10:00.317380 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzlkq\" (UniqueName: \"kubernetes.io/projected/b650f074-0767-497f-8a4d-d7e477b759f6-kube-api-access-mzlkq\") pod \"b650f074-0767-497f-8a4d-d7e477b759f6\" (UID: \"b650f074-0767-497f-8a4d-d7e477b759f6\") " Nov 25 11:10:00 crc kubenswrapper[4776]: I1125 11:10:00.317640 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b650f074-0767-497f-8a4d-d7e477b759f6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b650f074-0767-497f-8a4d-d7e477b759f6" (UID: "b650f074-0767-497f-8a4d-d7e477b759f6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:10:00 crc kubenswrapper[4776]: I1125 11:10:00.317641 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/559c0e1c-d5ad-46f7-b9c5-3597858cc244-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "559c0e1c-d5ad-46f7-b9c5-3597858cc244" (UID: "559c0e1c-d5ad-46f7-b9c5-3597858cc244"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:10:00 crc kubenswrapper[4776]: I1125 11:10:00.318227 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/559c0e1c-d5ad-46f7-b9c5-3597858cc244-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:10:00 crc kubenswrapper[4776]: I1125 11:10:00.318245 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b650f074-0767-497f-8a4d-d7e477b759f6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:10:00 crc kubenswrapper[4776]: I1125 11:10:00.322875 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b650f074-0767-497f-8a4d-d7e477b759f6-kube-api-access-mzlkq" (OuterVolumeSpecName: "kube-api-access-mzlkq") pod "b650f074-0767-497f-8a4d-d7e477b759f6" (UID: "b650f074-0767-497f-8a4d-d7e477b759f6"). InnerVolumeSpecName "kube-api-access-mzlkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:10:00 crc kubenswrapper[4776]: I1125 11:10:00.323370 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/559c0e1c-d5ad-46f7-b9c5-3597858cc244-kube-api-access-pbq99" (OuterVolumeSpecName: "kube-api-access-pbq99") pod "559c0e1c-d5ad-46f7-b9c5-3597858cc244" (UID: "559c0e1c-d5ad-46f7-b9c5-3597858cc244"). InnerVolumeSpecName "kube-api-access-pbq99". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:10:00 crc kubenswrapper[4776]: I1125 11:10:00.419854 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbq99\" (UniqueName: \"kubernetes.io/projected/559c0e1c-d5ad-46f7-b9c5-3597858cc244-kube-api-access-pbq99\") on node \"crc\" DevicePath \"\"" Nov 25 11:10:00 crc kubenswrapper[4776]: I1125 11:10:00.419900 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzlkq\" (UniqueName: \"kubernetes.io/projected/b650f074-0767-497f-8a4d-d7e477b759f6-kube-api-access-mzlkq\") on node \"crc\" DevicePath \"\"" Nov 25 11:10:00 crc kubenswrapper[4776]: I1125 11:10:00.690431 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-2jqfn" event={"ID":"559c0e1c-d5ad-46f7-b9c5-3597858cc244","Type":"ContainerDied","Data":"9e2197045c32b925122ce979e7e9342c107fcb3684bf0deb833b6100d196ece4"} Nov 25 11:10:00 crc kubenswrapper[4776]: I1125 11:10:00.690475 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e2197045c32b925122ce979e7e9342c107fcb3684bf0deb833b6100d196ece4" Nov 25 11:10:00 crc kubenswrapper[4776]: I1125 11:10:00.690529 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-2jqfn" Nov 25 11:10:00 crc kubenswrapper[4776]: I1125 11:10:00.704222 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-695d-account-create-fq4l2" event={"ID":"b650f074-0767-497f-8a4d-d7e477b759f6","Type":"ContainerDied","Data":"d8207908f348b47a43ead3f4754e9f10a85997fa38f6f5e0752e9d472815b0e2"} Nov 25 11:10:00 crc kubenswrapper[4776]: I1125 11:10:00.704259 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8207908f348b47a43ead3f4754e9f10a85997fa38f6f5e0752e9d472815b0e2" Nov 25 11:10:00 crc kubenswrapper[4776]: I1125 11:10:00.704304 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-695d-account-create-fq4l2" Nov 25 11:10:01 crc kubenswrapper[4776]: I1125 11:10:01.673062 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-vw6x4"] Nov 25 11:10:01 crc kubenswrapper[4776]: E1125 11:10:01.673746 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="559c0e1c-d5ad-46f7-b9c5-3597858cc244" containerName="mariadb-database-create" Nov 25 11:10:01 crc kubenswrapper[4776]: I1125 11:10:01.673758 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="559c0e1c-d5ad-46f7-b9c5-3597858cc244" containerName="mariadb-database-create" Nov 25 11:10:01 crc kubenswrapper[4776]: E1125 11:10:01.673787 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b650f074-0767-497f-8a4d-d7e477b759f6" containerName="mariadb-account-create" Nov 25 11:10:01 crc kubenswrapper[4776]: I1125 11:10:01.673794 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="b650f074-0767-497f-8a4d-d7e477b759f6" containerName="mariadb-account-create" Nov 25 11:10:01 crc kubenswrapper[4776]: I1125 11:10:01.673978 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="b650f074-0767-497f-8a4d-d7e477b759f6" containerName="mariadb-account-create" Nov 25 11:10:01 crc kubenswrapper[4776]: I1125 11:10:01.673998 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="559c0e1c-d5ad-46f7-b9c5-3597858cc244" containerName="mariadb-database-create" Nov 25 11:10:01 crc kubenswrapper[4776]: I1125 11:10:01.674727 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-vw6x4" Nov 25 11:10:01 crc kubenswrapper[4776]: I1125 11:10:01.678108 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-9lzlh" Nov 25 11:10:01 crc kubenswrapper[4776]: I1125 11:10:01.684453 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Nov 25 11:10:01 crc kubenswrapper[4776]: I1125 11:10:01.685143 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-vw6x4"] Nov 25 11:10:01 crc kubenswrapper[4776]: I1125 11:10:01.847964 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0-combined-ca-bundle\") pod \"heat-db-sync-vw6x4\" (UID: \"7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0\") " pod="openstack/heat-db-sync-vw6x4" Nov 25 11:10:01 crc kubenswrapper[4776]: I1125 11:10:01.848420 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5nhn\" (UniqueName: \"kubernetes.io/projected/7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0-kube-api-access-r5nhn\") pod \"heat-db-sync-vw6x4\" (UID: \"7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0\") " pod="openstack/heat-db-sync-vw6x4" Nov 25 11:10:01 crc kubenswrapper[4776]: I1125 11:10:01.848478 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0-config-data\") pod \"heat-db-sync-vw6x4\" (UID: \"7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0\") " pod="openstack/heat-db-sync-vw6x4" Nov 25 11:10:01 crc kubenswrapper[4776]: I1125 11:10:01.950561 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0-config-data\") pod \"heat-db-sync-vw6x4\" (UID: \"7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0\") " pod="openstack/heat-db-sync-vw6x4" Nov 25 11:10:01 crc kubenswrapper[4776]: I1125 11:10:01.950994 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0-combined-ca-bundle\") pod \"heat-db-sync-vw6x4\" (UID: \"7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0\") " pod="openstack/heat-db-sync-vw6x4" Nov 25 11:10:01 crc kubenswrapper[4776]: I1125 11:10:01.951173 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5nhn\" (UniqueName: \"kubernetes.io/projected/7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0-kube-api-access-r5nhn\") pod \"heat-db-sync-vw6x4\" (UID: \"7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0\") " pod="openstack/heat-db-sync-vw6x4" Nov 25 11:10:01 crc kubenswrapper[4776]: I1125 11:10:01.957867 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0-config-data\") pod \"heat-db-sync-vw6x4\" (UID: \"7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0\") " pod="openstack/heat-db-sync-vw6x4" Nov 25 11:10:01 crc kubenswrapper[4776]: I1125 11:10:01.957875 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0-combined-ca-bundle\") pod \"heat-db-sync-vw6x4\" (UID: \"7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0\") " pod="openstack/heat-db-sync-vw6x4" Nov 25 11:10:01 crc kubenswrapper[4776]: I1125 11:10:01.968661 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5nhn\" (UniqueName: \"kubernetes.io/projected/7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0-kube-api-access-r5nhn\") pod \"heat-db-sync-vw6x4\" (UID: \"7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0\") " pod="openstack/heat-db-sync-vw6x4" Nov 25 11:10:01 crc kubenswrapper[4776]: I1125 11:10:01.995472 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-vw6x4" Nov 25 11:10:02 crc kubenswrapper[4776]: W1125 11:10:02.503042 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ed9669a_44b6_4995_89a2_5b4bc6d7f4a0.slice/crio-7cc3f0608bf72ba439dd672f85d2188094403e534ded3e43cb74c59e6065652a WatchSource:0}: Error finding container 7cc3f0608bf72ba439dd672f85d2188094403e534ded3e43cb74c59e6065652a: Status 404 returned error can't find the container with id 7cc3f0608bf72ba439dd672f85d2188094403e534ded3e43cb74c59e6065652a Nov 25 11:10:02 crc kubenswrapper[4776]: I1125 11:10:02.505608 4776 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 11:10:02 crc kubenswrapper[4776]: I1125 11:10:02.505638 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-vw6x4"] Nov 25 11:10:02 crc kubenswrapper[4776]: I1125 11:10:02.721444 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-vw6x4" event={"ID":"7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0","Type":"ContainerStarted","Data":"7cc3f0608bf72ba439dd672f85d2188094403e534ded3e43cb74c59e6065652a"} Nov 25 11:10:05 crc kubenswrapper[4776]: I1125 11:10:05.235091 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-f4f475bb-4hkzw" Nov 25 11:10:05 crc kubenswrapper[4776]: I1125 11:10:05.235651 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-f4f475bb-4hkzw" Nov 25 11:10:14 crc kubenswrapper[4776]: I1125 11:10:14.496609 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5rplh"] Nov 25 11:10:14 crc kubenswrapper[4776]: I1125 11:10:14.499262 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5rplh" Nov 25 11:10:14 crc kubenswrapper[4776]: I1125 11:10:14.508547 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5rplh"] Nov 25 11:10:14 crc kubenswrapper[4776]: I1125 11:10:14.693320 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87b2d93f-af01-49e9-9e3e-0ac6a41f4a93-utilities\") pod \"redhat-marketplace-5rplh\" (UID: \"87b2d93f-af01-49e9-9e3e-0ac6a41f4a93\") " pod="openshift-marketplace/redhat-marketplace-5rplh" Nov 25 11:10:14 crc kubenswrapper[4776]: I1125 11:10:14.693430 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87b2d93f-af01-49e9-9e3e-0ac6a41f4a93-catalog-content\") pod \"redhat-marketplace-5rplh\" (UID: \"87b2d93f-af01-49e9-9e3e-0ac6a41f4a93\") " pod="openshift-marketplace/redhat-marketplace-5rplh" Nov 25 11:10:14 crc kubenswrapper[4776]: I1125 11:10:14.693771 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7qhb\" (UniqueName: \"kubernetes.io/projected/87b2d93f-af01-49e9-9e3e-0ac6a41f4a93-kube-api-access-j7qhb\") pod \"redhat-marketplace-5rplh\" (UID: \"87b2d93f-af01-49e9-9e3e-0ac6a41f4a93\") " pod="openshift-marketplace/redhat-marketplace-5rplh" Nov 25 11:10:14 crc kubenswrapper[4776]: I1125 11:10:14.795233 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7qhb\" (UniqueName: \"kubernetes.io/projected/87b2d93f-af01-49e9-9e3e-0ac6a41f4a93-kube-api-access-j7qhb\") pod \"redhat-marketplace-5rplh\" (UID: \"87b2d93f-af01-49e9-9e3e-0ac6a41f4a93\") " pod="openshift-marketplace/redhat-marketplace-5rplh" Nov 25 11:10:14 crc kubenswrapper[4776]: I1125 11:10:14.795385 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87b2d93f-af01-49e9-9e3e-0ac6a41f4a93-utilities\") pod \"redhat-marketplace-5rplh\" (UID: \"87b2d93f-af01-49e9-9e3e-0ac6a41f4a93\") " pod="openshift-marketplace/redhat-marketplace-5rplh" Nov 25 11:10:14 crc kubenswrapper[4776]: I1125 11:10:14.795497 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87b2d93f-af01-49e9-9e3e-0ac6a41f4a93-catalog-content\") pod \"redhat-marketplace-5rplh\" (UID: \"87b2d93f-af01-49e9-9e3e-0ac6a41f4a93\") " pod="openshift-marketplace/redhat-marketplace-5rplh" Nov 25 11:10:14 crc kubenswrapper[4776]: I1125 11:10:14.795913 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87b2d93f-af01-49e9-9e3e-0ac6a41f4a93-catalog-content\") pod \"redhat-marketplace-5rplh\" (UID: \"87b2d93f-af01-49e9-9e3e-0ac6a41f4a93\") " pod="openshift-marketplace/redhat-marketplace-5rplh" Nov 25 11:10:14 crc kubenswrapper[4776]: I1125 11:10:14.796011 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87b2d93f-af01-49e9-9e3e-0ac6a41f4a93-utilities\") pod \"redhat-marketplace-5rplh\" (UID: \"87b2d93f-af01-49e9-9e3e-0ac6a41f4a93\") " pod="openshift-marketplace/redhat-marketplace-5rplh" Nov 25 11:10:14 crc kubenswrapper[4776]: I1125 11:10:14.853557 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7qhb\" (UniqueName: \"kubernetes.io/projected/87b2d93f-af01-49e9-9e3e-0ac6a41f4a93-kube-api-access-j7qhb\") pod \"redhat-marketplace-5rplh\" (UID: \"87b2d93f-af01-49e9-9e3e-0ac6a41f4a93\") " pod="openshift-marketplace/redhat-marketplace-5rplh" Nov 25 11:10:14 crc kubenswrapper[4776]: I1125 11:10:14.858305 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5rplh" Nov 25 11:10:14 crc kubenswrapper[4776]: E1125 11:10:14.981256 4776 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-heat-engine@sha256:5e03376bd895346dc8f627ca15ded942526ed8b5e92872f453ce272e694d18d4" Nov 25 11:10:14 crc kubenswrapper[4776]: E1125 11:10:14.981726 4776 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.io/podified-antelope-centos9/openstack-heat-engine@sha256:5e03376bd895346dc8f627ca15ded942526ed8b5e92872f453ce272e694d18d4,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r5nhn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-vw6x4_openstack(7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 11:10:14 crc kubenswrapper[4776]: E1125 11:10:14.983646 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-vw6x4" podUID="7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0" Nov 25 11:10:15 crc kubenswrapper[4776]: I1125 11:10:15.237509 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-f4f475bb-4hkzw" podUID="029b9a31-76a8-446d-be2c-53760cb3980a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.128:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.128:8443: connect: connection refused" Nov 25 11:10:15 crc kubenswrapper[4776]: I1125 11:10:15.428468 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5rplh"] Nov 25 11:10:15 crc kubenswrapper[4776]: I1125 11:10:15.867649 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rplh" event={"ID":"87b2d93f-af01-49e9-9e3e-0ac6a41f4a93","Type":"ContainerStarted","Data":"16251086e1c0dc31a77107c8a53582b0eda45961c585ac85cd74bf3b94a51983"} Nov 25 11:10:15 crc kubenswrapper[4776]: I1125 11:10:15.868183 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rplh" event={"ID":"87b2d93f-af01-49e9-9e3e-0ac6a41f4a93","Type":"ContainerStarted","Data":"e5156d052383c91c1cea89eece8334c3f3a13fd6d9afa2937d49e1416f438796"} Nov 25 11:10:15 crc kubenswrapper[4776]: E1125 11:10:15.869662 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-heat-engine@sha256:5e03376bd895346dc8f627ca15ded942526ed8b5e92872f453ce272e694d18d4\\\"\"" pod="openstack/heat-db-sync-vw6x4" podUID="7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0" Nov 25 11:10:16 crc kubenswrapper[4776]: I1125 11:10:16.877885 4776 generic.go:334] "Generic (PLEG): container finished" podID="87b2d93f-af01-49e9-9e3e-0ac6a41f4a93" containerID="16251086e1c0dc31a77107c8a53582b0eda45961c585ac85cd74bf3b94a51983" exitCode=0 Nov 25 11:10:16 crc kubenswrapper[4776]: I1125 11:10:16.877983 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rplh" event={"ID":"87b2d93f-af01-49e9-9e3e-0ac6a41f4a93","Type":"ContainerDied","Data":"16251086e1c0dc31a77107c8a53582b0eda45961c585ac85cd74bf3b94a51983"} Nov 25 11:10:19 crc kubenswrapper[4776]: I1125 11:10:19.912485 4776 generic.go:334] "Generic (PLEG): container finished" podID="87b2d93f-af01-49e9-9e3e-0ac6a41f4a93" containerID="04d298534b9e544bab0caccfc6341f125b434fab7c3d19b8b54b1cc565a5cd0e" exitCode=0 Nov 25 11:10:19 crc kubenswrapper[4776]: I1125 11:10:19.912722 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rplh" event={"ID":"87b2d93f-af01-49e9-9e3e-0ac6a41f4a93","Type":"ContainerDied","Data":"04d298534b9e544bab0caccfc6341f125b434fab7c3d19b8b54b1cc565a5cd0e"} Nov 25 11:10:21 crc kubenswrapper[4776]: I1125 11:10:21.941189 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rplh" event={"ID":"87b2d93f-af01-49e9-9e3e-0ac6a41f4a93","Type":"ContainerStarted","Data":"6ddb00b9639f78b04566c42b4040bc52075b684dcb70ce8f976077e3fd34a1d6"} Nov 25 11:10:22 crc kubenswrapper[4776]: I1125 11:10:22.969971 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5rplh" podStartSLOduration=4.417347816 podStartE2EDuration="8.969953211s" podCreationTimestamp="2025-11-25 11:10:14 +0000 UTC" firstStartedPulling="2025-11-25 11:10:16.879908592 +0000 UTC m=+6361.920968145" lastFinishedPulling="2025-11-25 11:10:21.432513987 +0000 UTC m=+6366.473573540" observedRunningTime="2025-11-25 11:10:22.966872734 +0000 UTC m=+6368.007932287" watchObservedRunningTime="2025-11-25 11:10:22.969953211 +0000 UTC m=+6368.011012764" Nov 25 11:10:24 crc kubenswrapper[4776]: I1125 11:10:24.859152 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5rplh" Nov 25 11:10:24 crc kubenswrapper[4776]: I1125 11:10:24.859428 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5rplh" Nov 25 11:10:24 crc kubenswrapper[4776]: I1125 11:10:24.926257 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5rplh" Nov 25 11:10:30 crc kubenswrapper[4776]: I1125 11:10:30.239282 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-f4f475bb-4hkzw" podUID="029b9a31-76a8-446d-be2c-53760cb3980a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.128:8443/dashboard/auth/login/?next=/dashboard/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 11:10:33 crc kubenswrapper[4776]: I1125 11:10:33.050823 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-vw6x4" event={"ID":"7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0","Type":"ContainerStarted","Data":"c9adcfb517f34ad34e65da5b5067c75734c643d40cc5e3419dad41bf3caa54bd"} Nov 25 11:10:33 crc kubenswrapper[4776]: I1125 11:10:33.090278 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-vw6x4" podStartSLOduration=3.223067288 podStartE2EDuration="32.090258925s" podCreationTimestamp="2025-11-25 11:10:01 +0000 UTC" firstStartedPulling="2025-11-25 11:10:02.505311977 +0000 UTC m=+6347.546371530" lastFinishedPulling="2025-11-25 11:10:31.372503594 +0000 UTC m=+6376.413563167" observedRunningTime="2025-11-25 11:10:33.082230283 +0000 UTC m=+6378.123289846" watchObservedRunningTime="2025-11-25 11:10:33.090258925 +0000 UTC m=+6378.131318478" Nov 25 11:10:34 crc kubenswrapper[4776]: I1125 11:10:34.542936 4776 scope.go:117] "RemoveContainer" containerID="2b63f488b275331fa7a54b18739ebdbff1487a47277829a841d45990dd323cb6" Nov 25 11:10:34 crc kubenswrapper[4776]: I1125 11:10:34.915220 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5rplh" Nov 25 11:10:34 crc kubenswrapper[4776]: I1125 11:10:34.972003 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5rplh"] Nov 25 11:10:35 crc kubenswrapper[4776]: I1125 11:10:35.073739 4776 generic.go:334] "Generic (PLEG): container finished" podID="7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0" containerID="c9adcfb517f34ad34e65da5b5067c75734c643d40cc5e3419dad41bf3caa54bd" exitCode=0 Nov 25 11:10:35 crc kubenswrapper[4776]: I1125 11:10:35.073957 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5rplh" podUID="87b2d93f-af01-49e9-9e3e-0ac6a41f4a93" containerName="registry-server" containerID="cri-o://6ddb00b9639f78b04566c42b4040bc52075b684dcb70ce8f976077e3fd34a1d6" gracePeriod=2 Nov 25 11:10:35 crc kubenswrapper[4776]: I1125 11:10:35.074229 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-vw6x4" event={"ID":"7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0","Type":"ContainerDied","Data":"c9adcfb517f34ad34e65da5b5067c75734c643d40cc5e3419dad41bf3caa54bd"} Nov 25 11:10:36 crc kubenswrapper[4776]: I1125 11:10:36.084457 4776 generic.go:334] "Generic (PLEG): container finished" podID="87b2d93f-af01-49e9-9e3e-0ac6a41f4a93" containerID="6ddb00b9639f78b04566c42b4040bc52075b684dcb70ce8f976077e3fd34a1d6" exitCode=0 Nov 25 11:10:36 crc kubenswrapper[4776]: I1125 11:10:36.084553 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rplh" event={"ID":"87b2d93f-af01-49e9-9e3e-0ac6a41f4a93","Type":"ContainerDied","Data":"6ddb00b9639f78b04566c42b4040bc52075b684dcb70ce8f976077e3fd34a1d6"} Nov 25 11:10:36 crc kubenswrapper[4776]: I1125 11:10:36.906059 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-vw6x4" Nov 25 11:10:36 crc kubenswrapper[4776]: I1125 11:10:36.920248 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5rplh" Nov 25 11:10:36 crc kubenswrapper[4776]: I1125 11:10:36.975025 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-f4f475bb-4hkzw" Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.012781 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0-combined-ca-bundle\") pod \"7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0\" (UID: \"7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0\") " Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.012918 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87b2d93f-af01-49e9-9e3e-0ac6a41f4a93-catalog-content\") pod \"87b2d93f-af01-49e9-9e3e-0ac6a41f4a93\" (UID: \"87b2d93f-af01-49e9-9e3e-0ac6a41f4a93\") " Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.012960 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0-config-data\") pod \"7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0\" (UID: \"7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0\") " Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.013192 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7qhb\" (UniqueName: \"kubernetes.io/projected/87b2d93f-af01-49e9-9e3e-0ac6a41f4a93-kube-api-access-j7qhb\") pod \"87b2d93f-af01-49e9-9e3e-0ac6a41f4a93\" (UID: \"87b2d93f-af01-49e9-9e3e-0ac6a41f4a93\") " Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.013317 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5nhn\" (UniqueName: \"kubernetes.io/projected/7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0-kube-api-access-r5nhn\") pod \"7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0\" (UID: \"7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0\") " Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.013463 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87b2d93f-af01-49e9-9e3e-0ac6a41f4a93-utilities\") pod \"87b2d93f-af01-49e9-9e3e-0ac6a41f4a93\" (UID: \"87b2d93f-af01-49e9-9e3e-0ac6a41f4a93\") " Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.014253 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87b2d93f-af01-49e9-9e3e-0ac6a41f4a93-utilities" (OuterVolumeSpecName: "utilities") pod "87b2d93f-af01-49e9-9e3e-0ac6a41f4a93" (UID: "87b2d93f-af01-49e9-9e3e-0ac6a41f4a93"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.024366 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87b2d93f-af01-49e9-9e3e-0ac6a41f4a93-kube-api-access-j7qhb" (OuterVolumeSpecName: "kube-api-access-j7qhb") pod "87b2d93f-af01-49e9-9e3e-0ac6a41f4a93" (UID: "87b2d93f-af01-49e9-9e3e-0ac6a41f4a93"). InnerVolumeSpecName "kube-api-access-j7qhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.029290 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0-kube-api-access-r5nhn" (OuterVolumeSpecName: "kube-api-access-r5nhn") pod "7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0" (UID: "7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0"). InnerVolumeSpecName "kube-api-access-r5nhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.041352 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87b2d93f-af01-49e9-9e3e-0ac6a41f4a93-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "87b2d93f-af01-49e9-9e3e-0ac6a41f4a93" (UID: "87b2d93f-af01-49e9-9e3e-0ac6a41f4a93"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.050292 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0" (UID: "7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.112587 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5rplh" event={"ID":"87b2d93f-af01-49e9-9e3e-0ac6a41f4a93","Type":"ContainerDied","Data":"e5156d052383c91c1cea89eece8334c3f3a13fd6d9afa2937d49e1416f438796"} Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.113114 4776 scope.go:117] "RemoveContainer" containerID="6ddb00b9639f78b04566c42b4040bc52075b684dcb70ce8f976077e3fd34a1d6" Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.112597 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5rplh" Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.121243 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7qhb\" (UniqueName: \"kubernetes.io/projected/87b2d93f-af01-49e9-9e3e-0ac6a41f4a93-kube-api-access-j7qhb\") on node \"crc\" DevicePath \"\"" Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.121589 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5nhn\" (UniqueName: \"kubernetes.io/projected/7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0-kube-api-access-r5nhn\") on node \"crc\" DevicePath \"\"" Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.121614 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87b2d93f-af01-49e9-9e3e-0ac6a41f4a93-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.121627 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.121641 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87b2d93f-af01-49e9-9e3e-0ac6a41f4a93-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.125835 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-vw6x4" event={"ID":"7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0","Type":"ContainerDied","Data":"7cc3f0608bf72ba439dd672f85d2188094403e534ded3e43cb74c59e6065652a"} Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.125877 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7cc3f0608bf72ba439dd672f85d2188094403e534ded3e43cb74c59e6065652a" Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.125952 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-vw6x4" Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.149519 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0-config-data" (OuterVolumeSpecName: "config-data") pod "7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0" (UID: "7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.226530 4776 scope.go:117] "RemoveContainer" containerID="04d298534b9e544bab0caccfc6341f125b434fab7c3d19b8b54b1cc565a5cd0e" Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.227616 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.235313 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5rplh"] Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.245782 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5rplh"] Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.262881 4776 scope.go:117] "RemoveContainer" containerID="16251086e1c0dc31a77107c8a53582b0eda45961c585ac85cd74bf3b94a51983" Nov 25 11:10:37 crc kubenswrapper[4776]: I1125 11:10:37.683701 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87b2d93f-af01-49e9-9e3e-0ac6a41f4a93" path="/var/lib/kubelet/pods/87b2d93f-af01-49e9-9e3e-0ac6a41f4a93/volumes" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.520415 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-5dc4c495b-tmsg5"] Nov 25 11:10:38 crc kubenswrapper[4776]: E1125 11:10:38.520917 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0" containerName="heat-db-sync" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.520935 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0" containerName="heat-db-sync" Nov 25 11:10:38 crc kubenswrapper[4776]: E1125 11:10:38.520964 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87b2d93f-af01-49e9-9e3e-0ac6a41f4a93" containerName="extract-utilities" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.520973 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="87b2d93f-af01-49e9-9e3e-0ac6a41f4a93" containerName="extract-utilities" Nov 25 11:10:38 crc kubenswrapper[4776]: E1125 11:10:38.520998 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87b2d93f-af01-49e9-9e3e-0ac6a41f4a93" containerName="extract-content" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.521007 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="87b2d93f-af01-49e9-9e3e-0ac6a41f4a93" containerName="extract-content" Nov 25 11:10:38 crc kubenswrapper[4776]: E1125 11:10:38.521034 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87b2d93f-af01-49e9-9e3e-0ac6a41f4a93" containerName="registry-server" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.521043 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="87b2d93f-af01-49e9-9e3e-0ac6a41f4a93" containerName="registry-server" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.521309 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0" containerName="heat-db-sync" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.521333 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="87b2d93f-af01-49e9-9e3e-0ac6a41f4a93" containerName="registry-server" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.522165 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5dc4c495b-tmsg5" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.526486 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.526531 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.530842 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-9lzlh" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.531170 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-5dc4c495b-tmsg5"] Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.673846 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/352f32f4-36a2-4a3c-858d-52b176e296b1-combined-ca-bundle\") pod \"heat-engine-5dc4c495b-tmsg5\" (UID: \"352f32f4-36a2-4a3c-858d-52b176e296b1\") " pod="openstack/heat-engine-5dc4c495b-tmsg5" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.674203 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/352f32f4-36a2-4a3c-858d-52b176e296b1-config-data-custom\") pod \"heat-engine-5dc4c495b-tmsg5\" (UID: \"352f32f4-36a2-4a3c-858d-52b176e296b1\") " pod="openstack/heat-engine-5dc4c495b-tmsg5" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.674328 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/352f32f4-36a2-4a3c-858d-52b176e296b1-config-data\") pod \"heat-engine-5dc4c495b-tmsg5\" (UID: \"352f32f4-36a2-4a3c-858d-52b176e296b1\") " pod="openstack/heat-engine-5dc4c495b-tmsg5" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.674561 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r962s\" (UniqueName: \"kubernetes.io/projected/352f32f4-36a2-4a3c-858d-52b176e296b1-kube-api-access-r962s\") pod \"heat-engine-5dc4c495b-tmsg5\" (UID: \"352f32f4-36a2-4a3c-858d-52b176e296b1\") " pod="openstack/heat-engine-5dc4c495b-tmsg5" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.717574 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-f4f475bb-4hkzw" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.778737 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/352f32f4-36a2-4a3c-858d-52b176e296b1-config-data\") pod \"heat-engine-5dc4c495b-tmsg5\" (UID: \"352f32f4-36a2-4a3c-858d-52b176e296b1\") " pod="openstack/heat-engine-5dc4c495b-tmsg5" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.778917 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r962s\" (UniqueName: \"kubernetes.io/projected/352f32f4-36a2-4a3c-858d-52b176e296b1-kube-api-access-r962s\") pod \"heat-engine-5dc4c495b-tmsg5\" (UID: \"352f32f4-36a2-4a3c-858d-52b176e296b1\") " pod="openstack/heat-engine-5dc4c495b-tmsg5" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.779060 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/352f32f4-36a2-4a3c-858d-52b176e296b1-combined-ca-bundle\") pod \"heat-engine-5dc4c495b-tmsg5\" (UID: \"352f32f4-36a2-4a3c-858d-52b176e296b1\") " pod="openstack/heat-engine-5dc4c495b-tmsg5" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.779114 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/352f32f4-36a2-4a3c-858d-52b176e296b1-config-data-custom\") pod \"heat-engine-5dc4c495b-tmsg5\" (UID: \"352f32f4-36a2-4a3c-858d-52b176e296b1\") " pod="openstack/heat-engine-5dc4c495b-tmsg5" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.788542 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-656cc574-dstj4"] Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.788836 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-656cc574-dstj4" podUID="219a7347-5102-4e6c-9b38-deea6275f715" containerName="horizon-log" containerID="cri-o://75624a8110711b2f1d37d5065d9a5ab2035226b6e3ac4acf037118c35cb7d28c" gracePeriod=30 Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.789319 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-656cc574-dstj4" podUID="219a7347-5102-4e6c-9b38-deea6275f715" containerName="horizon" containerID="cri-o://f5dd0b1f4ea70853679a58ed439e71548cd36135ef8443036178352353139632" gracePeriod=30 Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.808890 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/352f32f4-36a2-4a3c-858d-52b176e296b1-config-data\") pod \"heat-engine-5dc4c495b-tmsg5\" (UID: \"352f32f4-36a2-4a3c-858d-52b176e296b1\") " pod="openstack/heat-engine-5dc4c495b-tmsg5" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.812086 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/352f32f4-36a2-4a3c-858d-52b176e296b1-config-data-custom\") pod \"heat-engine-5dc4c495b-tmsg5\" (UID: \"352f32f4-36a2-4a3c-858d-52b176e296b1\") " pod="openstack/heat-engine-5dc4c495b-tmsg5" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.814550 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r962s\" (UniqueName: \"kubernetes.io/projected/352f32f4-36a2-4a3c-858d-52b176e296b1-kube-api-access-r962s\") pod \"heat-engine-5dc4c495b-tmsg5\" (UID: \"352f32f4-36a2-4a3c-858d-52b176e296b1\") " pod="openstack/heat-engine-5dc4c495b-tmsg5" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.818535 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/352f32f4-36a2-4a3c-858d-52b176e296b1-combined-ca-bundle\") pod \"heat-engine-5dc4c495b-tmsg5\" (UID: \"352f32f4-36a2-4a3c-858d-52b176e296b1\") " pod="openstack/heat-engine-5dc4c495b-tmsg5" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.844568 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5dc4c495b-tmsg5" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.901438 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-648b897fb8-xj9jc"] Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.907573 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-648b897fb8-xj9jc" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.920462 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.941255 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-6b88954b6d-fm4rs"] Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.942623 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6b88954b6d-fm4rs" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.945680 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.946431 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-648b897fb8-xj9jc"] Nov 25 11:10:38 crc kubenswrapper[4776]: I1125 11:10:38.973545 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6b88954b6d-fm4rs"] Nov 25 11:10:39 crc kubenswrapper[4776]: I1125 11:10:39.101124 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a0ee8963-fd22-4acf-a2dc-871aa6cd3c54-config-data-custom\") pod \"heat-cfnapi-648b897fb8-xj9jc\" (UID: \"a0ee8963-fd22-4acf-a2dc-871aa6cd3c54\") " pod="openstack/heat-cfnapi-648b897fb8-xj9jc" Nov 25 11:10:39 crc kubenswrapper[4776]: I1125 11:10:39.101532 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7vqw\" (UniqueName: \"kubernetes.io/projected/e031e3dc-4e68-4b61-bf4d-4eb0246f6643-kube-api-access-x7vqw\") pod \"heat-api-6b88954b6d-fm4rs\" (UID: \"e031e3dc-4e68-4b61-bf4d-4eb0246f6643\") " pod="openstack/heat-api-6b88954b6d-fm4rs" Nov 25 11:10:39 crc kubenswrapper[4776]: I1125 11:10:39.101910 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e031e3dc-4e68-4b61-bf4d-4eb0246f6643-combined-ca-bundle\") pod \"heat-api-6b88954b6d-fm4rs\" (UID: \"e031e3dc-4e68-4b61-bf4d-4eb0246f6643\") " pod="openstack/heat-api-6b88954b6d-fm4rs" Nov 25 11:10:39 crc kubenswrapper[4776]: I1125 11:10:39.102078 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ee8963-fd22-4acf-a2dc-871aa6cd3c54-combined-ca-bundle\") pod \"heat-cfnapi-648b897fb8-xj9jc\" (UID: \"a0ee8963-fd22-4acf-a2dc-871aa6cd3c54\") " pod="openstack/heat-cfnapi-648b897fb8-xj9jc" Nov 25 11:10:39 crc kubenswrapper[4776]: I1125 11:10:39.102128 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e031e3dc-4e68-4b61-bf4d-4eb0246f6643-config-data-custom\") pod \"heat-api-6b88954b6d-fm4rs\" (UID: \"e031e3dc-4e68-4b61-bf4d-4eb0246f6643\") " pod="openstack/heat-api-6b88954b6d-fm4rs" Nov 25 11:10:39 crc kubenswrapper[4776]: I1125 11:10:39.102171 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e031e3dc-4e68-4b61-bf4d-4eb0246f6643-config-data\") pod \"heat-api-6b88954b6d-fm4rs\" (UID: \"e031e3dc-4e68-4b61-bf4d-4eb0246f6643\") " pod="openstack/heat-api-6b88954b6d-fm4rs" Nov 25 11:10:39 crc kubenswrapper[4776]: I1125 11:10:39.102195 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0ee8963-fd22-4acf-a2dc-871aa6cd3c54-config-data\") pod \"heat-cfnapi-648b897fb8-xj9jc\" (UID: \"a0ee8963-fd22-4acf-a2dc-871aa6cd3c54\") " pod="openstack/heat-cfnapi-648b897fb8-xj9jc" Nov 25 11:10:39 crc kubenswrapper[4776]: I1125 11:10:39.102221 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7vml\" (UniqueName: \"kubernetes.io/projected/a0ee8963-fd22-4acf-a2dc-871aa6cd3c54-kube-api-access-w7vml\") pod \"heat-cfnapi-648b897fb8-xj9jc\" (UID: \"a0ee8963-fd22-4acf-a2dc-871aa6cd3c54\") " pod="openstack/heat-cfnapi-648b897fb8-xj9jc" Nov 25 11:10:39 crc kubenswrapper[4776]: I1125 11:10:39.203696 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a0ee8963-fd22-4acf-a2dc-871aa6cd3c54-config-data-custom\") pod \"heat-cfnapi-648b897fb8-xj9jc\" (UID: \"a0ee8963-fd22-4acf-a2dc-871aa6cd3c54\") " pod="openstack/heat-cfnapi-648b897fb8-xj9jc" Nov 25 11:10:39 crc kubenswrapper[4776]: I1125 11:10:39.203761 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7vqw\" (UniqueName: \"kubernetes.io/projected/e031e3dc-4e68-4b61-bf4d-4eb0246f6643-kube-api-access-x7vqw\") pod \"heat-api-6b88954b6d-fm4rs\" (UID: \"e031e3dc-4e68-4b61-bf4d-4eb0246f6643\") " pod="openstack/heat-api-6b88954b6d-fm4rs" Nov 25 11:10:39 crc kubenswrapper[4776]: I1125 11:10:39.203853 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e031e3dc-4e68-4b61-bf4d-4eb0246f6643-combined-ca-bundle\") pod \"heat-api-6b88954b6d-fm4rs\" (UID: \"e031e3dc-4e68-4b61-bf4d-4eb0246f6643\") " pod="openstack/heat-api-6b88954b6d-fm4rs" Nov 25 11:10:39 crc kubenswrapper[4776]: I1125 11:10:39.203893 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ee8963-fd22-4acf-a2dc-871aa6cd3c54-combined-ca-bundle\") pod \"heat-cfnapi-648b897fb8-xj9jc\" (UID: \"a0ee8963-fd22-4acf-a2dc-871aa6cd3c54\") " pod="openstack/heat-cfnapi-648b897fb8-xj9jc" Nov 25 11:10:39 crc kubenswrapper[4776]: I1125 11:10:39.203914 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e031e3dc-4e68-4b61-bf4d-4eb0246f6643-config-data-custom\") pod \"heat-api-6b88954b6d-fm4rs\" (UID: \"e031e3dc-4e68-4b61-bf4d-4eb0246f6643\") " pod="openstack/heat-api-6b88954b6d-fm4rs" Nov 25 11:10:39 crc kubenswrapper[4776]: I1125 11:10:39.203940 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e031e3dc-4e68-4b61-bf4d-4eb0246f6643-config-data\") pod \"heat-api-6b88954b6d-fm4rs\" (UID: \"e031e3dc-4e68-4b61-bf4d-4eb0246f6643\") " pod="openstack/heat-api-6b88954b6d-fm4rs" Nov 25 11:10:39 crc kubenswrapper[4776]: I1125 11:10:39.203959 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0ee8963-fd22-4acf-a2dc-871aa6cd3c54-config-data\") pod \"heat-cfnapi-648b897fb8-xj9jc\" (UID: \"a0ee8963-fd22-4acf-a2dc-871aa6cd3c54\") " pod="openstack/heat-cfnapi-648b897fb8-xj9jc" Nov 25 11:10:39 crc kubenswrapper[4776]: I1125 11:10:39.203977 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7vml\" (UniqueName: \"kubernetes.io/projected/a0ee8963-fd22-4acf-a2dc-871aa6cd3c54-kube-api-access-w7vml\") pod \"heat-cfnapi-648b897fb8-xj9jc\" (UID: \"a0ee8963-fd22-4acf-a2dc-871aa6cd3c54\") " pod="openstack/heat-cfnapi-648b897fb8-xj9jc" Nov 25 11:10:39 crc kubenswrapper[4776]: I1125 11:10:39.218061 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0ee8963-fd22-4acf-a2dc-871aa6cd3c54-config-data\") pod \"heat-cfnapi-648b897fb8-xj9jc\" (UID: \"a0ee8963-fd22-4acf-a2dc-871aa6cd3c54\") " pod="openstack/heat-cfnapi-648b897fb8-xj9jc" Nov 25 11:10:39 crc kubenswrapper[4776]: I1125 11:10:39.220216 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e031e3dc-4e68-4b61-bf4d-4eb0246f6643-config-data\") pod \"heat-api-6b88954b6d-fm4rs\" (UID: \"e031e3dc-4e68-4b61-bf4d-4eb0246f6643\") " pod="openstack/heat-api-6b88954b6d-fm4rs" Nov 25 11:10:39 crc kubenswrapper[4776]: I1125 11:10:39.221551 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a0ee8963-fd22-4acf-a2dc-871aa6cd3c54-config-data-custom\") pod \"heat-cfnapi-648b897fb8-xj9jc\" (UID: \"a0ee8963-fd22-4acf-a2dc-871aa6cd3c54\") " pod="openstack/heat-cfnapi-648b897fb8-xj9jc" Nov 25 11:10:39 crc kubenswrapper[4776]: I1125 11:10:39.221673 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ee8963-fd22-4acf-a2dc-871aa6cd3c54-combined-ca-bundle\") pod \"heat-cfnapi-648b897fb8-xj9jc\" (UID: \"a0ee8963-fd22-4acf-a2dc-871aa6cd3c54\") " pod="openstack/heat-cfnapi-648b897fb8-xj9jc" Nov 25 11:10:39 crc kubenswrapper[4776]: I1125 11:10:39.223939 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e031e3dc-4e68-4b61-bf4d-4eb0246f6643-config-data-custom\") pod \"heat-api-6b88954b6d-fm4rs\" (UID: \"e031e3dc-4e68-4b61-bf4d-4eb0246f6643\") " pod="openstack/heat-api-6b88954b6d-fm4rs" Nov 25 11:10:39 crc kubenswrapper[4776]: I1125 11:10:39.224100 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e031e3dc-4e68-4b61-bf4d-4eb0246f6643-combined-ca-bundle\") pod \"heat-api-6b88954b6d-fm4rs\" (UID: \"e031e3dc-4e68-4b61-bf4d-4eb0246f6643\") " pod="openstack/heat-api-6b88954b6d-fm4rs" Nov 25 11:10:39 crc kubenswrapper[4776]: I1125 11:10:39.232007 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7vml\" (UniqueName: \"kubernetes.io/projected/a0ee8963-fd22-4acf-a2dc-871aa6cd3c54-kube-api-access-w7vml\") pod \"heat-cfnapi-648b897fb8-xj9jc\" (UID: \"a0ee8963-fd22-4acf-a2dc-871aa6cd3c54\") " pod="openstack/heat-cfnapi-648b897fb8-xj9jc" Nov 25 11:10:39 crc kubenswrapper[4776]: I1125 11:10:39.232287 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7vqw\" (UniqueName: \"kubernetes.io/projected/e031e3dc-4e68-4b61-bf4d-4eb0246f6643-kube-api-access-x7vqw\") pod \"heat-api-6b88954b6d-fm4rs\" (UID: \"e031e3dc-4e68-4b61-bf4d-4eb0246f6643\") " pod="openstack/heat-api-6b88954b6d-fm4rs" Nov 25 11:10:39 crc kubenswrapper[4776]: I1125 11:10:39.253896 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-648b897fb8-xj9jc" Nov 25 11:10:39 crc kubenswrapper[4776]: I1125 11:10:39.269176 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6b88954b6d-fm4rs" Nov 25 11:10:39 crc kubenswrapper[4776]: I1125 11:10:39.963945 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-5dc4c495b-tmsg5"] Nov 25 11:10:40 crc kubenswrapper[4776]: I1125 11:10:40.065953 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6b88954b6d-fm4rs"] Nov 25 11:10:40 crc kubenswrapper[4776]: W1125 11:10:40.081374 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode031e3dc_4e68_4b61_bf4d_4eb0246f6643.slice/crio-d696c6ba72bdcd014bcc1db07c7df4404295185df7116c8723fba6ff71c23fc4 WatchSource:0}: Error finding container d696c6ba72bdcd014bcc1db07c7df4404295185df7116c8723fba6ff71c23fc4: Status 404 returned error can't find the container with id d696c6ba72bdcd014bcc1db07c7df4404295185df7116c8723fba6ff71c23fc4 Nov 25 11:10:40 crc kubenswrapper[4776]: W1125 11:10:40.085842 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0ee8963_fd22_4acf_a2dc_871aa6cd3c54.slice/crio-587e80ad69607a39eb917c6ffca1dcc9a2d49b64728890777d07c3d6cddbaea1 WatchSource:0}: Error finding container 587e80ad69607a39eb917c6ffca1dcc9a2d49b64728890777d07c3d6cddbaea1: Status 404 returned error can't find the container with id 587e80ad69607a39eb917c6ffca1dcc9a2d49b64728890777d07c3d6cddbaea1 Nov 25 11:10:40 crc kubenswrapper[4776]: I1125 11:10:40.088981 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-648b897fb8-xj9jc"] Nov 25 11:10:40 crc kubenswrapper[4776]: I1125 11:10:40.158845 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-648b897fb8-xj9jc" event={"ID":"a0ee8963-fd22-4acf-a2dc-871aa6cd3c54","Type":"ContainerStarted","Data":"587e80ad69607a39eb917c6ffca1dcc9a2d49b64728890777d07c3d6cddbaea1"} Nov 25 11:10:40 crc kubenswrapper[4776]: I1125 11:10:40.159946 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6b88954b6d-fm4rs" event={"ID":"e031e3dc-4e68-4b61-bf4d-4eb0246f6643","Type":"ContainerStarted","Data":"d696c6ba72bdcd014bcc1db07c7df4404295185df7116c8723fba6ff71c23fc4"} Nov 25 11:10:40 crc kubenswrapper[4776]: I1125 11:10:40.162571 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5dc4c495b-tmsg5" event={"ID":"352f32f4-36a2-4a3c-858d-52b176e296b1","Type":"ContainerStarted","Data":"ed0dacf50cae3a12b3469bfbbbe9ea65125d54885f0235307b07ad8e0ccdfba0"} Nov 25 11:10:41 crc kubenswrapper[4776]: I1125 11:10:41.181492 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5dc4c495b-tmsg5" event={"ID":"352f32f4-36a2-4a3c-858d-52b176e296b1","Type":"ContainerStarted","Data":"57929ea0709123456be6f27638c6b9e2666c84483d585344a7d644ec936d1f55"} Nov 25 11:10:41 crc kubenswrapper[4776]: I1125 11:10:41.182233 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-5dc4c495b-tmsg5" Nov 25 11:10:41 crc kubenswrapper[4776]: I1125 11:10:41.205995 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-5dc4c495b-tmsg5" podStartSLOduration=3.205971301 podStartE2EDuration="3.205971301s" podCreationTimestamp="2025-11-25 11:10:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:10:41.198974815 +0000 UTC m=+6386.240034358" watchObservedRunningTime="2025-11-25 11:10:41.205971301 +0000 UTC m=+6386.247030854" Nov 25 11:10:42 crc kubenswrapper[4776]: I1125 11:10:42.065021 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-656cc574-dstj4" podUID="219a7347-5102-4e6c-9b38-deea6275f715" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.125:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:35360->10.217.1.125:8443: read: connection reset by peer" Nov 25 11:10:42 crc kubenswrapper[4776]: I1125 11:10:42.197262 4776 generic.go:334] "Generic (PLEG): container finished" podID="219a7347-5102-4e6c-9b38-deea6275f715" containerID="f5dd0b1f4ea70853679a58ed439e71548cd36135ef8443036178352353139632" exitCode=0 Nov 25 11:10:42 crc kubenswrapper[4776]: I1125 11:10:42.198286 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-656cc574-dstj4" event={"ID":"219a7347-5102-4e6c-9b38-deea6275f715","Type":"ContainerDied","Data":"f5dd0b1f4ea70853679a58ed439e71548cd36135ef8443036178352353139632"} Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.442804 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-5954f4557-4hz28"] Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.450167 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-5954f4557-4hz28"] Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.451023 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5954f4557-4hz28" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.547130 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-6f8d5975f9-zsrqs"] Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.548890 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6f8d5975f9-zsrqs" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.560521 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-69c47bdc88-5ngxw"] Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.564532 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-69c47bdc88-5ngxw" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.570805 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4291e940-29c4-4c08-98ea-227e96cd687f-combined-ca-bundle\") pod \"heat-engine-5954f4557-4hz28\" (UID: \"4291e940-29c4-4c08-98ea-227e96cd687f\") " pod="openstack/heat-engine-5954f4557-4hz28" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.571149 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4291e940-29c4-4c08-98ea-227e96cd687f-config-data-custom\") pod \"heat-engine-5954f4557-4hz28\" (UID: \"4291e940-29c4-4c08-98ea-227e96cd687f\") " pod="openstack/heat-engine-5954f4557-4hz28" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.571299 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4g9z\" (UniqueName: \"kubernetes.io/projected/4291e940-29c4-4c08-98ea-227e96cd687f-kube-api-access-n4g9z\") pod \"heat-engine-5954f4557-4hz28\" (UID: \"4291e940-29c4-4c08-98ea-227e96cd687f\") " pod="openstack/heat-engine-5954f4557-4hz28" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.571567 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4291e940-29c4-4c08-98ea-227e96cd687f-config-data\") pod \"heat-engine-5954f4557-4hz28\" (UID: \"4291e940-29c4-4c08-98ea-227e96cd687f\") " pod="openstack/heat-engine-5954f4557-4hz28" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.575433 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-6f8d5975f9-zsrqs"] Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.589109 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-69c47bdc88-5ngxw"] Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.673319 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4g9z\" (UniqueName: \"kubernetes.io/projected/4291e940-29c4-4c08-98ea-227e96cd687f-kube-api-access-n4g9z\") pod \"heat-engine-5954f4557-4hz28\" (UID: \"4291e940-29c4-4c08-98ea-227e96cd687f\") " pod="openstack/heat-engine-5954f4557-4hz28" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.673429 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a989b8a6-3b6b-4841-9f5b-0e23cfac4cba-config-data-custom\") pod \"heat-api-69c47bdc88-5ngxw\" (UID: \"a989b8a6-3b6b-4841-9f5b-0e23cfac4cba\") " pod="openstack/heat-api-69c47bdc88-5ngxw" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.673466 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2eb1d65a-b5ca-4278-837d-cb255d764840-config-data-custom\") pod \"heat-cfnapi-6f8d5975f9-zsrqs\" (UID: \"2eb1d65a-b5ca-4278-837d-cb255d764840\") " pod="openstack/heat-cfnapi-6f8d5975f9-zsrqs" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.673513 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a989b8a6-3b6b-4841-9f5b-0e23cfac4cba-combined-ca-bundle\") pod \"heat-api-69c47bdc88-5ngxw\" (UID: \"a989b8a6-3b6b-4841-9f5b-0e23cfac4cba\") " pod="openstack/heat-api-69c47bdc88-5ngxw" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.673539 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a989b8a6-3b6b-4841-9f5b-0e23cfac4cba-config-data\") pod \"heat-api-69c47bdc88-5ngxw\" (UID: \"a989b8a6-3b6b-4841-9f5b-0e23cfac4cba\") " pod="openstack/heat-api-69c47bdc88-5ngxw" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.673621 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eb1d65a-b5ca-4278-837d-cb255d764840-combined-ca-bundle\") pod \"heat-cfnapi-6f8d5975f9-zsrqs\" (UID: \"2eb1d65a-b5ca-4278-837d-cb255d764840\") " pod="openstack/heat-cfnapi-6f8d5975f9-zsrqs" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.673695 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4291e940-29c4-4c08-98ea-227e96cd687f-config-data\") pod \"heat-engine-5954f4557-4hz28\" (UID: \"4291e940-29c4-4c08-98ea-227e96cd687f\") " pod="openstack/heat-engine-5954f4557-4hz28" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.673731 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zthn\" (UniqueName: \"kubernetes.io/projected/2eb1d65a-b5ca-4278-837d-cb255d764840-kube-api-access-4zthn\") pod \"heat-cfnapi-6f8d5975f9-zsrqs\" (UID: \"2eb1d65a-b5ca-4278-837d-cb255d764840\") " pod="openstack/heat-cfnapi-6f8d5975f9-zsrqs" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.673782 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2eb1d65a-b5ca-4278-837d-cb255d764840-config-data\") pod \"heat-cfnapi-6f8d5975f9-zsrqs\" (UID: \"2eb1d65a-b5ca-4278-837d-cb255d764840\") " pod="openstack/heat-cfnapi-6f8d5975f9-zsrqs" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.673808 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wv72\" (UniqueName: \"kubernetes.io/projected/a989b8a6-3b6b-4841-9f5b-0e23cfac4cba-kube-api-access-7wv72\") pod \"heat-api-69c47bdc88-5ngxw\" (UID: \"a989b8a6-3b6b-4841-9f5b-0e23cfac4cba\") " pod="openstack/heat-api-69c47bdc88-5ngxw" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.673844 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4291e940-29c4-4c08-98ea-227e96cd687f-combined-ca-bundle\") pod \"heat-engine-5954f4557-4hz28\" (UID: \"4291e940-29c4-4c08-98ea-227e96cd687f\") " pod="openstack/heat-engine-5954f4557-4hz28" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.673914 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4291e940-29c4-4c08-98ea-227e96cd687f-config-data-custom\") pod \"heat-engine-5954f4557-4hz28\" (UID: \"4291e940-29c4-4c08-98ea-227e96cd687f\") " pod="openstack/heat-engine-5954f4557-4hz28" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.679390 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4291e940-29c4-4c08-98ea-227e96cd687f-combined-ca-bundle\") pod \"heat-engine-5954f4557-4hz28\" (UID: \"4291e940-29c4-4c08-98ea-227e96cd687f\") " pod="openstack/heat-engine-5954f4557-4hz28" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.679858 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4291e940-29c4-4c08-98ea-227e96cd687f-config-data-custom\") pod \"heat-engine-5954f4557-4hz28\" (UID: \"4291e940-29c4-4c08-98ea-227e96cd687f\") " pod="openstack/heat-engine-5954f4557-4hz28" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.686311 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4291e940-29c4-4c08-98ea-227e96cd687f-config-data\") pod \"heat-engine-5954f4557-4hz28\" (UID: \"4291e940-29c4-4c08-98ea-227e96cd687f\") " pod="openstack/heat-engine-5954f4557-4hz28" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.693285 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4g9z\" (UniqueName: \"kubernetes.io/projected/4291e940-29c4-4c08-98ea-227e96cd687f-kube-api-access-n4g9z\") pod \"heat-engine-5954f4557-4hz28\" (UID: \"4291e940-29c4-4c08-98ea-227e96cd687f\") " pod="openstack/heat-engine-5954f4557-4hz28" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.778524 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a989b8a6-3b6b-4841-9f5b-0e23cfac4cba-config-data-custom\") pod \"heat-api-69c47bdc88-5ngxw\" (UID: \"a989b8a6-3b6b-4841-9f5b-0e23cfac4cba\") " pod="openstack/heat-api-69c47bdc88-5ngxw" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.778586 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2eb1d65a-b5ca-4278-837d-cb255d764840-config-data-custom\") pod \"heat-cfnapi-6f8d5975f9-zsrqs\" (UID: \"2eb1d65a-b5ca-4278-837d-cb255d764840\") " pod="openstack/heat-cfnapi-6f8d5975f9-zsrqs" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.778633 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a989b8a6-3b6b-4841-9f5b-0e23cfac4cba-combined-ca-bundle\") pod \"heat-api-69c47bdc88-5ngxw\" (UID: \"a989b8a6-3b6b-4841-9f5b-0e23cfac4cba\") " pod="openstack/heat-api-69c47bdc88-5ngxw" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.778674 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a989b8a6-3b6b-4841-9f5b-0e23cfac4cba-config-data\") pod \"heat-api-69c47bdc88-5ngxw\" (UID: \"a989b8a6-3b6b-4841-9f5b-0e23cfac4cba\") " pod="openstack/heat-api-69c47bdc88-5ngxw" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.778770 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eb1d65a-b5ca-4278-837d-cb255d764840-combined-ca-bundle\") pod \"heat-cfnapi-6f8d5975f9-zsrqs\" (UID: \"2eb1d65a-b5ca-4278-837d-cb255d764840\") " pod="openstack/heat-cfnapi-6f8d5975f9-zsrqs" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.778845 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zthn\" (UniqueName: \"kubernetes.io/projected/2eb1d65a-b5ca-4278-837d-cb255d764840-kube-api-access-4zthn\") pod \"heat-cfnapi-6f8d5975f9-zsrqs\" (UID: \"2eb1d65a-b5ca-4278-837d-cb255d764840\") " pod="openstack/heat-cfnapi-6f8d5975f9-zsrqs" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.778869 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2eb1d65a-b5ca-4278-837d-cb255d764840-config-data\") pod \"heat-cfnapi-6f8d5975f9-zsrqs\" (UID: \"2eb1d65a-b5ca-4278-837d-cb255d764840\") " pod="openstack/heat-cfnapi-6f8d5975f9-zsrqs" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.778887 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wv72\" (UniqueName: \"kubernetes.io/projected/a989b8a6-3b6b-4841-9f5b-0e23cfac4cba-kube-api-access-7wv72\") pod \"heat-api-69c47bdc88-5ngxw\" (UID: \"a989b8a6-3b6b-4841-9f5b-0e23cfac4cba\") " pod="openstack/heat-api-69c47bdc88-5ngxw" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.783373 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2eb1d65a-b5ca-4278-837d-cb255d764840-config-data-custom\") pod \"heat-cfnapi-6f8d5975f9-zsrqs\" (UID: \"2eb1d65a-b5ca-4278-837d-cb255d764840\") " pod="openstack/heat-cfnapi-6f8d5975f9-zsrqs" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.784050 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a989b8a6-3b6b-4841-9f5b-0e23cfac4cba-combined-ca-bundle\") pod \"heat-api-69c47bdc88-5ngxw\" (UID: \"a989b8a6-3b6b-4841-9f5b-0e23cfac4cba\") " pod="openstack/heat-api-69c47bdc88-5ngxw" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.784388 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eb1d65a-b5ca-4278-837d-cb255d764840-combined-ca-bundle\") pod \"heat-cfnapi-6f8d5975f9-zsrqs\" (UID: \"2eb1d65a-b5ca-4278-837d-cb255d764840\") " pod="openstack/heat-cfnapi-6f8d5975f9-zsrqs" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.784841 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a989b8a6-3b6b-4841-9f5b-0e23cfac4cba-config-data\") pod \"heat-api-69c47bdc88-5ngxw\" (UID: \"a989b8a6-3b6b-4841-9f5b-0e23cfac4cba\") " pod="openstack/heat-api-69c47bdc88-5ngxw" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.789047 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a989b8a6-3b6b-4841-9f5b-0e23cfac4cba-config-data-custom\") pod \"heat-api-69c47bdc88-5ngxw\" (UID: \"a989b8a6-3b6b-4841-9f5b-0e23cfac4cba\") " pod="openstack/heat-api-69c47bdc88-5ngxw" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.797302 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2eb1d65a-b5ca-4278-837d-cb255d764840-config-data\") pod \"heat-cfnapi-6f8d5975f9-zsrqs\" (UID: \"2eb1d65a-b5ca-4278-837d-cb255d764840\") " pod="openstack/heat-cfnapi-6f8d5975f9-zsrqs" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.799171 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wv72\" (UniqueName: \"kubernetes.io/projected/a989b8a6-3b6b-4841-9f5b-0e23cfac4cba-kube-api-access-7wv72\") pod \"heat-api-69c47bdc88-5ngxw\" (UID: \"a989b8a6-3b6b-4841-9f5b-0e23cfac4cba\") " pod="openstack/heat-api-69c47bdc88-5ngxw" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.799926 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zthn\" (UniqueName: \"kubernetes.io/projected/2eb1d65a-b5ca-4278-837d-cb255d764840-kube-api-access-4zthn\") pod \"heat-cfnapi-6f8d5975f9-zsrqs\" (UID: \"2eb1d65a-b5ca-4278-837d-cb255d764840\") " pod="openstack/heat-cfnapi-6f8d5975f9-zsrqs" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.813755 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5954f4557-4hz28" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.864498 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6f8d5975f9-zsrqs" Nov 25 11:10:47 crc kubenswrapper[4776]: I1125 11:10:47.887996 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-69c47bdc88-5ngxw" Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.257415 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6b88954b6d-fm4rs" event={"ID":"e031e3dc-4e68-4b61-bf4d-4eb0246f6643","Type":"ContainerStarted","Data":"d4103054022308b045fb8118354421f420b58123187f375df80ea7e18c2f6e41"} Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.257990 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-6b88954b6d-fm4rs" Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.259368 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-648b897fb8-xj9jc" event={"ID":"a0ee8963-fd22-4acf-a2dc-871aa6cd3c54","Type":"ContainerStarted","Data":"d1a943b93761e2e6dc282c1823b575aed26148690c49f9e18e0128f12abe5533"} Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.259559 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-648b897fb8-xj9jc" Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.294879 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-6b88954b6d-fm4rs" podStartSLOduration=4.014085022 podStartE2EDuration="10.294857528s" podCreationTimestamp="2025-11-25 11:10:38 +0000 UTC" firstStartedPulling="2025-11-25 11:10:40.084181338 +0000 UTC m=+6385.125240891" lastFinishedPulling="2025-11-25 11:10:46.364953844 +0000 UTC m=+6391.406013397" observedRunningTime="2025-11-25 11:10:48.275295517 +0000 UTC m=+6393.316355080" watchObservedRunningTime="2025-11-25 11:10:48.294857528 +0000 UTC m=+6393.335917081" Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.315466 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-648b897fb8-xj9jc" podStartSLOduration=3.437010929 podStartE2EDuration="10.315442395s" podCreationTimestamp="2025-11-25 11:10:38 +0000 UTC" firstStartedPulling="2025-11-25 11:10:40.091918852 +0000 UTC m=+6385.132978395" lastFinishedPulling="2025-11-25 11:10:46.970350308 +0000 UTC m=+6392.011409861" observedRunningTime="2025-11-25 11:10:48.295495774 +0000 UTC m=+6393.336555327" watchObservedRunningTime="2025-11-25 11:10:48.315442395 +0000 UTC m=+6393.356501948" Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.483380 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-69c47bdc88-5ngxw"] Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.494260 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-5954f4557-4hz28"] Nov 25 11:10:48 crc kubenswrapper[4776]: W1125 11:10:48.494583 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4291e940_29c4_4c08_98ea_227e96cd687f.slice/crio-ecc192534343326d970067c0fdfb8cd37a8a140989d7bac50e2a5173979ef9de WatchSource:0}: Error finding container ecc192534343326d970067c0fdfb8cd37a8a140989d7bac50e2a5173979ef9de: Status 404 returned error can't find the container with id ecc192534343326d970067c0fdfb8cd37a8a140989d7bac50e2a5173979ef9de Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.623927 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-6f8d5975f9-zsrqs"] Nov 25 11:10:48 crc kubenswrapper[4776]: W1125 11:10:48.696620 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2eb1d65a_b5ca_4278_837d_cb255d764840.slice/crio-4f19307456553bf5b28a62d6566e5637437a40b354dd2d752e95f3e0eccb5d76 WatchSource:0}: Error finding container 4f19307456553bf5b28a62d6566e5637437a40b354dd2d752e95f3e0eccb5d76: Status 404 returned error can't find the container with id 4f19307456553bf5b28a62d6566e5637437a40b354dd2d752e95f3e0eccb5d76 Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.747954 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-648b897fb8-xj9jc"] Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.773131 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-6b88954b6d-fm4rs"] Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.805279 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-58c77f6f47-2zhwf"] Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.807010 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-58c77f6f47-2zhwf" Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.814079 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.814113 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.828136 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-75966d5674-cmtbw"] Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.829789 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-75966d5674-cmtbw" Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.832848 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.833265 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.853554 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-58c77f6f47-2zhwf"] Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.871016 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-75966d5674-cmtbw"] Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.913625 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7441fab-e3ef-4668-939a-b8d012872bdb-public-tls-certs\") pod \"heat-api-75966d5674-cmtbw\" (UID: \"e7441fab-e3ef-4668-939a-b8d012872bdb\") " pod="openstack/heat-api-75966d5674-cmtbw" Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.913708 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7441fab-e3ef-4668-939a-b8d012872bdb-internal-tls-certs\") pod \"heat-api-75966d5674-cmtbw\" (UID: \"e7441fab-e3ef-4668-939a-b8d012872bdb\") " pod="openstack/heat-api-75966d5674-cmtbw" Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.913778 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7441fab-e3ef-4668-939a-b8d012872bdb-config-data-custom\") pod \"heat-api-75966d5674-cmtbw\" (UID: \"e7441fab-e3ef-4668-939a-b8d012872bdb\") " pod="openstack/heat-api-75966d5674-cmtbw" Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.913800 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5aa7be27-73be-4401-a206-69b4d6a47a12-config-data\") pod \"heat-cfnapi-58c77f6f47-2zhwf\" (UID: \"5aa7be27-73be-4401-a206-69b4d6a47a12\") " pod="openstack/heat-cfnapi-58c77f6f47-2zhwf" Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.913832 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5aa7be27-73be-4401-a206-69b4d6a47a12-config-data-custom\") pod \"heat-cfnapi-58c77f6f47-2zhwf\" (UID: \"5aa7be27-73be-4401-a206-69b4d6a47a12\") " pod="openstack/heat-cfnapi-58c77f6f47-2zhwf" Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.913859 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aa7be27-73be-4401-a206-69b4d6a47a12-public-tls-certs\") pod \"heat-cfnapi-58c77f6f47-2zhwf\" (UID: \"5aa7be27-73be-4401-a206-69b4d6a47a12\") " pod="openstack/heat-cfnapi-58c77f6f47-2zhwf" Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.913893 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrnmn\" (UniqueName: \"kubernetes.io/projected/e7441fab-e3ef-4668-939a-b8d012872bdb-kube-api-access-xrnmn\") pod \"heat-api-75966d5674-cmtbw\" (UID: \"e7441fab-e3ef-4668-939a-b8d012872bdb\") " pod="openstack/heat-api-75966d5674-cmtbw" Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.913913 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7441fab-e3ef-4668-939a-b8d012872bdb-config-data\") pod \"heat-api-75966d5674-cmtbw\" (UID: \"e7441fab-e3ef-4668-939a-b8d012872bdb\") " pod="openstack/heat-api-75966d5674-cmtbw" Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.913935 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aa7be27-73be-4401-a206-69b4d6a47a12-internal-tls-certs\") pod \"heat-cfnapi-58c77f6f47-2zhwf\" (UID: \"5aa7be27-73be-4401-a206-69b4d6a47a12\") " pod="openstack/heat-cfnapi-58c77f6f47-2zhwf" Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.913961 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7441fab-e3ef-4668-939a-b8d012872bdb-combined-ca-bundle\") pod \"heat-api-75966d5674-cmtbw\" (UID: \"e7441fab-e3ef-4668-939a-b8d012872bdb\") " pod="openstack/heat-api-75966d5674-cmtbw" Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.913986 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa7be27-73be-4401-a206-69b4d6a47a12-combined-ca-bundle\") pod \"heat-cfnapi-58c77f6f47-2zhwf\" (UID: \"5aa7be27-73be-4401-a206-69b4d6a47a12\") " pod="openstack/heat-cfnapi-58c77f6f47-2zhwf" Nov 25 11:10:48 crc kubenswrapper[4776]: I1125 11:10:48.914006 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdn9z\" (UniqueName: \"kubernetes.io/projected/5aa7be27-73be-4401-a206-69b4d6a47a12-kube-api-access-mdn9z\") pod \"heat-cfnapi-58c77f6f47-2zhwf\" (UID: \"5aa7be27-73be-4401-a206-69b4d6a47a12\") " pod="openstack/heat-cfnapi-58c77f6f47-2zhwf" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.015608 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrnmn\" (UniqueName: \"kubernetes.io/projected/e7441fab-e3ef-4668-939a-b8d012872bdb-kube-api-access-xrnmn\") pod \"heat-api-75966d5674-cmtbw\" (UID: \"e7441fab-e3ef-4668-939a-b8d012872bdb\") " pod="openstack/heat-api-75966d5674-cmtbw" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.015654 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7441fab-e3ef-4668-939a-b8d012872bdb-config-data\") pod \"heat-api-75966d5674-cmtbw\" (UID: \"e7441fab-e3ef-4668-939a-b8d012872bdb\") " pod="openstack/heat-api-75966d5674-cmtbw" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.015683 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aa7be27-73be-4401-a206-69b4d6a47a12-internal-tls-certs\") pod \"heat-cfnapi-58c77f6f47-2zhwf\" (UID: \"5aa7be27-73be-4401-a206-69b4d6a47a12\") " pod="openstack/heat-cfnapi-58c77f6f47-2zhwf" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.015712 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7441fab-e3ef-4668-939a-b8d012872bdb-combined-ca-bundle\") pod \"heat-api-75966d5674-cmtbw\" (UID: \"e7441fab-e3ef-4668-939a-b8d012872bdb\") " pod="openstack/heat-api-75966d5674-cmtbw" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.015737 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa7be27-73be-4401-a206-69b4d6a47a12-combined-ca-bundle\") pod \"heat-cfnapi-58c77f6f47-2zhwf\" (UID: \"5aa7be27-73be-4401-a206-69b4d6a47a12\") " pod="openstack/heat-cfnapi-58c77f6f47-2zhwf" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.015756 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdn9z\" (UniqueName: \"kubernetes.io/projected/5aa7be27-73be-4401-a206-69b4d6a47a12-kube-api-access-mdn9z\") pod \"heat-cfnapi-58c77f6f47-2zhwf\" (UID: \"5aa7be27-73be-4401-a206-69b4d6a47a12\") " pod="openstack/heat-cfnapi-58c77f6f47-2zhwf" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.015812 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7441fab-e3ef-4668-939a-b8d012872bdb-public-tls-certs\") pod \"heat-api-75966d5674-cmtbw\" (UID: \"e7441fab-e3ef-4668-939a-b8d012872bdb\") " pod="openstack/heat-api-75966d5674-cmtbw" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.015847 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7441fab-e3ef-4668-939a-b8d012872bdb-internal-tls-certs\") pod \"heat-api-75966d5674-cmtbw\" (UID: \"e7441fab-e3ef-4668-939a-b8d012872bdb\") " pod="openstack/heat-api-75966d5674-cmtbw" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.015900 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7441fab-e3ef-4668-939a-b8d012872bdb-config-data-custom\") pod \"heat-api-75966d5674-cmtbw\" (UID: \"e7441fab-e3ef-4668-939a-b8d012872bdb\") " pod="openstack/heat-api-75966d5674-cmtbw" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.015921 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5aa7be27-73be-4401-a206-69b4d6a47a12-config-data\") pod \"heat-cfnapi-58c77f6f47-2zhwf\" (UID: \"5aa7be27-73be-4401-a206-69b4d6a47a12\") " pod="openstack/heat-cfnapi-58c77f6f47-2zhwf" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.015948 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5aa7be27-73be-4401-a206-69b4d6a47a12-config-data-custom\") pod \"heat-cfnapi-58c77f6f47-2zhwf\" (UID: \"5aa7be27-73be-4401-a206-69b4d6a47a12\") " pod="openstack/heat-cfnapi-58c77f6f47-2zhwf" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.015974 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aa7be27-73be-4401-a206-69b4d6a47a12-public-tls-certs\") pod \"heat-cfnapi-58c77f6f47-2zhwf\" (UID: \"5aa7be27-73be-4401-a206-69b4d6a47a12\") " pod="openstack/heat-cfnapi-58c77f6f47-2zhwf" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.029643 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aa7be27-73be-4401-a206-69b4d6a47a12-internal-tls-certs\") pod \"heat-cfnapi-58c77f6f47-2zhwf\" (UID: \"5aa7be27-73be-4401-a206-69b4d6a47a12\") " pod="openstack/heat-cfnapi-58c77f6f47-2zhwf" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.030253 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5aa7be27-73be-4401-a206-69b4d6a47a12-config-data-custom\") pod \"heat-cfnapi-58c77f6f47-2zhwf\" (UID: \"5aa7be27-73be-4401-a206-69b4d6a47a12\") " pod="openstack/heat-cfnapi-58c77f6f47-2zhwf" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.032301 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7441fab-e3ef-4668-939a-b8d012872bdb-combined-ca-bundle\") pod \"heat-api-75966d5674-cmtbw\" (UID: \"e7441fab-e3ef-4668-939a-b8d012872bdb\") " pod="openstack/heat-api-75966d5674-cmtbw" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.036956 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrnmn\" (UniqueName: \"kubernetes.io/projected/e7441fab-e3ef-4668-939a-b8d012872bdb-kube-api-access-xrnmn\") pod \"heat-api-75966d5674-cmtbw\" (UID: \"e7441fab-e3ef-4668-939a-b8d012872bdb\") " pod="openstack/heat-api-75966d5674-cmtbw" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.042013 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aa7be27-73be-4401-a206-69b4d6a47a12-public-tls-certs\") pod \"heat-cfnapi-58c77f6f47-2zhwf\" (UID: \"5aa7be27-73be-4401-a206-69b4d6a47a12\") " pod="openstack/heat-cfnapi-58c77f6f47-2zhwf" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.043188 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdn9z\" (UniqueName: \"kubernetes.io/projected/5aa7be27-73be-4401-a206-69b4d6a47a12-kube-api-access-mdn9z\") pod \"heat-cfnapi-58c77f6f47-2zhwf\" (UID: \"5aa7be27-73be-4401-a206-69b4d6a47a12\") " pod="openstack/heat-cfnapi-58c77f6f47-2zhwf" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.043598 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7441fab-e3ef-4668-939a-b8d012872bdb-config-data\") pod \"heat-api-75966d5674-cmtbw\" (UID: \"e7441fab-e3ef-4668-939a-b8d012872bdb\") " pod="openstack/heat-api-75966d5674-cmtbw" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.044395 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa7be27-73be-4401-a206-69b4d6a47a12-combined-ca-bundle\") pod \"heat-cfnapi-58c77f6f47-2zhwf\" (UID: \"5aa7be27-73be-4401-a206-69b4d6a47a12\") " pod="openstack/heat-cfnapi-58c77f6f47-2zhwf" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.045187 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5aa7be27-73be-4401-a206-69b4d6a47a12-config-data\") pod \"heat-cfnapi-58c77f6f47-2zhwf\" (UID: \"5aa7be27-73be-4401-a206-69b4d6a47a12\") " pod="openstack/heat-cfnapi-58c77f6f47-2zhwf" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.047933 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7441fab-e3ef-4668-939a-b8d012872bdb-public-tls-certs\") pod \"heat-api-75966d5674-cmtbw\" (UID: \"e7441fab-e3ef-4668-939a-b8d012872bdb\") " pod="openstack/heat-api-75966d5674-cmtbw" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.048233 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e7441fab-e3ef-4668-939a-b8d012872bdb-config-data-custom\") pod \"heat-api-75966d5674-cmtbw\" (UID: \"e7441fab-e3ef-4668-939a-b8d012872bdb\") " pod="openstack/heat-api-75966d5674-cmtbw" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.048729 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7441fab-e3ef-4668-939a-b8d012872bdb-internal-tls-certs\") pod \"heat-api-75966d5674-cmtbw\" (UID: \"e7441fab-e3ef-4668-939a-b8d012872bdb\") " pod="openstack/heat-api-75966d5674-cmtbw" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.247974 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-58c77f6f47-2zhwf" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.258665 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-75966d5674-cmtbw" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.274060 4776 generic.go:334] "Generic (PLEG): container finished" podID="a989b8a6-3b6b-4841-9f5b-0e23cfac4cba" containerID="5513deae57eca2180f937d822d07688da5c2125b3679f7d0b61ff1c60b97bb93" exitCode=1 Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.274190 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69c47bdc88-5ngxw" event={"ID":"a989b8a6-3b6b-4841-9f5b-0e23cfac4cba","Type":"ContainerDied","Data":"5513deae57eca2180f937d822d07688da5c2125b3679f7d0b61ff1c60b97bb93"} Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.274229 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69c47bdc88-5ngxw" event={"ID":"a989b8a6-3b6b-4841-9f5b-0e23cfac4cba","Type":"ContainerStarted","Data":"3f36251a52289f25f46e472f74bafb35f9ba1c87b689bef45007a083add124d5"} Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.274968 4776 scope.go:117] "RemoveContainer" containerID="5513deae57eca2180f937d822d07688da5c2125b3679f7d0b61ff1c60b97bb93" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.278631 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5954f4557-4hz28" event={"ID":"4291e940-29c4-4c08-98ea-227e96cd687f","Type":"ContainerStarted","Data":"f3333dc8004a35ff1f6832c79afb972d0af5ff51c9ee456752c8ed8feb8ff0df"} Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.278664 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5954f4557-4hz28" event={"ID":"4291e940-29c4-4c08-98ea-227e96cd687f","Type":"ContainerStarted","Data":"ecc192534343326d970067c0fdfb8cd37a8a140989d7bac50e2a5173979ef9de"} Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.278763 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-5954f4557-4hz28" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.288705 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6f8d5975f9-zsrqs" event={"ID":"2eb1d65a-b5ca-4278-837d-cb255d764840","Type":"ContainerStarted","Data":"ee5599ee695e0232757a638d7422f7b58ac0fc1f3eccccb17c4c7a68d020eb27"} Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.288751 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6f8d5975f9-zsrqs" event={"ID":"2eb1d65a-b5ca-4278-837d-cb255d764840","Type":"ContainerStarted","Data":"4f19307456553bf5b28a62d6566e5637437a40b354dd2d752e95f3e0eccb5d76"} Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.290217 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-6f8d5975f9-zsrqs" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.340185 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-5954f4557-4hz28" podStartSLOduration=2.340162392 podStartE2EDuration="2.340162392s" podCreationTimestamp="2025-11-25 11:10:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:10:49.319821932 +0000 UTC m=+6394.360881485" watchObservedRunningTime="2025-11-25 11:10:49.340162392 +0000 UTC m=+6394.381221945" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.366856 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-6f8d5975f9-zsrqs" podStartSLOduration=2.366832271 podStartE2EDuration="2.366832271s" podCreationTimestamp="2025-11-25 11:10:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:10:49.341846824 +0000 UTC m=+6394.382906367" watchObservedRunningTime="2025-11-25 11:10:49.366832271 +0000 UTC m=+6394.407891824" Nov 25 11:10:49 crc kubenswrapper[4776]: I1125 11:10:49.920455 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-75966d5674-cmtbw"] Nov 25 11:10:50 crc kubenswrapper[4776]: I1125 11:10:50.064803 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-58c77f6f47-2zhwf"] Nov 25 11:10:50 crc kubenswrapper[4776]: I1125 11:10:50.301585 4776 generic.go:334] "Generic (PLEG): container finished" podID="a989b8a6-3b6b-4841-9f5b-0e23cfac4cba" containerID="e38211c0acab7215d123863ec8fc27231d34b3ebe6936a5a5e55fc1cc0abd81c" exitCode=1 Nov 25 11:10:50 crc kubenswrapper[4776]: I1125 11:10:50.301764 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69c47bdc88-5ngxw" event={"ID":"a989b8a6-3b6b-4841-9f5b-0e23cfac4cba","Type":"ContainerDied","Data":"e38211c0acab7215d123863ec8fc27231d34b3ebe6936a5a5e55fc1cc0abd81c"} Nov 25 11:10:50 crc kubenswrapper[4776]: I1125 11:10:50.301976 4776 scope.go:117] "RemoveContainer" containerID="5513deae57eca2180f937d822d07688da5c2125b3679f7d0b61ff1c60b97bb93" Nov 25 11:10:50 crc kubenswrapper[4776]: I1125 11:10:50.302388 4776 scope.go:117] "RemoveContainer" containerID="e38211c0acab7215d123863ec8fc27231d34b3ebe6936a5a5e55fc1cc0abd81c" Nov 25 11:10:50 crc kubenswrapper[4776]: E1125 11:10:50.302708 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-69c47bdc88-5ngxw_openstack(a989b8a6-3b6b-4841-9f5b-0e23cfac4cba)\"" pod="openstack/heat-api-69c47bdc88-5ngxw" podUID="a989b8a6-3b6b-4841-9f5b-0e23cfac4cba" Nov 25 11:10:50 crc kubenswrapper[4776]: I1125 11:10:50.304392 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-75966d5674-cmtbw" event={"ID":"e7441fab-e3ef-4668-939a-b8d012872bdb","Type":"ContainerStarted","Data":"5ed5098f644c0f0354c3a2ff88caa800d22d22b3e16407241fe7d16d9c5b040a"} Nov 25 11:10:50 crc kubenswrapper[4776]: I1125 11:10:50.305843 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-656cc574-dstj4" podUID="219a7347-5102-4e6c-9b38-deea6275f715" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.125:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.125:8443: connect: connection refused" Nov 25 11:10:50 crc kubenswrapper[4776]: I1125 11:10:50.309436 4776 generic.go:334] "Generic (PLEG): container finished" podID="2eb1d65a-b5ca-4278-837d-cb255d764840" containerID="ee5599ee695e0232757a638d7422f7b58ac0fc1f3eccccb17c4c7a68d020eb27" exitCode=1 Nov 25 11:10:50 crc kubenswrapper[4776]: I1125 11:10:50.309545 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6f8d5975f9-zsrqs" event={"ID":"2eb1d65a-b5ca-4278-837d-cb255d764840","Type":"ContainerDied","Data":"ee5599ee695e0232757a638d7422f7b58ac0fc1f3eccccb17c4c7a68d020eb27"} Nov 25 11:10:50 crc kubenswrapper[4776]: I1125 11:10:50.310196 4776 scope.go:117] "RemoveContainer" containerID="ee5599ee695e0232757a638d7422f7b58ac0fc1f3eccccb17c4c7a68d020eb27" Nov 25 11:10:50 crc kubenswrapper[4776]: I1125 11:10:50.323027 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-6b88954b6d-fm4rs" podUID="e031e3dc-4e68-4b61-bf4d-4eb0246f6643" containerName="heat-api" containerID="cri-o://d4103054022308b045fb8118354421f420b58123187f375df80ea7e18c2f6e41" gracePeriod=60 Nov 25 11:10:50 crc kubenswrapper[4776]: I1125 11:10:50.323405 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-648b897fb8-xj9jc" podUID="a0ee8963-fd22-4acf-a2dc-871aa6cd3c54" containerName="heat-cfnapi" containerID="cri-o://d1a943b93761e2e6dc282c1823b575aed26148690c49f9e18e0128f12abe5533" gracePeriod=60 Nov 25 11:10:50 crc kubenswrapper[4776]: I1125 11:10:50.323435 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-58c77f6f47-2zhwf" event={"ID":"5aa7be27-73be-4401-a206-69b4d6a47a12","Type":"ContainerStarted","Data":"cf45513543a2b4ca405a41dfa409cc780e008f18198d6c22df8dd912a0120bea"} Nov 25 11:10:50 crc kubenswrapper[4776]: I1125 11:10:50.323599 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-58c77f6f47-2zhwf" Nov 25 11:10:50 crc kubenswrapper[4776]: I1125 11:10:50.361989 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-58c77f6f47-2zhwf" podStartSLOduration=2.361967146 podStartE2EDuration="2.361967146s" podCreationTimestamp="2025-11-25 11:10:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:10:50.359586617 +0000 UTC m=+6395.400646170" watchObservedRunningTime="2025-11-25 11:10:50.361967146 +0000 UTC m=+6395.403026699" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.206056 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-648b897fb8-xj9jc" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.213431 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6b88954b6d-fm4rs" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.285943 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0ee8963-fd22-4acf-a2dc-871aa6cd3c54-config-data\") pod \"a0ee8963-fd22-4acf-a2dc-871aa6cd3c54\" (UID: \"a0ee8963-fd22-4acf-a2dc-871aa6cd3c54\") " Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.286224 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7vml\" (UniqueName: \"kubernetes.io/projected/a0ee8963-fd22-4acf-a2dc-871aa6cd3c54-kube-api-access-w7vml\") pod \"a0ee8963-fd22-4acf-a2dc-871aa6cd3c54\" (UID: \"a0ee8963-fd22-4acf-a2dc-871aa6cd3c54\") " Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.286284 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ee8963-fd22-4acf-a2dc-871aa6cd3c54-combined-ca-bundle\") pod \"a0ee8963-fd22-4acf-a2dc-871aa6cd3c54\" (UID: \"a0ee8963-fd22-4acf-a2dc-871aa6cd3c54\") " Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.286488 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a0ee8963-fd22-4acf-a2dc-871aa6cd3c54-config-data-custom\") pod \"a0ee8963-fd22-4acf-a2dc-871aa6cd3c54\" (UID: \"a0ee8963-fd22-4acf-a2dc-871aa6cd3c54\") " Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.305590 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0ee8963-fd22-4acf-a2dc-871aa6cd3c54-kube-api-access-w7vml" (OuterVolumeSpecName: "kube-api-access-w7vml") pod "a0ee8963-fd22-4acf-a2dc-871aa6cd3c54" (UID: "a0ee8963-fd22-4acf-a2dc-871aa6cd3c54"). InnerVolumeSpecName "kube-api-access-w7vml". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.314180 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0ee8963-fd22-4acf-a2dc-871aa6cd3c54-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a0ee8963-fd22-4acf-a2dc-871aa6cd3c54" (UID: "a0ee8963-fd22-4acf-a2dc-871aa6cd3c54"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.316718 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0ee8963-fd22-4acf-a2dc-871aa6cd3c54-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a0ee8963-fd22-4acf-a2dc-871aa6cd3c54" (UID: "a0ee8963-fd22-4acf-a2dc-871aa6cd3c54"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.338620 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-58c77f6f47-2zhwf" event={"ID":"5aa7be27-73be-4401-a206-69b4d6a47a12","Type":"ContainerStarted","Data":"8167f7af78ed122b0ac2e8f1b771d4dedc8e37c3f7503913932872dc537f24d3"} Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.342436 4776 scope.go:117] "RemoveContainer" containerID="e38211c0acab7215d123863ec8fc27231d34b3ebe6936a5a5e55fc1cc0abd81c" Nov 25 11:10:51 crc kubenswrapper[4776]: E1125 11:10:51.342713 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-69c47bdc88-5ngxw_openstack(a989b8a6-3b6b-4841-9f5b-0e23cfac4cba)\"" pod="openstack/heat-api-69c47bdc88-5ngxw" podUID="a989b8a6-3b6b-4841-9f5b-0e23cfac4cba" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.345032 4776 generic.go:334] "Generic (PLEG): container finished" podID="e031e3dc-4e68-4b61-bf4d-4eb0246f6643" containerID="d4103054022308b045fb8118354421f420b58123187f375df80ea7e18c2f6e41" exitCode=0 Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.345107 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6b88954b6d-fm4rs" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.345106 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6b88954b6d-fm4rs" event={"ID":"e031e3dc-4e68-4b61-bf4d-4eb0246f6643","Type":"ContainerDied","Data":"d4103054022308b045fb8118354421f420b58123187f375df80ea7e18c2f6e41"} Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.345144 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6b88954b6d-fm4rs" event={"ID":"e031e3dc-4e68-4b61-bf4d-4eb0246f6643","Type":"ContainerDied","Data":"d696c6ba72bdcd014bcc1db07c7df4404295185df7116c8723fba6ff71c23fc4"} Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.345178 4776 scope.go:117] "RemoveContainer" containerID="d4103054022308b045fb8118354421f420b58123187f375df80ea7e18c2f6e41" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.347521 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-75966d5674-cmtbw" event={"ID":"e7441fab-e3ef-4668-939a-b8d012872bdb","Type":"ContainerStarted","Data":"33dbe2c6ad0041728e7c48e73754a93c64b79fbaad7359d23c039167c0cf0ccc"} Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.347971 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-75966d5674-cmtbw" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.352159 4776 generic.go:334] "Generic (PLEG): container finished" podID="a0ee8963-fd22-4acf-a2dc-871aa6cd3c54" containerID="d1a943b93761e2e6dc282c1823b575aed26148690c49f9e18e0128f12abe5533" exitCode=0 Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.352258 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-648b897fb8-xj9jc" event={"ID":"a0ee8963-fd22-4acf-a2dc-871aa6cd3c54","Type":"ContainerDied","Data":"d1a943b93761e2e6dc282c1823b575aed26148690c49f9e18e0128f12abe5533"} Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.352293 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-648b897fb8-xj9jc" event={"ID":"a0ee8963-fd22-4acf-a2dc-871aa6cd3c54","Type":"ContainerDied","Data":"587e80ad69607a39eb917c6ffca1dcc9a2d49b64728890777d07c3d6cddbaea1"} Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.352599 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-648b897fb8-xj9jc" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.362844 4776 generic.go:334] "Generic (PLEG): container finished" podID="2eb1d65a-b5ca-4278-837d-cb255d764840" containerID="a97bdf099a328f68a95b469791789b6242bbc82a797b2efc55dca80f7a5468de" exitCode=1 Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.362892 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6f8d5975f9-zsrqs" event={"ID":"2eb1d65a-b5ca-4278-837d-cb255d764840","Type":"ContainerDied","Data":"a97bdf099a328f68a95b469791789b6242bbc82a797b2efc55dca80f7a5468de"} Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.363730 4776 scope.go:117] "RemoveContainer" containerID="a97bdf099a328f68a95b469791789b6242bbc82a797b2efc55dca80f7a5468de" Nov 25 11:10:51 crc kubenswrapper[4776]: E1125 11:10:51.363959 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-6f8d5975f9-zsrqs_openstack(2eb1d65a-b5ca-4278-837d-cb255d764840)\"" pod="openstack/heat-cfnapi-6f8d5975f9-zsrqs" podUID="2eb1d65a-b5ca-4278-837d-cb255d764840" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.389682 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e031e3dc-4e68-4b61-bf4d-4eb0246f6643-config-data-custom\") pod \"e031e3dc-4e68-4b61-bf4d-4eb0246f6643\" (UID: \"e031e3dc-4e68-4b61-bf4d-4eb0246f6643\") " Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.389896 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e031e3dc-4e68-4b61-bf4d-4eb0246f6643-config-data\") pod \"e031e3dc-4e68-4b61-bf4d-4eb0246f6643\" (UID: \"e031e3dc-4e68-4b61-bf4d-4eb0246f6643\") " Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.389965 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7vqw\" (UniqueName: \"kubernetes.io/projected/e031e3dc-4e68-4b61-bf4d-4eb0246f6643-kube-api-access-x7vqw\") pod \"e031e3dc-4e68-4b61-bf4d-4eb0246f6643\" (UID: \"e031e3dc-4e68-4b61-bf4d-4eb0246f6643\") " Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.390051 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e031e3dc-4e68-4b61-bf4d-4eb0246f6643-combined-ca-bundle\") pod \"e031e3dc-4e68-4b61-bf4d-4eb0246f6643\" (UID: \"e031e3dc-4e68-4b61-bf4d-4eb0246f6643\") " Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.390712 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7vml\" (UniqueName: \"kubernetes.io/projected/a0ee8963-fd22-4acf-a2dc-871aa6cd3c54-kube-api-access-w7vml\") on node \"crc\" DevicePath \"\"" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.390733 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ee8963-fd22-4acf-a2dc-871aa6cd3c54-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.390742 4776 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a0ee8963-fd22-4acf-a2dc-871aa6cd3c54-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.392199 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-75966d5674-cmtbw" podStartSLOduration=3.392173641 podStartE2EDuration="3.392173641s" podCreationTimestamp="2025-11-25 11:10:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:10:51.375143554 +0000 UTC m=+6396.416203127" watchObservedRunningTime="2025-11-25 11:10:51.392173641 +0000 UTC m=+6396.433233204" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.395891 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e031e3dc-4e68-4b61-bf4d-4eb0246f6643-kube-api-access-x7vqw" (OuterVolumeSpecName: "kube-api-access-x7vqw") pod "e031e3dc-4e68-4b61-bf4d-4eb0246f6643" (UID: "e031e3dc-4e68-4b61-bf4d-4eb0246f6643"). InnerVolumeSpecName "kube-api-access-x7vqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.400623 4776 scope.go:117] "RemoveContainer" containerID="d4103054022308b045fb8118354421f420b58123187f375df80ea7e18c2f6e41" Nov 25 11:10:51 crc kubenswrapper[4776]: E1125 11:10:51.402876 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4103054022308b045fb8118354421f420b58123187f375df80ea7e18c2f6e41\": container with ID starting with d4103054022308b045fb8118354421f420b58123187f375df80ea7e18c2f6e41 not found: ID does not exist" containerID="d4103054022308b045fb8118354421f420b58123187f375df80ea7e18c2f6e41" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.402911 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4103054022308b045fb8118354421f420b58123187f375df80ea7e18c2f6e41"} err="failed to get container status \"d4103054022308b045fb8118354421f420b58123187f375df80ea7e18c2f6e41\": rpc error: code = NotFound desc = could not find container \"d4103054022308b045fb8118354421f420b58123187f375df80ea7e18c2f6e41\": container with ID starting with d4103054022308b045fb8118354421f420b58123187f375df80ea7e18c2f6e41 not found: ID does not exist" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.402931 4776 scope.go:117] "RemoveContainer" containerID="d1a943b93761e2e6dc282c1823b575aed26148690c49f9e18e0128f12abe5533" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.402930 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e031e3dc-4e68-4b61-bf4d-4eb0246f6643-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e031e3dc-4e68-4b61-bf4d-4eb0246f6643" (UID: "e031e3dc-4e68-4b61-bf4d-4eb0246f6643"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.425629 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0ee8963-fd22-4acf-a2dc-871aa6cd3c54-config-data" (OuterVolumeSpecName: "config-data") pod "a0ee8963-fd22-4acf-a2dc-871aa6cd3c54" (UID: "a0ee8963-fd22-4acf-a2dc-871aa6cd3c54"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.437738 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e031e3dc-4e68-4b61-bf4d-4eb0246f6643-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e031e3dc-4e68-4b61-bf4d-4eb0246f6643" (UID: "e031e3dc-4e68-4b61-bf4d-4eb0246f6643"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.442961 4776 scope.go:117] "RemoveContainer" containerID="d1a943b93761e2e6dc282c1823b575aed26148690c49f9e18e0128f12abe5533" Nov 25 11:10:51 crc kubenswrapper[4776]: E1125 11:10:51.443740 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1a943b93761e2e6dc282c1823b575aed26148690c49f9e18e0128f12abe5533\": container with ID starting with d1a943b93761e2e6dc282c1823b575aed26148690c49f9e18e0128f12abe5533 not found: ID does not exist" containerID="d1a943b93761e2e6dc282c1823b575aed26148690c49f9e18e0128f12abe5533" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.443787 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1a943b93761e2e6dc282c1823b575aed26148690c49f9e18e0128f12abe5533"} err="failed to get container status \"d1a943b93761e2e6dc282c1823b575aed26148690c49f9e18e0128f12abe5533\": rpc error: code = NotFound desc = could not find container \"d1a943b93761e2e6dc282c1823b575aed26148690c49f9e18e0128f12abe5533\": container with ID starting with d1a943b93761e2e6dc282c1823b575aed26148690c49f9e18e0128f12abe5533 not found: ID does not exist" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.443813 4776 scope.go:117] "RemoveContainer" containerID="ee5599ee695e0232757a638d7422f7b58ac0fc1f3eccccb17c4c7a68d020eb27" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.461574 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e031e3dc-4e68-4b61-bf4d-4eb0246f6643-config-data" (OuterVolumeSpecName: "config-data") pod "e031e3dc-4e68-4b61-bf4d-4eb0246f6643" (UID: "e031e3dc-4e68-4b61-bf4d-4eb0246f6643"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.492641 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e031e3dc-4e68-4b61-bf4d-4eb0246f6643-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.492677 4776 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e031e3dc-4e68-4b61-bf4d-4eb0246f6643-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.492687 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e031e3dc-4e68-4b61-bf4d-4eb0246f6643-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.492695 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7vqw\" (UniqueName: \"kubernetes.io/projected/e031e3dc-4e68-4b61-bf4d-4eb0246f6643-kube-api-access-x7vqw\") on node \"crc\" DevicePath \"\"" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.492707 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0ee8963-fd22-4acf-a2dc-871aa6cd3c54-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.685692 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-6b88954b6d-fm4rs"] Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.695043 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-6b88954b6d-fm4rs"] Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.703349 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-648b897fb8-xj9jc"] Nov 25 11:10:51 crc kubenswrapper[4776]: I1125 11:10:51.712948 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-648b897fb8-xj9jc"] Nov 25 11:10:52 crc kubenswrapper[4776]: I1125 11:10:52.375857 4776 scope.go:117] "RemoveContainer" containerID="a97bdf099a328f68a95b469791789b6242bbc82a797b2efc55dca80f7a5468de" Nov 25 11:10:52 crc kubenswrapper[4776]: E1125 11:10:52.376347 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-6f8d5975f9-zsrqs_openstack(2eb1d65a-b5ca-4278-837d-cb255d764840)\"" pod="openstack/heat-cfnapi-6f8d5975f9-zsrqs" podUID="2eb1d65a-b5ca-4278-837d-cb255d764840" Nov 25 11:10:52 crc kubenswrapper[4776]: I1125 11:10:52.865010 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-6f8d5975f9-zsrqs" Nov 25 11:10:52 crc kubenswrapper[4776]: I1125 11:10:52.865356 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-6f8d5975f9-zsrqs" Nov 25 11:10:52 crc kubenswrapper[4776]: I1125 11:10:52.889640 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-69c47bdc88-5ngxw" Nov 25 11:10:52 crc kubenswrapper[4776]: I1125 11:10:52.889686 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-69c47bdc88-5ngxw" Nov 25 11:10:52 crc kubenswrapper[4776]: I1125 11:10:52.890395 4776 scope.go:117] "RemoveContainer" containerID="e38211c0acab7215d123863ec8fc27231d34b3ebe6936a5a5e55fc1cc0abd81c" Nov 25 11:10:52 crc kubenswrapper[4776]: E1125 11:10:52.890625 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-69c47bdc88-5ngxw_openstack(a989b8a6-3b6b-4841-9f5b-0e23cfac4cba)\"" pod="openstack/heat-api-69c47bdc88-5ngxw" podUID="a989b8a6-3b6b-4841-9f5b-0e23cfac4cba" Nov 25 11:10:53 crc kubenswrapper[4776]: I1125 11:10:53.387437 4776 scope.go:117] "RemoveContainer" containerID="a97bdf099a328f68a95b469791789b6242bbc82a797b2efc55dca80f7a5468de" Nov 25 11:10:53 crc kubenswrapper[4776]: E1125 11:10:53.387690 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-6f8d5975f9-zsrqs_openstack(2eb1d65a-b5ca-4278-837d-cb255d764840)\"" pod="openstack/heat-cfnapi-6f8d5975f9-zsrqs" podUID="2eb1d65a-b5ca-4278-837d-cb255d764840" Nov 25 11:10:53 crc kubenswrapper[4776]: I1125 11:10:53.672478 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0ee8963-fd22-4acf-a2dc-871aa6cd3c54" path="/var/lib/kubelet/pods/a0ee8963-fd22-4acf-a2dc-871aa6cd3c54/volumes" Nov 25 11:10:53 crc kubenswrapper[4776]: I1125 11:10:53.673020 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e031e3dc-4e68-4b61-bf4d-4eb0246f6643" path="/var/lib/kubelet/pods/e031e3dc-4e68-4b61-bf4d-4eb0246f6643/volumes" Nov 25 11:10:58 crc kubenswrapper[4776]: I1125 11:10:58.877388 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-5dc4c495b-tmsg5" Nov 25 11:11:00 crc kubenswrapper[4776]: I1125 11:11:00.305896 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-656cc574-dstj4" podUID="219a7347-5102-4e6c-9b38-deea6275f715" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.125:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.125:8443: connect: connection refused" Nov 25 11:11:00 crc kubenswrapper[4776]: I1125 11:11:00.306532 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-656cc574-dstj4" Nov 25 11:11:00 crc kubenswrapper[4776]: I1125 11:11:00.685508 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-75966d5674-cmtbw" Nov 25 11:11:00 crc kubenswrapper[4776]: I1125 11:11:00.724420 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-58c77f6f47-2zhwf" Nov 25 11:11:00 crc kubenswrapper[4776]: I1125 11:11:00.746391 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-69c47bdc88-5ngxw"] Nov 25 11:11:00 crc kubenswrapper[4776]: I1125 11:11:00.805755 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-6f8d5975f9-zsrqs"] Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.271338 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-69c47bdc88-5ngxw" Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.284925 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6f8d5975f9-zsrqs" Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.416272 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2eb1d65a-b5ca-4278-837d-cb255d764840-config-data\") pod \"2eb1d65a-b5ca-4278-837d-cb255d764840\" (UID: \"2eb1d65a-b5ca-4278-837d-cb255d764840\") " Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.417122 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wv72\" (UniqueName: \"kubernetes.io/projected/a989b8a6-3b6b-4841-9f5b-0e23cfac4cba-kube-api-access-7wv72\") pod \"a989b8a6-3b6b-4841-9f5b-0e23cfac4cba\" (UID: \"a989b8a6-3b6b-4841-9f5b-0e23cfac4cba\") " Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.417199 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2eb1d65a-b5ca-4278-837d-cb255d764840-config-data-custom\") pod \"2eb1d65a-b5ca-4278-837d-cb255d764840\" (UID: \"2eb1d65a-b5ca-4278-837d-cb255d764840\") " Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.417237 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eb1d65a-b5ca-4278-837d-cb255d764840-combined-ca-bundle\") pod \"2eb1d65a-b5ca-4278-837d-cb255d764840\" (UID: \"2eb1d65a-b5ca-4278-837d-cb255d764840\") " Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.417297 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a989b8a6-3b6b-4841-9f5b-0e23cfac4cba-combined-ca-bundle\") pod \"a989b8a6-3b6b-4841-9f5b-0e23cfac4cba\" (UID: \"a989b8a6-3b6b-4841-9f5b-0e23cfac4cba\") " Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.417359 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zthn\" (UniqueName: \"kubernetes.io/projected/2eb1d65a-b5ca-4278-837d-cb255d764840-kube-api-access-4zthn\") pod \"2eb1d65a-b5ca-4278-837d-cb255d764840\" (UID: \"2eb1d65a-b5ca-4278-837d-cb255d764840\") " Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.417414 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a989b8a6-3b6b-4841-9f5b-0e23cfac4cba-config-data\") pod \"a989b8a6-3b6b-4841-9f5b-0e23cfac4cba\" (UID: \"a989b8a6-3b6b-4841-9f5b-0e23cfac4cba\") " Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.417503 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a989b8a6-3b6b-4841-9f5b-0e23cfac4cba-config-data-custom\") pod \"a989b8a6-3b6b-4841-9f5b-0e23cfac4cba\" (UID: \"a989b8a6-3b6b-4841-9f5b-0e23cfac4cba\") " Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.423680 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a989b8a6-3b6b-4841-9f5b-0e23cfac4cba-kube-api-access-7wv72" (OuterVolumeSpecName: "kube-api-access-7wv72") pod "a989b8a6-3b6b-4841-9f5b-0e23cfac4cba" (UID: "a989b8a6-3b6b-4841-9f5b-0e23cfac4cba"). InnerVolumeSpecName "kube-api-access-7wv72". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.424348 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2eb1d65a-b5ca-4278-837d-cb255d764840-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2eb1d65a-b5ca-4278-837d-cb255d764840" (UID: "2eb1d65a-b5ca-4278-837d-cb255d764840"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.424846 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a989b8a6-3b6b-4841-9f5b-0e23cfac4cba-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a989b8a6-3b6b-4841-9f5b-0e23cfac4cba" (UID: "a989b8a6-3b6b-4841-9f5b-0e23cfac4cba"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.425215 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2eb1d65a-b5ca-4278-837d-cb255d764840-kube-api-access-4zthn" (OuterVolumeSpecName: "kube-api-access-4zthn") pod "2eb1d65a-b5ca-4278-837d-cb255d764840" (UID: "2eb1d65a-b5ca-4278-837d-cb255d764840"). InnerVolumeSpecName "kube-api-access-4zthn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.448499 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2eb1d65a-b5ca-4278-837d-cb255d764840-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2eb1d65a-b5ca-4278-837d-cb255d764840" (UID: "2eb1d65a-b5ca-4278-837d-cb255d764840"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.448642 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a989b8a6-3b6b-4841-9f5b-0e23cfac4cba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a989b8a6-3b6b-4841-9f5b-0e23cfac4cba" (UID: "a989b8a6-3b6b-4841-9f5b-0e23cfac4cba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.459811 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6f8d5975f9-zsrqs" event={"ID":"2eb1d65a-b5ca-4278-837d-cb255d764840","Type":"ContainerDied","Data":"4f19307456553bf5b28a62d6566e5637437a40b354dd2d752e95f3e0eccb5d76"} Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.459874 4776 scope.go:117] "RemoveContainer" containerID="a97bdf099a328f68a95b469791789b6242bbc82a797b2efc55dca80f7a5468de" Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.459989 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6f8d5975f9-zsrqs" Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.462962 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69c47bdc88-5ngxw" event={"ID":"a989b8a6-3b6b-4841-9f5b-0e23cfac4cba","Type":"ContainerDied","Data":"3f36251a52289f25f46e472f74bafb35f9ba1c87b689bef45007a083add124d5"} Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.463033 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-69c47bdc88-5ngxw" Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.481335 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2eb1d65a-b5ca-4278-837d-cb255d764840-config-data" (OuterVolumeSpecName: "config-data") pod "2eb1d65a-b5ca-4278-837d-cb255d764840" (UID: "2eb1d65a-b5ca-4278-837d-cb255d764840"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.488243 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a989b8a6-3b6b-4841-9f5b-0e23cfac4cba-config-data" (OuterVolumeSpecName: "config-data") pod "a989b8a6-3b6b-4841-9f5b-0e23cfac4cba" (UID: "a989b8a6-3b6b-4841-9f5b-0e23cfac4cba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.520658 4776 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a989b8a6-3b6b-4841-9f5b-0e23cfac4cba-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.520698 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2eb1d65a-b5ca-4278-837d-cb255d764840-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.520714 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wv72\" (UniqueName: \"kubernetes.io/projected/a989b8a6-3b6b-4841-9f5b-0e23cfac4cba-kube-api-access-7wv72\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.520726 4776 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2eb1d65a-b5ca-4278-837d-cb255d764840-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.520740 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eb1d65a-b5ca-4278-837d-cb255d764840-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.520751 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a989b8a6-3b6b-4841-9f5b-0e23cfac4cba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.520763 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zthn\" (UniqueName: \"kubernetes.io/projected/2eb1d65a-b5ca-4278-837d-cb255d764840-kube-api-access-4zthn\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.520776 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a989b8a6-3b6b-4841-9f5b-0e23cfac4cba-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.589430 4776 scope.go:117] "RemoveContainer" containerID="e38211c0acab7215d123863ec8fc27231d34b3ebe6936a5a5e55fc1cc0abd81c" Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.790203 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-6f8d5975f9-zsrqs"] Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.798738 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-6f8d5975f9-zsrqs"] Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.806753 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-69c47bdc88-5ngxw"] Nov 25 11:11:01 crc kubenswrapper[4776]: I1125 11:11:01.814274 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-69c47bdc88-5ngxw"] Nov 25 11:11:03 crc kubenswrapper[4776]: I1125 11:11:03.674003 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2eb1d65a-b5ca-4278-837d-cb255d764840" path="/var/lib/kubelet/pods/2eb1d65a-b5ca-4278-837d-cb255d764840/volumes" Nov 25 11:11:03 crc kubenswrapper[4776]: I1125 11:11:03.675228 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a989b8a6-3b6b-4841-9f5b-0e23cfac4cba" path="/var/lib/kubelet/pods/a989b8a6-3b6b-4841-9f5b-0e23cfac4cba/volumes" Nov 25 11:11:07 crc kubenswrapper[4776]: I1125 11:11:07.850672 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-5954f4557-4hz28" Nov 25 11:11:07 crc kubenswrapper[4776]: I1125 11:11:07.911765 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-5dc4c495b-tmsg5"] Nov 25 11:11:07 crc kubenswrapper[4776]: I1125 11:11:07.911992 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-5dc4c495b-tmsg5" podUID="352f32f4-36a2-4a3c-858d-52b176e296b1" containerName="heat-engine" containerID="cri-o://57929ea0709123456be6f27638c6b9e2666c84483d585344a7d644ec936d1f55" gracePeriod=60 Nov 25 11:11:08 crc kubenswrapper[4776]: E1125 11:11:08.850586 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="57929ea0709123456be6f27638c6b9e2666c84483d585344a7d644ec936d1f55" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 25 11:11:08 crc kubenswrapper[4776]: E1125 11:11:08.855521 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="57929ea0709123456be6f27638c6b9e2666c84483d585344a7d644ec936d1f55" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 25 11:11:08 crc kubenswrapper[4776]: E1125 11:11:08.857115 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="57929ea0709123456be6f27638c6b9e2666c84483d585344a7d644ec936d1f55" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 25 11:11:08 crc kubenswrapper[4776]: E1125 11:11:08.857157 4776 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-5dc4c495b-tmsg5" podUID="352f32f4-36a2-4a3c-858d-52b176e296b1" containerName="heat-engine" Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.296801 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-656cc574-dstj4" Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.395775 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/219a7347-5102-4e6c-9b38-deea6275f715-combined-ca-bundle\") pod \"219a7347-5102-4e6c-9b38-deea6275f715\" (UID: \"219a7347-5102-4e6c-9b38-deea6275f715\") " Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.396168 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/219a7347-5102-4e6c-9b38-deea6275f715-horizon-secret-key\") pod \"219a7347-5102-4e6c-9b38-deea6275f715\" (UID: \"219a7347-5102-4e6c-9b38-deea6275f715\") " Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.396306 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/219a7347-5102-4e6c-9b38-deea6275f715-config-data\") pod \"219a7347-5102-4e6c-9b38-deea6275f715\" (UID: \"219a7347-5102-4e6c-9b38-deea6275f715\") " Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.396353 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/219a7347-5102-4e6c-9b38-deea6275f715-horizon-tls-certs\") pod \"219a7347-5102-4e6c-9b38-deea6275f715\" (UID: \"219a7347-5102-4e6c-9b38-deea6275f715\") " Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.396424 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/219a7347-5102-4e6c-9b38-deea6275f715-scripts\") pod \"219a7347-5102-4e6c-9b38-deea6275f715\" (UID: \"219a7347-5102-4e6c-9b38-deea6275f715\") " Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.396471 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkvvq\" (UniqueName: \"kubernetes.io/projected/219a7347-5102-4e6c-9b38-deea6275f715-kube-api-access-kkvvq\") pod \"219a7347-5102-4e6c-9b38-deea6275f715\" (UID: \"219a7347-5102-4e6c-9b38-deea6275f715\") " Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.396623 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/219a7347-5102-4e6c-9b38-deea6275f715-logs\") pod \"219a7347-5102-4e6c-9b38-deea6275f715\" (UID: \"219a7347-5102-4e6c-9b38-deea6275f715\") " Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.399042 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/219a7347-5102-4e6c-9b38-deea6275f715-logs" (OuterVolumeSpecName: "logs") pod "219a7347-5102-4e6c-9b38-deea6275f715" (UID: "219a7347-5102-4e6c-9b38-deea6275f715"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.407609 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/219a7347-5102-4e6c-9b38-deea6275f715-kube-api-access-kkvvq" (OuterVolumeSpecName: "kube-api-access-kkvvq") pod "219a7347-5102-4e6c-9b38-deea6275f715" (UID: "219a7347-5102-4e6c-9b38-deea6275f715"). InnerVolumeSpecName "kube-api-access-kkvvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.415256 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/219a7347-5102-4e6c-9b38-deea6275f715-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "219a7347-5102-4e6c-9b38-deea6275f715" (UID: "219a7347-5102-4e6c-9b38-deea6275f715"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.430758 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/219a7347-5102-4e6c-9b38-deea6275f715-scripts" (OuterVolumeSpecName: "scripts") pod "219a7347-5102-4e6c-9b38-deea6275f715" (UID: "219a7347-5102-4e6c-9b38-deea6275f715"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.444100 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/219a7347-5102-4e6c-9b38-deea6275f715-config-data" (OuterVolumeSpecName: "config-data") pod "219a7347-5102-4e6c-9b38-deea6275f715" (UID: "219a7347-5102-4e6c-9b38-deea6275f715"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.452335 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/219a7347-5102-4e6c-9b38-deea6275f715-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "219a7347-5102-4e6c-9b38-deea6275f715" (UID: "219a7347-5102-4e6c-9b38-deea6275f715"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.454180 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/219a7347-5102-4e6c-9b38-deea6275f715-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "219a7347-5102-4e6c-9b38-deea6275f715" (UID: "219a7347-5102-4e6c-9b38-deea6275f715"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.498994 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/219a7347-5102-4e6c-9b38-deea6275f715-logs\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.499036 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/219a7347-5102-4e6c-9b38-deea6275f715-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.499049 4776 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/219a7347-5102-4e6c-9b38-deea6275f715-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.499078 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/219a7347-5102-4e6c-9b38-deea6275f715-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.499091 4776 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/219a7347-5102-4e6c-9b38-deea6275f715-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.499101 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/219a7347-5102-4e6c-9b38-deea6275f715-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.499113 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkvvq\" (UniqueName: \"kubernetes.io/projected/219a7347-5102-4e6c-9b38-deea6275f715-kube-api-access-kkvvq\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.546441 4776 generic.go:334] "Generic (PLEG): container finished" podID="219a7347-5102-4e6c-9b38-deea6275f715" containerID="75624a8110711b2f1d37d5065d9a5ab2035226b6e3ac4acf037118c35cb7d28c" exitCode=137 Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.546517 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-656cc574-dstj4" event={"ID":"219a7347-5102-4e6c-9b38-deea6275f715","Type":"ContainerDied","Data":"75624a8110711b2f1d37d5065d9a5ab2035226b6e3ac4acf037118c35cb7d28c"} Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.546559 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-656cc574-dstj4" event={"ID":"219a7347-5102-4e6c-9b38-deea6275f715","Type":"ContainerDied","Data":"accf9af77c7a917385e318edd9989c36af54f20883603f01cbe8c6df33d9d891"} Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.546576 4776 scope.go:117] "RemoveContainer" containerID="f5dd0b1f4ea70853679a58ed439e71548cd36135ef8443036178352353139632" Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.546699 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-656cc574-dstj4" Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.597206 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-656cc574-dstj4"] Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.610113 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-656cc574-dstj4"] Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.683545 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="219a7347-5102-4e6c-9b38-deea6275f715" path="/var/lib/kubelet/pods/219a7347-5102-4e6c-9b38-deea6275f715/volumes" Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.742991 4776 scope.go:117] "RemoveContainer" containerID="75624a8110711b2f1d37d5065d9a5ab2035226b6e3ac4acf037118c35cb7d28c" Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.765626 4776 scope.go:117] "RemoveContainer" containerID="f5dd0b1f4ea70853679a58ed439e71548cd36135ef8443036178352353139632" Nov 25 11:11:09 crc kubenswrapper[4776]: E1125 11:11:09.766051 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5dd0b1f4ea70853679a58ed439e71548cd36135ef8443036178352353139632\": container with ID starting with f5dd0b1f4ea70853679a58ed439e71548cd36135ef8443036178352353139632 not found: ID does not exist" containerID="f5dd0b1f4ea70853679a58ed439e71548cd36135ef8443036178352353139632" Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.766103 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5dd0b1f4ea70853679a58ed439e71548cd36135ef8443036178352353139632"} err="failed to get container status \"f5dd0b1f4ea70853679a58ed439e71548cd36135ef8443036178352353139632\": rpc error: code = NotFound desc = could not find container \"f5dd0b1f4ea70853679a58ed439e71548cd36135ef8443036178352353139632\": container with ID starting with f5dd0b1f4ea70853679a58ed439e71548cd36135ef8443036178352353139632 not found: ID does not exist" Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.766146 4776 scope.go:117] "RemoveContainer" containerID="75624a8110711b2f1d37d5065d9a5ab2035226b6e3ac4acf037118c35cb7d28c" Nov 25 11:11:09 crc kubenswrapper[4776]: E1125 11:11:09.766520 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75624a8110711b2f1d37d5065d9a5ab2035226b6e3ac4acf037118c35cb7d28c\": container with ID starting with 75624a8110711b2f1d37d5065d9a5ab2035226b6e3ac4acf037118c35cb7d28c not found: ID does not exist" containerID="75624a8110711b2f1d37d5065d9a5ab2035226b6e3ac4acf037118c35cb7d28c" Nov 25 11:11:09 crc kubenswrapper[4776]: I1125 11:11:09.766545 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75624a8110711b2f1d37d5065d9a5ab2035226b6e3ac4acf037118c35cb7d28c"} err="failed to get container status \"75624a8110711b2f1d37d5065d9a5ab2035226b6e3ac4acf037118c35cb7d28c\": rpc error: code = NotFound desc = could not find container \"75624a8110711b2f1d37d5065d9a5ab2035226b6e3ac4acf037118c35cb7d28c\": container with ID starting with 75624a8110711b2f1d37d5065d9a5ab2035226b6e3ac4acf037118c35cb7d28c not found: ID does not exist" Nov 25 11:11:17 crc kubenswrapper[4776]: I1125 11:11:17.037140 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-pnmqh"] Nov 25 11:11:17 crc kubenswrapper[4776]: I1125 11:11:17.048648 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-f6ca-account-create-qsbv2"] Nov 25 11:11:17 crc kubenswrapper[4776]: I1125 11:11:17.059544 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-pnmqh"] Nov 25 11:11:17 crc kubenswrapper[4776]: I1125 11:11:17.070299 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-f6ca-account-create-qsbv2"] Nov 25 11:11:17 crc kubenswrapper[4776]: I1125 11:11:17.676830 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e356929a-2335-4a51-b410-39521ac9f4d1" path="/var/lib/kubelet/pods/e356929a-2335-4a51-b410-39521ac9f4d1/volumes" Nov 25 11:11:17 crc kubenswrapper[4776]: I1125 11:11:17.679288 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5e277db-7467-4a4f-9715-6854e235cac1" path="/var/lib/kubelet/pods/e5e277db-7467-4a4f-9715-6854e235cac1/volumes" Nov 25 11:11:18 crc kubenswrapper[4776]: E1125 11:11:18.848896 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="57929ea0709123456be6f27638c6b9e2666c84483d585344a7d644ec936d1f55" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 25 11:11:18 crc kubenswrapper[4776]: E1125 11:11:18.851143 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="57929ea0709123456be6f27638c6b9e2666c84483d585344a7d644ec936d1f55" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 25 11:11:18 crc kubenswrapper[4776]: E1125 11:11:18.855905 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="57929ea0709123456be6f27638c6b9e2666c84483d585344a7d644ec936d1f55" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Nov 25 11:11:18 crc kubenswrapper[4776]: E1125 11:11:18.855994 4776 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-5dc4c495b-tmsg5" podUID="352f32f4-36a2-4a3c-858d-52b176e296b1" containerName="heat-engine" Nov 25 11:11:19 crc kubenswrapper[4776]: I1125 11:11:19.531960 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5dc4c495b-tmsg5" Nov 25 11:11:19 crc kubenswrapper[4776]: I1125 11:11:19.625165 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r962s\" (UniqueName: \"kubernetes.io/projected/352f32f4-36a2-4a3c-858d-52b176e296b1-kube-api-access-r962s\") pod \"352f32f4-36a2-4a3c-858d-52b176e296b1\" (UID: \"352f32f4-36a2-4a3c-858d-52b176e296b1\") " Nov 25 11:11:19 crc kubenswrapper[4776]: I1125 11:11:19.625535 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/352f32f4-36a2-4a3c-858d-52b176e296b1-config-data\") pod \"352f32f4-36a2-4a3c-858d-52b176e296b1\" (UID: \"352f32f4-36a2-4a3c-858d-52b176e296b1\") " Nov 25 11:11:19 crc kubenswrapper[4776]: I1125 11:11:19.625650 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/352f32f4-36a2-4a3c-858d-52b176e296b1-combined-ca-bundle\") pod \"352f32f4-36a2-4a3c-858d-52b176e296b1\" (UID: \"352f32f4-36a2-4a3c-858d-52b176e296b1\") " Nov 25 11:11:19 crc kubenswrapper[4776]: I1125 11:11:19.625684 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/352f32f4-36a2-4a3c-858d-52b176e296b1-config-data-custom\") pod \"352f32f4-36a2-4a3c-858d-52b176e296b1\" (UID: \"352f32f4-36a2-4a3c-858d-52b176e296b1\") " Nov 25 11:11:19 crc kubenswrapper[4776]: I1125 11:11:19.635453 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/352f32f4-36a2-4a3c-858d-52b176e296b1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "352f32f4-36a2-4a3c-858d-52b176e296b1" (UID: "352f32f4-36a2-4a3c-858d-52b176e296b1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:11:19 crc kubenswrapper[4776]: I1125 11:11:19.656971 4776 generic.go:334] "Generic (PLEG): container finished" podID="352f32f4-36a2-4a3c-858d-52b176e296b1" containerID="57929ea0709123456be6f27638c6b9e2666c84483d585344a7d644ec936d1f55" exitCode=0 Nov 25 11:11:19 crc kubenswrapper[4776]: I1125 11:11:19.657038 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5dc4c495b-tmsg5" Nov 25 11:11:19 crc kubenswrapper[4776]: I1125 11:11:19.657041 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5dc4c495b-tmsg5" event={"ID":"352f32f4-36a2-4a3c-858d-52b176e296b1","Type":"ContainerDied","Data":"57929ea0709123456be6f27638c6b9e2666c84483d585344a7d644ec936d1f55"} Nov 25 11:11:19 crc kubenswrapper[4776]: I1125 11:11:19.657107 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5dc4c495b-tmsg5" event={"ID":"352f32f4-36a2-4a3c-858d-52b176e296b1","Type":"ContainerDied","Data":"ed0dacf50cae3a12b3469bfbbbe9ea65125d54885f0235307b07ad8e0ccdfba0"} Nov 25 11:11:19 crc kubenswrapper[4776]: I1125 11:11:19.657126 4776 scope.go:117] "RemoveContainer" containerID="57929ea0709123456be6f27638c6b9e2666c84483d585344a7d644ec936d1f55" Nov 25 11:11:19 crc kubenswrapper[4776]: I1125 11:11:19.658483 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/352f32f4-36a2-4a3c-858d-52b176e296b1-kube-api-access-r962s" (OuterVolumeSpecName: "kube-api-access-r962s") pod "352f32f4-36a2-4a3c-858d-52b176e296b1" (UID: "352f32f4-36a2-4a3c-858d-52b176e296b1"). InnerVolumeSpecName "kube-api-access-r962s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:11:19 crc kubenswrapper[4776]: I1125 11:11:19.670403 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/352f32f4-36a2-4a3c-858d-52b176e296b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "352f32f4-36a2-4a3c-858d-52b176e296b1" (UID: "352f32f4-36a2-4a3c-858d-52b176e296b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:11:19 crc kubenswrapper[4776]: I1125 11:11:19.693548 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/352f32f4-36a2-4a3c-858d-52b176e296b1-config-data" (OuterVolumeSpecName: "config-data") pod "352f32f4-36a2-4a3c-858d-52b176e296b1" (UID: "352f32f4-36a2-4a3c-858d-52b176e296b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:11:19 crc kubenswrapper[4776]: I1125 11:11:19.731388 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/352f32f4-36a2-4a3c-858d-52b176e296b1-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:19 crc kubenswrapper[4776]: I1125 11:11:19.731425 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/352f32f4-36a2-4a3c-858d-52b176e296b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:19 crc kubenswrapper[4776]: I1125 11:11:19.731442 4776 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/352f32f4-36a2-4a3c-858d-52b176e296b1-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:19 crc kubenswrapper[4776]: I1125 11:11:19.731454 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r962s\" (UniqueName: \"kubernetes.io/projected/352f32f4-36a2-4a3c-858d-52b176e296b1-kube-api-access-r962s\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:19 crc kubenswrapper[4776]: I1125 11:11:19.764371 4776 scope.go:117] "RemoveContainer" containerID="57929ea0709123456be6f27638c6b9e2666c84483d585344a7d644ec936d1f55" Nov 25 11:11:19 crc kubenswrapper[4776]: E1125 11:11:19.764984 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57929ea0709123456be6f27638c6b9e2666c84483d585344a7d644ec936d1f55\": container with ID starting with 57929ea0709123456be6f27638c6b9e2666c84483d585344a7d644ec936d1f55 not found: ID does not exist" containerID="57929ea0709123456be6f27638c6b9e2666c84483d585344a7d644ec936d1f55" Nov 25 11:11:19 crc kubenswrapper[4776]: I1125 11:11:19.765010 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57929ea0709123456be6f27638c6b9e2666c84483d585344a7d644ec936d1f55"} err="failed to get container status \"57929ea0709123456be6f27638c6b9e2666c84483d585344a7d644ec936d1f55\": rpc error: code = NotFound desc = could not find container \"57929ea0709123456be6f27638c6b9e2666c84483d585344a7d644ec936d1f55\": container with ID starting with 57929ea0709123456be6f27638c6b9e2666c84483d585344a7d644ec936d1f55 not found: ID does not exist" Nov 25 11:11:19 crc kubenswrapper[4776]: I1125 11:11:19.992388 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-5dc4c495b-tmsg5"] Nov 25 11:11:20 crc kubenswrapper[4776]: I1125 11:11:20.003528 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-5dc4c495b-tmsg5"] Nov 25 11:11:21 crc kubenswrapper[4776]: I1125 11:11:21.672930 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="352f32f4-36a2-4a3c-858d-52b176e296b1" path="/var/lib/kubelet/pods/352f32f4-36a2-4a3c-858d-52b176e296b1/volumes" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.700472 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz"] Nov 25 11:11:22 crc kubenswrapper[4776]: E1125 11:11:22.700990 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a989b8a6-3b6b-4841-9f5b-0e23cfac4cba" containerName="heat-api" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.701007 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a989b8a6-3b6b-4841-9f5b-0e23cfac4cba" containerName="heat-api" Nov 25 11:11:22 crc kubenswrapper[4776]: E1125 11:11:22.701028 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e031e3dc-4e68-4b61-bf4d-4eb0246f6643" containerName="heat-api" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.701035 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="e031e3dc-4e68-4b61-bf4d-4eb0246f6643" containerName="heat-api" Nov 25 11:11:22 crc kubenswrapper[4776]: E1125 11:11:22.701052 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eb1d65a-b5ca-4278-837d-cb255d764840" containerName="heat-cfnapi" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.701058 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eb1d65a-b5ca-4278-837d-cb255d764840" containerName="heat-cfnapi" Nov 25 11:11:22 crc kubenswrapper[4776]: E1125 11:11:22.701094 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0ee8963-fd22-4acf-a2dc-871aa6cd3c54" containerName="heat-cfnapi" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.701100 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0ee8963-fd22-4acf-a2dc-871aa6cd3c54" containerName="heat-cfnapi" Nov 25 11:11:22 crc kubenswrapper[4776]: E1125 11:11:22.701114 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="219a7347-5102-4e6c-9b38-deea6275f715" containerName="horizon" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.701120 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="219a7347-5102-4e6c-9b38-deea6275f715" containerName="horizon" Nov 25 11:11:22 crc kubenswrapper[4776]: E1125 11:11:22.701131 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="352f32f4-36a2-4a3c-858d-52b176e296b1" containerName="heat-engine" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.701137 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="352f32f4-36a2-4a3c-858d-52b176e296b1" containerName="heat-engine" Nov 25 11:11:22 crc kubenswrapper[4776]: E1125 11:11:22.701151 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="219a7347-5102-4e6c-9b38-deea6275f715" containerName="horizon-log" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.701157 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="219a7347-5102-4e6c-9b38-deea6275f715" containerName="horizon-log" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.701373 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="a989b8a6-3b6b-4841-9f5b-0e23cfac4cba" containerName="heat-api" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.701388 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="2eb1d65a-b5ca-4278-837d-cb255d764840" containerName="heat-cfnapi" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.701410 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="352f32f4-36a2-4a3c-858d-52b176e296b1" containerName="heat-engine" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.701420 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="2eb1d65a-b5ca-4278-837d-cb255d764840" containerName="heat-cfnapi" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.701430 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0ee8963-fd22-4acf-a2dc-871aa6cd3c54" containerName="heat-cfnapi" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.701445 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="e031e3dc-4e68-4b61-bf4d-4eb0246f6643" containerName="heat-api" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.701458 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="219a7347-5102-4e6c-9b38-deea6275f715" containerName="horizon" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.701476 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="219a7347-5102-4e6c-9b38-deea6275f715" containerName="horizon-log" Nov 25 11:11:22 crc kubenswrapper[4776]: E1125 11:11:22.701672 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a989b8a6-3b6b-4841-9f5b-0e23cfac4cba" containerName="heat-api" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.701680 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a989b8a6-3b6b-4841-9f5b-0e23cfac4cba" containerName="heat-api" Nov 25 11:11:22 crc kubenswrapper[4776]: E1125 11:11:22.701690 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eb1d65a-b5ca-4278-837d-cb255d764840" containerName="heat-cfnapi" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.701696 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eb1d65a-b5ca-4278-837d-cb255d764840" containerName="heat-cfnapi" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.701871 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="a989b8a6-3b6b-4841-9f5b-0e23cfac4cba" containerName="heat-api" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.703188 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.706379 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.710932 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz"] Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.786622 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dbc20081-c609-4ddf-a321-2ae092fd67e9-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz\" (UID: \"dbc20081-c609-4ddf-a321-2ae092fd67e9\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.787318 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dbc20081-c609-4ddf-a321-2ae092fd67e9-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz\" (UID: \"dbc20081-c609-4ddf-a321-2ae092fd67e9\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.787487 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhp98\" (UniqueName: \"kubernetes.io/projected/dbc20081-c609-4ddf-a321-2ae092fd67e9-kube-api-access-vhp98\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz\" (UID: \"dbc20081-c609-4ddf-a321-2ae092fd67e9\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.889608 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dbc20081-c609-4ddf-a321-2ae092fd67e9-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz\" (UID: \"dbc20081-c609-4ddf-a321-2ae092fd67e9\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.889999 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhp98\" (UniqueName: \"kubernetes.io/projected/dbc20081-c609-4ddf-a321-2ae092fd67e9-kube-api-access-vhp98\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz\" (UID: \"dbc20081-c609-4ddf-a321-2ae092fd67e9\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.890048 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dbc20081-c609-4ddf-a321-2ae092fd67e9-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz\" (UID: \"dbc20081-c609-4ddf-a321-2ae092fd67e9\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.890114 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dbc20081-c609-4ddf-a321-2ae092fd67e9-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz\" (UID: \"dbc20081-c609-4ddf-a321-2ae092fd67e9\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.890475 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dbc20081-c609-4ddf-a321-2ae092fd67e9-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz\" (UID: \"dbc20081-c609-4ddf-a321-2ae092fd67e9\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz" Nov 25 11:11:22 crc kubenswrapper[4776]: I1125 11:11:22.910406 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhp98\" (UniqueName: \"kubernetes.io/projected/dbc20081-c609-4ddf-a321-2ae092fd67e9-kube-api-access-vhp98\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz\" (UID: \"dbc20081-c609-4ddf-a321-2ae092fd67e9\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz" Nov 25 11:11:23 crc kubenswrapper[4776]: I1125 11:11:23.028723 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz" Nov 25 11:11:23 crc kubenswrapper[4776]: I1125 11:11:23.478674 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz"] Nov 25 11:11:23 crc kubenswrapper[4776]: W1125 11:11:23.484480 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddbc20081_c609_4ddf_a321_2ae092fd67e9.slice/crio-76d1e886d75fbf7a19eacca8486828606f99c8e2875cc2e1682f6f5fa656fc18 WatchSource:0}: Error finding container 76d1e886d75fbf7a19eacca8486828606f99c8e2875cc2e1682f6f5fa656fc18: Status 404 returned error can't find the container with id 76d1e886d75fbf7a19eacca8486828606f99c8e2875cc2e1682f6f5fa656fc18 Nov 25 11:11:23 crc kubenswrapper[4776]: I1125 11:11:23.706590 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz" event={"ID":"dbc20081-c609-4ddf-a321-2ae092fd67e9","Type":"ContainerStarted","Data":"83f4601d6a39ca5a91f05da229cffe609f47f0aa24683fa725a5b8527ddf0d93"} Nov 25 11:11:23 crc kubenswrapper[4776]: I1125 11:11:23.706646 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz" event={"ID":"dbc20081-c609-4ddf-a321-2ae092fd67e9","Type":"ContainerStarted","Data":"76d1e886d75fbf7a19eacca8486828606f99c8e2875cc2e1682f6f5fa656fc18"} Nov 25 11:11:24 crc kubenswrapper[4776]: I1125 11:11:24.718471 4776 generic.go:334] "Generic (PLEG): container finished" podID="dbc20081-c609-4ddf-a321-2ae092fd67e9" containerID="83f4601d6a39ca5a91f05da229cffe609f47f0aa24683fa725a5b8527ddf0d93" exitCode=0 Nov 25 11:11:24 crc kubenswrapper[4776]: I1125 11:11:24.718599 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz" event={"ID":"dbc20081-c609-4ddf-a321-2ae092fd67e9","Type":"ContainerDied","Data":"83f4601d6a39ca5a91f05da229cffe609f47f0aa24683fa725a5b8527ddf0d93"} Nov 25 11:11:25 crc kubenswrapper[4776]: I1125 11:11:25.035772 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-kz9jg"] Nov 25 11:11:25 crc kubenswrapper[4776]: I1125 11:11:25.047292 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-kz9jg"] Nov 25 11:11:25 crc kubenswrapper[4776]: I1125 11:11:25.674118 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcb0fc87-e3b2-475c-a203-aeda2c4e21da" path="/var/lib/kubelet/pods/fcb0fc87-e3b2-475c-a203-aeda2c4e21da/volumes" Nov 25 11:11:26 crc kubenswrapper[4776]: I1125 11:11:26.738542 4776 generic.go:334] "Generic (PLEG): container finished" podID="dbc20081-c609-4ddf-a321-2ae092fd67e9" containerID="9e76198c11920041d0d89aeaabb94b480847fbdf06b0d3c585ed12239d6c4e1a" exitCode=0 Nov 25 11:11:26 crc kubenswrapper[4776]: I1125 11:11:26.738654 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz" event={"ID":"dbc20081-c609-4ddf-a321-2ae092fd67e9","Type":"ContainerDied","Data":"9e76198c11920041d0d89aeaabb94b480847fbdf06b0d3c585ed12239d6c4e1a"} Nov 25 11:11:27 crc kubenswrapper[4776]: I1125 11:11:27.749335 4776 generic.go:334] "Generic (PLEG): container finished" podID="dbc20081-c609-4ddf-a321-2ae092fd67e9" containerID="41ece4cff2b34a1709898440e554f958c79678855e7382e56633791aa16ad374" exitCode=0 Nov 25 11:11:27 crc kubenswrapper[4776]: I1125 11:11:27.749463 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz" event={"ID":"dbc20081-c609-4ddf-a321-2ae092fd67e9","Type":"ContainerDied","Data":"41ece4cff2b34a1709898440e554f958c79678855e7382e56633791aa16ad374"} Nov 25 11:11:29 crc kubenswrapper[4776]: I1125 11:11:29.144034 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz" Nov 25 11:11:29 crc kubenswrapper[4776]: I1125 11:11:29.213926 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dbc20081-c609-4ddf-a321-2ae092fd67e9-util\") pod \"dbc20081-c609-4ddf-a321-2ae092fd67e9\" (UID: \"dbc20081-c609-4ddf-a321-2ae092fd67e9\") " Nov 25 11:11:29 crc kubenswrapper[4776]: I1125 11:11:29.214019 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dbc20081-c609-4ddf-a321-2ae092fd67e9-bundle\") pod \"dbc20081-c609-4ddf-a321-2ae092fd67e9\" (UID: \"dbc20081-c609-4ddf-a321-2ae092fd67e9\") " Nov 25 11:11:29 crc kubenswrapper[4776]: I1125 11:11:29.214129 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhp98\" (UniqueName: \"kubernetes.io/projected/dbc20081-c609-4ddf-a321-2ae092fd67e9-kube-api-access-vhp98\") pod \"dbc20081-c609-4ddf-a321-2ae092fd67e9\" (UID: \"dbc20081-c609-4ddf-a321-2ae092fd67e9\") " Nov 25 11:11:29 crc kubenswrapper[4776]: I1125 11:11:29.216617 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbc20081-c609-4ddf-a321-2ae092fd67e9-bundle" (OuterVolumeSpecName: "bundle") pod "dbc20081-c609-4ddf-a321-2ae092fd67e9" (UID: "dbc20081-c609-4ddf-a321-2ae092fd67e9"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:11:29 crc kubenswrapper[4776]: I1125 11:11:29.220141 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbc20081-c609-4ddf-a321-2ae092fd67e9-kube-api-access-vhp98" (OuterVolumeSpecName: "kube-api-access-vhp98") pod "dbc20081-c609-4ddf-a321-2ae092fd67e9" (UID: "dbc20081-c609-4ddf-a321-2ae092fd67e9"). InnerVolumeSpecName "kube-api-access-vhp98". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:11:29 crc kubenswrapper[4776]: I1125 11:11:29.316526 4776 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dbc20081-c609-4ddf-a321-2ae092fd67e9-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:29 crc kubenswrapper[4776]: I1125 11:11:29.316570 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhp98\" (UniqueName: \"kubernetes.io/projected/dbc20081-c609-4ddf-a321-2ae092fd67e9-kube-api-access-vhp98\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:29 crc kubenswrapper[4776]: I1125 11:11:29.478698 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbc20081-c609-4ddf-a321-2ae092fd67e9-util" (OuterVolumeSpecName: "util") pod "dbc20081-c609-4ddf-a321-2ae092fd67e9" (UID: "dbc20081-c609-4ddf-a321-2ae092fd67e9"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:11:29 crc kubenswrapper[4776]: I1125 11:11:29.520341 4776 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dbc20081-c609-4ddf-a321-2ae092fd67e9-util\") on node \"crc\" DevicePath \"\"" Nov 25 11:11:29 crc kubenswrapper[4776]: I1125 11:11:29.767145 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz" event={"ID":"dbc20081-c609-4ddf-a321-2ae092fd67e9","Type":"ContainerDied","Data":"76d1e886d75fbf7a19eacca8486828606f99c8e2875cc2e1682f6f5fa656fc18"} Nov 25 11:11:29 crc kubenswrapper[4776]: I1125 11:11:29.767445 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76d1e886d75fbf7a19eacca8486828606f99c8e2875cc2e1682f6f5fa656fc18" Nov 25 11:11:29 crc kubenswrapper[4776]: I1125 11:11:29.767200 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz" Nov 25 11:11:34 crc kubenswrapper[4776]: I1125 11:11:34.955118 4776 scope.go:117] "RemoveContainer" containerID="9cdb2b9569c3e16ed00198fe74fc270e2aedadd09e725000eb60c8f716c2afcc" Nov 25 11:11:34 crc kubenswrapper[4776]: I1125 11:11:34.989845 4776 scope.go:117] "RemoveContainer" containerID="fd5d2241a662f5b682bc3191a53df6f5849ca0002fef8d45ee81d122a9b632d3" Nov 25 11:11:35 crc kubenswrapper[4776]: I1125 11:11:35.065742 4776 scope.go:117] "RemoveContainer" containerID="ac78b904326fa8bea78f7c60190b901320665430e39ce84ba4d6a49f0cbc8473" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.570964 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-n9rjm"] Nov 25 11:11:40 crc kubenswrapper[4776]: E1125 11:11:40.571970 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbc20081-c609-4ddf-a321-2ae092fd67e9" containerName="extract" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.571984 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbc20081-c609-4ddf-a321-2ae092fd67e9" containerName="extract" Nov 25 11:11:40 crc kubenswrapper[4776]: E1125 11:11:40.571998 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbc20081-c609-4ddf-a321-2ae092fd67e9" containerName="util" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.572005 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbc20081-c609-4ddf-a321-2ae092fd67e9" containerName="util" Nov 25 11:11:40 crc kubenswrapper[4776]: E1125 11:11:40.572026 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbc20081-c609-4ddf-a321-2ae092fd67e9" containerName="pull" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.572034 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbc20081-c609-4ddf-a321-2ae092fd67e9" containerName="pull" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.572507 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbc20081-c609-4ddf-a321-2ae092fd67e9" containerName="extract" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.573359 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-n9rjm" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.580539 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.580680 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.580760 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-mgnpw" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.623490 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-n9rjm"] Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.652218 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5f4zd\" (UniqueName: \"kubernetes.io/projected/05d15308-0620-46cf-838e-1c8da434b69f-kube-api-access-5f4zd\") pod \"obo-prometheus-operator-668cf9dfbb-n9rjm\" (UID: \"05d15308-0620-46cf-838e-1c8da434b69f\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-n9rjm" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.706934 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-544ff7697f-l2d6r"] Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.708345 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-544ff7697f-l2d6r" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.710101 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-b95cn" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.721644 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.725760 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-544ff7697f-mdz99"] Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.729492 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-544ff7697f-mdz99" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.738495 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-544ff7697f-l2d6r"] Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.751437 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-544ff7697f-mdz99"] Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.754314 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/404823ba-8531-4881-a133-e3900d9bd6c8-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-544ff7697f-l2d6r\" (UID: \"404823ba-8531-4881-a133-e3900d9bd6c8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-544ff7697f-l2d6r" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.754649 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/404823ba-8531-4881-a133-e3900d9bd6c8-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-544ff7697f-l2d6r\" (UID: \"404823ba-8531-4881-a133-e3900d9bd6c8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-544ff7697f-l2d6r" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.754723 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/981f87ca-c762-4f02-b2bf-b22732edfd22-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-544ff7697f-mdz99\" (UID: \"981f87ca-c762-4f02-b2bf-b22732edfd22\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-544ff7697f-mdz99" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.754842 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5f4zd\" (UniqueName: \"kubernetes.io/projected/05d15308-0620-46cf-838e-1c8da434b69f-kube-api-access-5f4zd\") pod \"obo-prometheus-operator-668cf9dfbb-n9rjm\" (UID: \"05d15308-0620-46cf-838e-1c8da434b69f\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-n9rjm" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.758722 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/981f87ca-c762-4f02-b2bf-b22732edfd22-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-544ff7697f-mdz99\" (UID: \"981f87ca-c762-4f02-b2bf-b22732edfd22\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-544ff7697f-mdz99" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.800781 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5f4zd\" (UniqueName: \"kubernetes.io/projected/05d15308-0620-46cf-838e-1c8da434b69f-kube-api-access-5f4zd\") pod \"obo-prometheus-operator-668cf9dfbb-n9rjm\" (UID: \"05d15308-0620-46cf-838e-1c8da434b69f\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-n9rjm" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.860430 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/404823ba-8531-4881-a133-e3900d9bd6c8-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-544ff7697f-l2d6r\" (UID: \"404823ba-8531-4881-a133-e3900d9bd6c8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-544ff7697f-l2d6r" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.860543 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/404823ba-8531-4881-a133-e3900d9bd6c8-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-544ff7697f-l2d6r\" (UID: \"404823ba-8531-4881-a133-e3900d9bd6c8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-544ff7697f-l2d6r" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.860572 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/981f87ca-c762-4f02-b2bf-b22732edfd22-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-544ff7697f-mdz99\" (UID: \"981f87ca-c762-4f02-b2bf-b22732edfd22\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-544ff7697f-mdz99" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.860675 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/981f87ca-c762-4f02-b2bf-b22732edfd22-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-544ff7697f-mdz99\" (UID: \"981f87ca-c762-4f02-b2bf-b22732edfd22\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-544ff7697f-mdz99" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.864424 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/981f87ca-c762-4f02-b2bf-b22732edfd22-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-544ff7697f-mdz99\" (UID: \"981f87ca-c762-4f02-b2bf-b22732edfd22\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-544ff7697f-mdz99" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.864424 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/981f87ca-c762-4f02-b2bf-b22732edfd22-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-544ff7697f-mdz99\" (UID: \"981f87ca-c762-4f02-b2bf-b22732edfd22\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-544ff7697f-mdz99" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.864893 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/404823ba-8531-4881-a133-e3900d9bd6c8-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-544ff7697f-l2d6r\" (UID: \"404823ba-8531-4881-a133-e3900d9bd6c8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-544ff7697f-l2d6r" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.889295 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/404823ba-8531-4881-a133-e3900d9bd6c8-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-544ff7697f-l2d6r\" (UID: \"404823ba-8531-4881-a133-e3900d9bd6c8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-544ff7697f-l2d6r" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.904771 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-n9rjm" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.906250 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-6hnbp"] Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.907487 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-6hnbp" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.911256 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.911398 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-qnk6v" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.928107 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-6hnbp"] Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.964261 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/79fbf35b-52fb-4f94-89b9-d5b257b86de6-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-6hnbp\" (UID: \"79fbf35b-52fb-4f94-89b9-d5b257b86de6\") " pod="openshift-operators/observability-operator-d8bb48f5d-6hnbp" Nov 25 11:11:40 crc kubenswrapper[4776]: I1125 11:11:40.964362 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dk2fz\" (UniqueName: \"kubernetes.io/projected/79fbf35b-52fb-4f94-89b9-d5b257b86de6-kube-api-access-dk2fz\") pod \"observability-operator-d8bb48f5d-6hnbp\" (UID: \"79fbf35b-52fb-4f94-89b9-d5b257b86de6\") " pod="openshift-operators/observability-operator-d8bb48f5d-6hnbp" Nov 25 11:11:41 crc kubenswrapper[4776]: I1125 11:11:41.032505 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-544ff7697f-l2d6r" Nov 25 11:11:41 crc kubenswrapper[4776]: I1125 11:11:41.057210 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-544ff7697f-mdz99" Nov 25 11:11:41 crc kubenswrapper[4776]: I1125 11:11:41.069328 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/79fbf35b-52fb-4f94-89b9-d5b257b86de6-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-6hnbp\" (UID: \"79fbf35b-52fb-4f94-89b9-d5b257b86de6\") " pod="openshift-operators/observability-operator-d8bb48f5d-6hnbp" Nov 25 11:11:41 crc kubenswrapper[4776]: I1125 11:11:41.069487 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dk2fz\" (UniqueName: \"kubernetes.io/projected/79fbf35b-52fb-4f94-89b9-d5b257b86de6-kube-api-access-dk2fz\") pod \"observability-operator-d8bb48f5d-6hnbp\" (UID: \"79fbf35b-52fb-4f94-89b9-d5b257b86de6\") " pod="openshift-operators/observability-operator-d8bb48f5d-6hnbp" Nov 25 11:11:41 crc kubenswrapper[4776]: I1125 11:11:41.106497 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dk2fz\" (UniqueName: \"kubernetes.io/projected/79fbf35b-52fb-4f94-89b9-d5b257b86de6-kube-api-access-dk2fz\") pod \"observability-operator-d8bb48f5d-6hnbp\" (UID: \"79fbf35b-52fb-4f94-89b9-d5b257b86de6\") " pod="openshift-operators/observability-operator-d8bb48f5d-6hnbp" Nov 25 11:11:41 crc kubenswrapper[4776]: I1125 11:11:41.129750 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/79fbf35b-52fb-4f94-89b9-d5b257b86de6-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-6hnbp\" (UID: \"79fbf35b-52fb-4f94-89b9-d5b257b86de6\") " pod="openshift-operators/observability-operator-d8bb48f5d-6hnbp" Nov 25 11:11:41 crc kubenswrapper[4776]: I1125 11:11:41.153893 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-zr5h9"] Nov 25 11:11:41 crc kubenswrapper[4776]: I1125 11:11:41.155327 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-zr5h9" Nov 25 11:11:41 crc kubenswrapper[4776]: I1125 11:11:41.160932 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-x99zv" Nov 25 11:11:41 crc kubenswrapper[4776]: I1125 11:11:41.212426 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-zr5h9"] Nov 25 11:11:41 crc kubenswrapper[4776]: I1125 11:11:41.284401 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/06bd1f63-0261-44b4-91ca-483b379c21d8-openshift-service-ca\") pod \"perses-operator-5446b9c989-zr5h9\" (UID: \"06bd1f63-0261-44b4-91ca-483b379c21d8\") " pod="openshift-operators/perses-operator-5446b9c989-zr5h9" Nov 25 11:11:41 crc kubenswrapper[4776]: I1125 11:11:41.284455 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsfdk\" (UniqueName: \"kubernetes.io/projected/06bd1f63-0261-44b4-91ca-483b379c21d8-kube-api-access-tsfdk\") pod \"perses-operator-5446b9c989-zr5h9\" (UID: \"06bd1f63-0261-44b4-91ca-483b379c21d8\") " pod="openshift-operators/perses-operator-5446b9c989-zr5h9" Nov 25 11:11:41 crc kubenswrapper[4776]: I1125 11:11:41.379001 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-6hnbp" Nov 25 11:11:41 crc kubenswrapper[4776]: I1125 11:11:41.387557 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/06bd1f63-0261-44b4-91ca-483b379c21d8-openshift-service-ca\") pod \"perses-operator-5446b9c989-zr5h9\" (UID: \"06bd1f63-0261-44b4-91ca-483b379c21d8\") " pod="openshift-operators/perses-operator-5446b9c989-zr5h9" Nov 25 11:11:41 crc kubenswrapper[4776]: I1125 11:11:41.387630 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsfdk\" (UniqueName: \"kubernetes.io/projected/06bd1f63-0261-44b4-91ca-483b379c21d8-kube-api-access-tsfdk\") pod \"perses-operator-5446b9c989-zr5h9\" (UID: \"06bd1f63-0261-44b4-91ca-483b379c21d8\") " pod="openshift-operators/perses-operator-5446b9c989-zr5h9" Nov 25 11:11:41 crc kubenswrapper[4776]: I1125 11:11:41.389472 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/06bd1f63-0261-44b4-91ca-483b379c21d8-openshift-service-ca\") pod \"perses-operator-5446b9c989-zr5h9\" (UID: \"06bd1f63-0261-44b4-91ca-483b379c21d8\") " pod="openshift-operators/perses-operator-5446b9c989-zr5h9" Nov 25 11:11:41 crc kubenswrapper[4776]: I1125 11:11:41.420872 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsfdk\" (UniqueName: \"kubernetes.io/projected/06bd1f63-0261-44b4-91ca-483b379c21d8-kube-api-access-tsfdk\") pod \"perses-operator-5446b9c989-zr5h9\" (UID: \"06bd1f63-0261-44b4-91ca-483b379c21d8\") " pod="openshift-operators/perses-operator-5446b9c989-zr5h9" Nov 25 11:11:41 crc kubenswrapper[4776]: W1125 11:11:41.514046 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05d15308_0620_46cf_838e_1c8da434b69f.slice/crio-312e659e4fc44820c46482eb7efe141a5def7cf446081a8b28ead0fe8f512436 WatchSource:0}: Error finding container 312e659e4fc44820c46482eb7efe141a5def7cf446081a8b28ead0fe8f512436: Status 404 returned error can't find the container with id 312e659e4fc44820c46482eb7efe141a5def7cf446081a8b28ead0fe8f512436 Nov 25 11:11:41 crc kubenswrapper[4776]: I1125 11:11:41.514606 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-zr5h9" Nov 25 11:11:41 crc kubenswrapper[4776]: I1125 11:11:41.514628 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-n9rjm"] Nov 25 11:11:41 crc kubenswrapper[4776]: W1125 11:11:41.709746 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod404823ba_8531_4881_a133_e3900d9bd6c8.slice/crio-41fc28c57c019fc0b65fceaec3daf024e795c0dd33f9525e2d8a68cb1a3e265f WatchSource:0}: Error finding container 41fc28c57c019fc0b65fceaec3daf024e795c0dd33f9525e2d8a68cb1a3e265f: Status 404 returned error can't find the container with id 41fc28c57c019fc0b65fceaec3daf024e795c0dd33f9525e2d8a68cb1a3e265f Nov 25 11:11:41 crc kubenswrapper[4776]: I1125 11:11:41.725103 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-544ff7697f-mdz99"] Nov 25 11:11:41 crc kubenswrapper[4776]: I1125 11:11:41.749313 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-544ff7697f-l2d6r"] Nov 25 11:11:41 crc kubenswrapper[4776]: I1125 11:11:41.957497 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-544ff7697f-mdz99" event={"ID":"981f87ca-c762-4f02-b2bf-b22732edfd22","Type":"ContainerStarted","Data":"e95b8e705a4782a0e0e3ec622fddbcefc33dc3a8bf2f7858315c3bd6583115b3"} Nov 25 11:11:41 crc kubenswrapper[4776]: I1125 11:11:41.959928 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-n9rjm" event={"ID":"05d15308-0620-46cf-838e-1c8da434b69f","Type":"ContainerStarted","Data":"312e659e4fc44820c46482eb7efe141a5def7cf446081a8b28ead0fe8f512436"} Nov 25 11:11:41 crc kubenswrapper[4776]: I1125 11:11:41.962100 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-544ff7697f-l2d6r" event={"ID":"404823ba-8531-4881-a133-e3900d9bd6c8","Type":"ContainerStarted","Data":"41fc28c57c019fc0b65fceaec3daf024e795c0dd33f9525e2d8a68cb1a3e265f"} Nov 25 11:11:42 crc kubenswrapper[4776]: I1125 11:11:42.003320 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-6hnbp"] Nov 25 11:11:42 crc kubenswrapper[4776]: I1125 11:11:42.176341 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-zr5h9"] Nov 25 11:11:42 crc kubenswrapper[4776]: W1125 11:11:42.178531 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06bd1f63_0261_44b4_91ca_483b379c21d8.slice/crio-ad1c3688c9e174dffcb2c321d26cbfef7b5ee22082bf32cabce9b430de7aef30 WatchSource:0}: Error finding container ad1c3688c9e174dffcb2c321d26cbfef7b5ee22082bf32cabce9b430de7aef30: Status 404 returned error can't find the container with id ad1c3688c9e174dffcb2c321d26cbfef7b5ee22082bf32cabce9b430de7aef30 Nov 25 11:11:42 crc kubenswrapper[4776]: I1125 11:11:42.996727 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-6hnbp" event={"ID":"79fbf35b-52fb-4f94-89b9-d5b257b86de6","Type":"ContainerStarted","Data":"8a435eca749c3c11bef6bed986e4edfd6c51fa7fd01cab99a72d77b4516d0c64"} Nov 25 11:11:42 crc kubenswrapper[4776]: I1125 11:11:42.998670 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-zr5h9" event={"ID":"06bd1f63-0261-44b4-91ca-483b379c21d8","Type":"ContainerStarted","Data":"ad1c3688c9e174dffcb2c321d26cbfef7b5ee22082bf32cabce9b430de7aef30"} Nov 25 11:11:47 crc kubenswrapper[4776]: I1125 11:11:47.821482 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:11:47 crc kubenswrapper[4776]: I1125 11:11:47.821941 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:11:51 crc kubenswrapper[4776]: I1125 11:11:51.124416 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-544ff7697f-l2d6r" event={"ID":"404823ba-8531-4881-a133-e3900d9bd6c8","Type":"ContainerStarted","Data":"dc1691aeccf048f6e52c317dac3e45cec4b6e9be013737fa93d06bee2d68639b"} Nov 25 11:11:51 crc kubenswrapper[4776]: I1125 11:11:51.127885 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-544ff7697f-mdz99" event={"ID":"981f87ca-c762-4f02-b2bf-b22732edfd22","Type":"ContainerStarted","Data":"e9247305141a80bee6eb6179c8d128a007f3de5944d7620047ff12cff301d40c"} Nov 25 11:11:51 crc kubenswrapper[4776]: I1125 11:11:51.130949 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-n9rjm" event={"ID":"05d15308-0620-46cf-838e-1c8da434b69f","Type":"ContainerStarted","Data":"ee7dd381d9b3ba751f9002e267b0d7a4e832ed2011c6d3744a90f5241b1357ed"} Nov 25 11:11:51 crc kubenswrapper[4776]: I1125 11:11:51.132788 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-zr5h9" event={"ID":"06bd1f63-0261-44b4-91ca-483b379c21d8","Type":"ContainerStarted","Data":"fea0bdc58690104bbd52cd18fe8799beaf68a7dd26f5b55c043f57876dd28e06"} Nov 25 11:11:51 crc kubenswrapper[4776]: I1125 11:11:51.133092 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-zr5h9" Nov 25 11:11:51 crc kubenswrapper[4776]: I1125 11:11:51.147489 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-544ff7697f-l2d6r" podStartSLOduration=2.563981344 podStartE2EDuration="11.147471469s" podCreationTimestamp="2025-11-25 11:11:40 +0000 UTC" firstStartedPulling="2025-11-25 11:11:41.714011112 +0000 UTC m=+6446.755070665" lastFinishedPulling="2025-11-25 11:11:50.297501237 +0000 UTC m=+6455.338560790" observedRunningTime="2025-11-25 11:11:51.141207502 +0000 UTC m=+6456.182267055" watchObservedRunningTime="2025-11-25 11:11:51.147471469 +0000 UTC m=+6456.188531022" Nov 25 11:11:51 crc kubenswrapper[4776]: I1125 11:11:51.184210 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-zr5h9" podStartSLOduration=2.012700953 podStartE2EDuration="10.18418933s" podCreationTimestamp="2025-11-25 11:11:41 +0000 UTC" firstStartedPulling="2025-11-25 11:11:42.187984155 +0000 UTC m=+6447.229043708" lastFinishedPulling="2025-11-25 11:11:50.359472532 +0000 UTC m=+6455.400532085" observedRunningTime="2025-11-25 11:11:51.17898514 +0000 UTC m=+6456.220044703" watchObservedRunningTime="2025-11-25 11:11:51.18418933 +0000 UTC m=+6456.225248893" Nov 25 11:11:51 crc kubenswrapper[4776]: I1125 11:11:51.206750 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-544ff7697f-mdz99" podStartSLOduration=2.634125084 podStartE2EDuration="11.206729126s" podCreationTimestamp="2025-11-25 11:11:40 +0000 UTC" firstStartedPulling="2025-11-25 11:11:41.723977662 +0000 UTC m=+6446.765037215" lastFinishedPulling="2025-11-25 11:11:50.296581704 +0000 UTC m=+6455.337641257" observedRunningTime="2025-11-25 11:11:51.199780631 +0000 UTC m=+6456.240840204" watchObservedRunningTime="2025-11-25 11:11:51.206729126 +0000 UTC m=+6456.247788679" Nov 25 11:11:52 crc kubenswrapper[4776]: I1125 11:11:52.178713 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-n9rjm" podStartSLOduration=3.435637959 podStartE2EDuration="12.178695099s" podCreationTimestamp="2025-11-25 11:11:40 +0000 UTC" firstStartedPulling="2025-11-25 11:11:41.553451412 +0000 UTC m=+6446.594510965" lastFinishedPulling="2025-11-25 11:11:50.296508542 +0000 UTC m=+6455.337568105" observedRunningTime="2025-11-25 11:11:52.174047963 +0000 UTC m=+6457.215107516" watchObservedRunningTime="2025-11-25 11:11:52.178695099 +0000 UTC m=+6457.219754652" Nov 25 11:11:53 crc kubenswrapper[4776]: I1125 11:11:53.162038 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-6hnbp" event={"ID":"79fbf35b-52fb-4f94-89b9-d5b257b86de6","Type":"ContainerStarted","Data":"786d65e472a6e99aa8259e0128eb9cbd89e0bbad3771940032fc5bca9777ba75"} Nov 25 11:11:53 crc kubenswrapper[4776]: I1125 11:11:53.164238 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-6hnbp" Nov 25 11:11:53 crc kubenswrapper[4776]: I1125 11:11:53.187004 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-6hnbp" podStartSLOduration=3.413916001 podStartE2EDuration="13.186984333s" podCreationTimestamp="2025-11-25 11:11:40 +0000 UTC" firstStartedPulling="2025-11-25 11:11:42.019953448 +0000 UTC m=+6447.061013001" lastFinishedPulling="2025-11-25 11:11:51.79302178 +0000 UTC m=+6456.834081333" observedRunningTime="2025-11-25 11:11:53.181768762 +0000 UTC m=+6458.222828315" watchObservedRunningTime="2025-11-25 11:11:53.186984333 +0000 UTC m=+6458.228043886" Nov 25 11:11:53 crc kubenswrapper[4776]: I1125 11:11:53.257243 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-6hnbp" Nov 25 11:11:55 crc kubenswrapper[4776]: I1125 11:11:55.048516 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-4595-account-create-rrgbl"] Nov 25 11:11:55 crc kubenswrapper[4776]: I1125 11:11:55.060359 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-78bsr"] Nov 25 11:11:55 crc kubenswrapper[4776]: I1125 11:11:55.070681 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-4595-account-create-rrgbl"] Nov 25 11:11:55 crc kubenswrapper[4776]: I1125 11:11:55.081789 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-78bsr"] Nov 25 11:11:55 crc kubenswrapper[4776]: I1125 11:11:55.675887 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59ac55fd-b5a0-47f9-aae2-9d1384594315" path="/var/lib/kubelet/pods/59ac55fd-b5a0-47f9-aae2-9d1384594315/volumes" Nov 25 11:11:55 crc kubenswrapper[4776]: I1125 11:11:55.677687 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="865436be-d605-4159-a85e-38a7403703e9" path="/var/lib/kubelet/pods/865436be-d605-4159-a85e-38a7403703e9/volumes" Nov 25 11:12:01 crc kubenswrapper[4776]: I1125 11:12:01.518682 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-zr5h9" Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.102958 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.103603 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="ee61762a-4aa0-4ae9-9107-fca99430ccfa" containerName="openstackclient" containerID="cri-o://b3605f75d8bd3af002c79d3388d36f17448d7027115745579f13ee502a0e5ded" gracePeriod=2 Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.119938 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.203041 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 25 11:12:04 crc kubenswrapper[4776]: E1125 11:12:04.203571 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee61762a-4aa0-4ae9-9107-fca99430ccfa" containerName="openstackclient" Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.203590 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee61762a-4aa0-4ae9-9107-fca99430ccfa" containerName="openstackclient" Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.203871 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee61762a-4aa0-4ae9-9107-fca99430ccfa" containerName="openstackclient" Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.209006 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.220584 4776 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="ee61762a-4aa0-4ae9-9107-fca99430ccfa" podUID="dae628f6-9079-4f73-b208-0f8fd6d43af0" Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.222656 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.259167 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dae628f6-9079-4f73-b208-0f8fd6d43af0-openstack-config\") pod \"openstackclient\" (UID: \"dae628f6-9079-4f73-b208-0f8fd6d43af0\") " pod="openstack/openstackclient" Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.259250 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dae628f6-9079-4f73-b208-0f8fd6d43af0-openstack-config-secret\") pod \"openstackclient\" (UID: \"dae628f6-9079-4f73-b208-0f8fd6d43af0\") " pod="openstack/openstackclient" Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.259292 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dae628f6-9079-4f73-b208-0f8fd6d43af0-combined-ca-bundle\") pod \"openstackclient\" (UID: \"dae628f6-9079-4f73-b208-0f8fd6d43af0\") " pod="openstack/openstackclient" Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.259342 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4tcj\" (UniqueName: \"kubernetes.io/projected/dae628f6-9079-4f73-b208-0f8fd6d43af0-kube-api-access-r4tcj\") pod \"openstackclient\" (UID: \"dae628f6-9079-4f73-b208-0f8fd6d43af0\") " pod="openstack/openstackclient" Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.355145 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.356673 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.361221 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dae628f6-9079-4f73-b208-0f8fd6d43af0-openstack-config\") pod \"openstackclient\" (UID: \"dae628f6-9079-4f73-b208-0f8fd6d43af0\") " pod="openstack/openstackclient" Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.361303 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dae628f6-9079-4f73-b208-0f8fd6d43af0-openstack-config-secret\") pod \"openstackclient\" (UID: \"dae628f6-9079-4f73-b208-0f8fd6d43af0\") " pod="openstack/openstackclient" Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.361346 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dae628f6-9079-4f73-b208-0f8fd6d43af0-combined-ca-bundle\") pod \"openstackclient\" (UID: \"dae628f6-9079-4f73-b208-0f8fd6d43af0\") " pod="openstack/openstackclient" Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.361386 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4tcj\" (UniqueName: \"kubernetes.io/projected/dae628f6-9079-4f73-b208-0f8fd6d43af0-kube-api-access-r4tcj\") pod \"openstackclient\" (UID: \"dae628f6-9079-4f73-b208-0f8fd6d43af0\") " pod="openstack/openstackclient" Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.361449 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-5pqqp" Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.362903 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dae628f6-9079-4f73-b208-0f8fd6d43af0-openstack-config\") pod \"openstackclient\" (UID: \"dae628f6-9079-4f73-b208-0f8fd6d43af0\") " pod="openstack/openstackclient" Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.370236 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dae628f6-9079-4f73-b208-0f8fd6d43af0-openstack-config-secret\") pod \"openstackclient\" (UID: \"dae628f6-9079-4f73-b208-0f8fd6d43af0\") " pod="openstack/openstackclient" Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.375811 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dae628f6-9079-4f73-b208-0f8fd6d43af0-combined-ca-bundle\") pod \"openstackclient\" (UID: \"dae628f6-9079-4f73-b208-0f8fd6d43af0\") " pod="openstack/openstackclient" Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.406212 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.465183 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fljtn\" (UniqueName: \"kubernetes.io/projected/1e161da1-d800-4a51-a70a-cfb8974b11b9-kube-api-access-fljtn\") pod \"kube-state-metrics-0\" (UID: \"1e161da1-d800-4a51-a70a-cfb8974b11b9\") " pod="openstack/kube-state-metrics-0" Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.476614 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4tcj\" (UniqueName: \"kubernetes.io/projected/dae628f6-9079-4f73-b208-0f8fd6d43af0-kube-api-access-r4tcj\") pod \"openstackclient\" (UID: \"dae628f6-9079-4f73-b208-0f8fd6d43af0\") " pod="openstack/openstackclient" Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.562604 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.566934 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fljtn\" (UniqueName: \"kubernetes.io/projected/1e161da1-d800-4a51-a70a-cfb8974b11b9-kube-api-access-fljtn\") pod \"kube-state-metrics-0\" (UID: \"1e161da1-d800-4a51-a70a-cfb8974b11b9\") " pod="openstack/kube-state-metrics-0" Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.662478 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fljtn\" (UniqueName: \"kubernetes.io/projected/1e161da1-d800-4a51-a70a-cfb8974b11b9-kube-api-access-fljtn\") pod \"kube-state-metrics-0\" (UID: \"1e161da1-d800-4a51-a70a-cfb8974b11b9\") " pod="openstack/kube-state-metrics-0" Nov 25 11:12:04 crc kubenswrapper[4776]: I1125 11:12:04.847942 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 11:12:05 crc kubenswrapper[4776]: I1125 11:12:05.160018 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-gdfkv"] Nov 25 11:12:05 crc kubenswrapper[4776]: I1125 11:12:05.186573 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-gdfkv"] Nov 25 11:12:05 crc kubenswrapper[4776]: I1125 11:12:05.752930 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5b03bdd-3ede-487b-8403-073190b9f68b" path="/var/lib/kubelet/pods/b5b03bdd-3ede-487b-8403-073190b9f68b/volumes" Nov 25 11:12:05 crc kubenswrapper[4776]: I1125 11:12:05.809292 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 25 11:12:05 crc kubenswrapper[4776]: I1125 11:12:05.862523 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Nov 25 11:12:05 crc kubenswrapper[4776]: I1125 11:12:05.865785 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Nov 25 11:12:05 crc kubenswrapper[4776]: I1125 11:12:05.868041 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-cluster-tls-config" Nov 25 11:12:05 crc kubenswrapper[4776]: I1125 11:12:05.868222 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Nov 25 11:12:05 crc kubenswrapper[4776]: I1125 11:12:05.868489 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-btshg" Nov 25 11:12:05 crc kubenswrapper[4776]: I1125 11:12:05.868822 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Nov 25 11:12:05 crc kubenswrapper[4776]: I1125 11:12:05.868986 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Nov 25 11:12:05 crc kubenswrapper[4776]: I1125 11:12:05.930282 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Nov 25 11:12:05 crc kubenswrapper[4776]: I1125 11:12:05.969221 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8l28\" (UniqueName: \"kubernetes.io/projected/464f639d-26e5-4d21-9d9b-c84585fbd8e2-kube-api-access-k8l28\") pod \"alertmanager-metric-storage-0\" (UID: \"464f639d-26e5-4d21-9d9b-c84585fbd8e2\") " pod="openstack/alertmanager-metric-storage-0" Nov 25 11:12:05 crc kubenswrapper[4776]: I1125 11:12:05.969280 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/464f639d-26e5-4d21-9d9b-c84585fbd8e2-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"464f639d-26e5-4d21-9d9b-c84585fbd8e2\") " pod="openstack/alertmanager-metric-storage-0" Nov 25 11:12:05 crc kubenswrapper[4776]: I1125 11:12:05.969308 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/464f639d-26e5-4d21-9d9b-c84585fbd8e2-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"464f639d-26e5-4d21-9d9b-c84585fbd8e2\") " pod="openstack/alertmanager-metric-storage-0" Nov 25 11:12:05 crc kubenswrapper[4776]: I1125 11:12:05.969362 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/464f639d-26e5-4d21-9d9b-c84585fbd8e2-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"464f639d-26e5-4d21-9d9b-c84585fbd8e2\") " pod="openstack/alertmanager-metric-storage-0" Nov 25 11:12:05 crc kubenswrapper[4776]: I1125 11:12:05.969425 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/464f639d-26e5-4d21-9d9b-c84585fbd8e2-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"464f639d-26e5-4d21-9d9b-c84585fbd8e2\") " pod="openstack/alertmanager-metric-storage-0" Nov 25 11:12:05 crc kubenswrapper[4776]: I1125 11:12:05.969478 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/464f639d-26e5-4d21-9d9b-c84585fbd8e2-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"464f639d-26e5-4d21-9d9b-c84585fbd8e2\") " pod="openstack/alertmanager-metric-storage-0" Nov 25 11:12:05 crc kubenswrapper[4776]: I1125 11:12:05.969534 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/464f639d-26e5-4d21-9d9b-c84585fbd8e2-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"464f639d-26e5-4d21-9d9b-c84585fbd8e2\") " pod="openstack/alertmanager-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.078454 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/464f639d-26e5-4d21-9d9b-c84585fbd8e2-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"464f639d-26e5-4d21-9d9b-c84585fbd8e2\") " pod="openstack/alertmanager-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.078584 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/464f639d-26e5-4d21-9d9b-c84585fbd8e2-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"464f639d-26e5-4d21-9d9b-c84585fbd8e2\") " pod="openstack/alertmanager-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.078652 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/464f639d-26e5-4d21-9d9b-c84585fbd8e2-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"464f639d-26e5-4d21-9d9b-c84585fbd8e2\") " pod="openstack/alertmanager-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.078730 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/464f639d-26e5-4d21-9d9b-c84585fbd8e2-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"464f639d-26e5-4d21-9d9b-c84585fbd8e2\") " pod="openstack/alertmanager-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.078762 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8l28\" (UniqueName: \"kubernetes.io/projected/464f639d-26e5-4d21-9d9b-c84585fbd8e2-kube-api-access-k8l28\") pod \"alertmanager-metric-storage-0\" (UID: \"464f639d-26e5-4d21-9d9b-c84585fbd8e2\") " pod="openstack/alertmanager-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.078786 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/464f639d-26e5-4d21-9d9b-c84585fbd8e2-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"464f639d-26e5-4d21-9d9b-c84585fbd8e2\") " pod="openstack/alertmanager-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.078809 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/464f639d-26e5-4d21-9d9b-c84585fbd8e2-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"464f639d-26e5-4d21-9d9b-c84585fbd8e2\") " pod="openstack/alertmanager-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.100408 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/464f639d-26e5-4d21-9d9b-c84585fbd8e2-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"464f639d-26e5-4d21-9d9b-c84585fbd8e2\") " pod="openstack/alertmanager-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.102937 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/464f639d-26e5-4d21-9d9b-c84585fbd8e2-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"464f639d-26e5-4d21-9d9b-c84585fbd8e2\") " pod="openstack/alertmanager-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.125024 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/464f639d-26e5-4d21-9d9b-c84585fbd8e2-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"464f639d-26e5-4d21-9d9b-c84585fbd8e2\") " pod="openstack/alertmanager-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.125024 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/464f639d-26e5-4d21-9d9b-c84585fbd8e2-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"464f639d-26e5-4d21-9d9b-c84585fbd8e2\") " pod="openstack/alertmanager-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.129768 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/464f639d-26e5-4d21-9d9b-c84585fbd8e2-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"464f639d-26e5-4d21-9d9b-c84585fbd8e2\") " pod="openstack/alertmanager-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.131372 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/464f639d-26e5-4d21-9d9b-c84585fbd8e2-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"464f639d-26e5-4d21-9d9b-c84585fbd8e2\") " pod="openstack/alertmanager-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.178634 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8l28\" (UniqueName: \"kubernetes.io/projected/464f639d-26e5-4d21-9d9b-c84585fbd8e2-kube-api-access-k8l28\") pod \"alertmanager-metric-storage-0\" (UID: \"464f639d-26e5-4d21-9d9b-c84585fbd8e2\") " pod="openstack/alertmanager-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.307571 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.330564 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"dae628f6-9079-4f73-b208-0f8fd6d43af0","Type":"ContainerStarted","Data":"e63b41e7650a6f84feb1f471bb3cf1be9f1ff6c513d4207bca72f4056b577a60"} Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.336702 4776 generic.go:334] "Generic (PLEG): container finished" podID="ee61762a-4aa0-4ae9-9107-fca99430ccfa" containerID="b3605f75d8bd3af002c79d3388d36f17448d7027115745579f13ee502a0e5ded" exitCode=137 Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.387659 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.411267 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.474452 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.474688 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.501205 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.504924 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-g6rm2" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.505213 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.505399 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.505588 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.505866 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.634335 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2c22f76e-686f-485b-aca5-953cb16bb9f8-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.634443 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2c22f76e-686f-485b-aca5-953cb16bb9f8-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.634490 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2c22f76e-686f-485b-aca5-953cb16bb9f8-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.634593 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2c22f76e-686f-485b-aca5-953cb16bb9f8-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.634629 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2c22f76e-686f-485b-aca5-953cb16bb9f8-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.634659 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2c22f76e-686f-485b-aca5-953cb16bb9f8-config\") pod \"prometheus-metric-storage-0\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.634729 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4zrf\" (UniqueName: \"kubernetes.io/projected/2c22f76e-686f-485b-aca5-953cb16bb9f8-kube-api-access-w4zrf\") pod \"prometheus-metric-storage-0\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.634842 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-472db1ee-c7d3-4f22-a8cf-04390f346bd1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-472db1ee-c7d3-4f22-a8cf-04390f346bd1\") pod \"prometheus-metric-storage-0\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.736855 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2c22f76e-686f-485b-aca5-953cb16bb9f8-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.736987 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2c22f76e-686f-485b-aca5-953cb16bb9f8-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.737026 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2c22f76e-686f-485b-aca5-953cb16bb9f8-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.737056 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2c22f76e-686f-485b-aca5-953cb16bb9f8-config\") pod \"prometheus-metric-storage-0\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.737132 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4zrf\" (UniqueName: \"kubernetes.io/projected/2c22f76e-686f-485b-aca5-953cb16bb9f8-kube-api-access-w4zrf\") pod \"prometheus-metric-storage-0\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.737247 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-472db1ee-c7d3-4f22-a8cf-04390f346bd1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-472db1ee-c7d3-4f22-a8cf-04390f346bd1\") pod \"prometheus-metric-storage-0\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.737301 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2c22f76e-686f-485b-aca5-953cb16bb9f8-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.737342 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2c22f76e-686f-485b-aca5-953cb16bb9f8-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.738349 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2c22f76e-686f-485b-aca5-953cb16bb9f8-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.742074 4776 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.742109 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-472db1ee-c7d3-4f22-a8cf-04390f346bd1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-472db1ee-c7d3-4f22-a8cf-04390f346bd1\") pod \"prometheus-metric-storage-0\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ff6e355e361fb5843939103249f848baecb449fc7ba76f9864fb5e4e8612b051/globalmount\"" pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.747480 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2c22f76e-686f-485b-aca5-953cb16bb9f8-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.751767 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2c22f76e-686f-485b-aca5-953cb16bb9f8-config\") pod \"prometheus-metric-storage-0\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.754336 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2c22f76e-686f-485b-aca5-953cb16bb9f8-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.800973 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2c22f76e-686f-485b-aca5-953cb16bb9f8-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.801050 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2c22f76e-686f-485b-aca5-953cb16bb9f8-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.812056 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4zrf\" (UniqueName: \"kubernetes.io/projected/2c22f76e-686f-485b-aca5-953cb16bb9f8-kube-api-access-w4zrf\") pod \"prometheus-metric-storage-0\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.917187 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.934445 4776 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="ee61762a-4aa0-4ae9-9107-fca99430ccfa" podUID="dae628f6-9079-4f73-b208-0f8fd6d43af0" Nov 25 11:12:06 crc kubenswrapper[4776]: I1125 11:12:06.970708 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-472db1ee-c7d3-4f22-a8cf-04390f346bd1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-472db1ee-c7d3-4f22-a8cf-04390f346bd1\") pod \"prometheus-metric-storage-0\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:07 crc kubenswrapper[4776]: I1125 11:12:07.044839 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ee61762a-4aa0-4ae9-9107-fca99430ccfa-openstack-config\") pod \"ee61762a-4aa0-4ae9-9107-fca99430ccfa\" (UID: \"ee61762a-4aa0-4ae9-9107-fca99430ccfa\") " Nov 25 11:12:07 crc kubenswrapper[4776]: I1125 11:12:07.045205 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ee61762a-4aa0-4ae9-9107-fca99430ccfa-openstack-config-secret\") pod \"ee61762a-4aa0-4ae9-9107-fca99430ccfa\" (UID: \"ee61762a-4aa0-4ae9-9107-fca99430ccfa\") " Nov 25 11:12:07 crc kubenswrapper[4776]: I1125 11:12:07.045382 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkz8f\" (UniqueName: \"kubernetes.io/projected/ee61762a-4aa0-4ae9-9107-fca99430ccfa-kube-api-access-lkz8f\") pod \"ee61762a-4aa0-4ae9-9107-fca99430ccfa\" (UID: \"ee61762a-4aa0-4ae9-9107-fca99430ccfa\") " Nov 25 11:12:07 crc kubenswrapper[4776]: I1125 11:12:07.045572 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee61762a-4aa0-4ae9-9107-fca99430ccfa-combined-ca-bundle\") pod \"ee61762a-4aa0-4ae9-9107-fca99430ccfa\" (UID: \"ee61762a-4aa0-4ae9-9107-fca99430ccfa\") " Nov 25 11:12:07 crc kubenswrapper[4776]: I1125 11:12:07.063221 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee61762a-4aa0-4ae9-9107-fca99430ccfa-kube-api-access-lkz8f" (OuterVolumeSpecName: "kube-api-access-lkz8f") pod "ee61762a-4aa0-4ae9-9107-fca99430ccfa" (UID: "ee61762a-4aa0-4ae9-9107-fca99430ccfa"). InnerVolumeSpecName "kube-api-access-lkz8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:12:07 crc kubenswrapper[4776]: I1125 11:12:07.109927 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee61762a-4aa0-4ae9-9107-fca99430ccfa-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "ee61762a-4aa0-4ae9-9107-fca99430ccfa" (UID: "ee61762a-4aa0-4ae9-9107-fca99430ccfa"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:12:07 crc kubenswrapper[4776]: I1125 11:12:07.119310 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee61762a-4aa0-4ae9-9107-fca99430ccfa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee61762a-4aa0-4ae9-9107-fca99430ccfa" (UID: "ee61762a-4aa0-4ae9-9107-fca99430ccfa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:12:07 crc kubenswrapper[4776]: I1125 11:12:07.153947 4776 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ee61762a-4aa0-4ae9-9107-fca99430ccfa-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:07 crc kubenswrapper[4776]: I1125 11:12:07.153998 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkz8f\" (UniqueName: \"kubernetes.io/projected/ee61762a-4aa0-4ae9-9107-fca99430ccfa-kube-api-access-lkz8f\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:07 crc kubenswrapper[4776]: I1125 11:12:07.154010 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee61762a-4aa0-4ae9-9107-fca99430ccfa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:07 crc kubenswrapper[4776]: I1125 11:12:07.154345 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee61762a-4aa0-4ae9-9107-fca99430ccfa-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "ee61762a-4aa0-4ae9-9107-fca99430ccfa" (UID: "ee61762a-4aa0-4ae9-9107-fca99430ccfa"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:12:07 crc kubenswrapper[4776]: I1125 11:12:07.176075 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Nov 25 11:12:07 crc kubenswrapper[4776]: I1125 11:12:07.204684 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:07 crc kubenswrapper[4776]: I1125 11:12:07.255952 4776 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ee61762a-4aa0-4ae9-9107-fca99430ccfa-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:07 crc kubenswrapper[4776]: I1125 11:12:07.355029 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1e161da1-d800-4a51-a70a-cfb8974b11b9","Type":"ContainerStarted","Data":"1d25d5d9a90aeb464ebd9a63c3addbcaedd4399ec37f6e2af5d754642931c9bb"} Nov 25 11:12:07 crc kubenswrapper[4776]: I1125 11:12:07.355262 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1e161da1-d800-4a51-a70a-cfb8974b11b9","Type":"ContainerStarted","Data":"5d898c84f191a7760fa1d779e129a922f8f4ef6078ded130e37a2a16c28db459"} Nov 25 11:12:07 crc kubenswrapper[4776]: I1125 11:12:07.356540 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 25 11:12:07 crc kubenswrapper[4776]: I1125 11:12:07.363624 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 11:12:07 crc kubenswrapper[4776]: I1125 11:12:07.363935 4776 scope.go:117] "RemoveContainer" containerID="b3605f75d8bd3af002c79d3388d36f17448d7027115745579f13ee502a0e5ded" Nov 25 11:12:07 crc kubenswrapper[4776]: I1125 11:12:07.372257 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"464f639d-26e5-4d21-9d9b-c84585fbd8e2","Type":"ContainerStarted","Data":"9a58a18533624fb8077c22e0e1776aebfea9e1f18308c78ef1698fab91696fac"} Nov 25 11:12:07 crc kubenswrapper[4776]: I1125 11:12:07.392145 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"dae628f6-9079-4f73-b208-0f8fd6d43af0","Type":"ContainerStarted","Data":"d8b57a933c20e558755b1388d4f9eda2cc171be3b0311d191c5e675f78e903e2"} Nov 25 11:12:07 crc kubenswrapper[4776]: I1125 11:12:07.421649 4776 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="ee61762a-4aa0-4ae9-9107-fca99430ccfa" podUID="dae628f6-9079-4f73-b208-0f8fd6d43af0" Nov 25 11:12:07 crc kubenswrapper[4776]: I1125 11:12:07.432593 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.908909419 podStartE2EDuration="3.43256156s" podCreationTimestamp="2025-11-25 11:12:04 +0000 UTC" firstStartedPulling="2025-11-25 11:12:06.542172405 +0000 UTC m=+6471.583231958" lastFinishedPulling="2025-11-25 11:12:07.065824546 +0000 UTC m=+6472.106884099" observedRunningTime="2025-11-25 11:12:07.410477495 +0000 UTC m=+6472.451537048" watchObservedRunningTime="2025-11-25 11:12:07.43256156 +0000 UTC m=+6472.473621113" Nov 25 11:12:07 crc kubenswrapper[4776]: I1125 11:12:07.485707 4776 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="ee61762a-4aa0-4ae9-9107-fca99430ccfa" podUID="dae628f6-9079-4f73-b208-0f8fd6d43af0" Nov 25 11:12:07 crc kubenswrapper[4776]: I1125 11:12:07.492013 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.491986081 podStartE2EDuration="3.491986081s" podCreationTimestamp="2025-11-25 11:12:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:12:07.438487998 +0000 UTC m=+6472.479547551" watchObservedRunningTime="2025-11-25 11:12:07.491986081 +0000 UTC m=+6472.533045634" Nov 25 11:12:07 crc kubenswrapper[4776]: I1125 11:12:07.681974 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee61762a-4aa0-4ae9-9107-fca99430ccfa" path="/var/lib/kubelet/pods/ee61762a-4aa0-4ae9-9107-fca99430ccfa/volumes" Nov 25 11:12:07 crc kubenswrapper[4776]: I1125 11:12:07.826014 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 25 11:12:08 crc kubenswrapper[4776]: I1125 11:12:08.405856 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2c22f76e-686f-485b-aca5-953cb16bb9f8","Type":"ContainerStarted","Data":"93345fabd70f94efa89630ae9d9e3e140216222237609877f2aee9e8d8295cfa"} Nov 25 11:12:14 crc kubenswrapper[4776]: I1125 11:12:14.853036 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 25 11:12:17 crc kubenswrapper[4776]: I1125 11:12:17.057961 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-k4dgh"] Nov 25 11:12:17 crc kubenswrapper[4776]: I1125 11:12:17.062080 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k4dgh" Nov 25 11:12:17 crc kubenswrapper[4776]: I1125 11:12:17.087222 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k4dgh"] Nov 25 11:12:17 crc kubenswrapper[4776]: I1125 11:12:17.090990 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bcd8737-c5cc-47be-9e76-42ec7756e333-utilities\") pod \"certified-operators-k4dgh\" (UID: \"9bcd8737-c5cc-47be-9e76-42ec7756e333\") " pod="openshift-marketplace/certified-operators-k4dgh" Nov 25 11:12:17 crc kubenswrapper[4776]: I1125 11:12:17.091098 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bcd8737-c5cc-47be-9e76-42ec7756e333-catalog-content\") pod \"certified-operators-k4dgh\" (UID: \"9bcd8737-c5cc-47be-9e76-42ec7756e333\") " pod="openshift-marketplace/certified-operators-k4dgh" Nov 25 11:12:17 crc kubenswrapper[4776]: I1125 11:12:17.091239 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-597f5\" (UniqueName: \"kubernetes.io/projected/9bcd8737-c5cc-47be-9e76-42ec7756e333-kube-api-access-597f5\") pod \"certified-operators-k4dgh\" (UID: \"9bcd8737-c5cc-47be-9e76-42ec7756e333\") " pod="openshift-marketplace/certified-operators-k4dgh" Nov 25 11:12:17 crc kubenswrapper[4776]: I1125 11:12:17.192638 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-597f5\" (UniqueName: \"kubernetes.io/projected/9bcd8737-c5cc-47be-9e76-42ec7756e333-kube-api-access-597f5\") pod \"certified-operators-k4dgh\" (UID: \"9bcd8737-c5cc-47be-9e76-42ec7756e333\") " pod="openshift-marketplace/certified-operators-k4dgh" Nov 25 11:12:17 crc kubenswrapper[4776]: I1125 11:12:17.192968 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bcd8737-c5cc-47be-9e76-42ec7756e333-utilities\") pod \"certified-operators-k4dgh\" (UID: \"9bcd8737-c5cc-47be-9e76-42ec7756e333\") " pod="openshift-marketplace/certified-operators-k4dgh" Nov 25 11:12:17 crc kubenswrapper[4776]: I1125 11:12:17.193141 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bcd8737-c5cc-47be-9e76-42ec7756e333-catalog-content\") pod \"certified-operators-k4dgh\" (UID: \"9bcd8737-c5cc-47be-9e76-42ec7756e333\") " pod="openshift-marketplace/certified-operators-k4dgh" Nov 25 11:12:17 crc kubenswrapper[4776]: I1125 11:12:17.193407 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bcd8737-c5cc-47be-9e76-42ec7756e333-utilities\") pod \"certified-operators-k4dgh\" (UID: \"9bcd8737-c5cc-47be-9e76-42ec7756e333\") " pod="openshift-marketplace/certified-operators-k4dgh" Nov 25 11:12:17 crc kubenswrapper[4776]: I1125 11:12:17.193450 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bcd8737-c5cc-47be-9e76-42ec7756e333-catalog-content\") pod \"certified-operators-k4dgh\" (UID: \"9bcd8737-c5cc-47be-9e76-42ec7756e333\") " pod="openshift-marketplace/certified-operators-k4dgh" Nov 25 11:12:17 crc kubenswrapper[4776]: I1125 11:12:17.223997 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-597f5\" (UniqueName: \"kubernetes.io/projected/9bcd8737-c5cc-47be-9e76-42ec7756e333-kube-api-access-597f5\") pod \"certified-operators-k4dgh\" (UID: \"9bcd8737-c5cc-47be-9e76-42ec7756e333\") " pod="openshift-marketplace/certified-operators-k4dgh" Nov 25 11:12:17 crc kubenswrapper[4776]: I1125 11:12:17.379910 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k4dgh" Nov 25 11:12:17 crc kubenswrapper[4776]: I1125 11:12:17.544936 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"464f639d-26e5-4d21-9d9b-c84585fbd8e2","Type":"ContainerStarted","Data":"46ebb53a695ecb512326089bd6eba64f7625ad98e4da08a34fb06963e202fbfc"} Nov 25 11:12:17 crc kubenswrapper[4776]: I1125 11:12:17.560928 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2c22f76e-686f-485b-aca5-953cb16bb9f8","Type":"ContainerStarted","Data":"a5c6e299e3d5f114ab1ec22f96d9da9be9aa6e0ed52170b485116a0011062378"} Nov 25 11:12:17 crc kubenswrapper[4776]: I1125 11:12:17.821560 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:12:17 crc kubenswrapper[4776]: I1125 11:12:17.821613 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:12:17 crc kubenswrapper[4776]: I1125 11:12:17.955291 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k4dgh"] Nov 25 11:12:17 crc kubenswrapper[4776]: W1125 11:12:17.963992 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9bcd8737_c5cc_47be_9e76_42ec7756e333.slice/crio-b1a151fd5d06609451fd8b43eed9454c9db48803bdb5a3445dbbc61a3dd72005 WatchSource:0}: Error finding container b1a151fd5d06609451fd8b43eed9454c9db48803bdb5a3445dbbc61a3dd72005: Status 404 returned error can't find the container with id b1a151fd5d06609451fd8b43eed9454c9db48803bdb5a3445dbbc61a3dd72005 Nov 25 11:12:18 crc kubenswrapper[4776]: E1125 11:12:18.475165 4776 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9bcd8737_c5cc_47be_9e76_42ec7756e333.slice/crio-2facd521f13f62360c7d0fa9a89b7033638ccee7d40d16d9cfd536e659abeec0.scope\": RecentStats: unable to find data in memory cache]" Nov 25 11:12:18 crc kubenswrapper[4776]: I1125 11:12:18.570432 4776 generic.go:334] "Generic (PLEG): container finished" podID="9bcd8737-c5cc-47be-9e76-42ec7756e333" containerID="2facd521f13f62360c7d0fa9a89b7033638ccee7d40d16d9cfd536e659abeec0" exitCode=0 Nov 25 11:12:18 crc kubenswrapper[4776]: I1125 11:12:18.570514 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k4dgh" event={"ID":"9bcd8737-c5cc-47be-9e76-42ec7756e333","Type":"ContainerDied","Data":"2facd521f13f62360c7d0fa9a89b7033638ccee7d40d16d9cfd536e659abeec0"} Nov 25 11:12:18 crc kubenswrapper[4776]: I1125 11:12:18.570551 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k4dgh" event={"ID":"9bcd8737-c5cc-47be-9e76-42ec7756e333","Type":"ContainerStarted","Data":"b1a151fd5d06609451fd8b43eed9454c9db48803bdb5a3445dbbc61a3dd72005"} Nov 25 11:12:21 crc kubenswrapper[4776]: I1125 11:12:21.603819 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k4dgh" event={"ID":"9bcd8737-c5cc-47be-9e76-42ec7756e333","Type":"ContainerStarted","Data":"2ce904bcdd49994cdfc8b7e4ab7816fe34cb12056e74c83b75e0faad52d32d01"} Nov 25 11:12:23 crc kubenswrapper[4776]: I1125 11:12:23.622529 4776 generic.go:334] "Generic (PLEG): container finished" podID="2c22f76e-686f-485b-aca5-953cb16bb9f8" containerID="a5c6e299e3d5f114ab1ec22f96d9da9be9aa6e0ed52170b485116a0011062378" exitCode=0 Nov 25 11:12:23 crc kubenswrapper[4776]: I1125 11:12:23.622622 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2c22f76e-686f-485b-aca5-953cb16bb9f8","Type":"ContainerDied","Data":"a5c6e299e3d5f114ab1ec22f96d9da9be9aa6e0ed52170b485116a0011062378"} Nov 25 11:12:24 crc kubenswrapper[4776]: I1125 11:12:24.635918 4776 generic.go:334] "Generic (PLEG): container finished" podID="464f639d-26e5-4d21-9d9b-c84585fbd8e2" containerID="46ebb53a695ecb512326089bd6eba64f7625ad98e4da08a34fb06963e202fbfc" exitCode=0 Nov 25 11:12:24 crc kubenswrapper[4776]: I1125 11:12:24.636037 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"464f639d-26e5-4d21-9d9b-c84585fbd8e2","Type":"ContainerDied","Data":"46ebb53a695ecb512326089bd6eba64f7625ad98e4da08a34fb06963e202fbfc"} Nov 25 11:12:26 crc kubenswrapper[4776]: I1125 11:12:26.668222 4776 generic.go:334] "Generic (PLEG): container finished" podID="9bcd8737-c5cc-47be-9e76-42ec7756e333" containerID="2ce904bcdd49994cdfc8b7e4ab7816fe34cb12056e74c83b75e0faad52d32d01" exitCode=0 Nov 25 11:12:26 crc kubenswrapper[4776]: I1125 11:12:26.668305 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k4dgh" event={"ID":"9bcd8737-c5cc-47be-9e76-42ec7756e333","Type":"ContainerDied","Data":"2ce904bcdd49994cdfc8b7e4ab7816fe34cb12056e74c83b75e0faad52d32d01"} Nov 25 11:12:35 crc kubenswrapper[4776]: I1125 11:12:35.418962 4776 scope.go:117] "RemoveContainer" containerID="e036f950b949a43764acf5a11e58886283eb69f531f34baf6e18e175c1ab7588" Nov 25 11:12:37 crc kubenswrapper[4776]: I1125 11:12:37.754072 4776 scope.go:117] "RemoveContainer" containerID="0bead501d42c1f9defd4db9ce9dbd117c8813e94be8fe3f158d2f400855361cd" Nov 25 11:12:37 crc kubenswrapper[4776]: I1125 11:12:37.782656 4776 scope.go:117] "RemoveContainer" containerID="084262977fe869689b12fd1e2ff0368c44c32915ea0383fc5dd5c733933bcf6c" Nov 25 11:12:37 crc kubenswrapper[4776]: I1125 11:12:37.987295 4776 scope.go:117] "RemoveContainer" containerID="6478752a998200191ccc75ba8bbac28e78c153aa2bac4f02c428d64aa8be44fe" Nov 25 11:12:38 crc kubenswrapper[4776]: I1125 11:12:38.010817 4776 scope.go:117] "RemoveContainer" containerID="3e5e58c668fec192a0a49e5e7e1abd071ab2d9e0be25bb4a9f499a2eafd26898" Nov 25 11:12:38 crc kubenswrapper[4776]: I1125 11:12:38.786862 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2c22f76e-686f-485b-aca5-953cb16bb9f8","Type":"ContainerStarted","Data":"7f4b79ed91bbd26f7e991ea152d411c3c65c2f4a5d540c19584b810f0b2da628"} Nov 25 11:12:38 crc kubenswrapper[4776]: I1125 11:12:38.789363 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k4dgh" event={"ID":"9bcd8737-c5cc-47be-9e76-42ec7756e333","Type":"ContainerStarted","Data":"f61c7413f541b66d84d8bc85373a9b905d1ca04bf5606ae3b3b6cbee1510c0f2"} Nov 25 11:12:38 crc kubenswrapper[4776]: I1125 11:12:38.793731 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"464f639d-26e5-4d21-9d9b-c84585fbd8e2","Type":"ContainerStarted","Data":"c5552ee91b860599595457935e1ba4fbcdd05c3ee78a4de2ba7a8c67694f6293"} Nov 25 11:12:38 crc kubenswrapper[4776]: I1125 11:12:38.822602 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-k4dgh" podStartSLOduration=11.878166214 podStartE2EDuration="21.822581275s" podCreationTimestamp="2025-11-25 11:12:17 +0000 UTC" firstStartedPulling="2025-11-25 11:12:18.57429753 +0000 UTC m=+6483.615357073" lastFinishedPulling="2025-11-25 11:12:28.518712581 +0000 UTC m=+6493.559772134" observedRunningTime="2025-11-25 11:12:38.808826289 +0000 UTC m=+6503.849885852" watchObservedRunningTime="2025-11-25 11:12:38.822581275 +0000 UTC m=+6503.863640828" Nov 25 11:12:42 crc kubenswrapper[4776]: I1125 11:12:42.840043 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2c22f76e-686f-485b-aca5-953cb16bb9f8","Type":"ContainerStarted","Data":"e042028628e4b4e55e4e8b5856d23911587782bb0084a8dd5f20cf8d5e59a020"} Nov 25 11:12:42 crc kubenswrapper[4776]: I1125 11:12:42.842488 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"464f639d-26e5-4d21-9d9b-c84585fbd8e2","Type":"ContainerStarted","Data":"4b13797e8e00d58b171ea9d850c0dee33f5e0ac45388bfc42459e36be67cc40e"} Nov 25 11:12:42 crc kubenswrapper[4776]: I1125 11:12:42.842831 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Nov 25 11:12:42 crc kubenswrapper[4776]: I1125 11:12:42.845115 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Nov 25 11:12:42 crc kubenswrapper[4776]: I1125 11:12:42.878914 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=7.295756679 podStartE2EDuration="37.878896174s" podCreationTimestamp="2025-11-25 11:12:05 +0000 UTC" firstStartedPulling="2025-11-25 11:12:07.201044359 +0000 UTC m=+6472.242103912" lastFinishedPulling="2025-11-25 11:12:37.784183854 +0000 UTC m=+6502.825243407" observedRunningTime="2025-11-25 11:12:42.861675331 +0000 UTC m=+6507.902734874" watchObservedRunningTime="2025-11-25 11:12:42.878896174 +0000 UTC m=+6507.919955727" Nov 25 11:12:46 crc kubenswrapper[4776]: I1125 11:12:46.890541 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2c22f76e-686f-485b-aca5-953cb16bb9f8","Type":"ContainerStarted","Data":"6c6a9570809a6163b94faa1d8799e71d2876afb79c7ed4d488abe35819ee5252"} Nov 25 11:12:46 crc kubenswrapper[4776]: I1125 11:12:46.921496 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.811994577 podStartE2EDuration="41.921460688s" podCreationTimestamp="2025-11-25 11:12:05 +0000 UTC" firstStartedPulling="2025-11-25 11:12:07.838854546 +0000 UTC m=+6472.879914109" lastFinishedPulling="2025-11-25 11:12:45.948320667 +0000 UTC m=+6510.989380220" observedRunningTime="2025-11-25 11:12:46.919271174 +0000 UTC m=+6511.960330727" watchObservedRunningTime="2025-11-25 11:12:46.921460688 +0000 UTC m=+6511.962520241" Nov 25 11:12:47 crc kubenswrapper[4776]: I1125 11:12:47.205408 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:47 crc kubenswrapper[4776]: I1125 11:12:47.380442 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-k4dgh" Nov 25 11:12:47 crc kubenswrapper[4776]: I1125 11:12:47.380506 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-k4dgh" Nov 25 11:12:47 crc kubenswrapper[4776]: I1125 11:12:47.432752 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-k4dgh" Nov 25 11:12:47 crc kubenswrapper[4776]: I1125 11:12:47.817972 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:12:47 crc kubenswrapper[4776]: I1125 11:12:47.818031 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:12:47 crc kubenswrapper[4776]: I1125 11:12:47.818096 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 11:12:47 crc kubenswrapper[4776]: I1125 11:12:47.818929 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d9823ef55a496477edc68cef40859ee8ddcd3f8a6fd6685be1f1eda0a305897e"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 11:12:47 crc kubenswrapper[4776]: I1125 11:12:47.818998 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://d9823ef55a496477edc68cef40859ee8ddcd3f8a6fd6685be1f1eda0a305897e" gracePeriod=600 Nov 25 11:12:47 crc kubenswrapper[4776]: I1125 11:12:47.970872 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-k4dgh" Nov 25 11:12:48 crc kubenswrapper[4776]: I1125 11:12:48.033783 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k4dgh"] Nov 25 11:12:48 crc kubenswrapper[4776]: I1125 11:12:48.912795 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="d9823ef55a496477edc68cef40859ee8ddcd3f8a6fd6685be1f1eda0a305897e" exitCode=0 Nov 25 11:12:48 crc kubenswrapper[4776]: I1125 11:12:48.912882 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"d9823ef55a496477edc68cef40859ee8ddcd3f8a6fd6685be1f1eda0a305897e"} Nov 25 11:12:48 crc kubenswrapper[4776]: I1125 11:12:48.913683 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564"} Nov 25 11:12:48 crc kubenswrapper[4776]: I1125 11:12:48.913751 4776 scope.go:117] "RemoveContainer" containerID="8565987c20d8795e20eba174c69eb3cf4accfcca5002d20fcd9eba3de1cd1698" Nov 25 11:12:49 crc kubenswrapper[4776]: I1125 11:12:49.923766 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-k4dgh" podUID="9bcd8737-c5cc-47be-9e76-42ec7756e333" containerName="registry-server" containerID="cri-o://f61c7413f541b66d84d8bc85373a9b905d1ca04bf5606ae3b3b6cbee1510c0f2" gracePeriod=2 Nov 25 11:12:50 crc kubenswrapper[4776]: I1125 11:12:50.522053 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k4dgh" Nov 25 11:12:50 crc kubenswrapper[4776]: I1125 11:12:50.660511 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bcd8737-c5cc-47be-9e76-42ec7756e333-catalog-content\") pod \"9bcd8737-c5cc-47be-9e76-42ec7756e333\" (UID: \"9bcd8737-c5cc-47be-9e76-42ec7756e333\") " Nov 25 11:12:50 crc kubenswrapper[4776]: I1125 11:12:50.660808 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bcd8737-c5cc-47be-9e76-42ec7756e333-utilities\") pod \"9bcd8737-c5cc-47be-9e76-42ec7756e333\" (UID: \"9bcd8737-c5cc-47be-9e76-42ec7756e333\") " Nov 25 11:12:50 crc kubenswrapper[4776]: I1125 11:12:50.660906 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-597f5\" (UniqueName: \"kubernetes.io/projected/9bcd8737-c5cc-47be-9e76-42ec7756e333-kube-api-access-597f5\") pod \"9bcd8737-c5cc-47be-9e76-42ec7756e333\" (UID: \"9bcd8737-c5cc-47be-9e76-42ec7756e333\") " Nov 25 11:12:50 crc kubenswrapper[4776]: I1125 11:12:50.662172 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9bcd8737-c5cc-47be-9e76-42ec7756e333-utilities" (OuterVolumeSpecName: "utilities") pod "9bcd8737-c5cc-47be-9e76-42ec7756e333" (UID: "9bcd8737-c5cc-47be-9e76-42ec7756e333"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:12:50 crc kubenswrapper[4776]: I1125 11:12:50.671567 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bcd8737-c5cc-47be-9e76-42ec7756e333-kube-api-access-597f5" (OuterVolumeSpecName: "kube-api-access-597f5") pod "9bcd8737-c5cc-47be-9e76-42ec7756e333" (UID: "9bcd8737-c5cc-47be-9e76-42ec7756e333"). InnerVolumeSpecName "kube-api-access-597f5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:12:50 crc kubenswrapper[4776]: I1125 11:12:50.718284 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9bcd8737-c5cc-47be-9e76-42ec7756e333-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9bcd8737-c5cc-47be-9e76-42ec7756e333" (UID: "9bcd8737-c5cc-47be-9e76-42ec7756e333"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:12:50 crc kubenswrapper[4776]: I1125 11:12:50.764105 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bcd8737-c5cc-47be-9e76-42ec7756e333-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:50 crc kubenswrapper[4776]: I1125 11:12:50.764707 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-597f5\" (UniqueName: \"kubernetes.io/projected/9bcd8737-c5cc-47be-9e76-42ec7756e333-kube-api-access-597f5\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:50 crc kubenswrapper[4776]: I1125 11:12:50.764724 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bcd8737-c5cc-47be-9e76-42ec7756e333-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:50 crc kubenswrapper[4776]: I1125 11:12:50.936877 4776 generic.go:334] "Generic (PLEG): container finished" podID="9bcd8737-c5cc-47be-9e76-42ec7756e333" containerID="f61c7413f541b66d84d8bc85373a9b905d1ca04bf5606ae3b3b6cbee1510c0f2" exitCode=0 Nov 25 11:12:50 crc kubenswrapper[4776]: I1125 11:12:50.937108 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k4dgh" event={"ID":"9bcd8737-c5cc-47be-9e76-42ec7756e333","Type":"ContainerDied","Data":"f61c7413f541b66d84d8bc85373a9b905d1ca04bf5606ae3b3b6cbee1510c0f2"} Nov 25 11:12:50 crc kubenswrapper[4776]: I1125 11:12:50.937586 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k4dgh" event={"ID":"9bcd8737-c5cc-47be-9e76-42ec7756e333","Type":"ContainerDied","Data":"b1a151fd5d06609451fd8b43eed9454c9db48803bdb5a3445dbbc61a3dd72005"} Nov 25 11:12:50 crc kubenswrapper[4776]: I1125 11:12:50.937205 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k4dgh" Nov 25 11:12:50 crc kubenswrapper[4776]: I1125 11:12:50.937617 4776 scope.go:117] "RemoveContainer" containerID="f61c7413f541b66d84d8bc85373a9b905d1ca04bf5606ae3b3b6cbee1510c0f2" Nov 25 11:12:50 crc kubenswrapper[4776]: I1125 11:12:50.977392 4776 scope.go:117] "RemoveContainer" containerID="2ce904bcdd49994cdfc8b7e4ab7816fe34cb12056e74c83b75e0faad52d32d01" Nov 25 11:12:50 crc kubenswrapper[4776]: I1125 11:12:50.990708 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k4dgh"] Nov 25 11:12:51 crc kubenswrapper[4776]: I1125 11:12:51.002158 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-k4dgh"] Nov 25 11:12:51 crc kubenswrapper[4776]: I1125 11:12:51.010709 4776 scope.go:117] "RemoveContainer" containerID="2facd521f13f62360c7d0fa9a89b7033638ccee7d40d16d9cfd536e659abeec0" Nov 25 11:12:51 crc kubenswrapper[4776]: I1125 11:12:51.047847 4776 scope.go:117] "RemoveContainer" containerID="f61c7413f541b66d84d8bc85373a9b905d1ca04bf5606ae3b3b6cbee1510c0f2" Nov 25 11:12:51 crc kubenswrapper[4776]: E1125 11:12:51.048423 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f61c7413f541b66d84d8bc85373a9b905d1ca04bf5606ae3b3b6cbee1510c0f2\": container with ID starting with f61c7413f541b66d84d8bc85373a9b905d1ca04bf5606ae3b3b6cbee1510c0f2 not found: ID does not exist" containerID="f61c7413f541b66d84d8bc85373a9b905d1ca04bf5606ae3b3b6cbee1510c0f2" Nov 25 11:12:51 crc kubenswrapper[4776]: I1125 11:12:51.048467 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f61c7413f541b66d84d8bc85373a9b905d1ca04bf5606ae3b3b6cbee1510c0f2"} err="failed to get container status \"f61c7413f541b66d84d8bc85373a9b905d1ca04bf5606ae3b3b6cbee1510c0f2\": rpc error: code = NotFound desc = could not find container \"f61c7413f541b66d84d8bc85373a9b905d1ca04bf5606ae3b3b6cbee1510c0f2\": container with ID starting with f61c7413f541b66d84d8bc85373a9b905d1ca04bf5606ae3b3b6cbee1510c0f2 not found: ID does not exist" Nov 25 11:12:51 crc kubenswrapper[4776]: I1125 11:12:51.048496 4776 scope.go:117] "RemoveContainer" containerID="2ce904bcdd49994cdfc8b7e4ab7816fe34cb12056e74c83b75e0faad52d32d01" Nov 25 11:12:51 crc kubenswrapper[4776]: E1125 11:12:51.048978 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ce904bcdd49994cdfc8b7e4ab7816fe34cb12056e74c83b75e0faad52d32d01\": container with ID starting with 2ce904bcdd49994cdfc8b7e4ab7816fe34cb12056e74c83b75e0faad52d32d01 not found: ID does not exist" containerID="2ce904bcdd49994cdfc8b7e4ab7816fe34cb12056e74c83b75e0faad52d32d01" Nov 25 11:12:51 crc kubenswrapper[4776]: I1125 11:12:51.048999 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ce904bcdd49994cdfc8b7e4ab7816fe34cb12056e74c83b75e0faad52d32d01"} err="failed to get container status \"2ce904bcdd49994cdfc8b7e4ab7816fe34cb12056e74c83b75e0faad52d32d01\": rpc error: code = NotFound desc = could not find container \"2ce904bcdd49994cdfc8b7e4ab7816fe34cb12056e74c83b75e0faad52d32d01\": container with ID starting with 2ce904bcdd49994cdfc8b7e4ab7816fe34cb12056e74c83b75e0faad52d32d01 not found: ID does not exist" Nov 25 11:12:51 crc kubenswrapper[4776]: I1125 11:12:51.049011 4776 scope.go:117] "RemoveContainer" containerID="2facd521f13f62360c7d0fa9a89b7033638ccee7d40d16d9cfd536e659abeec0" Nov 25 11:12:51 crc kubenswrapper[4776]: E1125 11:12:51.049372 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2facd521f13f62360c7d0fa9a89b7033638ccee7d40d16d9cfd536e659abeec0\": container with ID starting with 2facd521f13f62360c7d0fa9a89b7033638ccee7d40d16d9cfd536e659abeec0 not found: ID does not exist" containerID="2facd521f13f62360c7d0fa9a89b7033638ccee7d40d16d9cfd536e659abeec0" Nov 25 11:12:51 crc kubenswrapper[4776]: I1125 11:12:51.049409 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2facd521f13f62360c7d0fa9a89b7033638ccee7d40d16d9cfd536e659abeec0"} err="failed to get container status \"2facd521f13f62360c7d0fa9a89b7033638ccee7d40d16d9cfd536e659abeec0\": rpc error: code = NotFound desc = could not find container \"2facd521f13f62360c7d0fa9a89b7033638ccee7d40d16d9cfd536e659abeec0\": container with ID starting with 2facd521f13f62360c7d0fa9a89b7033638ccee7d40d16d9cfd536e659abeec0 not found: ID does not exist" Nov 25 11:12:51 crc kubenswrapper[4776]: I1125 11:12:51.674390 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bcd8737-c5cc-47be-9e76-42ec7756e333" path="/var/lib/kubelet/pods/9bcd8737-c5cc-47be-9e76-42ec7756e333/volumes" Nov 25 11:12:52 crc kubenswrapper[4776]: I1125 11:12:52.205463 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:52 crc kubenswrapper[4776]: I1125 11:12:52.207351 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:52 crc kubenswrapper[4776]: I1125 11:12:52.962573 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:54 crc kubenswrapper[4776]: I1125 11:12:54.471171 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Nov 25 11:12:54 crc kubenswrapper[4776]: I1125 11:12:54.471922 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="dae628f6-9079-4f73-b208-0f8fd6d43af0" containerName="openstackclient" containerID="cri-o://d8b57a933c20e558755b1388d4f9eda2cc171be3b0311d191c5e675f78e903e2" gracePeriod=2 Nov 25 11:12:54 crc kubenswrapper[4776]: I1125 11:12:54.485290 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Nov 25 11:12:54 crc kubenswrapper[4776]: I1125 11:12:54.509694 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 25 11:12:54 crc kubenswrapper[4776]: E1125 11:12:54.510164 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dae628f6-9079-4f73-b208-0f8fd6d43af0" containerName="openstackclient" Nov 25 11:12:54 crc kubenswrapper[4776]: I1125 11:12:54.510186 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="dae628f6-9079-4f73-b208-0f8fd6d43af0" containerName="openstackclient" Nov 25 11:12:54 crc kubenswrapper[4776]: E1125 11:12:54.510218 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bcd8737-c5cc-47be-9e76-42ec7756e333" containerName="registry-server" Nov 25 11:12:54 crc kubenswrapper[4776]: I1125 11:12:54.510230 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bcd8737-c5cc-47be-9e76-42ec7756e333" containerName="registry-server" Nov 25 11:12:54 crc kubenswrapper[4776]: E1125 11:12:54.510255 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bcd8737-c5cc-47be-9e76-42ec7756e333" containerName="extract-utilities" Nov 25 11:12:54 crc kubenswrapper[4776]: I1125 11:12:54.510264 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bcd8737-c5cc-47be-9e76-42ec7756e333" containerName="extract-utilities" Nov 25 11:12:54 crc kubenswrapper[4776]: E1125 11:12:54.510288 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bcd8737-c5cc-47be-9e76-42ec7756e333" containerName="extract-content" Nov 25 11:12:54 crc kubenswrapper[4776]: I1125 11:12:54.510294 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bcd8737-c5cc-47be-9e76-42ec7756e333" containerName="extract-content" Nov 25 11:12:54 crc kubenswrapper[4776]: I1125 11:12:54.510508 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="dae628f6-9079-4f73-b208-0f8fd6d43af0" containerName="openstackclient" Nov 25 11:12:54 crc kubenswrapper[4776]: I1125 11:12:54.510526 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bcd8737-c5cc-47be-9e76-42ec7756e333" containerName="registry-server" Nov 25 11:12:54 crc kubenswrapper[4776]: I1125 11:12:54.511254 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 11:12:54 crc kubenswrapper[4776]: I1125 11:12:54.522944 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 25 11:12:54 crc kubenswrapper[4776]: I1125 11:12:54.547931 4776 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="dae628f6-9079-4f73-b208-0f8fd6d43af0" podUID="96898351-caa6-4edc-8080-898f9fff0f56" Nov 25 11:12:54 crc kubenswrapper[4776]: I1125 11:12:54.654548 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/96898351-caa6-4edc-8080-898f9fff0f56-openstack-config\") pod \"openstackclient\" (UID: \"96898351-caa6-4edc-8080-898f9fff0f56\") " pod="openstack/openstackclient" Nov 25 11:12:54 crc kubenswrapper[4776]: I1125 11:12:54.654709 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/96898351-caa6-4edc-8080-898f9fff0f56-openstack-config-secret\") pod \"openstackclient\" (UID: \"96898351-caa6-4edc-8080-898f9fff0f56\") " pod="openstack/openstackclient" Nov 25 11:12:54 crc kubenswrapper[4776]: I1125 11:12:54.654861 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96898351-caa6-4edc-8080-898f9fff0f56-combined-ca-bundle\") pod \"openstackclient\" (UID: \"96898351-caa6-4edc-8080-898f9fff0f56\") " pod="openstack/openstackclient" Nov 25 11:12:54 crc kubenswrapper[4776]: I1125 11:12:54.654895 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrqfc\" (UniqueName: \"kubernetes.io/projected/96898351-caa6-4edc-8080-898f9fff0f56-kube-api-access-mrqfc\") pod \"openstackclient\" (UID: \"96898351-caa6-4edc-8080-898f9fff0f56\") " pod="openstack/openstackclient" Nov 25 11:12:54 crc kubenswrapper[4776]: I1125 11:12:54.757058 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96898351-caa6-4edc-8080-898f9fff0f56-combined-ca-bundle\") pod \"openstackclient\" (UID: \"96898351-caa6-4edc-8080-898f9fff0f56\") " pod="openstack/openstackclient" Nov 25 11:12:54 crc kubenswrapper[4776]: I1125 11:12:54.757134 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrqfc\" (UniqueName: \"kubernetes.io/projected/96898351-caa6-4edc-8080-898f9fff0f56-kube-api-access-mrqfc\") pod \"openstackclient\" (UID: \"96898351-caa6-4edc-8080-898f9fff0f56\") " pod="openstack/openstackclient" Nov 25 11:12:54 crc kubenswrapper[4776]: I1125 11:12:54.757210 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/96898351-caa6-4edc-8080-898f9fff0f56-openstack-config\") pod \"openstackclient\" (UID: \"96898351-caa6-4edc-8080-898f9fff0f56\") " pod="openstack/openstackclient" Nov 25 11:12:54 crc kubenswrapper[4776]: I1125 11:12:54.757359 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/96898351-caa6-4edc-8080-898f9fff0f56-openstack-config-secret\") pod \"openstackclient\" (UID: \"96898351-caa6-4edc-8080-898f9fff0f56\") " pod="openstack/openstackclient" Nov 25 11:12:54 crc kubenswrapper[4776]: I1125 11:12:54.760095 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/96898351-caa6-4edc-8080-898f9fff0f56-openstack-config\") pod \"openstackclient\" (UID: \"96898351-caa6-4edc-8080-898f9fff0f56\") " pod="openstack/openstackclient" Nov 25 11:12:54 crc kubenswrapper[4776]: I1125 11:12:54.766042 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96898351-caa6-4edc-8080-898f9fff0f56-combined-ca-bundle\") pod \"openstackclient\" (UID: \"96898351-caa6-4edc-8080-898f9fff0f56\") " pod="openstack/openstackclient" Nov 25 11:12:54 crc kubenswrapper[4776]: I1125 11:12:54.769766 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/96898351-caa6-4edc-8080-898f9fff0f56-openstack-config-secret\") pod \"openstackclient\" (UID: \"96898351-caa6-4edc-8080-898f9fff0f56\") " pod="openstack/openstackclient" Nov 25 11:12:54 crc kubenswrapper[4776]: I1125 11:12:54.781954 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrqfc\" (UniqueName: \"kubernetes.io/projected/96898351-caa6-4edc-8080-898f9fff0f56-kube-api-access-mrqfc\") pod \"openstackclient\" (UID: \"96898351-caa6-4edc-8080-898f9fff0f56\") " pod="openstack/openstackclient" Nov 25 11:12:54 crc kubenswrapper[4776]: I1125 11:12:54.848579 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 11:12:55 crc kubenswrapper[4776]: I1125 11:12:55.481345 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 25 11:12:55 crc kubenswrapper[4776]: I1125 11:12:55.911979 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 25 11:12:55 crc kubenswrapper[4776]: I1125 11:12:55.912546 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="2c22f76e-686f-485b-aca5-953cb16bb9f8" containerName="prometheus" containerID="cri-o://7f4b79ed91bbd26f7e991ea152d411c3c65c2f4a5d540c19584b810f0b2da628" gracePeriod=600 Nov 25 11:12:55 crc kubenswrapper[4776]: I1125 11:12:55.912864 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="2c22f76e-686f-485b-aca5-953cb16bb9f8" containerName="thanos-sidecar" containerID="cri-o://6c6a9570809a6163b94faa1d8799e71d2876afb79c7ed4d488abe35819ee5252" gracePeriod=600 Nov 25 11:12:55 crc kubenswrapper[4776]: I1125 11:12:55.913063 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="2c22f76e-686f-485b-aca5-953cb16bb9f8" containerName="config-reloader" containerID="cri-o://e042028628e4b4e55e4e8b5856d23911587782bb0084a8dd5f20cf8d5e59a020" gracePeriod=600 Nov 25 11:12:56 crc kubenswrapper[4776]: I1125 11:12:56.036967 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"96898351-caa6-4edc-8080-898f9fff0f56","Type":"ContainerStarted","Data":"b16b2cf822d1b74ebc25e8e3eb36d0453d1430025125df293a3d5bc230a9058a"} Nov 25 11:12:56 crc kubenswrapper[4776]: I1125 11:12:56.037009 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"96898351-caa6-4edc-8080-898f9fff0f56","Type":"ContainerStarted","Data":"1cfb3f1dcadf8cffaa921ac02e6883d2840e85c3da931fccf34cfc32c292740d"} Nov 25 11:12:56 crc kubenswrapper[4776]: I1125 11:12:56.070176 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.070151251 podStartE2EDuration="2.070151251s" podCreationTimestamp="2025-11-25 11:12:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:12:56.057540894 +0000 UTC m=+6521.098600457" watchObservedRunningTime="2025-11-25 11:12:56.070151251 +0000 UTC m=+6521.111210824" Nov 25 11:12:56 crc kubenswrapper[4776]: I1125 11:12:56.753509 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 11:12:56 crc kubenswrapper[4776]: I1125 11:12:56.763659 4776 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="dae628f6-9079-4f73-b208-0f8fd6d43af0" podUID="96898351-caa6-4edc-8080-898f9fff0f56" Nov 25 11:12:56 crc kubenswrapper[4776]: I1125 11:12:56.817580 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dae628f6-9079-4f73-b208-0f8fd6d43af0-openstack-config\") pod \"dae628f6-9079-4f73-b208-0f8fd6d43af0\" (UID: \"dae628f6-9079-4f73-b208-0f8fd6d43af0\") " Nov 25 11:12:56 crc kubenswrapper[4776]: I1125 11:12:56.817638 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4tcj\" (UniqueName: \"kubernetes.io/projected/dae628f6-9079-4f73-b208-0f8fd6d43af0-kube-api-access-r4tcj\") pod \"dae628f6-9079-4f73-b208-0f8fd6d43af0\" (UID: \"dae628f6-9079-4f73-b208-0f8fd6d43af0\") " Nov 25 11:12:56 crc kubenswrapper[4776]: I1125 11:12:56.817826 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dae628f6-9079-4f73-b208-0f8fd6d43af0-combined-ca-bundle\") pod \"dae628f6-9079-4f73-b208-0f8fd6d43af0\" (UID: \"dae628f6-9079-4f73-b208-0f8fd6d43af0\") " Nov 25 11:12:56 crc kubenswrapper[4776]: I1125 11:12:56.817879 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dae628f6-9079-4f73-b208-0f8fd6d43af0-openstack-config-secret\") pod \"dae628f6-9079-4f73-b208-0f8fd6d43af0\" (UID: \"dae628f6-9079-4f73-b208-0f8fd6d43af0\") " Nov 25 11:12:56 crc kubenswrapper[4776]: I1125 11:12:56.836736 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dae628f6-9079-4f73-b208-0f8fd6d43af0-kube-api-access-r4tcj" (OuterVolumeSpecName: "kube-api-access-r4tcj") pod "dae628f6-9079-4f73-b208-0f8fd6d43af0" (UID: "dae628f6-9079-4f73-b208-0f8fd6d43af0"). InnerVolumeSpecName "kube-api-access-r4tcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:12:56 crc kubenswrapper[4776]: I1125 11:12:56.880489 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dae628f6-9079-4f73-b208-0f8fd6d43af0-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "dae628f6-9079-4f73-b208-0f8fd6d43af0" (UID: "dae628f6-9079-4f73-b208-0f8fd6d43af0"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:12:56 crc kubenswrapper[4776]: I1125 11:12:56.914729 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dae628f6-9079-4f73-b208-0f8fd6d43af0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dae628f6-9079-4f73-b208-0f8fd6d43af0" (UID: "dae628f6-9079-4f73-b208-0f8fd6d43af0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:12:56 crc kubenswrapper[4776]: I1125 11:12:56.925621 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dae628f6-9079-4f73-b208-0f8fd6d43af0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:56 crc kubenswrapper[4776]: I1125 11:12:56.925673 4776 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/dae628f6-9079-4f73-b208-0f8fd6d43af0-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:56 crc kubenswrapper[4776]: I1125 11:12:56.925687 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4tcj\" (UniqueName: \"kubernetes.io/projected/dae628f6-9079-4f73-b208-0f8fd6d43af0-kube-api-access-r4tcj\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:56 crc kubenswrapper[4776]: I1125 11:12:56.932885 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dae628f6-9079-4f73-b208-0f8fd6d43af0-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "dae628f6-9079-4f73-b208-0f8fd6d43af0" (UID: "dae628f6-9079-4f73-b208-0f8fd6d43af0"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.000717 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.028297 4776 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/dae628f6-9079-4f73-b208-0f8fd6d43af0-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.052984 4776 generic.go:334] "Generic (PLEG): container finished" podID="2c22f76e-686f-485b-aca5-953cb16bb9f8" containerID="6c6a9570809a6163b94faa1d8799e71d2876afb79c7ed4d488abe35819ee5252" exitCode=0 Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.053040 4776 generic.go:334] "Generic (PLEG): container finished" podID="2c22f76e-686f-485b-aca5-953cb16bb9f8" containerID="e042028628e4b4e55e4e8b5856d23911587782bb0084a8dd5f20cf8d5e59a020" exitCode=0 Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.053052 4776 generic.go:334] "Generic (PLEG): container finished" podID="2c22f76e-686f-485b-aca5-953cb16bb9f8" containerID="7f4b79ed91bbd26f7e991ea152d411c3c65c2f4a5d540c19584b810f0b2da628" exitCode=0 Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.058272 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2c22f76e-686f-485b-aca5-953cb16bb9f8","Type":"ContainerDied","Data":"6c6a9570809a6163b94faa1d8799e71d2876afb79c7ed4d488abe35819ee5252"} Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.058326 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2c22f76e-686f-485b-aca5-953cb16bb9f8","Type":"ContainerDied","Data":"e042028628e4b4e55e4e8b5856d23911587782bb0084a8dd5f20cf8d5e59a020"} Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.058341 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2c22f76e-686f-485b-aca5-953cb16bb9f8","Type":"ContainerDied","Data":"7f4b79ed91bbd26f7e991ea152d411c3c65c2f4a5d540c19584b810f0b2da628"} Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.058390 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2c22f76e-686f-485b-aca5-953cb16bb9f8","Type":"ContainerDied","Data":"93345fabd70f94efa89630ae9d9e3e140216222237609877f2aee9e8d8295cfa"} Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.058416 4776 scope.go:117] "RemoveContainer" containerID="6c6a9570809a6163b94faa1d8799e71d2876afb79c7ed4d488abe35819ee5252" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.058701 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.081491 4776 generic.go:334] "Generic (PLEG): container finished" podID="dae628f6-9079-4f73-b208-0f8fd6d43af0" containerID="d8b57a933c20e558755b1388d4f9eda2cc171be3b0311d191c5e675f78e903e2" exitCode=137 Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.082134 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.085515 4776 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="dae628f6-9079-4f73-b208-0f8fd6d43af0" podUID="96898351-caa6-4edc-8080-898f9fff0f56" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.108577 4776 scope.go:117] "RemoveContainer" containerID="e042028628e4b4e55e4e8b5856d23911587782bb0084a8dd5f20cf8d5e59a020" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.124369 4776 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="dae628f6-9079-4f73-b208-0f8fd6d43af0" podUID="96898351-caa6-4edc-8080-898f9fff0f56" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.130103 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2c22f76e-686f-485b-aca5-953cb16bb9f8-thanos-prometheus-http-client-file\") pod \"2c22f76e-686f-485b-aca5-953cb16bb9f8\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.130482 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-472db1ee-c7d3-4f22-a8cf-04390f346bd1\") pod \"2c22f76e-686f-485b-aca5-953cb16bb9f8\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.130548 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2c22f76e-686f-485b-aca5-953cb16bb9f8-tls-assets\") pod \"2c22f76e-686f-485b-aca5-953cb16bb9f8\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.130749 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2c22f76e-686f-485b-aca5-953cb16bb9f8-prometheus-metric-storage-rulefiles-0\") pod \"2c22f76e-686f-485b-aca5-953cb16bb9f8\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.130801 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2c22f76e-686f-485b-aca5-953cb16bb9f8-config-out\") pod \"2c22f76e-686f-485b-aca5-953cb16bb9f8\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.130885 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4zrf\" (UniqueName: \"kubernetes.io/projected/2c22f76e-686f-485b-aca5-953cb16bb9f8-kube-api-access-w4zrf\") pod \"2c22f76e-686f-485b-aca5-953cb16bb9f8\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.130965 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2c22f76e-686f-485b-aca5-953cb16bb9f8-config\") pod \"2c22f76e-686f-485b-aca5-953cb16bb9f8\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.131040 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2c22f76e-686f-485b-aca5-953cb16bb9f8-web-config\") pod \"2c22f76e-686f-485b-aca5-953cb16bb9f8\" (UID: \"2c22f76e-686f-485b-aca5-953cb16bb9f8\") " Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.131738 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c22f76e-686f-485b-aca5-953cb16bb9f8-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "2c22f76e-686f-485b-aca5-953cb16bb9f8" (UID: "2c22f76e-686f-485b-aca5-953cb16bb9f8"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.132853 4776 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2c22f76e-686f-485b-aca5-953cb16bb9f8-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.138921 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c22f76e-686f-485b-aca5-953cb16bb9f8-config-out" (OuterVolumeSpecName: "config-out") pod "2c22f76e-686f-485b-aca5-953cb16bb9f8" (UID: "2c22f76e-686f-485b-aca5-953cb16bb9f8"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.138946 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c22f76e-686f-485b-aca5-953cb16bb9f8-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "2c22f76e-686f-485b-aca5-953cb16bb9f8" (UID: "2c22f76e-686f-485b-aca5-953cb16bb9f8"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.141247 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c22f76e-686f-485b-aca5-953cb16bb9f8-config" (OuterVolumeSpecName: "config") pod "2c22f76e-686f-485b-aca5-953cb16bb9f8" (UID: "2c22f76e-686f-485b-aca5-953cb16bb9f8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.149204 4776 scope.go:117] "RemoveContainer" containerID="7f4b79ed91bbd26f7e991ea152d411c3c65c2f4a5d540c19584b810f0b2da628" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.162525 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c22f76e-686f-485b-aca5-953cb16bb9f8-kube-api-access-w4zrf" (OuterVolumeSpecName: "kube-api-access-w4zrf") pod "2c22f76e-686f-485b-aca5-953cb16bb9f8" (UID: "2c22f76e-686f-485b-aca5-953cb16bb9f8"). InnerVolumeSpecName "kube-api-access-w4zrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.164040 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c22f76e-686f-485b-aca5-953cb16bb9f8-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "2c22f76e-686f-485b-aca5-953cb16bb9f8" (UID: "2c22f76e-686f-485b-aca5-953cb16bb9f8"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.184895 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c22f76e-686f-485b-aca5-953cb16bb9f8-web-config" (OuterVolumeSpecName: "web-config") pod "2c22f76e-686f-485b-aca5-953cb16bb9f8" (UID: "2c22f76e-686f-485b-aca5-953cb16bb9f8"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.230485 4776 scope.go:117] "RemoveContainer" containerID="a5c6e299e3d5f114ab1ec22f96d9da9be9aa6e0ed52170b485116a0011062378" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.234933 4776 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2c22f76e-686f-485b-aca5-953cb16bb9f8-tls-assets\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.235185 4776 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2c22f76e-686f-485b-aca5-953cb16bb9f8-config-out\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.235301 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4zrf\" (UniqueName: \"kubernetes.io/projected/2c22f76e-686f-485b-aca5-953cb16bb9f8-kube-api-access-w4zrf\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.235391 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/2c22f76e-686f-485b-aca5-953cb16bb9f8-config\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.235450 4776 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2c22f76e-686f-485b-aca5-953cb16bb9f8-web-config\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.235510 4776 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2c22f76e-686f-485b-aca5-953cb16bb9f8-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.259433 4776 scope.go:117] "RemoveContainer" containerID="6c6a9570809a6163b94faa1d8799e71d2876afb79c7ed4d488abe35819ee5252" Nov 25 11:12:57 crc kubenswrapper[4776]: E1125 11:12:57.259846 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c6a9570809a6163b94faa1d8799e71d2876afb79c7ed4d488abe35819ee5252\": container with ID starting with 6c6a9570809a6163b94faa1d8799e71d2876afb79c7ed4d488abe35819ee5252 not found: ID does not exist" containerID="6c6a9570809a6163b94faa1d8799e71d2876afb79c7ed4d488abe35819ee5252" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.259884 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c6a9570809a6163b94faa1d8799e71d2876afb79c7ed4d488abe35819ee5252"} err="failed to get container status \"6c6a9570809a6163b94faa1d8799e71d2876afb79c7ed4d488abe35819ee5252\": rpc error: code = NotFound desc = could not find container \"6c6a9570809a6163b94faa1d8799e71d2876afb79c7ed4d488abe35819ee5252\": container with ID starting with 6c6a9570809a6163b94faa1d8799e71d2876afb79c7ed4d488abe35819ee5252 not found: ID does not exist" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.259909 4776 scope.go:117] "RemoveContainer" containerID="e042028628e4b4e55e4e8b5856d23911587782bb0084a8dd5f20cf8d5e59a020" Nov 25 11:12:57 crc kubenswrapper[4776]: E1125 11:12:57.261095 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e042028628e4b4e55e4e8b5856d23911587782bb0084a8dd5f20cf8d5e59a020\": container with ID starting with e042028628e4b4e55e4e8b5856d23911587782bb0084a8dd5f20cf8d5e59a020 not found: ID does not exist" containerID="e042028628e4b4e55e4e8b5856d23911587782bb0084a8dd5f20cf8d5e59a020" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.261122 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e042028628e4b4e55e4e8b5856d23911587782bb0084a8dd5f20cf8d5e59a020"} err="failed to get container status \"e042028628e4b4e55e4e8b5856d23911587782bb0084a8dd5f20cf8d5e59a020\": rpc error: code = NotFound desc = could not find container \"e042028628e4b4e55e4e8b5856d23911587782bb0084a8dd5f20cf8d5e59a020\": container with ID starting with e042028628e4b4e55e4e8b5856d23911587782bb0084a8dd5f20cf8d5e59a020 not found: ID does not exist" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.261139 4776 scope.go:117] "RemoveContainer" containerID="7f4b79ed91bbd26f7e991ea152d411c3c65c2f4a5d540c19584b810f0b2da628" Nov 25 11:12:57 crc kubenswrapper[4776]: E1125 11:12:57.261627 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f4b79ed91bbd26f7e991ea152d411c3c65c2f4a5d540c19584b810f0b2da628\": container with ID starting with 7f4b79ed91bbd26f7e991ea152d411c3c65c2f4a5d540c19584b810f0b2da628 not found: ID does not exist" containerID="7f4b79ed91bbd26f7e991ea152d411c3c65c2f4a5d540c19584b810f0b2da628" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.261657 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f4b79ed91bbd26f7e991ea152d411c3c65c2f4a5d540c19584b810f0b2da628"} err="failed to get container status \"7f4b79ed91bbd26f7e991ea152d411c3c65c2f4a5d540c19584b810f0b2da628\": rpc error: code = NotFound desc = could not find container \"7f4b79ed91bbd26f7e991ea152d411c3c65c2f4a5d540c19584b810f0b2da628\": container with ID starting with 7f4b79ed91bbd26f7e991ea152d411c3c65c2f4a5d540c19584b810f0b2da628 not found: ID does not exist" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.261672 4776 scope.go:117] "RemoveContainer" containerID="a5c6e299e3d5f114ab1ec22f96d9da9be9aa6e0ed52170b485116a0011062378" Nov 25 11:12:57 crc kubenswrapper[4776]: E1125 11:12:57.262154 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5c6e299e3d5f114ab1ec22f96d9da9be9aa6e0ed52170b485116a0011062378\": container with ID starting with a5c6e299e3d5f114ab1ec22f96d9da9be9aa6e0ed52170b485116a0011062378 not found: ID does not exist" containerID="a5c6e299e3d5f114ab1ec22f96d9da9be9aa6e0ed52170b485116a0011062378" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.262194 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5c6e299e3d5f114ab1ec22f96d9da9be9aa6e0ed52170b485116a0011062378"} err="failed to get container status \"a5c6e299e3d5f114ab1ec22f96d9da9be9aa6e0ed52170b485116a0011062378\": rpc error: code = NotFound desc = could not find container \"a5c6e299e3d5f114ab1ec22f96d9da9be9aa6e0ed52170b485116a0011062378\": container with ID starting with a5c6e299e3d5f114ab1ec22f96d9da9be9aa6e0ed52170b485116a0011062378 not found: ID does not exist" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.262219 4776 scope.go:117] "RemoveContainer" containerID="6c6a9570809a6163b94faa1d8799e71d2876afb79c7ed4d488abe35819ee5252" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.262503 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c6a9570809a6163b94faa1d8799e71d2876afb79c7ed4d488abe35819ee5252"} err="failed to get container status \"6c6a9570809a6163b94faa1d8799e71d2876afb79c7ed4d488abe35819ee5252\": rpc error: code = NotFound desc = could not find container \"6c6a9570809a6163b94faa1d8799e71d2876afb79c7ed4d488abe35819ee5252\": container with ID starting with 6c6a9570809a6163b94faa1d8799e71d2876afb79c7ed4d488abe35819ee5252 not found: ID does not exist" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.262527 4776 scope.go:117] "RemoveContainer" containerID="e042028628e4b4e55e4e8b5856d23911587782bb0084a8dd5f20cf8d5e59a020" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.265629 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e042028628e4b4e55e4e8b5856d23911587782bb0084a8dd5f20cf8d5e59a020"} err="failed to get container status \"e042028628e4b4e55e4e8b5856d23911587782bb0084a8dd5f20cf8d5e59a020\": rpc error: code = NotFound desc = could not find container \"e042028628e4b4e55e4e8b5856d23911587782bb0084a8dd5f20cf8d5e59a020\": container with ID starting with e042028628e4b4e55e4e8b5856d23911587782bb0084a8dd5f20cf8d5e59a020 not found: ID does not exist" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.265784 4776 scope.go:117] "RemoveContainer" containerID="7f4b79ed91bbd26f7e991ea152d411c3c65c2f4a5d540c19584b810f0b2da628" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.266336 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f4b79ed91bbd26f7e991ea152d411c3c65c2f4a5d540c19584b810f0b2da628"} err="failed to get container status \"7f4b79ed91bbd26f7e991ea152d411c3c65c2f4a5d540c19584b810f0b2da628\": rpc error: code = NotFound desc = could not find container \"7f4b79ed91bbd26f7e991ea152d411c3c65c2f4a5d540c19584b810f0b2da628\": container with ID starting with 7f4b79ed91bbd26f7e991ea152d411c3c65c2f4a5d540c19584b810f0b2da628 not found: ID does not exist" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.266429 4776 scope.go:117] "RemoveContainer" containerID="a5c6e299e3d5f114ab1ec22f96d9da9be9aa6e0ed52170b485116a0011062378" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.266733 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5c6e299e3d5f114ab1ec22f96d9da9be9aa6e0ed52170b485116a0011062378"} err="failed to get container status \"a5c6e299e3d5f114ab1ec22f96d9da9be9aa6e0ed52170b485116a0011062378\": rpc error: code = NotFound desc = could not find container \"a5c6e299e3d5f114ab1ec22f96d9da9be9aa6e0ed52170b485116a0011062378\": container with ID starting with a5c6e299e3d5f114ab1ec22f96d9da9be9aa6e0ed52170b485116a0011062378 not found: ID does not exist" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.266826 4776 scope.go:117] "RemoveContainer" containerID="6c6a9570809a6163b94faa1d8799e71d2876afb79c7ed4d488abe35819ee5252" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.275730 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c6a9570809a6163b94faa1d8799e71d2876afb79c7ed4d488abe35819ee5252"} err="failed to get container status \"6c6a9570809a6163b94faa1d8799e71d2876afb79c7ed4d488abe35819ee5252\": rpc error: code = NotFound desc = could not find container \"6c6a9570809a6163b94faa1d8799e71d2876afb79c7ed4d488abe35819ee5252\": container with ID starting with 6c6a9570809a6163b94faa1d8799e71d2876afb79c7ed4d488abe35819ee5252 not found: ID does not exist" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.275784 4776 scope.go:117] "RemoveContainer" containerID="e042028628e4b4e55e4e8b5856d23911587782bb0084a8dd5f20cf8d5e59a020" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.276419 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e042028628e4b4e55e4e8b5856d23911587782bb0084a8dd5f20cf8d5e59a020"} err="failed to get container status \"e042028628e4b4e55e4e8b5856d23911587782bb0084a8dd5f20cf8d5e59a020\": rpc error: code = NotFound desc = could not find container \"e042028628e4b4e55e4e8b5856d23911587782bb0084a8dd5f20cf8d5e59a020\": container with ID starting with e042028628e4b4e55e4e8b5856d23911587782bb0084a8dd5f20cf8d5e59a020 not found: ID does not exist" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.276453 4776 scope.go:117] "RemoveContainer" containerID="7f4b79ed91bbd26f7e991ea152d411c3c65c2f4a5d540c19584b810f0b2da628" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.279610 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f4b79ed91bbd26f7e991ea152d411c3c65c2f4a5d540c19584b810f0b2da628"} err="failed to get container status \"7f4b79ed91bbd26f7e991ea152d411c3c65c2f4a5d540c19584b810f0b2da628\": rpc error: code = NotFound desc = could not find container \"7f4b79ed91bbd26f7e991ea152d411c3c65c2f4a5d540c19584b810f0b2da628\": container with ID starting with 7f4b79ed91bbd26f7e991ea152d411c3c65c2f4a5d540c19584b810f0b2da628 not found: ID does not exist" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.279644 4776 scope.go:117] "RemoveContainer" containerID="a5c6e299e3d5f114ab1ec22f96d9da9be9aa6e0ed52170b485116a0011062378" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.280005 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5c6e299e3d5f114ab1ec22f96d9da9be9aa6e0ed52170b485116a0011062378"} err="failed to get container status \"a5c6e299e3d5f114ab1ec22f96d9da9be9aa6e0ed52170b485116a0011062378\": rpc error: code = NotFound desc = could not find container \"a5c6e299e3d5f114ab1ec22f96d9da9be9aa6e0ed52170b485116a0011062378\": container with ID starting with a5c6e299e3d5f114ab1ec22f96d9da9be9aa6e0ed52170b485116a0011062378 not found: ID does not exist" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.280047 4776 scope.go:117] "RemoveContainer" containerID="d8b57a933c20e558755b1388d4f9eda2cc171be3b0311d191c5e675f78e903e2" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.303054 4776 scope.go:117] "RemoveContainer" containerID="d8b57a933c20e558755b1388d4f9eda2cc171be3b0311d191c5e675f78e903e2" Nov 25 11:12:57 crc kubenswrapper[4776]: E1125 11:12:57.303562 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8b57a933c20e558755b1388d4f9eda2cc171be3b0311d191c5e675f78e903e2\": container with ID starting with d8b57a933c20e558755b1388d4f9eda2cc171be3b0311d191c5e675f78e903e2 not found: ID does not exist" containerID="d8b57a933c20e558755b1388d4f9eda2cc171be3b0311d191c5e675f78e903e2" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.303592 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8b57a933c20e558755b1388d4f9eda2cc171be3b0311d191c5e675f78e903e2"} err="failed to get container status \"d8b57a933c20e558755b1388d4f9eda2cc171be3b0311d191c5e675f78e903e2\": rpc error: code = NotFound desc = could not find container \"d8b57a933c20e558755b1388d4f9eda2cc171be3b0311d191c5e675f78e903e2\": container with ID starting with d8b57a933c20e558755b1388d4f9eda2cc171be3b0311d191c5e675f78e903e2 not found: ID does not exist" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.696855 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dae628f6-9079-4f73-b208-0f8fd6d43af0" path="/var/lib/kubelet/pods/dae628f6-9079-4f73-b208-0f8fd6d43af0/volumes" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.708536 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-472db1ee-c7d3-4f22-a8cf-04390f346bd1" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "2c22f76e-686f-485b-aca5-953cb16bb9f8" (UID: "2c22f76e-686f-485b-aca5-953cb16bb9f8"). InnerVolumeSpecName "pvc-472db1ee-c7d3-4f22-a8cf-04390f346bd1". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.747465 4776 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-472db1ee-c7d3-4f22-a8cf-04390f346bd1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-472db1ee-c7d3-4f22-a8cf-04390f346bd1\") on node \"crc\" " Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.810349 4776 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.810571 4776 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-472db1ee-c7d3-4f22-a8cf-04390f346bd1" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-472db1ee-c7d3-4f22-a8cf-04390f346bd1") on node "crc" Nov 25 11:12:57 crc kubenswrapper[4776]: I1125 11:12:57.850842 4776 reconciler_common.go:293] "Volume detached for volume \"pvc-472db1ee-c7d3-4f22-a8cf-04390f346bd1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-472db1ee-c7d3-4f22-a8cf-04390f346bd1\") on node \"crc\" DevicePath \"\"" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.009113 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.018666 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.034980 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 25 11:12:58 crc kubenswrapper[4776]: E1125 11:12:58.035544 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c22f76e-686f-485b-aca5-953cb16bb9f8" containerName="init-config-reloader" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.035568 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c22f76e-686f-485b-aca5-953cb16bb9f8" containerName="init-config-reloader" Nov 25 11:12:58 crc kubenswrapper[4776]: E1125 11:12:58.035584 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c22f76e-686f-485b-aca5-953cb16bb9f8" containerName="thanos-sidecar" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.035592 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c22f76e-686f-485b-aca5-953cb16bb9f8" containerName="thanos-sidecar" Nov 25 11:12:58 crc kubenswrapper[4776]: E1125 11:12:58.035637 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c22f76e-686f-485b-aca5-953cb16bb9f8" containerName="config-reloader" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.035645 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c22f76e-686f-485b-aca5-953cb16bb9f8" containerName="config-reloader" Nov 25 11:12:58 crc kubenswrapper[4776]: E1125 11:12:58.035657 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c22f76e-686f-485b-aca5-953cb16bb9f8" containerName="prometheus" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.035665 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c22f76e-686f-485b-aca5-953cb16bb9f8" containerName="prometheus" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.035924 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c22f76e-686f-485b-aca5-953cb16bb9f8" containerName="thanos-sidecar" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.035952 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c22f76e-686f-485b-aca5-953cb16bb9f8" containerName="prometheus" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.035970 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c22f76e-686f-485b-aca5-953cb16bb9f8" containerName="config-reloader" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.038295 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.042544 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.042770 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.042991 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.043847 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.044083 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-g6rm2" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.044357 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.052631 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.103137 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.169515 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/31358dae-00d6-47d3-b9aa-151088320199-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.169701 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/31358dae-00d6-47d3-b9aa-151088320199-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.169870 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-472db1ee-c7d3-4f22-a8cf-04390f346bd1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-472db1ee-c7d3-4f22-a8cf-04390f346bd1\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.169929 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/31358dae-00d6-47d3-b9aa-151088320199-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.170176 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/31358dae-00d6-47d3-b9aa-151088320199-config\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.170271 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/31358dae-00d6-47d3-b9aa-151088320199-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.170321 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/31358dae-00d6-47d3-b9aa-151088320199-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.170385 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccwvh\" (UniqueName: \"kubernetes.io/projected/31358dae-00d6-47d3-b9aa-151088320199-kube-api-access-ccwvh\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.170527 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/31358dae-00d6-47d3-b9aa-151088320199-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.170557 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31358dae-00d6-47d3-b9aa-151088320199-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.170704 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/31358dae-00d6-47d3-b9aa-151088320199-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.272449 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/31358dae-00d6-47d3-b9aa-151088320199-config\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.272537 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/31358dae-00d6-47d3-b9aa-151088320199-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.272567 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/31358dae-00d6-47d3-b9aa-151088320199-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.272613 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccwvh\" (UniqueName: \"kubernetes.io/projected/31358dae-00d6-47d3-b9aa-151088320199-kube-api-access-ccwvh\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.272681 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/31358dae-00d6-47d3-b9aa-151088320199-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.272709 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31358dae-00d6-47d3-b9aa-151088320199-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.272757 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/31358dae-00d6-47d3-b9aa-151088320199-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.272789 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/31358dae-00d6-47d3-b9aa-151088320199-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.272868 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/31358dae-00d6-47d3-b9aa-151088320199-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.272910 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-472db1ee-c7d3-4f22-a8cf-04390f346bd1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-472db1ee-c7d3-4f22-a8cf-04390f346bd1\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.272934 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/31358dae-00d6-47d3-b9aa-151088320199-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.275517 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/31358dae-00d6-47d3-b9aa-151088320199-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.279288 4776 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.279334 4776 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-472db1ee-c7d3-4f22-a8cf-04390f346bd1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-472db1ee-c7d3-4f22-a8cf-04390f346bd1\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ff6e355e361fb5843939103249f848baecb449fc7ba76f9864fb5e4e8612b051/globalmount\"" pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.279462 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/31358dae-00d6-47d3-b9aa-151088320199-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.280029 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/31358dae-00d6-47d3-b9aa-151088320199-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.280460 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/31358dae-00d6-47d3-b9aa-151088320199-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.281445 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/31358dae-00d6-47d3-b9aa-151088320199-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.281450 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/31358dae-00d6-47d3-b9aa-151088320199-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.281473 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/31358dae-00d6-47d3-b9aa-151088320199-config\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.281774 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/31358dae-00d6-47d3-b9aa-151088320199-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.282588 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31358dae-00d6-47d3-b9aa-151088320199-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.300028 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccwvh\" (UniqueName: \"kubernetes.io/projected/31358dae-00d6-47d3-b9aa-151088320199-kube-api-access-ccwvh\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.327215 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-472db1ee-c7d3-4f22-a8cf-04390f346bd1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-472db1ee-c7d3-4f22-a8cf-04390f346bd1\") pod \"prometheus-metric-storage-0\" (UID: \"31358dae-00d6-47d3-b9aa-151088320199\") " pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.366459 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.489982 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.505755 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.506221 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.510507 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.510855 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.579503 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9ea0919-b2a4-4c0e-9952-086726eab7b0-config-data\") pod \"ceilometer-0\" (UID: \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\") " pod="openstack/ceilometer-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.579552 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9ea0919-b2a4-4c0e-9952-086726eab7b0-scripts\") pod \"ceilometer-0\" (UID: \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\") " pod="openstack/ceilometer-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.579578 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nmbm\" (UniqueName: \"kubernetes.io/projected/a9ea0919-b2a4-4c0e-9952-086726eab7b0-kube-api-access-8nmbm\") pod \"ceilometer-0\" (UID: \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\") " pod="openstack/ceilometer-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.579644 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9ea0919-b2a4-4c0e-9952-086726eab7b0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\") " pod="openstack/ceilometer-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.579664 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9ea0919-b2a4-4c0e-9952-086726eab7b0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\") " pod="openstack/ceilometer-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.579839 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9ea0919-b2a4-4c0e-9952-086726eab7b0-run-httpd\") pod \"ceilometer-0\" (UID: \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\") " pod="openstack/ceilometer-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.580080 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9ea0919-b2a4-4c0e-9952-086726eab7b0-log-httpd\") pod \"ceilometer-0\" (UID: \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\") " pod="openstack/ceilometer-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.682059 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9ea0919-b2a4-4c0e-9952-086726eab7b0-config-data\") pod \"ceilometer-0\" (UID: \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\") " pod="openstack/ceilometer-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.682137 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9ea0919-b2a4-4c0e-9952-086726eab7b0-scripts\") pod \"ceilometer-0\" (UID: \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\") " pod="openstack/ceilometer-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.682175 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nmbm\" (UniqueName: \"kubernetes.io/projected/a9ea0919-b2a4-4c0e-9952-086726eab7b0-kube-api-access-8nmbm\") pod \"ceilometer-0\" (UID: \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\") " pod="openstack/ceilometer-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.682226 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9ea0919-b2a4-4c0e-9952-086726eab7b0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\") " pod="openstack/ceilometer-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.682249 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9ea0919-b2a4-4c0e-9952-086726eab7b0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\") " pod="openstack/ceilometer-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.682330 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9ea0919-b2a4-4c0e-9952-086726eab7b0-run-httpd\") pod \"ceilometer-0\" (UID: \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\") " pod="openstack/ceilometer-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.682420 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9ea0919-b2a4-4c0e-9952-086726eab7b0-log-httpd\") pod \"ceilometer-0\" (UID: \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\") " pod="openstack/ceilometer-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.683151 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9ea0919-b2a4-4c0e-9952-086726eab7b0-log-httpd\") pod \"ceilometer-0\" (UID: \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\") " pod="openstack/ceilometer-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.685148 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9ea0919-b2a4-4c0e-9952-086726eab7b0-run-httpd\") pod \"ceilometer-0\" (UID: \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\") " pod="openstack/ceilometer-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.688614 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9ea0919-b2a4-4c0e-9952-086726eab7b0-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\") " pod="openstack/ceilometer-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.707849 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9ea0919-b2a4-4c0e-9952-086726eab7b0-config-data\") pod \"ceilometer-0\" (UID: \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\") " pod="openstack/ceilometer-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.707927 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9ea0919-b2a4-4c0e-9952-086726eab7b0-scripts\") pod \"ceilometer-0\" (UID: \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\") " pod="openstack/ceilometer-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.708601 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9ea0919-b2a4-4c0e-9952-086726eab7b0-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\") " pod="openstack/ceilometer-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.714185 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nmbm\" (UniqueName: \"kubernetes.io/projected/a9ea0919-b2a4-4c0e-9952-086726eab7b0-kube-api-access-8nmbm\") pod \"ceilometer-0\" (UID: \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\") " pod="openstack/ceilometer-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.840482 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 11:12:58 crc kubenswrapper[4776]: I1125 11:12:58.950735 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Nov 25 11:12:59 crc kubenswrapper[4776]: I1125 11:12:59.120865 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"31358dae-00d6-47d3-b9aa-151088320199","Type":"ContainerStarted","Data":"13fa503c7083d99d9ac073b238f12f6944f8f655972aa0d8bfd59ec9ea0299c2"} Nov 25 11:12:59 crc kubenswrapper[4776]: I1125 11:12:59.371134 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 11:12:59 crc kubenswrapper[4776]: W1125 11:12:59.379441 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9ea0919_b2a4_4c0e_9952_086726eab7b0.slice/crio-2a188e6540a5beb503c4eca5d06ccdea4716fc22721f93bffaa5d43b44472d35 WatchSource:0}: Error finding container 2a188e6540a5beb503c4eca5d06ccdea4716fc22721f93bffaa5d43b44472d35: Status 404 returned error can't find the container with id 2a188e6540a5beb503c4eca5d06ccdea4716fc22721f93bffaa5d43b44472d35 Nov 25 11:12:59 crc kubenswrapper[4776]: I1125 11:12:59.676536 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c22f76e-686f-485b-aca5-953cb16bb9f8" path="/var/lib/kubelet/pods/2c22f76e-686f-485b-aca5-953cb16bb9f8/volumes" Nov 25 11:13:00 crc kubenswrapper[4776]: I1125 11:13:00.131396 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9ea0919-b2a4-4c0e-9952-086726eab7b0","Type":"ContainerStarted","Data":"2a188e6540a5beb503c4eca5d06ccdea4716fc22721f93bffaa5d43b44472d35"} Nov 25 11:13:01 crc kubenswrapper[4776]: I1125 11:13:01.141575 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9ea0919-b2a4-4c0e-9952-086726eab7b0","Type":"ContainerStarted","Data":"846a3c6a75464bfdff065d3d35f3fdc38fae1085bfa6576dc2a277ba9e77dcc8"} Nov 25 11:13:02 crc kubenswrapper[4776]: I1125 11:13:02.057868 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-8flmt"] Nov 25 11:13:02 crc kubenswrapper[4776]: I1125 11:13:02.069705 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-8flmt"] Nov 25 11:13:03 crc kubenswrapper[4776]: I1125 11:13:03.028033 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-2x8x7"] Nov 25 11:13:03 crc kubenswrapper[4776]: I1125 11:13:03.044123 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-2x8x7"] Nov 25 11:13:03 crc kubenswrapper[4776]: I1125 11:13:03.164976 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"31358dae-00d6-47d3-b9aa-151088320199","Type":"ContainerStarted","Data":"dcf31a9d511d343f87106131c7058b8caa0734732a2476ce92b388f9edf87cf3"} Nov 25 11:13:03 crc kubenswrapper[4776]: I1125 11:13:03.168729 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9ea0919-b2a4-4c0e-9952-086726eab7b0","Type":"ContainerStarted","Data":"669f88bd2c5c9856836d44a330d923704adde9f7c2b67c1ad502fa363c05a90b"} Nov 25 11:13:03 crc kubenswrapper[4776]: I1125 11:13:03.678276 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4988892b-e67d-4595-abb3-c097fa0808b5" path="/var/lib/kubelet/pods/4988892b-e67d-4595-abb3-c097fa0808b5/volumes" Nov 25 11:13:03 crc kubenswrapper[4776]: I1125 11:13:03.679415 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6d7ec58-a5a8-48aa-920d-d7f81e603f76" path="/var/lib/kubelet/pods/d6d7ec58-a5a8-48aa-920d-d7f81e603f76/volumes" Nov 25 11:13:04 crc kubenswrapper[4776]: I1125 11:13:04.037572 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-e3f8-account-create-ht27m"] Nov 25 11:13:04 crc kubenswrapper[4776]: I1125 11:13:04.051638 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-jj7jb"] Nov 25 11:13:04 crc kubenswrapper[4776]: I1125 11:13:04.059762 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-173a-account-create-2vr9z"] Nov 25 11:13:04 crc kubenswrapper[4776]: I1125 11:13:04.073359 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-173a-account-create-2vr9z"] Nov 25 11:13:04 crc kubenswrapper[4776]: I1125 11:13:04.083259 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-0d1a-account-create-mhvkn"] Nov 25 11:13:04 crc kubenswrapper[4776]: I1125 11:13:04.092186 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-jj7jb"] Nov 25 11:13:04 crc kubenswrapper[4776]: I1125 11:13:04.101738 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-e3f8-account-create-ht27m"] Nov 25 11:13:04 crc kubenswrapper[4776]: I1125 11:13:04.110385 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-0d1a-account-create-mhvkn"] Nov 25 11:13:04 crc kubenswrapper[4776]: I1125 11:13:04.180742 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9ea0919-b2a4-4c0e-9952-086726eab7b0","Type":"ContainerStarted","Data":"1f37ba68d430d93cb9a009e023cf114dd8ef9b19168405138e43bbb64adb038a"} Nov 25 11:13:05 crc kubenswrapper[4776]: I1125 11:13:05.679318 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c7b6238-2ba3-4d7f-a7f2-8f74092ce260" path="/var/lib/kubelet/pods/2c7b6238-2ba3-4d7f-a7f2-8f74092ce260/volumes" Nov 25 11:13:05 crc kubenswrapper[4776]: I1125 11:13:05.683272 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38ec4382-1ef1-4cab-af5b-26ffecd9354a" path="/var/lib/kubelet/pods/38ec4382-1ef1-4cab-af5b-26ffecd9354a/volumes" Nov 25 11:13:05 crc kubenswrapper[4776]: I1125 11:13:05.694551 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96dba744-9bb8-480c-8498-c8eedd4fdb88" path="/var/lib/kubelet/pods/96dba744-9bb8-480c-8498-c8eedd4fdb88/volumes" Nov 25 11:13:05 crc kubenswrapper[4776]: I1125 11:13:05.696242 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9ae59af-54ec-4b1d-8285-a248f0b5f50a" path="/var/lib/kubelet/pods/a9ae59af-54ec-4b1d-8285-a248f0b5f50a/volumes" Nov 25 11:13:06 crc kubenswrapper[4776]: I1125 11:13:06.202120 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9ea0919-b2a4-4c0e-9952-086726eab7b0","Type":"ContainerStarted","Data":"cf92e477f0608de994f1aabfe21a726b96a191b6930d5a2828fd2da621ee1f7b"} Nov 25 11:13:06 crc kubenswrapper[4776]: I1125 11:13:06.202626 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 11:13:06 crc kubenswrapper[4776]: I1125 11:13:06.238362 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.888253442 podStartE2EDuration="8.238344019s" podCreationTimestamp="2025-11-25 11:12:58 +0000 UTC" firstStartedPulling="2025-11-25 11:12:59.385428293 +0000 UTC m=+6524.426487846" lastFinishedPulling="2025-11-25 11:13:05.73551886 +0000 UTC m=+6530.776578423" observedRunningTime="2025-11-25 11:13:06.232474112 +0000 UTC m=+6531.273533665" watchObservedRunningTime="2025-11-25 11:13:06.238344019 +0000 UTC m=+6531.279403572" Nov 25 11:13:10 crc kubenswrapper[4776]: I1125 11:13:10.870759 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-jh4vh"] Nov 25 11:13:10 crc kubenswrapper[4776]: I1125 11:13:10.873104 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-jh4vh" Nov 25 11:13:10 crc kubenswrapper[4776]: I1125 11:13:10.881768 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-jh4vh"] Nov 25 11:13:10 crc kubenswrapper[4776]: I1125 11:13:10.945787 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93752b39-730c-470f-bbd9-2c9d39abbb03-operator-scripts\") pod \"aodh-db-create-jh4vh\" (UID: \"93752b39-730c-470f-bbd9-2c9d39abbb03\") " pod="openstack/aodh-db-create-jh4vh" Nov 25 11:13:10 crc kubenswrapper[4776]: I1125 11:13:10.945911 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4f5b\" (UniqueName: \"kubernetes.io/projected/93752b39-730c-470f-bbd9-2c9d39abbb03-kube-api-access-q4f5b\") pod \"aodh-db-create-jh4vh\" (UID: \"93752b39-730c-470f-bbd9-2c9d39abbb03\") " pod="openstack/aodh-db-create-jh4vh" Nov 25 11:13:11 crc kubenswrapper[4776]: I1125 11:13:11.048227 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93752b39-730c-470f-bbd9-2c9d39abbb03-operator-scripts\") pod \"aodh-db-create-jh4vh\" (UID: \"93752b39-730c-470f-bbd9-2c9d39abbb03\") " pod="openstack/aodh-db-create-jh4vh" Nov 25 11:13:11 crc kubenswrapper[4776]: I1125 11:13:11.048328 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4f5b\" (UniqueName: \"kubernetes.io/projected/93752b39-730c-470f-bbd9-2c9d39abbb03-kube-api-access-q4f5b\") pod \"aodh-db-create-jh4vh\" (UID: \"93752b39-730c-470f-bbd9-2c9d39abbb03\") " pod="openstack/aodh-db-create-jh4vh" Nov 25 11:13:11 crc kubenswrapper[4776]: I1125 11:13:11.049504 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93752b39-730c-470f-bbd9-2c9d39abbb03-operator-scripts\") pod \"aodh-db-create-jh4vh\" (UID: \"93752b39-730c-470f-bbd9-2c9d39abbb03\") " pod="openstack/aodh-db-create-jh4vh" Nov 25 11:13:11 crc kubenswrapper[4776]: I1125 11:13:11.093473 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4f5b\" (UniqueName: \"kubernetes.io/projected/93752b39-730c-470f-bbd9-2c9d39abbb03-kube-api-access-q4f5b\") pod \"aodh-db-create-jh4vh\" (UID: \"93752b39-730c-470f-bbd9-2c9d39abbb03\") " pod="openstack/aodh-db-create-jh4vh" Nov 25 11:13:11 crc kubenswrapper[4776]: I1125 11:13:11.135267 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-d479-account-create-6qg9g"] Nov 25 11:13:11 crc kubenswrapper[4776]: I1125 11:13:11.137173 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-d479-account-create-6qg9g" Nov 25 11:13:11 crc kubenswrapper[4776]: I1125 11:13:11.142837 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Nov 25 11:13:11 crc kubenswrapper[4776]: I1125 11:13:11.158222 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-d479-account-create-6qg9g"] Nov 25 11:13:11 crc kubenswrapper[4776]: I1125 11:13:11.201143 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-jh4vh" Nov 25 11:13:11 crc kubenswrapper[4776]: I1125 11:13:11.252708 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2dfc4fe-0a7f-4670-9075-3712e8e9f775-operator-scripts\") pod \"aodh-d479-account-create-6qg9g\" (UID: \"c2dfc4fe-0a7f-4670-9075-3712e8e9f775\") " pod="openstack/aodh-d479-account-create-6qg9g" Nov 25 11:13:11 crc kubenswrapper[4776]: I1125 11:13:11.252773 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bfqp\" (UniqueName: \"kubernetes.io/projected/c2dfc4fe-0a7f-4670-9075-3712e8e9f775-kube-api-access-4bfqp\") pod \"aodh-d479-account-create-6qg9g\" (UID: \"c2dfc4fe-0a7f-4670-9075-3712e8e9f775\") " pod="openstack/aodh-d479-account-create-6qg9g" Nov 25 11:13:11 crc kubenswrapper[4776]: I1125 11:13:11.256558 4776 generic.go:334] "Generic (PLEG): container finished" podID="31358dae-00d6-47d3-b9aa-151088320199" containerID="dcf31a9d511d343f87106131c7058b8caa0734732a2476ce92b388f9edf87cf3" exitCode=0 Nov 25 11:13:11 crc kubenswrapper[4776]: I1125 11:13:11.256598 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"31358dae-00d6-47d3-b9aa-151088320199","Type":"ContainerDied","Data":"dcf31a9d511d343f87106131c7058b8caa0734732a2476ce92b388f9edf87cf3"} Nov 25 11:13:11 crc kubenswrapper[4776]: I1125 11:13:11.355011 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2dfc4fe-0a7f-4670-9075-3712e8e9f775-operator-scripts\") pod \"aodh-d479-account-create-6qg9g\" (UID: \"c2dfc4fe-0a7f-4670-9075-3712e8e9f775\") " pod="openstack/aodh-d479-account-create-6qg9g" Nov 25 11:13:11 crc kubenswrapper[4776]: I1125 11:13:11.355428 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bfqp\" (UniqueName: \"kubernetes.io/projected/c2dfc4fe-0a7f-4670-9075-3712e8e9f775-kube-api-access-4bfqp\") pod \"aodh-d479-account-create-6qg9g\" (UID: \"c2dfc4fe-0a7f-4670-9075-3712e8e9f775\") " pod="openstack/aodh-d479-account-create-6qg9g" Nov 25 11:13:11 crc kubenswrapper[4776]: I1125 11:13:11.356698 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2dfc4fe-0a7f-4670-9075-3712e8e9f775-operator-scripts\") pod \"aodh-d479-account-create-6qg9g\" (UID: \"c2dfc4fe-0a7f-4670-9075-3712e8e9f775\") " pod="openstack/aodh-d479-account-create-6qg9g" Nov 25 11:13:11 crc kubenswrapper[4776]: I1125 11:13:11.384166 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bfqp\" (UniqueName: \"kubernetes.io/projected/c2dfc4fe-0a7f-4670-9075-3712e8e9f775-kube-api-access-4bfqp\") pod \"aodh-d479-account-create-6qg9g\" (UID: \"c2dfc4fe-0a7f-4670-9075-3712e8e9f775\") " pod="openstack/aodh-d479-account-create-6qg9g" Nov 25 11:13:11 crc kubenswrapper[4776]: I1125 11:13:11.462681 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-d479-account-create-6qg9g" Nov 25 11:13:11 crc kubenswrapper[4776]: I1125 11:13:11.815081 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-jh4vh"] Nov 25 11:13:11 crc kubenswrapper[4776]: W1125 11:13:11.818490 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93752b39_730c_470f_bbd9_2c9d39abbb03.slice/crio-bcdda743dac82f6390b7220ea76b95efe87a71fa90db786e957d5c079f53322f WatchSource:0}: Error finding container bcdda743dac82f6390b7220ea76b95efe87a71fa90db786e957d5c079f53322f: Status 404 returned error can't find the container with id bcdda743dac82f6390b7220ea76b95efe87a71fa90db786e957d5c079f53322f Nov 25 11:13:12 crc kubenswrapper[4776]: I1125 11:13:12.035589 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-d479-account-create-6qg9g"] Nov 25 11:13:12 crc kubenswrapper[4776]: I1125 11:13:12.267670 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-d479-account-create-6qg9g" event={"ID":"c2dfc4fe-0a7f-4670-9075-3712e8e9f775","Type":"ContainerStarted","Data":"8ca4352b56c459d9aa818b03808acd3312b884d82f0e69001bf6cc97dc4362fe"} Nov 25 11:13:12 crc kubenswrapper[4776]: I1125 11:13:12.269229 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-jh4vh" event={"ID":"93752b39-730c-470f-bbd9-2c9d39abbb03","Type":"ContainerStarted","Data":"bcdda743dac82f6390b7220ea76b95efe87a71fa90db786e957d5c079f53322f"} Nov 25 11:13:12 crc kubenswrapper[4776]: I1125 11:13:12.271627 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"31358dae-00d6-47d3-b9aa-151088320199","Type":"ContainerStarted","Data":"7c0ea1eb0b8c4b06a7a70555e929ea391da80ae5b59da0bc732ad358410381d9"} Nov 25 11:13:13 crc kubenswrapper[4776]: I1125 11:13:13.061140 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gpz2x"] Nov 25 11:13:13 crc kubenswrapper[4776]: I1125 11:13:13.074294 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gpz2x"] Nov 25 11:13:13 crc kubenswrapper[4776]: I1125 11:13:13.294972 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-d479-account-create-6qg9g" event={"ID":"c2dfc4fe-0a7f-4670-9075-3712e8e9f775","Type":"ContainerStarted","Data":"d47a1f096eb2f6c24dfa1f5885acd4fa5ce7a5ede02d8f2b3b92c9690e5297e7"} Nov 25 11:13:13 crc kubenswrapper[4776]: I1125 11:13:13.297329 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-jh4vh" event={"ID":"93752b39-730c-470f-bbd9-2c9d39abbb03","Type":"ContainerStarted","Data":"bcd45529c80fb8ea89ea0f561d2248c2a779e17efa6e7528105e09d9cf4027aa"} Nov 25 11:13:13 crc kubenswrapper[4776]: I1125 11:13:13.315813 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-d479-account-create-6qg9g" podStartSLOduration=2.315793579 podStartE2EDuration="2.315793579s" podCreationTimestamp="2025-11-25 11:13:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:13:13.30743426 +0000 UTC m=+6538.348493823" watchObservedRunningTime="2025-11-25 11:13:13.315793579 +0000 UTC m=+6538.356853132" Nov 25 11:13:13 crc kubenswrapper[4776]: I1125 11:13:13.325794 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-create-jh4vh" podStartSLOduration=3.32577512 podStartE2EDuration="3.32577512s" podCreationTimestamp="2025-11-25 11:13:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:13:13.322779785 +0000 UTC m=+6538.363839348" watchObservedRunningTime="2025-11-25 11:13:13.32577512 +0000 UTC m=+6538.366834673" Nov 25 11:13:13 crc kubenswrapper[4776]: I1125 11:13:13.679348 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ad93639-113c-4055-8bc2-d4fb8a35d5bb" path="/var/lib/kubelet/pods/5ad93639-113c-4055-8bc2-d4fb8a35d5bb/volumes" Nov 25 11:13:14 crc kubenswrapper[4776]: I1125 11:13:14.308037 4776 generic.go:334] "Generic (PLEG): container finished" podID="c2dfc4fe-0a7f-4670-9075-3712e8e9f775" containerID="d47a1f096eb2f6c24dfa1f5885acd4fa5ce7a5ede02d8f2b3b92c9690e5297e7" exitCode=0 Nov 25 11:13:14 crc kubenswrapper[4776]: I1125 11:13:14.308105 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-d479-account-create-6qg9g" event={"ID":"c2dfc4fe-0a7f-4670-9075-3712e8e9f775","Type":"ContainerDied","Data":"d47a1f096eb2f6c24dfa1f5885acd4fa5ce7a5ede02d8f2b3b92c9690e5297e7"} Nov 25 11:13:15 crc kubenswrapper[4776]: I1125 11:13:15.320104 4776 generic.go:334] "Generic (PLEG): container finished" podID="93752b39-730c-470f-bbd9-2c9d39abbb03" containerID="bcd45529c80fb8ea89ea0f561d2248c2a779e17efa6e7528105e09d9cf4027aa" exitCode=0 Nov 25 11:13:15 crc kubenswrapper[4776]: I1125 11:13:15.320301 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-jh4vh" event={"ID":"93752b39-730c-470f-bbd9-2c9d39abbb03","Type":"ContainerDied","Data":"bcd45529c80fb8ea89ea0f561d2248c2a779e17efa6e7528105e09d9cf4027aa"} Nov 25 11:13:15 crc kubenswrapper[4776]: I1125 11:13:15.330542 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"31358dae-00d6-47d3-b9aa-151088320199","Type":"ContainerStarted","Data":"8644d95ad09bd16fd89963fa6474520bda42e3f625dd489cd0a558b8045c30c2"} Nov 25 11:13:15 crc kubenswrapper[4776]: I1125 11:13:15.744671 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-d479-account-create-6qg9g" Nov 25 11:13:15 crc kubenswrapper[4776]: I1125 11:13:15.872898 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2dfc4fe-0a7f-4670-9075-3712e8e9f775-operator-scripts\") pod \"c2dfc4fe-0a7f-4670-9075-3712e8e9f775\" (UID: \"c2dfc4fe-0a7f-4670-9075-3712e8e9f775\") " Nov 25 11:13:15 crc kubenswrapper[4776]: I1125 11:13:15.873029 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bfqp\" (UniqueName: \"kubernetes.io/projected/c2dfc4fe-0a7f-4670-9075-3712e8e9f775-kube-api-access-4bfqp\") pod \"c2dfc4fe-0a7f-4670-9075-3712e8e9f775\" (UID: \"c2dfc4fe-0a7f-4670-9075-3712e8e9f775\") " Nov 25 11:13:15 crc kubenswrapper[4776]: I1125 11:13:15.874666 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2dfc4fe-0a7f-4670-9075-3712e8e9f775-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c2dfc4fe-0a7f-4670-9075-3712e8e9f775" (UID: "c2dfc4fe-0a7f-4670-9075-3712e8e9f775"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:13:15 crc kubenswrapper[4776]: I1125 11:13:15.880841 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2dfc4fe-0a7f-4670-9075-3712e8e9f775-kube-api-access-4bfqp" (OuterVolumeSpecName: "kube-api-access-4bfqp") pod "c2dfc4fe-0a7f-4670-9075-3712e8e9f775" (UID: "c2dfc4fe-0a7f-4670-9075-3712e8e9f775"). InnerVolumeSpecName "kube-api-access-4bfqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:13:15 crc kubenswrapper[4776]: I1125 11:13:15.976484 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2dfc4fe-0a7f-4670-9075-3712e8e9f775-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:13:15 crc kubenswrapper[4776]: I1125 11:13:15.976522 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bfqp\" (UniqueName: \"kubernetes.io/projected/c2dfc4fe-0a7f-4670-9075-3712e8e9f775-kube-api-access-4bfqp\") on node \"crc\" DevicePath \"\"" Nov 25 11:13:16 crc kubenswrapper[4776]: I1125 11:13:16.341610 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-d479-account-create-6qg9g" Nov 25 11:13:16 crc kubenswrapper[4776]: I1125 11:13:16.341589 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-d479-account-create-6qg9g" event={"ID":"c2dfc4fe-0a7f-4670-9075-3712e8e9f775","Type":"ContainerDied","Data":"8ca4352b56c459d9aa818b03808acd3312b884d82f0e69001bf6cc97dc4362fe"} Nov 25 11:13:16 crc kubenswrapper[4776]: I1125 11:13:16.341754 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ca4352b56c459d9aa818b03808acd3312b884d82f0e69001bf6cc97dc4362fe" Nov 25 11:13:16 crc kubenswrapper[4776]: I1125 11:13:16.344825 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"31358dae-00d6-47d3-b9aa-151088320199","Type":"ContainerStarted","Data":"3fa951ed33c55c3877e8b0ecbf59326006ca520b27ac44cdf4cb24fcdc36e2cf"} Nov 25 11:13:16 crc kubenswrapper[4776]: I1125 11:13:16.409186 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=18.409166923 podStartE2EDuration="18.409166923s" podCreationTimestamp="2025-11-25 11:12:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:13:16.404452854 +0000 UTC m=+6541.445512437" watchObservedRunningTime="2025-11-25 11:13:16.409166923 +0000 UTC m=+6541.450226476" Nov 25 11:13:16 crc kubenswrapper[4776]: I1125 11:13:16.737470 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-jh4vh" Nov 25 11:13:16 crc kubenswrapper[4776]: I1125 11:13:16.897772 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4f5b\" (UniqueName: \"kubernetes.io/projected/93752b39-730c-470f-bbd9-2c9d39abbb03-kube-api-access-q4f5b\") pod \"93752b39-730c-470f-bbd9-2c9d39abbb03\" (UID: \"93752b39-730c-470f-bbd9-2c9d39abbb03\") " Nov 25 11:13:16 crc kubenswrapper[4776]: I1125 11:13:16.897832 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93752b39-730c-470f-bbd9-2c9d39abbb03-operator-scripts\") pod \"93752b39-730c-470f-bbd9-2c9d39abbb03\" (UID: \"93752b39-730c-470f-bbd9-2c9d39abbb03\") " Nov 25 11:13:16 crc kubenswrapper[4776]: I1125 11:13:16.898913 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93752b39-730c-470f-bbd9-2c9d39abbb03-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "93752b39-730c-470f-bbd9-2c9d39abbb03" (UID: "93752b39-730c-470f-bbd9-2c9d39abbb03"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:13:16 crc kubenswrapper[4776]: I1125 11:13:16.912825 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93752b39-730c-470f-bbd9-2c9d39abbb03-kube-api-access-q4f5b" (OuterVolumeSpecName: "kube-api-access-q4f5b") pod "93752b39-730c-470f-bbd9-2c9d39abbb03" (UID: "93752b39-730c-470f-bbd9-2c9d39abbb03"). InnerVolumeSpecName "kube-api-access-q4f5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:13:17 crc kubenswrapper[4776]: I1125 11:13:17.000525 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4f5b\" (UniqueName: \"kubernetes.io/projected/93752b39-730c-470f-bbd9-2c9d39abbb03-kube-api-access-q4f5b\") on node \"crc\" DevicePath \"\"" Nov 25 11:13:17 crc kubenswrapper[4776]: I1125 11:13:17.000566 4776 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/93752b39-730c-470f-bbd9-2c9d39abbb03-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:13:17 crc kubenswrapper[4776]: I1125 11:13:17.356708 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-jh4vh" Nov 25 11:13:17 crc kubenswrapper[4776]: I1125 11:13:17.356701 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-jh4vh" event={"ID":"93752b39-730c-470f-bbd9-2c9d39abbb03","Type":"ContainerDied","Data":"bcdda743dac82f6390b7220ea76b95efe87a71fa90db786e957d5c079f53322f"} Nov 25 11:13:17 crc kubenswrapper[4776]: I1125 11:13:17.356767 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bcdda743dac82f6390b7220ea76b95efe87a71fa90db786e957d5c079f53322f" Nov 25 11:13:18 crc kubenswrapper[4776]: I1125 11:13:18.367191 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Nov 25 11:13:21 crc kubenswrapper[4776]: I1125 11:13:21.472573 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-zqlhx"] Nov 25 11:13:21 crc kubenswrapper[4776]: E1125 11:13:21.473684 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93752b39-730c-470f-bbd9-2c9d39abbb03" containerName="mariadb-database-create" Nov 25 11:13:21 crc kubenswrapper[4776]: I1125 11:13:21.473706 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="93752b39-730c-470f-bbd9-2c9d39abbb03" containerName="mariadb-database-create" Nov 25 11:13:21 crc kubenswrapper[4776]: E1125 11:13:21.473771 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2dfc4fe-0a7f-4670-9075-3712e8e9f775" containerName="mariadb-account-create" Nov 25 11:13:21 crc kubenswrapper[4776]: I1125 11:13:21.473779 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2dfc4fe-0a7f-4670-9075-3712e8e9f775" containerName="mariadb-account-create" Nov 25 11:13:21 crc kubenswrapper[4776]: I1125 11:13:21.474019 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="93752b39-730c-470f-bbd9-2c9d39abbb03" containerName="mariadb-database-create" Nov 25 11:13:21 crc kubenswrapper[4776]: I1125 11:13:21.474054 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2dfc4fe-0a7f-4670-9075-3712e8e9f775" containerName="mariadb-account-create" Nov 25 11:13:21 crc kubenswrapper[4776]: I1125 11:13:21.475110 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-zqlhx" Nov 25 11:13:21 crc kubenswrapper[4776]: I1125 11:13:21.477654 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-k8tkz" Nov 25 11:13:21 crc kubenswrapper[4776]: I1125 11:13:21.480441 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 25 11:13:21 crc kubenswrapper[4776]: I1125 11:13:21.485489 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 25 11:13:21 crc kubenswrapper[4776]: I1125 11:13:21.485501 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 25 11:13:21 crc kubenswrapper[4776]: I1125 11:13:21.490522 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-zqlhx"] Nov 25 11:13:21 crc kubenswrapper[4776]: I1125 11:13:21.604949 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e759b1f-c2d3-4bf7-8e83-c18086f584b4-combined-ca-bundle\") pod \"aodh-db-sync-zqlhx\" (UID: \"3e759b1f-c2d3-4bf7-8e83-c18086f584b4\") " pod="openstack/aodh-db-sync-zqlhx" Nov 25 11:13:21 crc kubenswrapper[4776]: I1125 11:13:21.605129 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e759b1f-c2d3-4bf7-8e83-c18086f584b4-config-data\") pod \"aodh-db-sync-zqlhx\" (UID: \"3e759b1f-c2d3-4bf7-8e83-c18086f584b4\") " pod="openstack/aodh-db-sync-zqlhx" Nov 25 11:13:21 crc kubenswrapper[4776]: I1125 11:13:21.605359 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e759b1f-c2d3-4bf7-8e83-c18086f584b4-scripts\") pod \"aodh-db-sync-zqlhx\" (UID: \"3e759b1f-c2d3-4bf7-8e83-c18086f584b4\") " pod="openstack/aodh-db-sync-zqlhx" Nov 25 11:13:21 crc kubenswrapper[4776]: I1125 11:13:21.605595 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzjms\" (UniqueName: \"kubernetes.io/projected/3e759b1f-c2d3-4bf7-8e83-c18086f584b4-kube-api-access-pzjms\") pod \"aodh-db-sync-zqlhx\" (UID: \"3e759b1f-c2d3-4bf7-8e83-c18086f584b4\") " pod="openstack/aodh-db-sync-zqlhx" Nov 25 11:13:21 crc kubenswrapper[4776]: I1125 11:13:21.707317 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e759b1f-c2d3-4bf7-8e83-c18086f584b4-scripts\") pod \"aodh-db-sync-zqlhx\" (UID: \"3e759b1f-c2d3-4bf7-8e83-c18086f584b4\") " pod="openstack/aodh-db-sync-zqlhx" Nov 25 11:13:21 crc kubenswrapper[4776]: I1125 11:13:21.707657 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzjms\" (UniqueName: \"kubernetes.io/projected/3e759b1f-c2d3-4bf7-8e83-c18086f584b4-kube-api-access-pzjms\") pod \"aodh-db-sync-zqlhx\" (UID: \"3e759b1f-c2d3-4bf7-8e83-c18086f584b4\") " pod="openstack/aodh-db-sync-zqlhx" Nov 25 11:13:21 crc kubenswrapper[4776]: I1125 11:13:21.707788 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e759b1f-c2d3-4bf7-8e83-c18086f584b4-combined-ca-bundle\") pod \"aodh-db-sync-zqlhx\" (UID: \"3e759b1f-c2d3-4bf7-8e83-c18086f584b4\") " pod="openstack/aodh-db-sync-zqlhx" Nov 25 11:13:21 crc kubenswrapper[4776]: I1125 11:13:21.707908 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e759b1f-c2d3-4bf7-8e83-c18086f584b4-config-data\") pod \"aodh-db-sync-zqlhx\" (UID: \"3e759b1f-c2d3-4bf7-8e83-c18086f584b4\") " pod="openstack/aodh-db-sync-zqlhx" Nov 25 11:13:21 crc kubenswrapper[4776]: I1125 11:13:21.713940 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e759b1f-c2d3-4bf7-8e83-c18086f584b4-combined-ca-bundle\") pod \"aodh-db-sync-zqlhx\" (UID: \"3e759b1f-c2d3-4bf7-8e83-c18086f584b4\") " pod="openstack/aodh-db-sync-zqlhx" Nov 25 11:13:21 crc kubenswrapper[4776]: I1125 11:13:21.714296 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e759b1f-c2d3-4bf7-8e83-c18086f584b4-config-data\") pod \"aodh-db-sync-zqlhx\" (UID: \"3e759b1f-c2d3-4bf7-8e83-c18086f584b4\") " pod="openstack/aodh-db-sync-zqlhx" Nov 25 11:13:21 crc kubenswrapper[4776]: I1125 11:13:21.714567 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e759b1f-c2d3-4bf7-8e83-c18086f584b4-scripts\") pod \"aodh-db-sync-zqlhx\" (UID: \"3e759b1f-c2d3-4bf7-8e83-c18086f584b4\") " pod="openstack/aodh-db-sync-zqlhx" Nov 25 11:13:21 crc kubenswrapper[4776]: I1125 11:13:21.726997 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzjms\" (UniqueName: \"kubernetes.io/projected/3e759b1f-c2d3-4bf7-8e83-c18086f584b4-kube-api-access-pzjms\") pod \"aodh-db-sync-zqlhx\" (UID: \"3e759b1f-c2d3-4bf7-8e83-c18086f584b4\") " pod="openstack/aodh-db-sync-zqlhx" Nov 25 11:13:21 crc kubenswrapper[4776]: I1125 11:13:21.798777 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-zqlhx" Nov 25 11:13:22 crc kubenswrapper[4776]: I1125 11:13:22.368805 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-zqlhx"] Nov 25 11:13:22 crc kubenswrapper[4776]: I1125 11:13:22.404542 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-zqlhx" event={"ID":"3e759b1f-c2d3-4bf7-8e83-c18086f584b4","Type":"ContainerStarted","Data":"0d83e123667237e573764cb9e1d89c0596a333aa633b4beb8da24a994b52a70f"} Nov 25 11:13:26 crc kubenswrapper[4776]: I1125 11:13:26.040565 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5trlq"] Nov 25 11:13:26 crc kubenswrapper[4776]: I1125 11:13:26.052341 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-5trlq"] Nov 25 11:13:27 crc kubenswrapper[4776]: I1125 11:13:27.674866 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd2a4d0a-c3e5-47de-86ba-2ff48462f434" path="/var/lib/kubelet/pods/bd2a4d0a-c3e5-47de-86ba-2ff48462f434/volumes" Nov 25 11:13:28 crc kubenswrapper[4776]: I1125 11:13:28.369746 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Nov 25 11:13:28 crc kubenswrapper[4776]: I1125 11:13:28.376883 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Nov 25 11:13:28 crc kubenswrapper[4776]: I1125 11:13:28.488338 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-zqlhx" event={"ID":"3e759b1f-c2d3-4bf7-8e83-c18086f584b4","Type":"ContainerStarted","Data":"55d975b1580e0afc94112b2d087d4f8efa5133c8f8f0c05a104079e8a370abd3"} Nov 25 11:13:28 crc kubenswrapper[4776]: I1125 11:13:28.504061 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Nov 25 11:13:28 crc kubenswrapper[4776]: I1125 11:13:28.513517 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-zqlhx" podStartSLOduration=2.400609245 podStartE2EDuration="7.513499102s" podCreationTimestamp="2025-11-25 11:13:21 +0000 UTC" firstStartedPulling="2025-11-25 11:13:22.374451151 +0000 UTC m=+6547.415510704" lastFinishedPulling="2025-11-25 11:13:27.487341008 +0000 UTC m=+6552.528400561" observedRunningTime="2025-11-25 11:13:28.505248364 +0000 UTC m=+6553.546307917" watchObservedRunningTime="2025-11-25 11:13:28.513499102 +0000 UTC m=+6553.554558655" Nov 25 11:13:28 crc kubenswrapper[4776]: I1125 11:13:28.867507 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 25 11:13:29 crc kubenswrapper[4776]: I1125 11:13:29.029510 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-k465n"] Nov 25 11:13:29 crc kubenswrapper[4776]: I1125 11:13:29.039133 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-k465n"] Nov 25 11:13:29 crc kubenswrapper[4776]: I1125 11:13:29.682990 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c54defc-dcc0-416c-9d88-fbc5944aa0e3" path="/var/lib/kubelet/pods/9c54defc-dcc0-416c-9d88-fbc5944aa0e3/volumes" Nov 25 11:13:33 crc kubenswrapper[4776]: I1125 11:13:33.542758 4776 generic.go:334] "Generic (PLEG): container finished" podID="3e759b1f-c2d3-4bf7-8e83-c18086f584b4" containerID="55d975b1580e0afc94112b2d087d4f8efa5133c8f8f0c05a104079e8a370abd3" exitCode=0 Nov 25 11:13:33 crc kubenswrapper[4776]: I1125 11:13:33.542834 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-zqlhx" event={"ID":"3e759b1f-c2d3-4bf7-8e83-c18086f584b4","Type":"ContainerDied","Data":"55d975b1580e0afc94112b2d087d4f8efa5133c8f8f0c05a104079e8a370abd3"} Nov 25 11:13:33 crc kubenswrapper[4776]: I1125 11:13:33.596606 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 11:13:33 crc kubenswrapper[4776]: I1125 11:13:33.597186 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="1e161da1-d800-4a51-a70a-cfb8974b11b9" containerName="kube-state-metrics" containerID="cri-o://1d25d5d9a90aeb464ebd9a63c3addbcaedd4399ec37f6e2af5d754642931c9bb" gracePeriod=30 Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.167606 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.238937 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fljtn\" (UniqueName: \"kubernetes.io/projected/1e161da1-d800-4a51-a70a-cfb8974b11b9-kube-api-access-fljtn\") pod \"1e161da1-d800-4a51-a70a-cfb8974b11b9\" (UID: \"1e161da1-d800-4a51-a70a-cfb8974b11b9\") " Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.245427 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e161da1-d800-4a51-a70a-cfb8974b11b9-kube-api-access-fljtn" (OuterVolumeSpecName: "kube-api-access-fljtn") pod "1e161da1-d800-4a51-a70a-cfb8974b11b9" (UID: "1e161da1-d800-4a51-a70a-cfb8974b11b9"). InnerVolumeSpecName "kube-api-access-fljtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.342040 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fljtn\" (UniqueName: \"kubernetes.io/projected/1e161da1-d800-4a51-a70a-cfb8974b11b9-kube-api-access-fljtn\") on node \"crc\" DevicePath \"\"" Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.554274 4776 generic.go:334] "Generic (PLEG): container finished" podID="1e161da1-d800-4a51-a70a-cfb8974b11b9" containerID="1d25d5d9a90aeb464ebd9a63c3addbcaedd4399ec37f6e2af5d754642931c9bb" exitCode=2 Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.554353 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.554356 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1e161da1-d800-4a51-a70a-cfb8974b11b9","Type":"ContainerDied","Data":"1d25d5d9a90aeb464ebd9a63c3addbcaedd4399ec37f6e2af5d754642931c9bb"} Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.554487 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1e161da1-d800-4a51-a70a-cfb8974b11b9","Type":"ContainerDied","Data":"5d898c84f191a7760fa1d779e129a922f8f4ef6078ded130e37a2a16c28db459"} Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.554519 4776 scope.go:117] "RemoveContainer" containerID="1d25d5d9a90aeb464ebd9a63c3addbcaedd4399ec37f6e2af5d754642931c9bb" Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.592868 4776 scope.go:117] "RemoveContainer" containerID="1d25d5d9a90aeb464ebd9a63c3addbcaedd4399ec37f6e2af5d754642931c9bb" Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.593447 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 11:13:34 crc kubenswrapper[4776]: E1125 11:13:34.593450 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d25d5d9a90aeb464ebd9a63c3addbcaedd4399ec37f6e2af5d754642931c9bb\": container with ID starting with 1d25d5d9a90aeb464ebd9a63c3addbcaedd4399ec37f6e2af5d754642931c9bb not found: ID does not exist" containerID="1d25d5d9a90aeb464ebd9a63c3addbcaedd4399ec37f6e2af5d754642931c9bb" Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.593514 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d25d5d9a90aeb464ebd9a63c3addbcaedd4399ec37f6e2af5d754642931c9bb"} err="failed to get container status \"1d25d5d9a90aeb464ebd9a63c3addbcaedd4399ec37f6e2af5d754642931c9bb\": rpc error: code = NotFound desc = could not find container \"1d25d5d9a90aeb464ebd9a63c3addbcaedd4399ec37f6e2af5d754642931c9bb\": container with ID starting with 1d25d5d9a90aeb464ebd9a63c3addbcaedd4399ec37f6e2af5d754642931c9bb not found: ID does not exist" Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.606389 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.619062 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 11:13:34 crc kubenswrapper[4776]: E1125 11:13:34.619612 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e161da1-d800-4a51-a70a-cfb8974b11b9" containerName="kube-state-metrics" Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.619636 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e161da1-d800-4a51-a70a-cfb8974b11b9" containerName="kube-state-metrics" Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.619905 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e161da1-d800-4a51-a70a-cfb8974b11b9" containerName="kube-state-metrics" Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.620942 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.624285 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.625515 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.636443 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.749755 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a25bfd2c-28ed-4856-9805-3242c9b7492b-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a25bfd2c-28ed-4856-9805-3242c9b7492b\") " pod="openstack/kube-state-metrics-0" Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.750049 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47b4c\" (UniqueName: \"kubernetes.io/projected/a25bfd2c-28ed-4856-9805-3242c9b7492b-kube-api-access-47b4c\") pod \"kube-state-metrics-0\" (UID: \"a25bfd2c-28ed-4856-9805-3242c9b7492b\") " pod="openstack/kube-state-metrics-0" Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.750102 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a25bfd2c-28ed-4856-9805-3242c9b7492b-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a25bfd2c-28ed-4856-9805-3242c9b7492b\") " pod="openstack/kube-state-metrics-0" Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.750186 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a25bfd2c-28ed-4856-9805-3242c9b7492b-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a25bfd2c-28ed-4856-9805-3242c9b7492b\") " pod="openstack/kube-state-metrics-0" Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.852631 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a25bfd2c-28ed-4856-9805-3242c9b7492b-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a25bfd2c-28ed-4856-9805-3242c9b7492b\") " pod="openstack/kube-state-metrics-0" Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.852690 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47b4c\" (UniqueName: \"kubernetes.io/projected/a25bfd2c-28ed-4856-9805-3242c9b7492b-kube-api-access-47b4c\") pod \"kube-state-metrics-0\" (UID: \"a25bfd2c-28ed-4856-9805-3242c9b7492b\") " pod="openstack/kube-state-metrics-0" Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.852734 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a25bfd2c-28ed-4856-9805-3242c9b7492b-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a25bfd2c-28ed-4856-9805-3242c9b7492b\") " pod="openstack/kube-state-metrics-0" Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.852828 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a25bfd2c-28ed-4856-9805-3242c9b7492b-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a25bfd2c-28ed-4856-9805-3242c9b7492b\") " pod="openstack/kube-state-metrics-0" Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.858414 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a25bfd2c-28ed-4856-9805-3242c9b7492b-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a25bfd2c-28ed-4856-9805-3242c9b7492b\") " pod="openstack/kube-state-metrics-0" Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.858931 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a25bfd2c-28ed-4856-9805-3242c9b7492b-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a25bfd2c-28ed-4856-9805-3242c9b7492b\") " pod="openstack/kube-state-metrics-0" Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.862249 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a25bfd2c-28ed-4856-9805-3242c9b7492b-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a25bfd2c-28ed-4856-9805-3242c9b7492b\") " pod="openstack/kube-state-metrics-0" Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.879509 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47b4c\" (UniqueName: \"kubernetes.io/projected/a25bfd2c-28ed-4856-9805-3242c9b7492b-kube-api-access-47b4c\") pod \"kube-state-metrics-0\" (UID: \"a25bfd2c-28ed-4856-9805-3242c9b7492b\") " pod="openstack/kube-state-metrics-0" Nov 25 11:13:34 crc kubenswrapper[4776]: I1125 11:13:34.946087 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 11:13:35 crc kubenswrapper[4776]: I1125 11:13:35.078031 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-zqlhx" Nov 25 11:13:35 crc kubenswrapper[4776]: I1125 11:13:35.157895 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e759b1f-c2d3-4bf7-8e83-c18086f584b4-combined-ca-bundle\") pod \"3e759b1f-c2d3-4bf7-8e83-c18086f584b4\" (UID: \"3e759b1f-c2d3-4bf7-8e83-c18086f584b4\") " Nov 25 11:13:35 crc kubenswrapper[4776]: I1125 11:13:35.158029 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e759b1f-c2d3-4bf7-8e83-c18086f584b4-config-data\") pod \"3e759b1f-c2d3-4bf7-8e83-c18086f584b4\" (UID: \"3e759b1f-c2d3-4bf7-8e83-c18086f584b4\") " Nov 25 11:13:35 crc kubenswrapper[4776]: I1125 11:13:35.158060 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e759b1f-c2d3-4bf7-8e83-c18086f584b4-scripts\") pod \"3e759b1f-c2d3-4bf7-8e83-c18086f584b4\" (UID: \"3e759b1f-c2d3-4bf7-8e83-c18086f584b4\") " Nov 25 11:13:35 crc kubenswrapper[4776]: I1125 11:13:35.158369 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzjms\" (UniqueName: \"kubernetes.io/projected/3e759b1f-c2d3-4bf7-8e83-c18086f584b4-kube-api-access-pzjms\") pod \"3e759b1f-c2d3-4bf7-8e83-c18086f584b4\" (UID: \"3e759b1f-c2d3-4bf7-8e83-c18086f584b4\") " Nov 25 11:13:35 crc kubenswrapper[4776]: I1125 11:13:35.192539 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e759b1f-c2d3-4bf7-8e83-c18086f584b4-kube-api-access-pzjms" (OuterVolumeSpecName: "kube-api-access-pzjms") pod "3e759b1f-c2d3-4bf7-8e83-c18086f584b4" (UID: "3e759b1f-c2d3-4bf7-8e83-c18086f584b4"). InnerVolumeSpecName "kube-api-access-pzjms". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:13:35 crc kubenswrapper[4776]: I1125 11:13:35.192811 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e759b1f-c2d3-4bf7-8e83-c18086f584b4-scripts" (OuterVolumeSpecName: "scripts") pod "3e759b1f-c2d3-4bf7-8e83-c18086f584b4" (UID: "3e759b1f-c2d3-4bf7-8e83-c18086f584b4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:13:35 crc kubenswrapper[4776]: I1125 11:13:35.220859 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e759b1f-c2d3-4bf7-8e83-c18086f584b4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3e759b1f-c2d3-4bf7-8e83-c18086f584b4" (UID: "3e759b1f-c2d3-4bf7-8e83-c18086f584b4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:13:35 crc kubenswrapper[4776]: I1125 11:13:35.222585 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e759b1f-c2d3-4bf7-8e83-c18086f584b4-config-data" (OuterVolumeSpecName: "config-data") pod "3e759b1f-c2d3-4bf7-8e83-c18086f584b4" (UID: "3e759b1f-c2d3-4bf7-8e83-c18086f584b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:13:35 crc kubenswrapper[4776]: I1125 11:13:35.261337 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzjms\" (UniqueName: \"kubernetes.io/projected/3e759b1f-c2d3-4bf7-8e83-c18086f584b4-kube-api-access-pzjms\") on node \"crc\" DevicePath \"\"" Nov 25 11:13:35 crc kubenswrapper[4776]: I1125 11:13:35.261379 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e759b1f-c2d3-4bf7-8e83-c18086f584b4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:13:35 crc kubenswrapper[4776]: I1125 11:13:35.261394 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e759b1f-c2d3-4bf7-8e83-c18086f584b4-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:13:35 crc kubenswrapper[4776]: I1125 11:13:35.261408 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e759b1f-c2d3-4bf7-8e83-c18086f584b4-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:13:35 crc kubenswrapper[4776]: I1125 11:13:35.266765 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 11:13:35 crc kubenswrapper[4776]: I1125 11:13:35.542643 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 11:13:35 crc kubenswrapper[4776]: I1125 11:13:35.542929 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a9ea0919-b2a4-4c0e-9952-086726eab7b0" containerName="ceilometer-central-agent" containerID="cri-o://846a3c6a75464bfdff065d3d35f3fdc38fae1085bfa6576dc2a277ba9e77dcc8" gracePeriod=30 Nov 25 11:13:35 crc kubenswrapper[4776]: I1125 11:13:35.543001 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a9ea0919-b2a4-4c0e-9952-086726eab7b0" containerName="proxy-httpd" containerID="cri-o://cf92e477f0608de994f1aabfe21a726b96a191b6930d5a2828fd2da621ee1f7b" gracePeriod=30 Nov 25 11:13:35 crc kubenswrapper[4776]: I1125 11:13:35.543038 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a9ea0919-b2a4-4c0e-9952-086726eab7b0" containerName="sg-core" containerID="cri-o://1f37ba68d430d93cb9a009e023cf114dd8ef9b19168405138e43bbb64adb038a" gracePeriod=30 Nov 25 11:13:35 crc kubenswrapper[4776]: I1125 11:13:35.543038 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a9ea0919-b2a4-4c0e-9952-086726eab7b0" containerName="ceilometer-notification-agent" containerID="cri-o://669f88bd2c5c9856836d44a330d923704adde9f7c2b67c1ad502fa363c05a90b" gracePeriod=30 Nov 25 11:13:35 crc kubenswrapper[4776]: I1125 11:13:35.568226 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a25bfd2c-28ed-4856-9805-3242c9b7492b","Type":"ContainerStarted","Data":"c364b3f1167a3f6bdee8e007624f57c7f735b10d6eb34bb28a3e3f36a80d9e65"} Nov 25 11:13:35 crc kubenswrapper[4776]: I1125 11:13:35.570504 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-zqlhx" event={"ID":"3e759b1f-c2d3-4bf7-8e83-c18086f584b4","Type":"ContainerDied","Data":"0d83e123667237e573764cb9e1d89c0596a333aa633b4beb8da24a994b52a70f"} Nov 25 11:13:35 crc kubenswrapper[4776]: I1125 11:13:35.570548 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d83e123667237e573764cb9e1d89c0596a333aa633b4beb8da24a994b52a70f" Nov 25 11:13:35 crc kubenswrapper[4776]: I1125 11:13:35.570560 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-zqlhx" Nov 25 11:13:35 crc kubenswrapper[4776]: I1125 11:13:35.687571 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e161da1-d800-4a51-a70a-cfb8974b11b9" path="/var/lib/kubelet/pods/1e161da1-d800-4a51-a70a-cfb8974b11b9/volumes" Nov 25 11:13:36 crc kubenswrapper[4776]: I1125 11:13:36.586552 4776 generic.go:334] "Generic (PLEG): container finished" podID="a9ea0919-b2a4-4c0e-9952-086726eab7b0" containerID="cf92e477f0608de994f1aabfe21a726b96a191b6930d5a2828fd2da621ee1f7b" exitCode=0 Nov 25 11:13:36 crc kubenswrapper[4776]: I1125 11:13:36.586583 4776 generic.go:334] "Generic (PLEG): container finished" podID="a9ea0919-b2a4-4c0e-9952-086726eab7b0" containerID="1f37ba68d430d93cb9a009e023cf114dd8ef9b19168405138e43bbb64adb038a" exitCode=2 Nov 25 11:13:36 crc kubenswrapper[4776]: I1125 11:13:36.586590 4776 generic.go:334] "Generic (PLEG): container finished" podID="a9ea0919-b2a4-4c0e-9952-086726eab7b0" containerID="846a3c6a75464bfdff065d3d35f3fdc38fae1085bfa6576dc2a277ba9e77dcc8" exitCode=0 Nov 25 11:13:36 crc kubenswrapper[4776]: I1125 11:13:36.586608 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9ea0919-b2a4-4c0e-9952-086726eab7b0","Type":"ContainerDied","Data":"cf92e477f0608de994f1aabfe21a726b96a191b6930d5a2828fd2da621ee1f7b"} Nov 25 11:13:36 crc kubenswrapper[4776]: I1125 11:13:36.586631 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9ea0919-b2a4-4c0e-9952-086726eab7b0","Type":"ContainerDied","Data":"1f37ba68d430d93cb9a009e023cf114dd8ef9b19168405138e43bbb64adb038a"} Nov 25 11:13:36 crc kubenswrapper[4776]: I1125 11:13:36.586640 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9ea0919-b2a4-4c0e-9952-086726eab7b0","Type":"ContainerDied","Data":"846a3c6a75464bfdff065d3d35f3fdc38fae1085bfa6576dc2a277ba9e77dcc8"} Nov 25 11:13:36 crc kubenswrapper[4776]: I1125 11:13:36.738002 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Nov 25 11:13:36 crc kubenswrapper[4776]: E1125 11:13:36.738567 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e759b1f-c2d3-4bf7-8e83-c18086f584b4" containerName="aodh-db-sync" Nov 25 11:13:36 crc kubenswrapper[4776]: I1125 11:13:36.738596 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e759b1f-c2d3-4bf7-8e83-c18086f584b4" containerName="aodh-db-sync" Nov 25 11:13:36 crc kubenswrapper[4776]: I1125 11:13:36.738920 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e759b1f-c2d3-4bf7-8e83-c18086f584b4" containerName="aodh-db-sync" Nov 25 11:13:36 crc kubenswrapper[4776]: I1125 11:13:36.741748 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 25 11:13:36 crc kubenswrapper[4776]: I1125 11:13:36.746827 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-k8tkz" Nov 25 11:13:36 crc kubenswrapper[4776]: I1125 11:13:36.747167 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 25 11:13:36 crc kubenswrapper[4776]: I1125 11:13:36.750735 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 25 11:13:36 crc kubenswrapper[4776]: I1125 11:13:36.763003 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 25 11:13:36 crc kubenswrapper[4776]: I1125 11:13:36.800387 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd69b534-9d4f-4e44-96f7-83f47eeca259-scripts\") pod \"aodh-0\" (UID: \"cd69b534-9d4f-4e44-96f7-83f47eeca259\") " pod="openstack/aodh-0" Nov 25 11:13:36 crc kubenswrapper[4776]: I1125 11:13:36.800596 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd69b534-9d4f-4e44-96f7-83f47eeca259-config-data\") pod \"aodh-0\" (UID: \"cd69b534-9d4f-4e44-96f7-83f47eeca259\") " pod="openstack/aodh-0" Nov 25 11:13:36 crc kubenswrapper[4776]: I1125 11:13:36.800625 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd69b534-9d4f-4e44-96f7-83f47eeca259-combined-ca-bundle\") pod \"aodh-0\" (UID: \"cd69b534-9d4f-4e44-96f7-83f47eeca259\") " pod="openstack/aodh-0" Nov 25 11:13:36 crc kubenswrapper[4776]: I1125 11:13:36.800656 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8886m\" (UniqueName: \"kubernetes.io/projected/cd69b534-9d4f-4e44-96f7-83f47eeca259-kube-api-access-8886m\") pod \"aodh-0\" (UID: \"cd69b534-9d4f-4e44-96f7-83f47eeca259\") " pod="openstack/aodh-0" Nov 25 11:13:36 crc kubenswrapper[4776]: I1125 11:13:36.902319 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd69b534-9d4f-4e44-96f7-83f47eeca259-config-data\") pod \"aodh-0\" (UID: \"cd69b534-9d4f-4e44-96f7-83f47eeca259\") " pod="openstack/aodh-0" Nov 25 11:13:36 crc kubenswrapper[4776]: I1125 11:13:36.902603 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd69b534-9d4f-4e44-96f7-83f47eeca259-combined-ca-bundle\") pod \"aodh-0\" (UID: \"cd69b534-9d4f-4e44-96f7-83f47eeca259\") " pod="openstack/aodh-0" Nov 25 11:13:36 crc kubenswrapper[4776]: I1125 11:13:36.902745 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8886m\" (UniqueName: \"kubernetes.io/projected/cd69b534-9d4f-4e44-96f7-83f47eeca259-kube-api-access-8886m\") pod \"aodh-0\" (UID: \"cd69b534-9d4f-4e44-96f7-83f47eeca259\") " pod="openstack/aodh-0" Nov 25 11:13:36 crc kubenswrapper[4776]: I1125 11:13:36.903304 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd69b534-9d4f-4e44-96f7-83f47eeca259-scripts\") pod \"aodh-0\" (UID: \"cd69b534-9d4f-4e44-96f7-83f47eeca259\") " pod="openstack/aodh-0" Nov 25 11:13:36 crc kubenswrapper[4776]: I1125 11:13:36.907892 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd69b534-9d4f-4e44-96f7-83f47eeca259-scripts\") pod \"aodh-0\" (UID: \"cd69b534-9d4f-4e44-96f7-83f47eeca259\") " pod="openstack/aodh-0" Nov 25 11:13:36 crc kubenswrapper[4776]: I1125 11:13:36.908919 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd69b534-9d4f-4e44-96f7-83f47eeca259-config-data\") pod \"aodh-0\" (UID: \"cd69b534-9d4f-4e44-96f7-83f47eeca259\") " pod="openstack/aodh-0" Nov 25 11:13:36 crc kubenswrapper[4776]: I1125 11:13:36.915237 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd69b534-9d4f-4e44-96f7-83f47eeca259-combined-ca-bundle\") pod \"aodh-0\" (UID: \"cd69b534-9d4f-4e44-96f7-83f47eeca259\") " pod="openstack/aodh-0" Nov 25 11:13:36 crc kubenswrapper[4776]: I1125 11:13:36.921492 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8886m\" (UniqueName: \"kubernetes.io/projected/cd69b534-9d4f-4e44-96f7-83f47eeca259-kube-api-access-8886m\") pod \"aodh-0\" (UID: \"cd69b534-9d4f-4e44-96f7-83f47eeca259\") " pod="openstack/aodh-0" Nov 25 11:13:37 crc kubenswrapper[4776]: I1125 11:13:37.069484 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 25 11:13:38 crc kubenswrapper[4776]: I1125 11:13:38.229741 4776 scope.go:117] "RemoveContainer" containerID="a80cec02848f3a1706589c3f1ffe0c08307a0c0588d9539181831f6e8d7221d2" Nov 25 11:13:38 crc kubenswrapper[4776]: I1125 11:13:38.329248 4776 scope.go:117] "RemoveContainer" containerID="0dba43c5ebc402c61560b9689b802fb034db2eaaee31930a18c9f03c8b125ead" Nov 25 11:13:38 crc kubenswrapper[4776]: I1125 11:13:38.348351 4776 scope.go:117] "RemoveContainer" containerID="0de30baae2bdfa0611870b2de52c24d2c88d042214876580178e3cef4f812d78" Nov 25 11:13:38 crc kubenswrapper[4776]: I1125 11:13:38.414125 4776 scope.go:117] "RemoveContainer" containerID="373b463ab348319edd3b2608fb5610259960bbb7b5cc35a33261c75f8705420f" Nov 25 11:13:38 crc kubenswrapper[4776]: I1125 11:13:38.446061 4776 scope.go:117] "RemoveContainer" containerID="89b549806d7cd66875b6ebdec26f160b49e682a3bab883acb744e930545b06be" Nov 25 11:13:38 crc kubenswrapper[4776]: I1125 11:13:38.478039 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 25 11:13:38 crc kubenswrapper[4776]: I1125 11:13:38.494385 4776 scope.go:117] "RemoveContainer" containerID="9f39dda292ef1dcaae6182d4beeb52ce0f0134af0a782ad5e3a3ee0df193fa6f" Nov 25 11:13:38 crc kubenswrapper[4776]: I1125 11:13:38.524627 4776 scope.go:117] "RemoveContainer" containerID="7496c12a89099167c778dc437fb6b9930c8f9c9172a07ecfbbd8b770ada7995b" Nov 25 11:13:38 crc kubenswrapper[4776]: I1125 11:13:38.549212 4776 scope.go:117] "RemoveContainer" containerID="2c2269cf05113666a597abe1a179abb3caf504221ea666cf04c88dd2225db787" Nov 25 11:13:38 crc kubenswrapper[4776]: I1125 11:13:38.582902 4776 scope.go:117] "RemoveContainer" containerID="cc2712b0b16a7e0249e124ae15cf4c417ac55be025874fb5ac41712df54e1bfe" Nov 25 11:13:38 crc kubenswrapper[4776]: I1125 11:13:38.607722 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"cd69b534-9d4f-4e44-96f7-83f47eeca259","Type":"ContainerStarted","Data":"54a69142581f994d8e21df219a3e8b9d1017fb2bbf07a2be0608ef0c6c8f0766"} Nov 25 11:13:38 crc kubenswrapper[4776]: I1125 11:13:38.629907 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a25bfd2c-28ed-4856-9805-3242c9b7492b","Type":"ContainerStarted","Data":"0020a6d6dfaf40a5d78696b2204c9394089b48fb748610c6fe5cb2fcdaccf7f4"} Nov 25 11:13:38 crc kubenswrapper[4776]: I1125 11:13:38.630004 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 25 11:13:38 crc kubenswrapper[4776]: I1125 11:13:38.651063 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.069869782 podStartE2EDuration="4.65104443s" podCreationTimestamp="2025-11-25 11:13:34 +0000 UTC" firstStartedPulling="2025-11-25 11:13:35.265763901 +0000 UTC m=+6560.306823444" lastFinishedPulling="2025-11-25 11:13:37.846938539 +0000 UTC m=+6562.887998092" observedRunningTime="2025-11-25 11:13:38.642826834 +0000 UTC m=+6563.683886387" watchObservedRunningTime="2025-11-25 11:13:38.65104443 +0000 UTC m=+6563.692103983" Nov 25 11:13:39 crc kubenswrapper[4776]: I1125 11:13:39.648506 4776 generic.go:334] "Generic (PLEG): container finished" podID="a9ea0919-b2a4-4c0e-9952-086726eab7b0" containerID="669f88bd2c5c9856836d44a330d923704adde9f7c2b67c1ad502fa363c05a90b" exitCode=0 Nov 25 11:13:39 crc kubenswrapper[4776]: I1125 11:13:39.648594 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9ea0919-b2a4-4c0e-9952-086726eab7b0","Type":"ContainerDied","Data":"669f88bd2c5c9856836d44a330d923704adde9f7c2b67c1ad502fa363c05a90b"} Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.379014 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.491389 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9ea0919-b2a4-4c0e-9952-086726eab7b0-sg-core-conf-yaml\") pod \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\" (UID: \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\") " Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.491521 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nmbm\" (UniqueName: \"kubernetes.io/projected/a9ea0919-b2a4-4c0e-9952-086726eab7b0-kube-api-access-8nmbm\") pod \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\" (UID: \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\") " Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.491562 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9ea0919-b2a4-4c0e-9952-086726eab7b0-combined-ca-bundle\") pod \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\" (UID: \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\") " Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.491616 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9ea0919-b2a4-4c0e-9952-086726eab7b0-config-data\") pod \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\" (UID: \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\") " Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.491892 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9ea0919-b2a4-4c0e-9952-086726eab7b0-log-httpd\") pod \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\" (UID: \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\") " Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.491995 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9ea0919-b2a4-4c0e-9952-086726eab7b0-run-httpd\") pod \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\" (UID: \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\") " Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.492122 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9ea0919-b2a4-4c0e-9952-086726eab7b0-scripts\") pod \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\" (UID: \"a9ea0919-b2a4-4c0e-9952-086726eab7b0\") " Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.492389 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9ea0919-b2a4-4c0e-9952-086726eab7b0-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a9ea0919-b2a4-4c0e-9952-086726eab7b0" (UID: "a9ea0919-b2a4-4c0e-9952-086726eab7b0"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.492472 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9ea0919-b2a4-4c0e-9952-086726eab7b0-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a9ea0919-b2a4-4c0e-9952-086726eab7b0" (UID: "a9ea0919-b2a4-4c0e-9952-086726eab7b0"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.492897 4776 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9ea0919-b2a4-4c0e-9952-086726eab7b0-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.492920 4776 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9ea0919-b2a4-4c0e-9952-086726eab7b0-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.496423 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9ea0919-b2a4-4c0e-9952-086726eab7b0-kube-api-access-8nmbm" (OuterVolumeSpecName: "kube-api-access-8nmbm") pod "a9ea0919-b2a4-4c0e-9952-086726eab7b0" (UID: "a9ea0919-b2a4-4c0e-9952-086726eab7b0"). InnerVolumeSpecName "kube-api-access-8nmbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.498174 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9ea0919-b2a4-4c0e-9952-086726eab7b0-scripts" (OuterVolumeSpecName: "scripts") pod "a9ea0919-b2a4-4c0e-9952-086726eab7b0" (UID: "a9ea0919-b2a4-4c0e-9952-086726eab7b0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.524636 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9ea0919-b2a4-4c0e-9952-086726eab7b0-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a9ea0919-b2a4-4c0e-9952-086726eab7b0" (UID: "a9ea0919-b2a4-4c0e-9952-086726eab7b0"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.594567 4776 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9ea0919-b2a4-4c0e-9952-086726eab7b0-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.594599 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nmbm\" (UniqueName: \"kubernetes.io/projected/a9ea0919-b2a4-4c0e-9952-086726eab7b0-kube-api-access-8nmbm\") on node \"crc\" DevicePath \"\"" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.594610 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9ea0919-b2a4-4c0e-9952-086726eab7b0-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.612837 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9ea0919-b2a4-4c0e-9952-086726eab7b0-config-data" (OuterVolumeSpecName: "config-data") pod "a9ea0919-b2a4-4c0e-9952-086726eab7b0" (UID: "a9ea0919-b2a4-4c0e-9952-086726eab7b0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.612871 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9ea0919-b2a4-4c0e-9952-086726eab7b0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a9ea0919-b2a4-4c0e-9952-086726eab7b0" (UID: "a9ea0919-b2a4-4c0e-9952-086726eab7b0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.663723 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"cd69b534-9d4f-4e44-96f7-83f47eeca259","Type":"ContainerStarted","Data":"fe3a993b7070572081a256a06201ec6d24e70d192df8371258444640c6b0b9d9"} Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.668531 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9ea0919-b2a4-4c0e-9952-086726eab7b0","Type":"ContainerDied","Data":"2a188e6540a5beb503c4eca5d06ccdea4716fc22721f93bffaa5d43b44472d35"} Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.668594 4776 scope.go:117] "RemoveContainer" containerID="cf92e477f0608de994f1aabfe21a726b96a191b6930d5a2828fd2da621ee1f7b" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.668789 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.697410 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9ea0919-b2a4-4c0e-9952-086726eab7b0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.697435 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9ea0919-b2a4-4c0e-9952-086726eab7b0-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.737338 4776 scope.go:117] "RemoveContainer" containerID="1f37ba68d430d93cb9a009e023cf114dd8ef9b19168405138e43bbb64adb038a" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.743572 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.768724 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.774516 4776 scope.go:117] "RemoveContainer" containerID="669f88bd2c5c9856836d44a330d923704adde9f7c2b67c1ad502fa363c05a90b" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.789607 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 11:13:40 crc kubenswrapper[4776]: E1125 11:13:40.790156 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9ea0919-b2a4-4c0e-9952-086726eab7b0" containerName="proxy-httpd" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.790179 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9ea0919-b2a4-4c0e-9952-086726eab7b0" containerName="proxy-httpd" Nov 25 11:13:40 crc kubenswrapper[4776]: E1125 11:13:40.790222 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9ea0919-b2a4-4c0e-9952-086726eab7b0" containerName="sg-core" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.790233 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9ea0919-b2a4-4c0e-9952-086726eab7b0" containerName="sg-core" Nov 25 11:13:40 crc kubenswrapper[4776]: E1125 11:13:40.790251 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9ea0919-b2a4-4c0e-9952-086726eab7b0" containerName="ceilometer-central-agent" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.790260 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9ea0919-b2a4-4c0e-9952-086726eab7b0" containerName="ceilometer-central-agent" Nov 25 11:13:40 crc kubenswrapper[4776]: E1125 11:13:40.790282 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9ea0919-b2a4-4c0e-9952-086726eab7b0" containerName="ceilometer-notification-agent" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.790290 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9ea0919-b2a4-4c0e-9952-086726eab7b0" containerName="ceilometer-notification-agent" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.790518 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9ea0919-b2a4-4c0e-9952-086726eab7b0" containerName="ceilometer-central-agent" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.790542 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9ea0919-b2a4-4c0e-9952-086726eab7b0" containerName="sg-core" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.790557 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9ea0919-b2a4-4c0e-9952-086726eab7b0" containerName="ceilometer-notification-agent" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.790577 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9ea0919-b2a4-4c0e-9952-086726eab7b0" containerName="proxy-httpd" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.793287 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.805300 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.805355 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.805709 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.807809 4776 scope.go:117] "RemoveContainer" containerID="846a3c6a75464bfdff065d3d35f3fdc38fae1085bfa6576dc2a277ba9e77dcc8" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.865013 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.901352 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-scripts\") pod \"ceilometer-0\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " pod="openstack/ceilometer-0" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.901414 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " pod="openstack/ceilometer-0" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.901482 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-config-data\") pod \"ceilometer-0\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " pod="openstack/ceilometer-0" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.901626 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " pod="openstack/ceilometer-0" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.901737 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4bn2\" (UniqueName: \"kubernetes.io/projected/2536b3d0-342c-4e81-9bdc-8952af88cb7d-kube-api-access-m4bn2\") pod \"ceilometer-0\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " pod="openstack/ceilometer-0" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.902323 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2536b3d0-342c-4e81-9bdc-8952af88cb7d-log-httpd\") pod \"ceilometer-0\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " pod="openstack/ceilometer-0" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.902394 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " pod="openstack/ceilometer-0" Nov 25 11:13:40 crc kubenswrapper[4776]: I1125 11:13:40.902475 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2536b3d0-342c-4e81-9bdc-8952af88cb7d-run-httpd\") pod \"ceilometer-0\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " pod="openstack/ceilometer-0" Nov 25 11:13:40 crc kubenswrapper[4776]: E1125 11:13:40.995808 4776 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9ea0919_b2a4_4c0e_9952_086726eab7b0.slice\": RecentStats: unable to find data in memory cache]" Nov 25 11:13:41 crc kubenswrapper[4776]: I1125 11:13:41.005977 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-scripts\") pod \"ceilometer-0\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " pod="openstack/ceilometer-0" Nov 25 11:13:41 crc kubenswrapper[4776]: I1125 11:13:41.006022 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " pod="openstack/ceilometer-0" Nov 25 11:13:41 crc kubenswrapper[4776]: I1125 11:13:41.006084 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-config-data\") pod \"ceilometer-0\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " pod="openstack/ceilometer-0" Nov 25 11:13:41 crc kubenswrapper[4776]: I1125 11:13:41.006195 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " pod="openstack/ceilometer-0" Nov 25 11:13:41 crc kubenswrapper[4776]: I1125 11:13:41.006240 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4bn2\" (UniqueName: \"kubernetes.io/projected/2536b3d0-342c-4e81-9bdc-8952af88cb7d-kube-api-access-m4bn2\") pod \"ceilometer-0\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " pod="openstack/ceilometer-0" Nov 25 11:13:41 crc kubenswrapper[4776]: I1125 11:13:41.006267 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2536b3d0-342c-4e81-9bdc-8952af88cb7d-log-httpd\") pod \"ceilometer-0\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " pod="openstack/ceilometer-0" Nov 25 11:13:41 crc kubenswrapper[4776]: I1125 11:13:41.006303 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " pod="openstack/ceilometer-0" Nov 25 11:13:41 crc kubenswrapper[4776]: I1125 11:13:41.006338 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2536b3d0-342c-4e81-9bdc-8952af88cb7d-run-httpd\") pod \"ceilometer-0\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " pod="openstack/ceilometer-0" Nov 25 11:13:41 crc kubenswrapper[4776]: I1125 11:13:41.006940 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2536b3d0-342c-4e81-9bdc-8952af88cb7d-run-httpd\") pod \"ceilometer-0\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " pod="openstack/ceilometer-0" Nov 25 11:13:41 crc kubenswrapper[4776]: I1125 11:13:41.012886 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2536b3d0-342c-4e81-9bdc-8952af88cb7d-log-httpd\") pod \"ceilometer-0\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " pod="openstack/ceilometer-0" Nov 25 11:13:41 crc kubenswrapper[4776]: I1125 11:13:41.013207 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-scripts\") pod \"ceilometer-0\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " pod="openstack/ceilometer-0" Nov 25 11:13:41 crc kubenswrapper[4776]: I1125 11:13:41.023355 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " pod="openstack/ceilometer-0" Nov 25 11:13:41 crc kubenswrapper[4776]: I1125 11:13:41.024135 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-config-data\") pod \"ceilometer-0\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " pod="openstack/ceilometer-0" Nov 25 11:13:41 crc kubenswrapper[4776]: I1125 11:13:41.024615 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " pod="openstack/ceilometer-0" Nov 25 11:13:41 crc kubenswrapper[4776]: I1125 11:13:41.026030 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " pod="openstack/ceilometer-0" Nov 25 11:13:41 crc kubenswrapper[4776]: I1125 11:13:41.042267 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4bn2\" (UniqueName: \"kubernetes.io/projected/2536b3d0-342c-4e81-9bdc-8952af88cb7d-kube-api-access-m4bn2\") pod \"ceilometer-0\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " pod="openstack/ceilometer-0" Nov 25 11:13:41 crc kubenswrapper[4776]: I1125 11:13:41.145167 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 25 11:13:41 crc kubenswrapper[4776]: I1125 11:13:41.167465 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 11:13:41 crc kubenswrapper[4776]: W1125 11:13:41.685253 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2536b3d0_342c_4e81_9bdc_8952af88cb7d.slice/crio-18badfa6b571b75d36bf606fd69110d227d64f261b29c0b1281fe83147ef2070 WatchSource:0}: Error finding container 18badfa6b571b75d36bf606fd69110d227d64f261b29c0b1281fe83147ef2070: Status 404 returned error can't find the container with id 18badfa6b571b75d36bf606fd69110d227d64f261b29c0b1281fe83147ef2070 Nov 25 11:13:41 crc kubenswrapper[4776]: I1125 11:13:41.686287 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9ea0919-b2a4-4c0e-9952-086726eab7b0" path="/var/lib/kubelet/pods/a9ea0919-b2a4-4c0e-9952-086726eab7b0/volumes" Nov 25 11:13:41 crc kubenswrapper[4776]: I1125 11:13:41.687380 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 11:13:42 crc kubenswrapper[4776]: I1125 11:13:42.702003 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2536b3d0-342c-4e81-9bdc-8952af88cb7d","Type":"ContainerStarted","Data":"18badfa6b571b75d36bf606fd69110d227d64f261b29c0b1281fe83147ef2070"} Nov 25 11:13:43 crc kubenswrapper[4776]: I1125 11:13:43.532737 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 11:13:43 crc kubenswrapper[4776]: I1125 11:13:43.713851 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"cd69b534-9d4f-4e44-96f7-83f47eeca259","Type":"ContainerStarted","Data":"00d95e51686ebffab6c58c4e9ffdf4c917f7567955ab8d9a11b1fc46a5040455"} Nov 25 11:13:43 crc kubenswrapper[4776]: I1125 11:13:43.716013 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2536b3d0-342c-4e81-9bdc-8952af88cb7d","Type":"ContainerStarted","Data":"3499069dcac56dfa3cb074fc86a9f155c6011eed7a5ab8bae3c1b1362c8741e1"} Nov 25 11:13:44 crc kubenswrapper[4776]: I1125 11:13:44.725277 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2536b3d0-342c-4e81-9bdc-8952af88cb7d","Type":"ContainerStarted","Data":"215dcd66ceb104d650396ece07309e9ad0b9af3cd6e93ed65a1ce5816fc58d6d"} Nov 25 11:13:44 crc kubenswrapper[4776]: I1125 11:13:44.977104 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 25 11:13:45 crc kubenswrapper[4776]: I1125 11:13:45.740597 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"cd69b534-9d4f-4e44-96f7-83f47eeca259","Type":"ContainerStarted","Data":"65f396e6b7d2e6b7884409da8eafde41099a0a8ec17fa3d9e48cf2c3303fde83"} Nov 25 11:13:45 crc kubenswrapper[4776]: I1125 11:13:45.743757 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2536b3d0-342c-4e81-9bdc-8952af88cb7d","Type":"ContainerStarted","Data":"73c621211ccb92190d50b3114da26a6814224e2987bc9832306a6cb49afba575"} Nov 25 11:13:46 crc kubenswrapper[4776]: I1125 11:13:46.047383 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-bz5ht"] Nov 25 11:13:46 crc kubenswrapper[4776]: I1125 11:13:46.060288 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-bz5ht"] Nov 25 11:13:47 crc kubenswrapper[4776]: I1125 11:13:47.702612 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430" path="/var/lib/kubelet/pods/f4d5a9d7-c1d4-42e5-bd10-9eb0a1c3a430/volumes" Nov 25 11:13:50 crc kubenswrapper[4776]: I1125 11:13:50.802355 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2536b3d0-342c-4e81-9bdc-8952af88cb7d","Type":"ContainerStarted","Data":"edc95998a1f2399f996528e577ddf4bafd4502c4c58f48947b4ed62e3854f743"} Nov 25 11:13:50 crc kubenswrapper[4776]: I1125 11:13:50.802831 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 11:13:50 crc kubenswrapper[4776]: I1125 11:13:50.802544 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2536b3d0-342c-4e81-9bdc-8952af88cb7d" containerName="ceilometer-central-agent" containerID="cri-o://3499069dcac56dfa3cb074fc86a9f155c6011eed7a5ab8bae3c1b1362c8741e1" gracePeriod=30 Nov 25 11:13:50 crc kubenswrapper[4776]: I1125 11:13:50.802939 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2536b3d0-342c-4e81-9bdc-8952af88cb7d" containerName="proxy-httpd" containerID="cri-o://edc95998a1f2399f996528e577ddf4bafd4502c4c58f48947b4ed62e3854f743" gracePeriod=30 Nov 25 11:13:50 crc kubenswrapper[4776]: I1125 11:13:50.802996 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2536b3d0-342c-4e81-9bdc-8952af88cb7d" containerName="ceilometer-notification-agent" containerID="cri-o://215dcd66ceb104d650396ece07309e9ad0b9af3cd6e93ed65a1ce5816fc58d6d" gracePeriod=30 Nov 25 11:13:50 crc kubenswrapper[4776]: I1125 11:13:50.803024 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2536b3d0-342c-4e81-9bdc-8952af88cb7d" containerName="sg-core" containerID="cri-o://73c621211ccb92190d50b3114da26a6814224e2987bc9832306a6cb49afba575" gracePeriod=30 Nov 25 11:13:50 crc kubenswrapper[4776]: I1125 11:13:50.838652 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.153814657 podStartE2EDuration="10.838637378s" podCreationTimestamp="2025-11-25 11:13:40 +0000 UTC" firstStartedPulling="2025-11-25 11:13:41.690441829 +0000 UTC m=+6566.731501382" lastFinishedPulling="2025-11-25 11:13:50.37526455 +0000 UTC m=+6575.416324103" observedRunningTime="2025-11-25 11:13:50.835909149 +0000 UTC m=+6575.876968722" watchObservedRunningTime="2025-11-25 11:13:50.838637378 +0000 UTC m=+6575.879696941" Nov 25 11:13:51 crc kubenswrapper[4776]: E1125 11:13:51.290359 4776 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2536b3d0_342c_4e81_9bdc_8952af88cb7d.slice/crio-conmon-215dcd66ceb104d650396ece07309e9ad0b9af3cd6e93ed65a1ce5816fc58d6d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2536b3d0_342c_4e81_9bdc_8952af88cb7d.slice/crio-215dcd66ceb104d650396ece07309e9ad0b9af3cd6e93ed65a1ce5816fc58d6d.scope\": RecentStats: unable to find data in memory cache]" Nov 25 11:13:51 crc kubenswrapper[4776]: I1125 11:13:51.814476 4776 generic.go:334] "Generic (PLEG): container finished" podID="2536b3d0-342c-4e81-9bdc-8952af88cb7d" containerID="73c621211ccb92190d50b3114da26a6814224e2987bc9832306a6cb49afba575" exitCode=2 Nov 25 11:13:51 crc kubenswrapper[4776]: I1125 11:13:51.814511 4776 generic.go:334] "Generic (PLEG): container finished" podID="2536b3d0-342c-4e81-9bdc-8952af88cb7d" containerID="215dcd66ceb104d650396ece07309e9ad0b9af3cd6e93ed65a1ce5816fc58d6d" exitCode=0 Nov 25 11:13:51 crc kubenswrapper[4776]: I1125 11:13:51.814544 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2536b3d0-342c-4e81-9bdc-8952af88cb7d","Type":"ContainerDied","Data":"73c621211ccb92190d50b3114da26a6814224e2987bc9832306a6cb49afba575"} Nov 25 11:13:51 crc kubenswrapper[4776]: I1125 11:13:51.814584 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2536b3d0-342c-4e81-9bdc-8952af88cb7d","Type":"ContainerDied","Data":"215dcd66ceb104d650396ece07309e9ad0b9af3cd6e93ed65a1ce5816fc58d6d"} Nov 25 11:13:51 crc kubenswrapper[4776]: I1125 11:13:51.816261 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"cd69b534-9d4f-4e44-96f7-83f47eeca259","Type":"ContainerStarted","Data":"bdb22909d55c3808601595581af91a9adf094fd07bfe6c30160ac0a116b2cb0a"} Nov 25 11:13:51 crc kubenswrapper[4776]: I1125 11:13:51.816414 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="cd69b534-9d4f-4e44-96f7-83f47eeca259" containerName="aodh-api" containerID="cri-o://fe3a993b7070572081a256a06201ec6d24e70d192df8371258444640c6b0b9d9" gracePeriod=30 Nov 25 11:13:51 crc kubenswrapper[4776]: I1125 11:13:51.816455 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="cd69b534-9d4f-4e44-96f7-83f47eeca259" containerName="aodh-listener" containerID="cri-o://bdb22909d55c3808601595581af91a9adf094fd07bfe6c30160ac0a116b2cb0a" gracePeriod=30 Nov 25 11:13:51 crc kubenswrapper[4776]: I1125 11:13:51.816533 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="cd69b534-9d4f-4e44-96f7-83f47eeca259" containerName="aodh-notifier" containerID="cri-o://65f396e6b7d2e6b7884409da8eafde41099a0a8ec17fa3d9e48cf2c3303fde83" gracePeriod=30 Nov 25 11:13:51 crc kubenswrapper[4776]: I1125 11:13:51.816565 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="cd69b534-9d4f-4e44-96f7-83f47eeca259" containerName="aodh-evaluator" containerID="cri-o://00d95e51686ebffab6c58c4e9ffdf4c917f7567955ab8d9a11b1fc46a5040455" gracePeriod=30 Nov 25 11:13:51 crc kubenswrapper[4776]: I1125 11:13:51.858547 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=3.804716133 podStartE2EDuration="15.858523724s" podCreationTimestamp="2025-11-25 11:13:36 +0000 UTC" firstStartedPulling="2025-11-25 11:13:38.526130575 +0000 UTC m=+6563.567190138" lastFinishedPulling="2025-11-25 11:13:50.579938176 +0000 UTC m=+6575.620997729" observedRunningTime="2025-11-25 11:13:51.853771435 +0000 UTC m=+6576.894830988" watchObservedRunningTime="2025-11-25 11:13:51.858523724 +0000 UTC m=+6576.899583277" Nov 25 11:13:52 crc kubenswrapper[4776]: I1125 11:13:52.833831 4776 generic.go:334] "Generic (PLEG): container finished" podID="2536b3d0-342c-4e81-9bdc-8952af88cb7d" containerID="3499069dcac56dfa3cb074fc86a9f155c6011eed7a5ab8bae3c1b1362c8741e1" exitCode=0 Nov 25 11:13:52 crc kubenswrapper[4776]: I1125 11:13:52.834213 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2536b3d0-342c-4e81-9bdc-8952af88cb7d","Type":"ContainerDied","Data":"3499069dcac56dfa3cb074fc86a9f155c6011eed7a5ab8bae3c1b1362c8741e1"} Nov 25 11:13:52 crc kubenswrapper[4776]: I1125 11:13:52.837933 4776 generic.go:334] "Generic (PLEG): container finished" podID="cd69b534-9d4f-4e44-96f7-83f47eeca259" containerID="00d95e51686ebffab6c58c4e9ffdf4c917f7567955ab8d9a11b1fc46a5040455" exitCode=0 Nov 25 11:13:52 crc kubenswrapper[4776]: I1125 11:13:52.837965 4776 generic.go:334] "Generic (PLEG): container finished" podID="cd69b534-9d4f-4e44-96f7-83f47eeca259" containerID="fe3a993b7070572081a256a06201ec6d24e70d192df8371258444640c6b0b9d9" exitCode=0 Nov 25 11:13:52 crc kubenswrapper[4776]: I1125 11:13:52.837986 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"cd69b534-9d4f-4e44-96f7-83f47eeca259","Type":"ContainerDied","Data":"00d95e51686ebffab6c58c4e9ffdf4c917f7567955ab8d9a11b1fc46a5040455"} Nov 25 11:13:52 crc kubenswrapper[4776]: I1125 11:13:52.838011 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"cd69b534-9d4f-4e44-96f7-83f47eeca259","Type":"ContainerDied","Data":"fe3a993b7070572081a256a06201ec6d24e70d192df8371258444640c6b0b9d9"} Nov 25 11:13:56 crc kubenswrapper[4776]: I1125 11:13:56.884324 4776 generic.go:334] "Generic (PLEG): container finished" podID="cd69b534-9d4f-4e44-96f7-83f47eeca259" containerID="65f396e6b7d2e6b7884409da8eafde41099a0a8ec17fa3d9e48cf2c3303fde83" exitCode=0 Nov 25 11:13:56 crc kubenswrapper[4776]: I1125 11:13:56.884387 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"cd69b534-9d4f-4e44-96f7-83f47eeca259","Type":"ContainerDied","Data":"65f396e6b7d2e6b7884409da8eafde41099a0a8ec17fa3d9e48cf2c3303fde83"} Nov 25 11:14:11 crc kubenswrapper[4776]: I1125 11:14:11.179499 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="2536b3d0-342c-4e81-9bdc-8952af88cb7d" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 25 11:14:21 crc kubenswrapper[4776]: I1125 11:14:21.129755 4776 generic.go:334] "Generic (PLEG): container finished" podID="2536b3d0-342c-4e81-9bdc-8952af88cb7d" containerID="edc95998a1f2399f996528e577ddf4bafd4502c4c58f48947b4ed62e3854f743" exitCode=137 Nov 25 11:14:21 crc kubenswrapper[4776]: I1125 11:14:21.129956 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2536b3d0-342c-4e81-9bdc-8952af88cb7d","Type":"ContainerDied","Data":"edc95998a1f2399f996528e577ddf4bafd4502c4c58f48947b4ed62e3854f743"} Nov 25 11:14:21 crc kubenswrapper[4776]: I1125 11:14:21.732101 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 11:14:21 crc kubenswrapper[4776]: I1125 11:14:21.790382 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-config-data\") pod \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " Nov 25 11:14:21 crc kubenswrapper[4776]: I1125 11:14:21.790432 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-combined-ca-bundle\") pod \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " Nov 25 11:14:21 crc kubenswrapper[4776]: I1125 11:14:21.790478 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-sg-core-conf-yaml\") pod \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " Nov 25 11:14:21 crc kubenswrapper[4776]: I1125 11:14:21.790497 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2536b3d0-342c-4e81-9bdc-8952af88cb7d-log-httpd\") pod \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " Nov 25 11:14:21 crc kubenswrapper[4776]: I1125 11:14:21.790620 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-scripts\") pod \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " Nov 25 11:14:21 crc kubenswrapper[4776]: I1125 11:14:21.790694 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-ceilometer-tls-certs\") pod \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " Nov 25 11:14:21 crc kubenswrapper[4776]: I1125 11:14:21.790726 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4bn2\" (UniqueName: \"kubernetes.io/projected/2536b3d0-342c-4e81-9bdc-8952af88cb7d-kube-api-access-m4bn2\") pod \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " Nov 25 11:14:21 crc kubenswrapper[4776]: I1125 11:14:21.790846 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2536b3d0-342c-4e81-9bdc-8952af88cb7d-run-httpd\") pod \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\" (UID: \"2536b3d0-342c-4e81-9bdc-8952af88cb7d\") " Nov 25 11:14:21 crc kubenswrapper[4776]: I1125 11:14:21.791774 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2536b3d0-342c-4e81-9bdc-8952af88cb7d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2536b3d0-342c-4e81-9bdc-8952af88cb7d" (UID: "2536b3d0-342c-4e81-9bdc-8952af88cb7d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:14:21 crc kubenswrapper[4776]: I1125 11:14:21.791889 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2536b3d0-342c-4e81-9bdc-8952af88cb7d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2536b3d0-342c-4e81-9bdc-8952af88cb7d" (UID: "2536b3d0-342c-4e81-9bdc-8952af88cb7d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:14:21 crc kubenswrapper[4776]: I1125 11:14:21.797293 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-scripts" (OuterVolumeSpecName: "scripts") pod "2536b3d0-342c-4e81-9bdc-8952af88cb7d" (UID: "2536b3d0-342c-4e81-9bdc-8952af88cb7d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:14:21 crc kubenswrapper[4776]: I1125 11:14:21.798280 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2536b3d0-342c-4e81-9bdc-8952af88cb7d-kube-api-access-m4bn2" (OuterVolumeSpecName: "kube-api-access-m4bn2") pod "2536b3d0-342c-4e81-9bdc-8952af88cb7d" (UID: "2536b3d0-342c-4e81-9bdc-8952af88cb7d"). InnerVolumeSpecName "kube-api-access-m4bn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:14:21 crc kubenswrapper[4776]: I1125 11:14:21.818216 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2536b3d0-342c-4e81-9bdc-8952af88cb7d" (UID: "2536b3d0-342c-4e81-9bdc-8952af88cb7d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:14:21 crc kubenswrapper[4776]: I1125 11:14:21.843936 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "2536b3d0-342c-4e81-9bdc-8952af88cb7d" (UID: "2536b3d0-342c-4e81-9bdc-8952af88cb7d"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:14:21 crc kubenswrapper[4776]: I1125 11:14:21.883917 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2536b3d0-342c-4e81-9bdc-8952af88cb7d" (UID: "2536b3d0-342c-4e81-9bdc-8952af88cb7d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:14:21 crc kubenswrapper[4776]: I1125 11:14:21.892860 4776 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2536b3d0-342c-4e81-9bdc-8952af88cb7d-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:21 crc kubenswrapper[4776]: I1125 11:14:21.892896 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:21 crc kubenswrapper[4776]: I1125 11:14:21.892905 4776 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:21 crc kubenswrapper[4776]: I1125 11:14:21.892914 4776 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2536b3d0-342c-4e81-9bdc-8952af88cb7d-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:21 crc kubenswrapper[4776]: I1125 11:14:21.892924 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:21 crc kubenswrapper[4776]: I1125 11:14:21.892933 4776 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:21 crc kubenswrapper[4776]: I1125 11:14:21.892942 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4bn2\" (UniqueName: \"kubernetes.io/projected/2536b3d0-342c-4e81-9bdc-8952af88cb7d-kube-api-access-m4bn2\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:21 crc kubenswrapper[4776]: I1125 11:14:21.939288 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-config-data" (OuterVolumeSpecName: "config-data") pod "2536b3d0-342c-4e81-9bdc-8952af88cb7d" (UID: "2536b3d0-342c-4e81-9bdc-8952af88cb7d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:14:21 crc kubenswrapper[4776]: I1125 11:14:21.994587 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2536b3d0-342c-4e81-9bdc-8952af88cb7d-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.142353 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2536b3d0-342c-4e81-9bdc-8952af88cb7d","Type":"ContainerDied","Data":"18badfa6b571b75d36bf606fd69110d227d64f261b29c0b1281fe83147ef2070"} Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.142393 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.142402 4776 scope.go:117] "RemoveContainer" containerID="edc95998a1f2399f996528e577ddf4bafd4502c4c58f48947b4ed62e3854f743" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.147182 4776 generic.go:334] "Generic (PLEG): container finished" podID="cd69b534-9d4f-4e44-96f7-83f47eeca259" containerID="bdb22909d55c3808601595581af91a9adf094fd07bfe6c30160ac0a116b2cb0a" exitCode=137 Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.147225 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"cd69b534-9d4f-4e44-96f7-83f47eeca259","Type":"ContainerDied","Data":"bdb22909d55c3808601595581af91a9adf094fd07bfe6c30160ac0a116b2cb0a"} Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.147249 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"cd69b534-9d4f-4e44-96f7-83f47eeca259","Type":"ContainerDied","Data":"54a69142581f994d8e21df219a3e8b9d1017fb2bbf07a2be0608ef0c6c8f0766"} Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.147258 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54a69142581f994d8e21df219a3e8b9d1017fb2bbf07a2be0608ef0c6c8f0766" Nov 25 11:14:22 crc kubenswrapper[4776]: E1125 11:14:22.149627 4776 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd69b534_9d4f_4e44_96f7_83f47eeca259.slice/crio-bdb22909d55c3808601595581af91a9adf094fd07bfe6c30160ac0a116b2cb0a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd69b534_9d4f_4e44_96f7_83f47eeca259.slice/crio-conmon-bdb22909d55c3808601595581af91a9adf094fd07bfe6c30160ac0a116b2cb0a.scope\": RecentStats: unable to find data in memory cache]" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.225748 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.227130 4776 scope.go:117] "RemoveContainer" containerID="73c621211ccb92190d50b3114da26a6814224e2987bc9832306a6cb49afba575" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.256163 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.260761 4776 scope.go:117] "RemoveContainer" containerID="215dcd66ceb104d650396ece07309e9ad0b9af3cd6e93ed65a1ce5816fc58d6d" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.288399 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.295886 4776 scope.go:117] "RemoveContainer" containerID="3499069dcac56dfa3cb074fc86a9f155c6011eed7a5ab8bae3c1b1362c8741e1" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.300575 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8886m\" (UniqueName: \"kubernetes.io/projected/cd69b534-9d4f-4e44-96f7-83f47eeca259-kube-api-access-8886m\") pod \"cd69b534-9d4f-4e44-96f7-83f47eeca259\" (UID: \"cd69b534-9d4f-4e44-96f7-83f47eeca259\") " Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.300688 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd69b534-9d4f-4e44-96f7-83f47eeca259-config-data\") pod \"cd69b534-9d4f-4e44-96f7-83f47eeca259\" (UID: \"cd69b534-9d4f-4e44-96f7-83f47eeca259\") " Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.300710 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd69b534-9d4f-4e44-96f7-83f47eeca259-scripts\") pod \"cd69b534-9d4f-4e44-96f7-83f47eeca259\" (UID: \"cd69b534-9d4f-4e44-96f7-83f47eeca259\") " Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.300792 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd69b534-9d4f-4e44-96f7-83f47eeca259-combined-ca-bundle\") pod \"cd69b534-9d4f-4e44-96f7-83f47eeca259\" (UID: \"cd69b534-9d4f-4e44-96f7-83f47eeca259\") " Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.304327 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd69b534-9d4f-4e44-96f7-83f47eeca259-scripts" (OuterVolumeSpecName: "scripts") pod "cd69b534-9d4f-4e44-96f7-83f47eeca259" (UID: "cd69b534-9d4f-4e44-96f7-83f47eeca259"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.305027 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd69b534-9d4f-4e44-96f7-83f47eeca259-kube-api-access-8886m" (OuterVolumeSpecName: "kube-api-access-8886m") pod "cd69b534-9d4f-4e44-96f7-83f47eeca259" (UID: "cd69b534-9d4f-4e44-96f7-83f47eeca259"). InnerVolumeSpecName "kube-api-access-8886m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.314868 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 11:14:22 crc kubenswrapper[4776]: E1125 11:14:22.315259 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd69b534-9d4f-4e44-96f7-83f47eeca259" containerName="aodh-evaluator" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.315276 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd69b534-9d4f-4e44-96f7-83f47eeca259" containerName="aodh-evaluator" Nov 25 11:14:22 crc kubenswrapper[4776]: E1125 11:14:22.315285 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd69b534-9d4f-4e44-96f7-83f47eeca259" containerName="aodh-api" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.315292 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd69b534-9d4f-4e44-96f7-83f47eeca259" containerName="aodh-api" Nov 25 11:14:22 crc kubenswrapper[4776]: E1125 11:14:22.315312 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd69b534-9d4f-4e44-96f7-83f47eeca259" containerName="aodh-notifier" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.315318 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd69b534-9d4f-4e44-96f7-83f47eeca259" containerName="aodh-notifier" Nov 25 11:14:22 crc kubenswrapper[4776]: E1125 11:14:22.315333 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd69b534-9d4f-4e44-96f7-83f47eeca259" containerName="aodh-listener" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.315339 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd69b534-9d4f-4e44-96f7-83f47eeca259" containerName="aodh-listener" Nov 25 11:14:22 crc kubenswrapper[4776]: E1125 11:14:22.315352 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2536b3d0-342c-4e81-9bdc-8952af88cb7d" containerName="ceilometer-notification-agent" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.315358 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2536b3d0-342c-4e81-9bdc-8952af88cb7d" containerName="ceilometer-notification-agent" Nov 25 11:14:22 crc kubenswrapper[4776]: E1125 11:14:22.315365 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2536b3d0-342c-4e81-9bdc-8952af88cb7d" containerName="sg-core" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.315371 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2536b3d0-342c-4e81-9bdc-8952af88cb7d" containerName="sg-core" Nov 25 11:14:22 crc kubenswrapper[4776]: E1125 11:14:22.315383 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2536b3d0-342c-4e81-9bdc-8952af88cb7d" containerName="ceilometer-central-agent" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.315389 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2536b3d0-342c-4e81-9bdc-8952af88cb7d" containerName="ceilometer-central-agent" Nov 25 11:14:22 crc kubenswrapper[4776]: E1125 11:14:22.315406 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2536b3d0-342c-4e81-9bdc-8952af88cb7d" containerName="proxy-httpd" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.315411 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2536b3d0-342c-4e81-9bdc-8952af88cb7d" containerName="proxy-httpd" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.315587 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="2536b3d0-342c-4e81-9bdc-8952af88cb7d" containerName="ceilometer-notification-agent" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.315603 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="2536b3d0-342c-4e81-9bdc-8952af88cb7d" containerName="ceilometer-central-agent" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.315612 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd69b534-9d4f-4e44-96f7-83f47eeca259" containerName="aodh-notifier" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.315622 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd69b534-9d4f-4e44-96f7-83f47eeca259" containerName="aodh-api" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.315632 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="2536b3d0-342c-4e81-9bdc-8952af88cb7d" containerName="sg-core" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.315644 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="2536b3d0-342c-4e81-9bdc-8952af88cb7d" containerName="proxy-httpd" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.315655 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd69b534-9d4f-4e44-96f7-83f47eeca259" containerName="aodh-evaluator" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.315666 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd69b534-9d4f-4e44-96f7-83f47eeca259" containerName="aodh-listener" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.317543 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.320646 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.320847 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.320951 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.329115 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.405534 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b11bc2ae-de8a-45a7-be5a-87a8d81180a2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b11bc2ae-de8a-45a7-be5a-87a8d81180a2\") " pod="openstack/ceilometer-0" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.405613 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b11bc2ae-de8a-45a7-be5a-87a8d81180a2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b11bc2ae-de8a-45a7-be5a-87a8d81180a2\") " pod="openstack/ceilometer-0" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.406063 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfdpt\" (UniqueName: \"kubernetes.io/projected/b11bc2ae-de8a-45a7-be5a-87a8d81180a2-kube-api-access-zfdpt\") pod \"ceilometer-0\" (UID: \"b11bc2ae-de8a-45a7-be5a-87a8d81180a2\") " pod="openstack/ceilometer-0" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.406216 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b11bc2ae-de8a-45a7-be5a-87a8d81180a2-config-data\") pod \"ceilometer-0\" (UID: \"b11bc2ae-de8a-45a7-be5a-87a8d81180a2\") " pod="openstack/ceilometer-0" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.406284 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b11bc2ae-de8a-45a7-be5a-87a8d81180a2-run-httpd\") pod \"ceilometer-0\" (UID: \"b11bc2ae-de8a-45a7-be5a-87a8d81180a2\") " pod="openstack/ceilometer-0" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.406375 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b11bc2ae-de8a-45a7-be5a-87a8d81180a2-log-httpd\") pod \"ceilometer-0\" (UID: \"b11bc2ae-de8a-45a7-be5a-87a8d81180a2\") " pod="openstack/ceilometer-0" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.406434 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b11bc2ae-de8a-45a7-be5a-87a8d81180a2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b11bc2ae-de8a-45a7-be5a-87a8d81180a2\") " pod="openstack/ceilometer-0" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.406524 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b11bc2ae-de8a-45a7-be5a-87a8d81180a2-scripts\") pod \"ceilometer-0\" (UID: \"b11bc2ae-de8a-45a7-be5a-87a8d81180a2\") " pod="openstack/ceilometer-0" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.406669 4776 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd69b534-9d4f-4e44-96f7-83f47eeca259-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.406691 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8886m\" (UniqueName: \"kubernetes.io/projected/cd69b534-9d4f-4e44-96f7-83f47eeca259-kube-api-access-8886m\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.433017 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd69b534-9d4f-4e44-96f7-83f47eeca259-config-data" (OuterVolumeSpecName: "config-data") pod "cd69b534-9d4f-4e44-96f7-83f47eeca259" (UID: "cd69b534-9d4f-4e44-96f7-83f47eeca259"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.456441 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd69b534-9d4f-4e44-96f7-83f47eeca259-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd69b534-9d4f-4e44-96f7-83f47eeca259" (UID: "cd69b534-9d4f-4e44-96f7-83f47eeca259"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.508342 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b11bc2ae-de8a-45a7-be5a-87a8d81180a2-scripts\") pod \"ceilometer-0\" (UID: \"b11bc2ae-de8a-45a7-be5a-87a8d81180a2\") " pod="openstack/ceilometer-0" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.508468 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b11bc2ae-de8a-45a7-be5a-87a8d81180a2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b11bc2ae-de8a-45a7-be5a-87a8d81180a2\") " pod="openstack/ceilometer-0" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.508518 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b11bc2ae-de8a-45a7-be5a-87a8d81180a2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b11bc2ae-de8a-45a7-be5a-87a8d81180a2\") " pod="openstack/ceilometer-0" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.508542 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfdpt\" (UniqueName: \"kubernetes.io/projected/b11bc2ae-de8a-45a7-be5a-87a8d81180a2-kube-api-access-zfdpt\") pod \"ceilometer-0\" (UID: \"b11bc2ae-de8a-45a7-be5a-87a8d81180a2\") " pod="openstack/ceilometer-0" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.508612 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b11bc2ae-de8a-45a7-be5a-87a8d81180a2-config-data\") pod \"ceilometer-0\" (UID: \"b11bc2ae-de8a-45a7-be5a-87a8d81180a2\") " pod="openstack/ceilometer-0" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.508655 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b11bc2ae-de8a-45a7-be5a-87a8d81180a2-run-httpd\") pod \"ceilometer-0\" (UID: \"b11bc2ae-de8a-45a7-be5a-87a8d81180a2\") " pod="openstack/ceilometer-0" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.508721 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b11bc2ae-de8a-45a7-be5a-87a8d81180a2-log-httpd\") pod \"ceilometer-0\" (UID: \"b11bc2ae-de8a-45a7-be5a-87a8d81180a2\") " pod="openstack/ceilometer-0" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.508755 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b11bc2ae-de8a-45a7-be5a-87a8d81180a2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b11bc2ae-de8a-45a7-be5a-87a8d81180a2\") " pod="openstack/ceilometer-0" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.508850 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd69b534-9d4f-4e44-96f7-83f47eeca259-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.509080 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd69b534-9d4f-4e44-96f7-83f47eeca259-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.511902 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b11bc2ae-de8a-45a7-be5a-87a8d81180a2-log-httpd\") pod \"ceilometer-0\" (UID: \"b11bc2ae-de8a-45a7-be5a-87a8d81180a2\") " pod="openstack/ceilometer-0" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.512607 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b11bc2ae-de8a-45a7-be5a-87a8d81180a2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b11bc2ae-de8a-45a7-be5a-87a8d81180a2\") " pod="openstack/ceilometer-0" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.513284 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b11bc2ae-de8a-45a7-be5a-87a8d81180a2-run-httpd\") pod \"ceilometer-0\" (UID: \"b11bc2ae-de8a-45a7-be5a-87a8d81180a2\") " pod="openstack/ceilometer-0" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.514544 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b11bc2ae-de8a-45a7-be5a-87a8d81180a2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b11bc2ae-de8a-45a7-be5a-87a8d81180a2\") " pod="openstack/ceilometer-0" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.522328 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b11bc2ae-de8a-45a7-be5a-87a8d81180a2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b11bc2ae-de8a-45a7-be5a-87a8d81180a2\") " pod="openstack/ceilometer-0" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.522825 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b11bc2ae-de8a-45a7-be5a-87a8d81180a2-config-data\") pod \"ceilometer-0\" (UID: \"b11bc2ae-de8a-45a7-be5a-87a8d81180a2\") " pod="openstack/ceilometer-0" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.523150 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b11bc2ae-de8a-45a7-be5a-87a8d81180a2-scripts\") pod \"ceilometer-0\" (UID: \"b11bc2ae-de8a-45a7-be5a-87a8d81180a2\") " pod="openstack/ceilometer-0" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.527959 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfdpt\" (UniqueName: \"kubernetes.io/projected/b11bc2ae-de8a-45a7-be5a-87a8d81180a2-kube-api-access-zfdpt\") pod \"ceilometer-0\" (UID: \"b11bc2ae-de8a-45a7-be5a-87a8d81180a2\") " pod="openstack/ceilometer-0" Nov 25 11:14:22 crc kubenswrapper[4776]: I1125 11:14:22.638307 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 11:14:23 crc kubenswrapper[4776]: W1125 11:14:23.113857 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb11bc2ae_de8a_45a7_be5a_87a8d81180a2.slice/crio-7433cae2129f5be67ecef4cad13df559a8bd9effd5b0b650ca211ce28245e659 WatchSource:0}: Error finding container 7433cae2129f5be67ecef4cad13df559a8bd9effd5b0b650ca211ce28245e659: Status 404 returned error can't find the container with id 7433cae2129f5be67ecef4cad13df559a8bd9effd5b0b650ca211ce28245e659 Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.114606 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.159006 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.159354 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b11bc2ae-de8a-45a7-be5a-87a8d81180a2","Type":"ContainerStarted","Data":"7433cae2129f5be67ecef4cad13df559a8bd9effd5b0b650ca211ce28245e659"} Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.211093 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.221307 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.237961 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.240568 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.243543 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.244292 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.244517 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.244729 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-k8tkz" Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.244779 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.248216 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.332897 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/81d6f945-72ee-4286-a46b-e6452508c428-internal-tls-certs\") pod \"aodh-0\" (UID: \"81d6f945-72ee-4286-a46b-e6452508c428\") " pod="openstack/aodh-0" Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.333421 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81d6f945-72ee-4286-a46b-e6452508c428-combined-ca-bundle\") pod \"aodh-0\" (UID: \"81d6f945-72ee-4286-a46b-e6452508c428\") " pod="openstack/aodh-0" Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.333576 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81d6f945-72ee-4286-a46b-e6452508c428-scripts\") pod \"aodh-0\" (UID: \"81d6f945-72ee-4286-a46b-e6452508c428\") " pod="openstack/aodh-0" Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.333745 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81d6f945-72ee-4286-a46b-e6452508c428-config-data\") pod \"aodh-0\" (UID: \"81d6f945-72ee-4286-a46b-e6452508c428\") " pod="openstack/aodh-0" Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.334376 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-258m6\" (UniqueName: \"kubernetes.io/projected/81d6f945-72ee-4286-a46b-e6452508c428-kube-api-access-258m6\") pod \"aodh-0\" (UID: \"81d6f945-72ee-4286-a46b-e6452508c428\") " pod="openstack/aodh-0" Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.334860 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/81d6f945-72ee-4286-a46b-e6452508c428-public-tls-certs\") pod \"aodh-0\" (UID: \"81d6f945-72ee-4286-a46b-e6452508c428\") " pod="openstack/aodh-0" Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.437647 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/81d6f945-72ee-4286-a46b-e6452508c428-public-tls-certs\") pod \"aodh-0\" (UID: \"81d6f945-72ee-4286-a46b-e6452508c428\") " pod="openstack/aodh-0" Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.437755 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/81d6f945-72ee-4286-a46b-e6452508c428-internal-tls-certs\") pod \"aodh-0\" (UID: \"81d6f945-72ee-4286-a46b-e6452508c428\") " pod="openstack/aodh-0" Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.437802 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81d6f945-72ee-4286-a46b-e6452508c428-combined-ca-bundle\") pod \"aodh-0\" (UID: \"81d6f945-72ee-4286-a46b-e6452508c428\") " pod="openstack/aodh-0" Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.437835 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81d6f945-72ee-4286-a46b-e6452508c428-scripts\") pod \"aodh-0\" (UID: \"81d6f945-72ee-4286-a46b-e6452508c428\") " pod="openstack/aodh-0" Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.437860 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81d6f945-72ee-4286-a46b-e6452508c428-config-data\") pod \"aodh-0\" (UID: \"81d6f945-72ee-4286-a46b-e6452508c428\") " pod="openstack/aodh-0" Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.437988 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-258m6\" (UniqueName: \"kubernetes.io/projected/81d6f945-72ee-4286-a46b-e6452508c428-kube-api-access-258m6\") pod \"aodh-0\" (UID: \"81d6f945-72ee-4286-a46b-e6452508c428\") " pod="openstack/aodh-0" Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.443516 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81d6f945-72ee-4286-a46b-e6452508c428-combined-ca-bundle\") pod \"aodh-0\" (UID: \"81d6f945-72ee-4286-a46b-e6452508c428\") " pod="openstack/aodh-0" Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.443651 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/81d6f945-72ee-4286-a46b-e6452508c428-public-tls-certs\") pod \"aodh-0\" (UID: \"81d6f945-72ee-4286-a46b-e6452508c428\") " pod="openstack/aodh-0" Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.443979 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/81d6f945-72ee-4286-a46b-e6452508c428-internal-tls-certs\") pod \"aodh-0\" (UID: \"81d6f945-72ee-4286-a46b-e6452508c428\") " pod="openstack/aodh-0" Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.446924 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/81d6f945-72ee-4286-a46b-e6452508c428-scripts\") pod \"aodh-0\" (UID: \"81d6f945-72ee-4286-a46b-e6452508c428\") " pod="openstack/aodh-0" Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.453797 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81d6f945-72ee-4286-a46b-e6452508c428-config-data\") pod \"aodh-0\" (UID: \"81d6f945-72ee-4286-a46b-e6452508c428\") " pod="openstack/aodh-0" Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.459818 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-258m6\" (UniqueName: \"kubernetes.io/projected/81d6f945-72ee-4286-a46b-e6452508c428-kube-api-access-258m6\") pod \"aodh-0\" (UID: \"81d6f945-72ee-4286-a46b-e6452508c428\") " pod="openstack/aodh-0" Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.561481 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.696547 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2536b3d0-342c-4e81-9bdc-8952af88cb7d" path="/var/lib/kubelet/pods/2536b3d0-342c-4e81-9bdc-8952af88cb7d/volumes" Nov 25 11:14:23 crc kubenswrapper[4776]: I1125 11:14:23.697862 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd69b534-9d4f-4e44-96f7-83f47eeca259" path="/var/lib/kubelet/pods/cd69b534-9d4f-4e44-96f7-83f47eeca259/volumes" Nov 25 11:14:24 crc kubenswrapper[4776]: I1125 11:14:24.110476 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Nov 25 11:14:24 crc kubenswrapper[4776]: I1125 11:14:24.173493 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b11bc2ae-de8a-45a7-be5a-87a8d81180a2","Type":"ContainerStarted","Data":"32d88bb70e4b363cfab5bc7abebaa554577aa8258f532bb357ae5ce8ff4590b2"} Nov 25 11:14:24 crc kubenswrapper[4776]: I1125 11:14:24.174389 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"81d6f945-72ee-4286-a46b-e6452508c428","Type":"ContainerStarted","Data":"508a326f1c37419d88948d268facc5ba98b2add1c3c4713461f75944258e9b38"} Nov 25 11:14:25 crc kubenswrapper[4776]: I1125 11:14:25.193912 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b11bc2ae-de8a-45a7-be5a-87a8d81180a2","Type":"ContainerStarted","Data":"34882a509c1c508cbd176c020ade005e33365517d26cc11cb364da55a88a802a"} Nov 25 11:14:25 crc kubenswrapper[4776]: I1125 11:14:25.198460 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"81d6f945-72ee-4286-a46b-e6452508c428","Type":"ContainerStarted","Data":"8c06fec87513db8de9f0e9559c721b07b295359a3c94d1be209eb85b5ab3ca86"} Nov 25 11:14:26 crc kubenswrapper[4776]: I1125 11:14:26.209358 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b11bc2ae-de8a-45a7-be5a-87a8d81180a2","Type":"ContainerStarted","Data":"1e95f878da9919863006198a16bb684b1d1ef7464f49efe34bb6f70bfd722e7f"} Nov 25 11:14:26 crc kubenswrapper[4776]: I1125 11:14:26.214787 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"81d6f945-72ee-4286-a46b-e6452508c428","Type":"ContainerStarted","Data":"108bd0a43b88debdac3dc86ad93913d2a62765d93fdd2c6f8ca9e43ad9fee81f"} Nov 25 11:14:27 crc kubenswrapper[4776]: I1125 11:14:27.230679 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"81d6f945-72ee-4286-a46b-e6452508c428","Type":"ContainerStarted","Data":"498f4c57ebd7c7270783a2c6d6abab5b9eb193c40deae40a23ae426a48557fc0"} Nov 25 11:14:27 crc kubenswrapper[4776]: I1125 11:14:27.231917 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"81d6f945-72ee-4286-a46b-e6452508c428","Type":"ContainerStarted","Data":"94867eb8515e6233ae9bd23a81e10b483968300c821ac78a5af01ef50abcc60a"} Nov 25 11:14:27 crc kubenswrapper[4776]: I1125 11:14:27.233731 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b11bc2ae-de8a-45a7-be5a-87a8d81180a2","Type":"ContainerStarted","Data":"239e6fd7532c36d2e0988288eb27046ba64a569ed7f19b26d9697dfa56de2b81"} Nov 25 11:14:27 crc kubenswrapper[4776]: I1125 11:14:27.234058 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 11:14:27 crc kubenswrapper[4776]: I1125 11:14:27.273996 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=1.681995454 podStartE2EDuration="4.273960224s" podCreationTimestamp="2025-11-25 11:14:23 +0000 UTC" firstStartedPulling="2025-11-25 11:14:24.114393139 +0000 UTC m=+6609.155452692" lastFinishedPulling="2025-11-25 11:14:26.706357909 +0000 UTC m=+6611.747417462" observedRunningTime="2025-11-25 11:14:27.252419444 +0000 UTC m=+6612.293479007" watchObservedRunningTime="2025-11-25 11:14:27.273960224 +0000 UTC m=+6612.315019767" Nov 25 11:14:27 crc kubenswrapper[4776]: I1125 11:14:27.284569 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.731583182 podStartE2EDuration="5.28454286s" podCreationTimestamp="2025-11-25 11:14:22 +0000 UTC" firstStartedPulling="2025-11-25 11:14:23.118238799 +0000 UTC m=+6608.159298352" lastFinishedPulling="2025-11-25 11:14:26.671198477 +0000 UTC m=+6611.712258030" observedRunningTime="2025-11-25 11:14:27.283005211 +0000 UTC m=+6612.324064764" watchObservedRunningTime="2025-11-25 11:14:27.28454286 +0000 UTC m=+6612.325602413" Nov 25 11:14:32 crc kubenswrapper[4776]: I1125 11:14:32.184591 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86cd554965-lzn6h"] Nov 25 11:14:32 crc kubenswrapper[4776]: I1125 11:14:32.187606 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86cd554965-lzn6h" Nov 25 11:14:32 crc kubenswrapper[4776]: I1125 11:14:32.190623 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1" Nov 25 11:14:32 crc kubenswrapper[4776]: I1125 11:14:32.217059 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86cd554965-lzn6h"] Nov 25 11:14:32 crc kubenswrapper[4776]: I1125 11:14:32.336267 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-ovsdbserver-sb\") pod \"dnsmasq-dns-86cd554965-lzn6h\" (UID: \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\") " pod="openstack/dnsmasq-dns-86cd554965-lzn6h" Nov 25 11:14:32 crc kubenswrapper[4776]: I1125 11:14:32.336319 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-openstack-cell1\") pod \"dnsmasq-dns-86cd554965-lzn6h\" (UID: \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\") " pod="openstack/dnsmasq-dns-86cd554965-lzn6h" Nov 25 11:14:32 crc kubenswrapper[4776]: I1125 11:14:32.336348 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lbfh\" (UniqueName: \"kubernetes.io/projected/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-kube-api-access-9lbfh\") pod \"dnsmasq-dns-86cd554965-lzn6h\" (UID: \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\") " pod="openstack/dnsmasq-dns-86cd554965-lzn6h" Nov 25 11:14:32 crc kubenswrapper[4776]: I1125 11:14:32.336534 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-ovsdbserver-nb\") pod \"dnsmasq-dns-86cd554965-lzn6h\" (UID: \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\") " pod="openstack/dnsmasq-dns-86cd554965-lzn6h" Nov 25 11:14:32 crc kubenswrapper[4776]: I1125 11:14:32.336714 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-dns-svc\") pod \"dnsmasq-dns-86cd554965-lzn6h\" (UID: \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\") " pod="openstack/dnsmasq-dns-86cd554965-lzn6h" Nov 25 11:14:32 crc kubenswrapper[4776]: I1125 11:14:32.336767 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-config\") pod \"dnsmasq-dns-86cd554965-lzn6h\" (UID: \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\") " pod="openstack/dnsmasq-dns-86cd554965-lzn6h" Nov 25 11:14:32 crc kubenswrapper[4776]: I1125 11:14:32.438670 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-ovsdbserver-sb\") pod \"dnsmasq-dns-86cd554965-lzn6h\" (UID: \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\") " pod="openstack/dnsmasq-dns-86cd554965-lzn6h" Nov 25 11:14:32 crc kubenswrapper[4776]: I1125 11:14:32.438731 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-openstack-cell1\") pod \"dnsmasq-dns-86cd554965-lzn6h\" (UID: \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\") " pod="openstack/dnsmasq-dns-86cd554965-lzn6h" Nov 25 11:14:32 crc kubenswrapper[4776]: I1125 11:14:32.438761 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lbfh\" (UniqueName: \"kubernetes.io/projected/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-kube-api-access-9lbfh\") pod \"dnsmasq-dns-86cd554965-lzn6h\" (UID: \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\") " pod="openstack/dnsmasq-dns-86cd554965-lzn6h" Nov 25 11:14:32 crc kubenswrapper[4776]: I1125 11:14:32.438947 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-ovsdbserver-nb\") pod \"dnsmasq-dns-86cd554965-lzn6h\" (UID: \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\") " pod="openstack/dnsmasq-dns-86cd554965-lzn6h" Nov 25 11:14:32 crc kubenswrapper[4776]: I1125 11:14:32.439086 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-dns-svc\") pod \"dnsmasq-dns-86cd554965-lzn6h\" (UID: \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\") " pod="openstack/dnsmasq-dns-86cd554965-lzn6h" Nov 25 11:14:32 crc kubenswrapper[4776]: I1125 11:14:32.439121 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-config\") pod \"dnsmasq-dns-86cd554965-lzn6h\" (UID: \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\") " pod="openstack/dnsmasq-dns-86cd554965-lzn6h" Nov 25 11:14:32 crc kubenswrapper[4776]: I1125 11:14:32.439768 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-openstack-cell1\") pod \"dnsmasq-dns-86cd554965-lzn6h\" (UID: \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\") " pod="openstack/dnsmasq-dns-86cd554965-lzn6h" Nov 25 11:14:32 crc kubenswrapper[4776]: I1125 11:14:32.439805 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-ovsdbserver-sb\") pod \"dnsmasq-dns-86cd554965-lzn6h\" (UID: \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\") " pod="openstack/dnsmasq-dns-86cd554965-lzn6h" Nov 25 11:14:32 crc kubenswrapper[4776]: I1125 11:14:32.440023 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-dns-svc\") pod \"dnsmasq-dns-86cd554965-lzn6h\" (UID: \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\") " pod="openstack/dnsmasq-dns-86cd554965-lzn6h" Nov 25 11:14:32 crc kubenswrapper[4776]: I1125 11:14:32.440669 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-ovsdbserver-nb\") pod \"dnsmasq-dns-86cd554965-lzn6h\" (UID: \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\") " pod="openstack/dnsmasq-dns-86cd554965-lzn6h" Nov 25 11:14:32 crc kubenswrapper[4776]: I1125 11:14:32.441170 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-config\") pod \"dnsmasq-dns-86cd554965-lzn6h\" (UID: \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\") " pod="openstack/dnsmasq-dns-86cd554965-lzn6h" Nov 25 11:14:32 crc kubenswrapper[4776]: I1125 11:14:32.457387 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lbfh\" (UniqueName: \"kubernetes.io/projected/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-kube-api-access-9lbfh\") pod \"dnsmasq-dns-86cd554965-lzn6h\" (UID: \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\") " pod="openstack/dnsmasq-dns-86cd554965-lzn6h" Nov 25 11:14:32 crc kubenswrapper[4776]: I1125 11:14:32.528603 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86cd554965-lzn6h" Nov 25 11:14:33 crc kubenswrapper[4776]: I1125 11:14:32.998919 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86cd554965-lzn6h"] Nov 25 11:14:33 crc kubenswrapper[4776]: I1125 11:14:33.302892 4776 generic.go:334] "Generic (PLEG): container finished" podID="54a80827-fae6-48e8-8ba0-cfd09a9c3cc4" containerID="7dc31647bb9fa1cb8d229efa7f6dd6722b4fb10bbf99791150f68a5c0e8adebe" exitCode=0 Nov 25 11:14:33 crc kubenswrapper[4776]: I1125 11:14:33.302984 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86cd554965-lzn6h" event={"ID":"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4","Type":"ContainerDied","Data":"7dc31647bb9fa1cb8d229efa7f6dd6722b4fb10bbf99791150f68a5c0e8adebe"} Nov 25 11:14:33 crc kubenswrapper[4776]: I1125 11:14:33.306360 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86cd554965-lzn6h" event={"ID":"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4","Type":"ContainerStarted","Data":"f752937c8fe681029d76416b084fcfc5b6ac14f7f216d6dccb735389d701dafe"} Nov 25 11:14:34 crc kubenswrapper[4776]: I1125 11:14:34.317858 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86cd554965-lzn6h" event={"ID":"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4","Type":"ContainerStarted","Data":"478a7dd64198c32cbe605e3bdbc14bcc9c6977d86fa2766db9f5b0811c5e67aa"} Nov 25 11:14:34 crc kubenswrapper[4776]: I1125 11:14:34.318230 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86cd554965-lzn6h" Nov 25 11:14:34 crc kubenswrapper[4776]: I1125 11:14:34.341835 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86cd554965-lzn6h" podStartSLOduration=2.341817354 podStartE2EDuration="2.341817354s" podCreationTimestamp="2025-11-25 11:14:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:14:34.333019344 +0000 UTC m=+6619.374078907" watchObservedRunningTime="2025-11-25 11:14:34.341817354 +0000 UTC m=+6619.382876907" Nov 25 11:14:39 crc kubenswrapper[4776]: I1125 11:14:39.089040 4776 scope.go:117] "RemoveContainer" containerID="c8decff072d39c84cc2226421867809629b8654d98030ccb6a1bbc5bbc277e53" Nov 25 11:14:42 crc kubenswrapper[4776]: I1125 11:14:42.530236 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86cd554965-lzn6h" Nov 25 11:14:42 crc kubenswrapper[4776]: I1125 11:14:42.590667 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86dbcbdfc9-dl692"] Nov 25 11:14:42 crc kubenswrapper[4776]: I1125 11:14:42.590985 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86dbcbdfc9-dl692" podUID="aeccbc20-2ea2-4507-924d-f81aab9d947a" containerName="dnsmasq-dns" containerID="cri-o://08b2752014bddb8246b5b237fdd56860d953bfc26af95426502f062d9c002b79" gracePeriod=10 Nov 25 11:14:42 crc kubenswrapper[4776]: I1125 11:14:42.766897 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-568cc6cf7c-6mkrs"] Nov 25 11:14:42 crc kubenswrapper[4776]: I1125 11:14:42.774531 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568cc6cf7c-6mkrs" Nov 25 11:14:42 crc kubenswrapper[4776]: I1125 11:14:42.794052 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-568cc6cf7c-6mkrs"] Nov 25 11:14:42 crc kubenswrapper[4776]: I1125 11:14:42.901344 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82c0a51a-57e4-4ab3-98b6-e2052e449bf0-ovsdbserver-nb\") pod \"dnsmasq-dns-568cc6cf7c-6mkrs\" (UID: \"82c0a51a-57e4-4ab3-98b6-e2052e449bf0\") " pod="openstack/dnsmasq-dns-568cc6cf7c-6mkrs" Nov 25 11:14:42 crc kubenswrapper[4776]: I1125 11:14:42.901799 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/82c0a51a-57e4-4ab3-98b6-e2052e449bf0-openstack-cell1\") pod \"dnsmasq-dns-568cc6cf7c-6mkrs\" (UID: \"82c0a51a-57e4-4ab3-98b6-e2052e449bf0\") " pod="openstack/dnsmasq-dns-568cc6cf7c-6mkrs" Nov 25 11:14:42 crc kubenswrapper[4776]: I1125 11:14:42.901833 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82c0a51a-57e4-4ab3-98b6-e2052e449bf0-dns-svc\") pod \"dnsmasq-dns-568cc6cf7c-6mkrs\" (UID: \"82c0a51a-57e4-4ab3-98b6-e2052e449bf0\") " pod="openstack/dnsmasq-dns-568cc6cf7c-6mkrs" Nov 25 11:14:42 crc kubenswrapper[4776]: I1125 11:14:42.902308 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82c0a51a-57e4-4ab3-98b6-e2052e449bf0-config\") pod \"dnsmasq-dns-568cc6cf7c-6mkrs\" (UID: \"82c0a51a-57e4-4ab3-98b6-e2052e449bf0\") " pod="openstack/dnsmasq-dns-568cc6cf7c-6mkrs" Nov 25 11:14:42 crc kubenswrapper[4776]: I1125 11:14:42.902386 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82c0a51a-57e4-4ab3-98b6-e2052e449bf0-ovsdbserver-sb\") pod \"dnsmasq-dns-568cc6cf7c-6mkrs\" (UID: \"82c0a51a-57e4-4ab3-98b6-e2052e449bf0\") " pod="openstack/dnsmasq-dns-568cc6cf7c-6mkrs" Nov 25 11:14:42 crc kubenswrapper[4776]: I1125 11:14:42.902449 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8v4s6\" (UniqueName: \"kubernetes.io/projected/82c0a51a-57e4-4ab3-98b6-e2052e449bf0-kube-api-access-8v4s6\") pod \"dnsmasq-dns-568cc6cf7c-6mkrs\" (UID: \"82c0a51a-57e4-4ab3-98b6-e2052e449bf0\") " pod="openstack/dnsmasq-dns-568cc6cf7c-6mkrs" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.005233 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82c0a51a-57e4-4ab3-98b6-e2052e449bf0-config\") pod \"dnsmasq-dns-568cc6cf7c-6mkrs\" (UID: \"82c0a51a-57e4-4ab3-98b6-e2052e449bf0\") " pod="openstack/dnsmasq-dns-568cc6cf7c-6mkrs" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.005279 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82c0a51a-57e4-4ab3-98b6-e2052e449bf0-ovsdbserver-sb\") pod \"dnsmasq-dns-568cc6cf7c-6mkrs\" (UID: \"82c0a51a-57e4-4ab3-98b6-e2052e449bf0\") " pod="openstack/dnsmasq-dns-568cc6cf7c-6mkrs" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.005301 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8v4s6\" (UniqueName: \"kubernetes.io/projected/82c0a51a-57e4-4ab3-98b6-e2052e449bf0-kube-api-access-8v4s6\") pod \"dnsmasq-dns-568cc6cf7c-6mkrs\" (UID: \"82c0a51a-57e4-4ab3-98b6-e2052e449bf0\") " pod="openstack/dnsmasq-dns-568cc6cf7c-6mkrs" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.005320 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82c0a51a-57e4-4ab3-98b6-e2052e449bf0-ovsdbserver-nb\") pod \"dnsmasq-dns-568cc6cf7c-6mkrs\" (UID: \"82c0a51a-57e4-4ab3-98b6-e2052e449bf0\") " pod="openstack/dnsmasq-dns-568cc6cf7c-6mkrs" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.005736 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/82c0a51a-57e4-4ab3-98b6-e2052e449bf0-openstack-cell1\") pod \"dnsmasq-dns-568cc6cf7c-6mkrs\" (UID: \"82c0a51a-57e4-4ab3-98b6-e2052e449bf0\") " pod="openstack/dnsmasq-dns-568cc6cf7c-6mkrs" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.005838 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82c0a51a-57e4-4ab3-98b6-e2052e449bf0-dns-svc\") pod \"dnsmasq-dns-568cc6cf7c-6mkrs\" (UID: \"82c0a51a-57e4-4ab3-98b6-e2052e449bf0\") " pod="openstack/dnsmasq-dns-568cc6cf7c-6mkrs" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.006595 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/82c0a51a-57e4-4ab3-98b6-e2052e449bf0-ovsdbserver-sb\") pod \"dnsmasq-dns-568cc6cf7c-6mkrs\" (UID: \"82c0a51a-57e4-4ab3-98b6-e2052e449bf0\") " pod="openstack/dnsmasq-dns-568cc6cf7c-6mkrs" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.006660 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/82c0a51a-57e4-4ab3-98b6-e2052e449bf0-ovsdbserver-nb\") pod \"dnsmasq-dns-568cc6cf7c-6mkrs\" (UID: \"82c0a51a-57e4-4ab3-98b6-e2052e449bf0\") " pod="openstack/dnsmasq-dns-568cc6cf7c-6mkrs" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.006707 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82c0a51a-57e4-4ab3-98b6-e2052e449bf0-config\") pod \"dnsmasq-dns-568cc6cf7c-6mkrs\" (UID: \"82c0a51a-57e4-4ab3-98b6-e2052e449bf0\") " pod="openstack/dnsmasq-dns-568cc6cf7c-6mkrs" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.006879 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/82c0a51a-57e4-4ab3-98b6-e2052e449bf0-openstack-cell1\") pod \"dnsmasq-dns-568cc6cf7c-6mkrs\" (UID: \"82c0a51a-57e4-4ab3-98b6-e2052e449bf0\") " pod="openstack/dnsmasq-dns-568cc6cf7c-6mkrs" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.007242 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/82c0a51a-57e4-4ab3-98b6-e2052e449bf0-dns-svc\") pod \"dnsmasq-dns-568cc6cf7c-6mkrs\" (UID: \"82c0a51a-57e4-4ab3-98b6-e2052e449bf0\") " pod="openstack/dnsmasq-dns-568cc6cf7c-6mkrs" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.031228 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8v4s6\" (UniqueName: \"kubernetes.io/projected/82c0a51a-57e4-4ab3-98b6-e2052e449bf0-kube-api-access-8v4s6\") pod \"dnsmasq-dns-568cc6cf7c-6mkrs\" (UID: \"82c0a51a-57e4-4ab3-98b6-e2052e449bf0\") " pod="openstack/dnsmasq-dns-568cc6cf7c-6mkrs" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.117477 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568cc6cf7c-6mkrs" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.262232 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86dbcbdfc9-dl692" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.413833 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29x8k\" (UniqueName: \"kubernetes.io/projected/aeccbc20-2ea2-4507-924d-f81aab9d947a-kube-api-access-29x8k\") pod \"aeccbc20-2ea2-4507-924d-f81aab9d947a\" (UID: \"aeccbc20-2ea2-4507-924d-f81aab9d947a\") " Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.413914 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aeccbc20-2ea2-4507-924d-f81aab9d947a-dns-svc\") pod \"aeccbc20-2ea2-4507-924d-f81aab9d947a\" (UID: \"aeccbc20-2ea2-4507-924d-f81aab9d947a\") " Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.414095 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aeccbc20-2ea2-4507-924d-f81aab9d947a-config\") pod \"aeccbc20-2ea2-4507-924d-f81aab9d947a\" (UID: \"aeccbc20-2ea2-4507-924d-f81aab9d947a\") " Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.414369 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aeccbc20-2ea2-4507-924d-f81aab9d947a-ovsdbserver-sb\") pod \"aeccbc20-2ea2-4507-924d-f81aab9d947a\" (UID: \"aeccbc20-2ea2-4507-924d-f81aab9d947a\") " Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.414446 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aeccbc20-2ea2-4507-924d-f81aab9d947a-ovsdbserver-nb\") pod \"aeccbc20-2ea2-4507-924d-f81aab9d947a\" (UID: \"aeccbc20-2ea2-4507-924d-f81aab9d947a\") " Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.419428 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aeccbc20-2ea2-4507-924d-f81aab9d947a-kube-api-access-29x8k" (OuterVolumeSpecName: "kube-api-access-29x8k") pod "aeccbc20-2ea2-4507-924d-f81aab9d947a" (UID: "aeccbc20-2ea2-4507-924d-f81aab9d947a"). InnerVolumeSpecName "kube-api-access-29x8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.424722 4776 generic.go:334] "Generic (PLEG): container finished" podID="aeccbc20-2ea2-4507-924d-f81aab9d947a" containerID="08b2752014bddb8246b5b237fdd56860d953bfc26af95426502f062d9c002b79" exitCode=0 Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.424766 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86dbcbdfc9-dl692" event={"ID":"aeccbc20-2ea2-4507-924d-f81aab9d947a","Type":"ContainerDied","Data":"08b2752014bddb8246b5b237fdd56860d953bfc26af95426502f062d9c002b79"} Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.424791 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86dbcbdfc9-dl692" event={"ID":"aeccbc20-2ea2-4507-924d-f81aab9d947a","Type":"ContainerDied","Data":"acd5be03ece01a4f28a5d1188113522f13fdb5ff3baf71cd74763b729c534b3c"} Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.424807 4776 scope.go:117] "RemoveContainer" containerID="08b2752014bddb8246b5b237fdd56860d953bfc26af95426502f062d9c002b79" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.424927 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86dbcbdfc9-dl692" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.449960 4776 scope.go:117] "RemoveContainer" containerID="3648b1978ec3bab3746f507d49deed95c584cb4515abd3c4a274b73b0453f620" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.471101 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aeccbc20-2ea2-4507-924d-f81aab9d947a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aeccbc20-2ea2-4507-924d-f81aab9d947a" (UID: "aeccbc20-2ea2-4507-924d-f81aab9d947a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.475525 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aeccbc20-2ea2-4507-924d-f81aab9d947a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "aeccbc20-2ea2-4507-924d-f81aab9d947a" (UID: "aeccbc20-2ea2-4507-924d-f81aab9d947a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.481087 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aeccbc20-2ea2-4507-924d-f81aab9d947a-config" (OuterVolumeSpecName: "config") pod "aeccbc20-2ea2-4507-924d-f81aab9d947a" (UID: "aeccbc20-2ea2-4507-924d-f81aab9d947a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.485130 4776 scope.go:117] "RemoveContainer" containerID="08b2752014bddb8246b5b237fdd56860d953bfc26af95426502f062d9c002b79" Nov 25 11:14:43 crc kubenswrapper[4776]: E1125 11:14:43.485543 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08b2752014bddb8246b5b237fdd56860d953bfc26af95426502f062d9c002b79\": container with ID starting with 08b2752014bddb8246b5b237fdd56860d953bfc26af95426502f062d9c002b79 not found: ID does not exist" containerID="08b2752014bddb8246b5b237fdd56860d953bfc26af95426502f062d9c002b79" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.485588 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08b2752014bddb8246b5b237fdd56860d953bfc26af95426502f062d9c002b79"} err="failed to get container status \"08b2752014bddb8246b5b237fdd56860d953bfc26af95426502f062d9c002b79\": rpc error: code = NotFound desc = could not find container \"08b2752014bddb8246b5b237fdd56860d953bfc26af95426502f062d9c002b79\": container with ID starting with 08b2752014bddb8246b5b237fdd56860d953bfc26af95426502f062d9c002b79 not found: ID does not exist" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.485613 4776 scope.go:117] "RemoveContainer" containerID="3648b1978ec3bab3746f507d49deed95c584cb4515abd3c4a274b73b0453f620" Nov 25 11:14:43 crc kubenswrapper[4776]: E1125 11:14:43.486055 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3648b1978ec3bab3746f507d49deed95c584cb4515abd3c4a274b73b0453f620\": container with ID starting with 3648b1978ec3bab3746f507d49deed95c584cb4515abd3c4a274b73b0453f620 not found: ID does not exist" containerID="3648b1978ec3bab3746f507d49deed95c584cb4515abd3c4a274b73b0453f620" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.486109 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3648b1978ec3bab3746f507d49deed95c584cb4515abd3c4a274b73b0453f620"} err="failed to get container status \"3648b1978ec3bab3746f507d49deed95c584cb4515abd3c4a274b73b0453f620\": rpc error: code = NotFound desc = could not find container \"3648b1978ec3bab3746f507d49deed95c584cb4515abd3c4a274b73b0453f620\": container with ID starting with 3648b1978ec3bab3746f507d49deed95c584cb4515abd3c4a274b73b0453f620 not found: ID does not exist" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.486511 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aeccbc20-2ea2-4507-924d-f81aab9d947a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aeccbc20-2ea2-4507-924d-f81aab9d947a" (UID: "aeccbc20-2ea2-4507-924d-f81aab9d947a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.516722 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aeccbc20-2ea2-4507-924d-f81aab9d947a-config\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.516764 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aeccbc20-2ea2-4507-924d-f81aab9d947a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.516776 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aeccbc20-2ea2-4507-924d-f81aab9d947a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.516789 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29x8k\" (UniqueName: \"kubernetes.io/projected/aeccbc20-2ea2-4507-924d-f81aab9d947a-kube-api-access-29x8k\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.516800 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aeccbc20-2ea2-4507-924d-f81aab9d947a-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:43 crc kubenswrapper[4776]: W1125 11:14:43.609944 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82c0a51a_57e4_4ab3_98b6_e2052e449bf0.slice/crio-3d587eba039795dfdcc38a9ba75906978e4c84d0cb068e456d17957af8d60db8 WatchSource:0}: Error finding container 3d587eba039795dfdcc38a9ba75906978e4c84d0cb068e456d17957af8d60db8: Status 404 returned error can't find the container with id 3d587eba039795dfdcc38a9ba75906978e4c84d0cb068e456d17957af8d60db8 Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.611310 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-568cc6cf7c-6mkrs"] Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.829755 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86dbcbdfc9-dl692"] Nov 25 11:14:43 crc kubenswrapper[4776]: I1125 11:14:43.841779 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86dbcbdfc9-dl692"] Nov 25 11:14:44 crc kubenswrapper[4776]: I1125 11:14:44.434665 4776 generic.go:334] "Generic (PLEG): container finished" podID="82c0a51a-57e4-4ab3-98b6-e2052e449bf0" containerID="3ac5c2daaead677421fe65935e9974c6ec60b5c7be3351f30078954c05f08a16" exitCode=0 Nov 25 11:14:44 crc kubenswrapper[4776]: I1125 11:14:44.434716 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568cc6cf7c-6mkrs" event={"ID":"82c0a51a-57e4-4ab3-98b6-e2052e449bf0","Type":"ContainerDied","Data":"3ac5c2daaead677421fe65935e9974c6ec60b5c7be3351f30078954c05f08a16"} Nov 25 11:14:44 crc kubenswrapper[4776]: I1125 11:14:44.435203 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568cc6cf7c-6mkrs" event={"ID":"82c0a51a-57e4-4ab3-98b6-e2052e449bf0","Type":"ContainerStarted","Data":"3d587eba039795dfdcc38a9ba75906978e4c84d0cb068e456d17957af8d60db8"} Nov 25 11:14:45 crc kubenswrapper[4776]: I1125 11:14:45.449197 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568cc6cf7c-6mkrs" event={"ID":"82c0a51a-57e4-4ab3-98b6-e2052e449bf0","Type":"ContainerStarted","Data":"ec4614cc24b7584e2ecd486a2d44de685a1469d63d8088005f40df1df0806a53"} Nov 25 11:14:45 crc kubenswrapper[4776]: I1125 11:14:45.449480 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-568cc6cf7c-6mkrs" Nov 25 11:14:45 crc kubenswrapper[4776]: I1125 11:14:45.470374 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-568cc6cf7c-6mkrs" podStartSLOduration=3.470354774 podStartE2EDuration="3.470354774s" podCreationTimestamp="2025-11-25 11:14:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:14:45.465227266 +0000 UTC m=+6630.506286819" watchObservedRunningTime="2025-11-25 11:14:45.470354774 +0000 UTC m=+6630.511414317" Nov 25 11:14:45 crc kubenswrapper[4776]: I1125 11:14:45.680151 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aeccbc20-2ea2-4507-924d-f81aab9d947a" path="/var/lib/kubelet/pods/aeccbc20-2ea2-4507-924d-f81aab9d947a/volumes" Nov 25 11:14:52 crc kubenswrapper[4776]: I1125 11:14:52.650005 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 25 11:14:53 crc kubenswrapper[4776]: I1125 11:14:53.119269 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-568cc6cf7c-6mkrs" Nov 25 11:14:53 crc kubenswrapper[4776]: I1125 11:14:53.203779 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86cd554965-lzn6h"] Nov 25 11:14:53 crc kubenswrapper[4776]: I1125 11:14:53.204609 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86cd554965-lzn6h" podUID="54a80827-fae6-48e8-8ba0-cfd09a9c3cc4" containerName="dnsmasq-dns" containerID="cri-o://478a7dd64198c32cbe605e3bdbc14bcc9c6977d86fa2766db9f5b0811c5e67aa" gracePeriod=10 Nov 25 11:14:53 crc kubenswrapper[4776]: I1125 11:14:53.537703 4776 generic.go:334] "Generic (PLEG): container finished" podID="54a80827-fae6-48e8-8ba0-cfd09a9c3cc4" containerID="478a7dd64198c32cbe605e3bdbc14bcc9c6977d86fa2766db9f5b0811c5e67aa" exitCode=0 Nov 25 11:14:53 crc kubenswrapper[4776]: I1125 11:14:53.537770 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86cd554965-lzn6h" event={"ID":"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4","Type":"ContainerDied","Data":"478a7dd64198c32cbe605e3bdbc14bcc9c6977d86fa2766db9f5b0811c5e67aa"} Nov 25 11:14:53 crc kubenswrapper[4776]: I1125 11:14:53.738752 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86cd554965-lzn6h" Nov 25 11:14:53 crc kubenswrapper[4776]: I1125 11:14:53.873716 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-ovsdbserver-sb\") pod \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\" (UID: \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\") " Nov 25 11:14:53 crc kubenswrapper[4776]: I1125 11:14:53.873880 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-config\") pod \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\" (UID: \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\") " Nov 25 11:14:53 crc kubenswrapper[4776]: I1125 11:14:53.873922 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-ovsdbserver-nb\") pod \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\" (UID: \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\") " Nov 25 11:14:53 crc kubenswrapper[4776]: I1125 11:14:53.873949 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-dns-svc\") pod \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\" (UID: \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\") " Nov 25 11:14:53 crc kubenswrapper[4776]: I1125 11:14:53.874098 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-openstack-cell1\") pod \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\" (UID: \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\") " Nov 25 11:14:53 crc kubenswrapper[4776]: I1125 11:14:53.874160 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lbfh\" (UniqueName: \"kubernetes.io/projected/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-kube-api-access-9lbfh\") pod \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\" (UID: \"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4\") " Nov 25 11:14:53 crc kubenswrapper[4776]: I1125 11:14:53.883425 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-kube-api-access-9lbfh" (OuterVolumeSpecName: "kube-api-access-9lbfh") pod "54a80827-fae6-48e8-8ba0-cfd09a9c3cc4" (UID: "54a80827-fae6-48e8-8ba0-cfd09a9c3cc4"). InnerVolumeSpecName "kube-api-access-9lbfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:14:53 crc kubenswrapper[4776]: I1125 11:14:53.929480 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "54a80827-fae6-48e8-8ba0-cfd09a9c3cc4" (UID: "54a80827-fae6-48e8-8ba0-cfd09a9c3cc4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:14:53 crc kubenswrapper[4776]: I1125 11:14:53.941859 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "54a80827-fae6-48e8-8ba0-cfd09a9c3cc4" (UID: "54a80827-fae6-48e8-8ba0-cfd09a9c3cc4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:14:53 crc kubenswrapper[4776]: I1125 11:14:53.945640 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "54a80827-fae6-48e8-8ba0-cfd09a9c3cc4" (UID: "54a80827-fae6-48e8-8ba0-cfd09a9c3cc4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:14:53 crc kubenswrapper[4776]: I1125 11:14:53.953746 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-config" (OuterVolumeSpecName: "config") pod "54a80827-fae6-48e8-8ba0-cfd09a9c3cc4" (UID: "54a80827-fae6-48e8-8ba0-cfd09a9c3cc4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:14:53 crc kubenswrapper[4776]: I1125 11:14:53.955895 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "54a80827-fae6-48e8-8ba0-cfd09a9c3cc4" (UID: "54a80827-fae6-48e8-8ba0-cfd09a9c3cc4"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:14:53 crc kubenswrapper[4776]: I1125 11:14:53.977249 4776 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-config\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:53 crc kubenswrapper[4776]: I1125 11:14:53.977285 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:53 crc kubenswrapper[4776]: I1125 11:14:53.977299 4776 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:53 crc kubenswrapper[4776]: I1125 11:14:53.977308 4776 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-openstack-cell1\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:53 crc kubenswrapper[4776]: I1125 11:14:53.977320 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lbfh\" (UniqueName: \"kubernetes.io/projected/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-kube-api-access-9lbfh\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:53 crc kubenswrapper[4776]: I1125 11:14:53.977331 4776 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 11:14:54 crc kubenswrapper[4776]: I1125 11:14:54.548778 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86cd554965-lzn6h" event={"ID":"54a80827-fae6-48e8-8ba0-cfd09a9c3cc4","Type":"ContainerDied","Data":"f752937c8fe681029d76416b084fcfc5b6ac14f7f216d6dccb735389d701dafe"} Nov 25 11:14:54 crc kubenswrapper[4776]: I1125 11:14:54.549106 4776 scope.go:117] "RemoveContainer" containerID="478a7dd64198c32cbe605e3bdbc14bcc9c6977d86fa2766db9f5b0811c5e67aa" Nov 25 11:14:54 crc kubenswrapper[4776]: I1125 11:14:54.548856 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86cd554965-lzn6h" Nov 25 11:14:54 crc kubenswrapper[4776]: I1125 11:14:54.578361 4776 scope.go:117] "RemoveContainer" containerID="7dc31647bb9fa1cb8d229efa7f6dd6722b4fb10bbf99791150f68a5c0e8adebe" Nov 25 11:14:54 crc kubenswrapper[4776]: I1125 11:14:54.593938 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86cd554965-lzn6h"] Nov 25 11:14:54 crc kubenswrapper[4776]: I1125 11:14:54.602384 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86cd554965-lzn6h"] Nov 25 11:14:55 crc kubenswrapper[4776]: I1125 11:14:55.675988 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54a80827-fae6-48e8-8ba0-cfd09a9c3cc4" path="/var/lib/kubelet/pods/54a80827-fae6-48e8-8ba0-cfd09a9c3cc4/volumes" Nov 25 11:15:00 crc kubenswrapper[4776]: I1125 11:15:00.157610 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401155-c8xlr"] Nov 25 11:15:00 crc kubenswrapper[4776]: E1125 11:15:00.158436 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeccbc20-2ea2-4507-924d-f81aab9d947a" containerName="init" Nov 25 11:15:00 crc kubenswrapper[4776]: I1125 11:15:00.158454 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeccbc20-2ea2-4507-924d-f81aab9d947a" containerName="init" Nov 25 11:15:00 crc kubenswrapper[4776]: E1125 11:15:00.158473 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeccbc20-2ea2-4507-924d-f81aab9d947a" containerName="dnsmasq-dns" Nov 25 11:15:00 crc kubenswrapper[4776]: I1125 11:15:00.158481 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeccbc20-2ea2-4507-924d-f81aab9d947a" containerName="dnsmasq-dns" Nov 25 11:15:00 crc kubenswrapper[4776]: E1125 11:15:00.158499 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54a80827-fae6-48e8-8ba0-cfd09a9c3cc4" containerName="dnsmasq-dns" Nov 25 11:15:00 crc kubenswrapper[4776]: I1125 11:15:00.158506 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="54a80827-fae6-48e8-8ba0-cfd09a9c3cc4" containerName="dnsmasq-dns" Nov 25 11:15:00 crc kubenswrapper[4776]: E1125 11:15:00.158528 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54a80827-fae6-48e8-8ba0-cfd09a9c3cc4" containerName="init" Nov 25 11:15:00 crc kubenswrapper[4776]: I1125 11:15:00.158533 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="54a80827-fae6-48e8-8ba0-cfd09a9c3cc4" containerName="init" Nov 25 11:15:00 crc kubenswrapper[4776]: I1125 11:15:00.158733 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeccbc20-2ea2-4507-924d-f81aab9d947a" containerName="dnsmasq-dns" Nov 25 11:15:00 crc kubenswrapper[4776]: I1125 11:15:00.158751 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="54a80827-fae6-48e8-8ba0-cfd09a9c3cc4" containerName="dnsmasq-dns" Nov 25 11:15:00 crc kubenswrapper[4776]: I1125 11:15:00.159545 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-c8xlr" Nov 25 11:15:00 crc kubenswrapper[4776]: I1125 11:15:00.164503 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 11:15:00 crc kubenswrapper[4776]: I1125 11:15:00.164624 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 11:15:00 crc kubenswrapper[4776]: I1125 11:15:00.168535 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401155-c8xlr"] Nov 25 11:15:00 crc kubenswrapper[4776]: I1125 11:15:00.217320 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fd8e5bf7-c842-4966-94d7-f33b21d20be3-secret-volume\") pod \"collect-profiles-29401155-c8xlr\" (UID: \"fd8e5bf7-c842-4966-94d7-f33b21d20be3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-c8xlr" Nov 25 11:15:00 crc kubenswrapper[4776]: I1125 11:15:00.217807 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdvkl\" (UniqueName: \"kubernetes.io/projected/fd8e5bf7-c842-4966-94d7-f33b21d20be3-kube-api-access-tdvkl\") pod \"collect-profiles-29401155-c8xlr\" (UID: \"fd8e5bf7-c842-4966-94d7-f33b21d20be3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-c8xlr" Nov 25 11:15:00 crc kubenswrapper[4776]: I1125 11:15:00.217933 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fd8e5bf7-c842-4966-94d7-f33b21d20be3-config-volume\") pod \"collect-profiles-29401155-c8xlr\" (UID: \"fd8e5bf7-c842-4966-94d7-f33b21d20be3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-c8xlr" Nov 25 11:15:00 crc kubenswrapper[4776]: I1125 11:15:00.319527 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdvkl\" (UniqueName: \"kubernetes.io/projected/fd8e5bf7-c842-4966-94d7-f33b21d20be3-kube-api-access-tdvkl\") pod \"collect-profiles-29401155-c8xlr\" (UID: \"fd8e5bf7-c842-4966-94d7-f33b21d20be3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-c8xlr" Nov 25 11:15:00 crc kubenswrapper[4776]: I1125 11:15:00.319896 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fd8e5bf7-c842-4966-94d7-f33b21d20be3-config-volume\") pod \"collect-profiles-29401155-c8xlr\" (UID: \"fd8e5bf7-c842-4966-94d7-f33b21d20be3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-c8xlr" Nov 25 11:15:00 crc kubenswrapper[4776]: I1125 11:15:00.320186 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fd8e5bf7-c842-4966-94d7-f33b21d20be3-secret-volume\") pod \"collect-profiles-29401155-c8xlr\" (UID: \"fd8e5bf7-c842-4966-94d7-f33b21d20be3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-c8xlr" Nov 25 11:15:00 crc kubenswrapper[4776]: I1125 11:15:00.321254 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fd8e5bf7-c842-4966-94d7-f33b21d20be3-config-volume\") pod \"collect-profiles-29401155-c8xlr\" (UID: \"fd8e5bf7-c842-4966-94d7-f33b21d20be3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-c8xlr" Nov 25 11:15:00 crc kubenswrapper[4776]: I1125 11:15:00.332006 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fd8e5bf7-c842-4966-94d7-f33b21d20be3-secret-volume\") pod \"collect-profiles-29401155-c8xlr\" (UID: \"fd8e5bf7-c842-4966-94d7-f33b21d20be3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-c8xlr" Nov 25 11:15:00 crc kubenswrapper[4776]: I1125 11:15:00.340845 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdvkl\" (UniqueName: \"kubernetes.io/projected/fd8e5bf7-c842-4966-94d7-f33b21d20be3-kube-api-access-tdvkl\") pod \"collect-profiles-29401155-c8xlr\" (UID: \"fd8e5bf7-c842-4966-94d7-f33b21d20be3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-c8xlr" Nov 25 11:15:00 crc kubenswrapper[4776]: I1125 11:15:00.490201 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-c8xlr" Nov 25 11:15:01 crc kubenswrapper[4776]: I1125 11:15:01.000934 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401155-c8xlr"] Nov 25 11:15:01 crc kubenswrapper[4776]: W1125 11:15:01.006982 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd8e5bf7_c842_4966_94d7_f33b21d20be3.slice/crio-3f773c9e05764575422931df78d3c6b78f86d81090a909367e44ef6232d89dff WatchSource:0}: Error finding container 3f773c9e05764575422931df78d3c6b78f86d81090a909367e44ef6232d89dff: Status 404 returned error can't find the container with id 3f773c9e05764575422931df78d3c6b78f86d81090a909367e44ef6232d89dff Nov 25 11:15:01 crc kubenswrapper[4776]: I1125 11:15:01.639615 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-c8xlr" event={"ID":"fd8e5bf7-c842-4966-94d7-f33b21d20be3","Type":"ContainerStarted","Data":"6d8f1800b7959498a8813e39a75ec0bdcb10bc2cab2d105342f6a5f186cf4107"} Nov 25 11:15:01 crc kubenswrapper[4776]: I1125 11:15:01.639675 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-c8xlr" event={"ID":"fd8e5bf7-c842-4966-94d7-f33b21d20be3","Type":"ContainerStarted","Data":"3f773c9e05764575422931df78d3c6b78f86d81090a909367e44ef6232d89dff"} Nov 25 11:15:01 crc kubenswrapper[4776]: I1125 11:15:01.655054 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-c8xlr" podStartSLOduration=1.655031315 podStartE2EDuration="1.655031315s" podCreationTimestamp="2025-11-25 11:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:15:01.653912507 +0000 UTC m=+6646.694972080" watchObservedRunningTime="2025-11-25 11:15:01.655031315 +0000 UTC m=+6646.696090868" Nov 25 11:15:02 crc kubenswrapper[4776]: I1125 11:15:02.648764 4776 generic.go:334] "Generic (PLEG): container finished" podID="fd8e5bf7-c842-4966-94d7-f33b21d20be3" containerID="6d8f1800b7959498a8813e39a75ec0bdcb10bc2cab2d105342f6a5f186cf4107" exitCode=0 Nov 25 11:15:02 crc kubenswrapper[4776]: I1125 11:15:02.648812 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-c8xlr" event={"ID":"fd8e5bf7-c842-4966-94d7-f33b21d20be3","Type":"ContainerDied","Data":"6d8f1800b7959498a8813e39a75ec0bdcb10bc2cab2d105342f6a5f186cf4107"} Nov 25 11:15:03 crc kubenswrapper[4776]: I1125 11:15:03.612370 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm"] Nov 25 11:15:03 crc kubenswrapper[4776]: I1125 11:15:03.614382 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm" Nov 25 11:15:03 crc kubenswrapper[4776]: I1125 11:15:03.617759 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:15:03 crc kubenswrapper[4776]: I1125 11:15:03.622294 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dknh7" Nov 25 11:15:03 crc kubenswrapper[4776]: I1125 11:15:03.622543 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 25 11:15:03 crc kubenswrapper[4776]: I1125 11:15:03.622733 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 25 11:15:03 crc kubenswrapper[4776]: I1125 11:15:03.629436 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm"] Nov 25 11:15:03 crc kubenswrapper[4776]: I1125 11:15:03.702300 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de94e490-fe33-4e32-ac8f-5dd704a298f1-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm\" (UID: \"de94e490-fe33-4e32-ac8f-5dd704a298f1\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm" Nov 25 11:15:03 crc kubenswrapper[4776]: I1125 11:15:03.702757 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de94e490-fe33-4e32-ac8f-5dd704a298f1-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm\" (UID: \"de94e490-fe33-4e32-ac8f-5dd704a298f1\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm" Nov 25 11:15:03 crc kubenswrapper[4776]: I1125 11:15:03.702809 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de94e490-fe33-4e32-ac8f-5dd704a298f1-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm\" (UID: \"de94e490-fe33-4e32-ac8f-5dd704a298f1\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm" Nov 25 11:15:03 crc kubenswrapper[4776]: I1125 11:15:03.702930 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drsck\" (UniqueName: \"kubernetes.io/projected/de94e490-fe33-4e32-ac8f-5dd704a298f1-kube-api-access-drsck\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm\" (UID: \"de94e490-fe33-4e32-ac8f-5dd704a298f1\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm" Nov 25 11:15:03 crc kubenswrapper[4776]: I1125 11:15:03.805199 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drsck\" (UniqueName: \"kubernetes.io/projected/de94e490-fe33-4e32-ac8f-5dd704a298f1-kube-api-access-drsck\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm\" (UID: \"de94e490-fe33-4e32-ac8f-5dd704a298f1\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm" Nov 25 11:15:03 crc kubenswrapper[4776]: I1125 11:15:03.805348 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de94e490-fe33-4e32-ac8f-5dd704a298f1-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm\" (UID: \"de94e490-fe33-4e32-ac8f-5dd704a298f1\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm" Nov 25 11:15:03 crc kubenswrapper[4776]: I1125 11:15:03.805507 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de94e490-fe33-4e32-ac8f-5dd704a298f1-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm\" (UID: \"de94e490-fe33-4e32-ac8f-5dd704a298f1\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm" Nov 25 11:15:03 crc kubenswrapper[4776]: I1125 11:15:03.805529 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de94e490-fe33-4e32-ac8f-5dd704a298f1-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm\" (UID: \"de94e490-fe33-4e32-ac8f-5dd704a298f1\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm" Nov 25 11:15:03 crc kubenswrapper[4776]: I1125 11:15:03.840976 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de94e490-fe33-4e32-ac8f-5dd704a298f1-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm\" (UID: \"de94e490-fe33-4e32-ac8f-5dd704a298f1\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm" Nov 25 11:15:03 crc kubenswrapper[4776]: I1125 11:15:03.842517 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de94e490-fe33-4e32-ac8f-5dd704a298f1-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm\" (UID: \"de94e490-fe33-4e32-ac8f-5dd704a298f1\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm" Nov 25 11:15:03 crc kubenswrapper[4776]: I1125 11:15:03.846126 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de94e490-fe33-4e32-ac8f-5dd704a298f1-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm\" (UID: \"de94e490-fe33-4e32-ac8f-5dd704a298f1\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm" Nov 25 11:15:03 crc kubenswrapper[4776]: I1125 11:15:03.860947 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drsck\" (UniqueName: \"kubernetes.io/projected/de94e490-fe33-4e32-ac8f-5dd704a298f1-kube-api-access-drsck\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm\" (UID: \"de94e490-fe33-4e32-ac8f-5dd704a298f1\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm" Nov 25 11:15:03 crc kubenswrapper[4776]: I1125 11:15:03.944953 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm" Nov 25 11:15:04 crc kubenswrapper[4776]: I1125 11:15:04.080686 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-c8xlr" Nov 25 11:15:04 crc kubenswrapper[4776]: I1125 11:15:04.212686 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fd8e5bf7-c842-4966-94d7-f33b21d20be3-config-volume\") pod \"fd8e5bf7-c842-4966-94d7-f33b21d20be3\" (UID: \"fd8e5bf7-c842-4966-94d7-f33b21d20be3\") " Nov 25 11:15:04 crc kubenswrapper[4776]: I1125 11:15:04.212882 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fd8e5bf7-c842-4966-94d7-f33b21d20be3-secret-volume\") pod \"fd8e5bf7-c842-4966-94d7-f33b21d20be3\" (UID: \"fd8e5bf7-c842-4966-94d7-f33b21d20be3\") " Nov 25 11:15:04 crc kubenswrapper[4776]: I1125 11:15:04.213031 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdvkl\" (UniqueName: \"kubernetes.io/projected/fd8e5bf7-c842-4966-94d7-f33b21d20be3-kube-api-access-tdvkl\") pod \"fd8e5bf7-c842-4966-94d7-f33b21d20be3\" (UID: \"fd8e5bf7-c842-4966-94d7-f33b21d20be3\") " Nov 25 11:15:04 crc kubenswrapper[4776]: I1125 11:15:04.213438 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd8e5bf7-c842-4966-94d7-f33b21d20be3-config-volume" (OuterVolumeSpecName: "config-volume") pod "fd8e5bf7-c842-4966-94d7-f33b21d20be3" (UID: "fd8e5bf7-c842-4966-94d7-f33b21d20be3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:15:04 crc kubenswrapper[4776]: I1125 11:15:04.213849 4776 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fd8e5bf7-c842-4966-94d7-f33b21d20be3-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 11:15:04 crc kubenswrapper[4776]: I1125 11:15:04.218552 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd8e5bf7-c842-4966-94d7-f33b21d20be3-kube-api-access-tdvkl" (OuterVolumeSpecName: "kube-api-access-tdvkl") pod "fd8e5bf7-c842-4966-94d7-f33b21d20be3" (UID: "fd8e5bf7-c842-4966-94d7-f33b21d20be3"). InnerVolumeSpecName "kube-api-access-tdvkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:15:04 crc kubenswrapper[4776]: I1125 11:15:04.218573 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd8e5bf7-c842-4966-94d7-f33b21d20be3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "fd8e5bf7-c842-4966-94d7-f33b21d20be3" (UID: "fd8e5bf7-c842-4966-94d7-f33b21d20be3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:15:04 crc kubenswrapper[4776]: I1125 11:15:04.315919 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdvkl\" (UniqueName: \"kubernetes.io/projected/fd8e5bf7-c842-4966-94d7-f33b21d20be3-kube-api-access-tdvkl\") on node \"crc\" DevicePath \"\"" Nov 25 11:15:04 crc kubenswrapper[4776]: I1125 11:15:04.315956 4776 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fd8e5bf7-c842-4966-94d7-f33b21d20be3-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 11:15:04 crc kubenswrapper[4776]: I1125 11:15:04.655613 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm"] Nov 25 11:15:04 crc kubenswrapper[4776]: W1125 11:15:04.661438 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde94e490_fe33_4e32_ac8f_5dd704a298f1.slice/crio-f93648c69220ad046c64c894f9ebee1d7ef86f7268423467b73e4a1c79e64bd5 WatchSource:0}: Error finding container f93648c69220ad046c64c894f9ebee1d7ef86f7268423467b73e4a1c79e64bd5: Status 404 returned error can't find the container with id f93648c69220ad046c64c894f9ebee1d7ef86f7268423467b73e4a1c79e64bd5 Nov 25 11:15:04 crc kubenswrapper[4776]: I1125 11:15:04.664172 4776 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 11:15:04 crc kubenswrapper[4776]: I1125 11:15:04.672702 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-c8xlr" event={"ID":"fd8e5bf7-c842-4966-94d7-f33b21d20be3","Type":"ContainerDied","Data":"3f773c9e05764575422931df78d3c6b78f86d81090a909367e44ef6232d89dff"} Nov 25 11:15:04 crc kubenswrapper[4776]: I1125 11:15:04.672750 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f773c9e05764575422931df78d3c6b78f86d81090a909367e44ef6232d89dff" Nov 25 11:15:04 crc kubenswrapper[4776]: I1125 11:15:04.672812 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401155-c8xlr" Nov 25 11:15:04 crc kubenswrapper[4776]: I1125 11:15:04.734198 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401110-zvmnb"] Nov 25 11:15:04 crc kubenswrapper[4776]: I1125 11:15:04.742883 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401110-zvmnb"] Nov 25 11:15:05 crc kubenswrapper[4776]: I1125 11:15:05.690561 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba" path="/var/lib/kubelet/pods/e25ccdc5-9273-47c9-8a8f-5a3e95d0bdba/volumes" Nov 25 11:15:05 crc kubenswrapper[4776]: I1125 11:15:05.710315 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm" event={"ID":"de94e490-fe33-4e32-ac8f-5dd704a298f1","Type":"ContainerStarted","Data":"f93648c69220ad046c64c894f9ebee1d7ef86f7268423467b73e4a1c79e64bd5"} Nov 25 11:15:15 crc kubenswrapper[4776]: I1125 11:15:15.784110 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm" event={"ID":"de94e490-fe33-4e32-ac8f-5dd704a298f1","Type":"ContainerStarted","Data":"1916157e538937f9d33e74f6c419457f745ab74143ca3d7ad9e6e3c05227167e"} Nov 25 11:15:15 crc kubenswrapper[4776]: I1125 11:15:15.807707 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm" podStartSLOduration=2.278654866 podStartE2EDuration="12.807684079s" podCreationTimestamp="2025-11-25 11:15:03 +0000 UTC" firstStartedPulling="2025-11-25 11:15:04.663937368 +0000 UTC m=+6649.704996921" lastFinishedPulling="2025-11-25 11:15:15.192966581 +0000 UTC m=+6660.234026134" observedRunningTime="2025-11-25 11:15:15.796459337 +0000 UTC m=+6660.837518900" watchObservedRunningTime="2025-11-25 11:15:15.807684079 +0000 UTC m=+6660.848743632" Nov 25 11:15:17 crc kubenswrapper[4776]: I1125 11:15:17.818095 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:15:17 crc kubenswrapper[4776]: I1125 11:15:17.818442 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:15:21 crc kubenswrapper[4776]: I1125 11:15:21.046944 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-create-cgttx"] Nov 25 11:15:21 crc kubenswrapper[4776]: I1125 11:15:21.057191 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-ac46-account-create-9jb68"] Nov 25 11:15:21 crc kubenswrapper[4776]: I1125 11:15:21.068468 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-create-cgttx"] Nov 25 11:15:21 crc kubenswrapper[4776]: I1125 11:15:21.078501 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-ac46-account-create-9jb68"] Nov 25 11:15:21 crc kubenswrapper[4776]: I1125 11:15:21.682357 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62d9656d-fca0-4900-a375-764c2567734e" path="/var/lib/kubelet/pods/62d9656d-fca0-4900-a375-764c2567734e/volumes" Nov 25 11:15:21 crc kubenswrapper[4776]: I1125 11:15:21.684848 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc137cb9-0788-455d-9f5f-2d7530c0c20e" path="/var/lib/kubelet/pods/dc137cb9-0788-455d-9f5f-2d7530c0c20e/volumes" Nov 25 11:15:27 crc kubenswrapper[4776]: I1125 11:15:27.961513 4776 generic.go:334] "Generic (PLEG): container finished" podID="de94e490-fe33-4e32-ac8f-5dd704a298f1" containerID="1916157e538937f9d33e74f6c419457f745ab74143ca3d7ad9e6e3c05227167e" exitCode=0 Nov 25 11:15:27 crc kubenswrapper[4776]: I1125 11:15:27.961595 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm" event={"ID":"de94e490-fe33-4e32-ac8f-5dd704a298f1","Type":"ContainerDied","Data":"1916157e538937f9d33e74f6c419457f745ab74143ca3d7ad9e6e3c05227167e"} Nov 25 11:15:29 crc kubenswrapper[4776]: I1125 11:15:29.377553 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm" Nov 25 11:15:29 crc kubenswrapper[4776]: I1125 11:15:29.489524 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de94e490-fe33-4e32-ac8f-5dd704a298f1-inventory\") pod \"de94e490-fe33-4e32-ac8f-5dd704a298f1\" (UID: \"de94e490-fe33-4e32-ac8f-5dd704a298f1\") " Nov 25 11:15:29 crc kubenswrapper[4776]: I1125 11:15:29.489739 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de94e490-fe33-4e32-ac8f-5dd704a298f1-pre-adoption-validation-combined-ca-bundle\") pod \"de94e490-fe33-4e32-ac8f-5dd704a298f1\" (UID: \"de94e490-fe33-4e32-ac8f-5dd704a298f1\") " Nov 25 11:15:29 crc kubenswrapper[4776]: I1125 11:15:29.489773 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drsck\" (UniqueName: \"kubernetes.io/projected/de94e490-fe33-4e32-ac8f-5dd704a298f1-kube-api-access-drsck\") pod \"de94e490-fe33-4e32-ac8f-5dd704a298f1\" (UID: \"de94e490-fe33-4e32-ac8f-5dd704a298f1\") " Nov 25 11:15:29 crc kubenswrapper[4776]: I1125 11:15:29.489843 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de94e490-fe33-4e32-ac8f-5dd704a298f1-ssh-key\") pod \"de94e490-fe33-4e32-ac8f-5dd704a298f1\" (UID: \"de94e490-fe33-4e32-ac8f-5dd704a298f1\") " Nov 25 11:15:29 crc kubenswrapper[4776]: I1125 11:15:29.497784 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de94e490-fe33-4e32-ac8f-5dd704a298f1-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "de94e490-fe33-4e32-ac8f-5dd704a298f1" (UID: "de94e490-fe33-4e32-ac8f-5dd704a298f1"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:15:29 crc kubenswrapper[4776]: I1125 11:15:29.497807 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de94e490-fe33-4e32-ac8f-5dd704a298f1-kube-api-access-drsck" (OuterVolumeSpecName: "kube-api-access-drsck") pod "de94e490-fe33-4e32-ac8f-5dd704a298f1" (UID: "de94e490-fe33-4e32-ac8f-5dd704a298f1"). InnerVolumeSpecName "kube-api-access-drsck". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:15:29 crc kubenswrapper[4776]: I1125 11:15:29.518098 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de94e490-fe33-4e32-ac8f-5dd704a298f1-inventory" (OuterVolumeSpecName: "inventory") pod "de94e490-fe33-4e32-ac8f-5dd704a298f1" (UID: "de94e490-fe33-4e32-ac8f-5dd704a298f1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:15:29 crc kubenswrapper[4776]: I1125 11:15:29.519654 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de94e490-fe33-4e32-ac8f-5dd704a298f1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "de94e490-fe33-4e32-ac8f-5dd704a298f1" (UID: "de94e490-fe33-4e32-ac8f-5dd704a298f1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:15:29 crc kubenswrapper[4776]: I1125 11:15:29.592546 4776 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de94e490-fe33-4e32-ac8f-5dd704a298f1-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:15:29 crc kubenswrapper[4776]: I1125 11:15:29.592588 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drsck\" (UniqueName: \"kubernetes.io/projected/de94e490-fe33-4e32-ac8f-5dd704a298f1-kube-api-access-drsck\") on node \"crc\" DevicePath \"\"" Nov 25 11:15:29 crc kubenswrapper[4776]: I1125 11:15:29.592600 4776 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/de94e490-fe33-4e32-ac8f-5dd704a298f1-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:15:29 crc kubenswrapper[4776]: I1125 11:15:29.592609 4776 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/de94e490-fe33-4e32-ac8f-5dd704a298f1-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:15:29 crc kubenswrapper[4776]: I1125 11:15:29.989416 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm" event={"ID":"de94e490-fe33-4e32-ac8f-5dd704a298f1","Type":"ContainerDied","Data":"f93648c69220ad046c64c894f9ebee1d7ef86f7268423467b73e4a1c79e64bd5"} Nov 25 11:15:29 crc kubenswrapper[4776]: I1125 11:15:29.989494 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f93648c69220ad046c64c894f9ebee1d7ef86f7268423467b73e4a1c79e64bd5" Nov 25 11:15:29 crc kubenswrapper[4776]: I1125 11:15:29.989609 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm" Nov 25 11:15:31 crc kubenswrapper[4776]: I1125 11:15:31.033509 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-persistence-db-create-gjqsq"] Nov 25 11:15:31 crc kubenswrapper[4776]: I1125 11:15:31.046693 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-persistence-db-create-gjqsq"] Nov 25 11:15:31 crc kubenswrapper[4776]: I1125 11:15:31.682829 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="703a295e-547f-4e35-a67a-dd2d77361161" path="/var/lib/kubelet/pods/703a295e-547f-4e35-a67a-dd2d77361161/volumes" Nov 25 11:15:32 crc kubenswrapper[4776]: I1125 11:15:32.041712 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-116d-account-create-jf6wx"] Nov 25 11:15:32 crc kubenswrapper[4776]: I1125 11:15:32.052560 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-116d-account-create-jf6wx"] Nov 25 11:15:33 crc kubenswrapper[4776]: I1125 11:15:33.686504 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53c3e53a-7c78-4607-923c-604eed89a9df" path="/var/lib/kubelet/pods/53c3e53a-7c78-4607-923c-604eed89a9df/volumes" Nov 25 11:15:36 crc kubenswrapper[4776]: I1125 11:15:36.278882 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s"] Nov 25 11:15:36 crc kubenswrapper[4776]: E1125 11:15:36.280408 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de94e490-fe33-4e32-ac8f-5dd704a298f1" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Nov 25 11:15:36 crc kubenswrapper[4776]: I1125 11:15:36.280428 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="de94e490-fe33-4e32-ac8f-5dd704a298f1" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Nov 25 11:15:36 crc kubenswrapper[4776]: E1125 11:15:36.280448 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd8e5bf7-c842-4966-94d7-f33b21d20be3" containerName="collect-profiles" Nov 25 11:15:36 crc kubenswrapper[4776]: I1125 11:15:36.280454 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd8e5bf7-c842-4966-94d7-f33b21d20be3" containerName="collect-profiles" Nov 25 11:15:36 crc kubenswrapper[4776]: I1125 11:15:36.280653 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd8e5bf7-c842-4966-94d7-f33b21d20be3" containerName="collect-profiles" Nov 25 11:15:36 crc kubenswrapper[4776]: I1125 11:15:36.280676 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="de94e490-fe33-4e32-ac8f-5dd704a298f1" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Nov 25 11:15:36 crc kubenswrapper[4776]: I1125 11:15:36.281489 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s" Nov 25 11:15:36 crc kubenswrapper[4776]: I1125 11:15:36.284219 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dknh7" Nov 25 11:15:36 crc kubenswrapper[4776]: I1125 11:15:36.284620 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 25 11:15:36 crc kubenswrapper[4776]: I1125 11:15:36.285611 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:15:36 crc kubenswrapper[4776]: I1125 11:15:36.291154 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 25 11:15:36 crc kubenswrapper[4776]: I1125 11:15:36.292848 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s"] Nov 25 11:15:36 crc kubenswrapper[4776]: I1125 11:15:36.470013 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s\" (UID: \"4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s" Nov 25 11:15:36 crc kubenswrapper[4776]: I1125 11:15:36.470264 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kls22\" (UniqueName: \"kubernetes.io/projected/4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8-kube-api-access-kls22\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s\" (UID: \"4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s" Nov 25 11:15:36 crc kubenswrapper[4776]: I1125 11:15:36.470477 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s\" (UID: \"4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s" Nov 25 11:15:36 crc kubenswrapper[4776]: I1125 11:15:36.470718 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s\" (UID: \"4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s" Nov 25 11:15:36 crc kubenswrapper[4776]: I1125 11:15:36.573782 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s\" (UID: \"4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s" Nov 25 11:15:36 crc kubenswrapper[4776]: I1125 11:15:36.573853 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kls22\" (UniqueName: \"kubernetes.io/projected/4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8-kube-api-access-kls22\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s\" (UID: \"4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s" Nov 25 11:15:36 crc kubenswrapper[4776]: I1125 11:15:36.573902 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s\" (UID: \"4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s" Nov 25 11:15:36 crc kubenswrapper[4776]: I1125 11:15:36.574007 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s\" (UID: \"4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s" Nov 25 11:15:36 crc kubenswrapper[4776]: I1125 11:15:36.579310 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s\" (UID: \"4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s" Nov 25 11:15:36 crc kubenswrapper[4776]: I1125 11:15:36.579417 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s\" (UID: \"4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s" Nov 25 11:15:36 crc kubenswrapper[4776]: I1125 11:15:36.583665 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s\" (UID: \"4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s" Nov 25 11:15:36 crc kubenswrapper[4776]: I1125 11:15:36.596127 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kls22\" (UniqueName: \"kubernetes.io/projected/4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8-kube-api-access-kls22\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s\" (UID: \"4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s" Nov 25 11:15:36 crc kubenswrapper[4776]: I1125 11:15:36.612701 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s" Nov 25 11:15:37 crc kubenswrapper[4776]: I1125 11:15:37.193789 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s"] Nov 25 11:15:37 crc kubenswrapper[4776]: W1125 11:15:37.203337 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ffa64e0_2cd8_4fe7_90d1_0e4ce4a88ba8.slice/crio-fdd7374ddbc98c24e09ef70d8405488b1d5242071efb1094ec5c66f436a8e725 WatchSource:0}: Error finding container fdd7374ddbc98c24e09ef70d8405488b1d5242071efb1094ec5c66f436a8e725: Status 404 returned error can't find the container with id fdd7374ddbc98c24e09ef70d8405488b1d5242071efb1094ec5c66f436a8e725 Nov 25 11:15:38 crc kubenswrapper[4776]: I1125 11:15:38.074270 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s" event={"ID":"4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8","Type":"ContainerStarted","Data":"12acfde4b5070a9f48760843816de0fe3aabf494f8eeb5c42ec02961dcc8dd30"} Nov 25 11:15:38 crc kubenswrapper[4776]: I1125 11:15:38.074870 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s" event={"ID":"4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8","Type":"ContainerStarted","Data":"fdd7374ddbc98c24e09ef70d8405488b1d5242071efb1094ec5c66f436a8e725"} Nov 25 11:15:38 crc kubenswrapper[4776]: I1125 11:15:38.104918 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s" podStartSLOduration=1.675900707 podStartE2EDuration="2.104892273s" podCreationTimestamp="2025-11-25 11:15:36 +0000 UTC" firstStartedPulling="2025-11-25 11:15:37.205217585 +0000 UTC m=+6682.246277138" lastFinishedPulling="2025-11-25 11:15:37.634209151 +0000 UTC m=+6682.675268704" observedRunningTime="2025-11-25 11:15:38.095999641 +0000 UTC m=+6683.137059194" watchObservedRunningTime="2025-11-25 11:15:38.104892273 +0000 UTC m=+6683.145951826" Nov 25 11:15:39 crc kubenswrapper[4776]: I1125 11:15:39.403022 4776 scope.go:117] "RemoveContainer" containerID="f025724c8ab81ee0a83b81d6cbcb981381614ada45b3b21ef81fbf4082a243d3" Nov 25 11:15:39 crc kubenswrapper[4776]: I1125 11:15:39.480959 4776 scope.go:117] "RemoveContainer" containerID="139ec3a7186c7a3dbacfefd2816155567ee5fa7b87a7852026635b3aaaff1030" Nov 25 11:15:39 crc kubenswrapper[4776]: I1125 11:15:39.566196 4776 scope.go:117] "RemoveContainer" containerID="7968e159f463c8259bd2c25b1735de14db0d97b3eacbf0f1fca4fa992873748d" Nov 25 11:15:39 crc kubenswrapper[4776]: I1125 11:15:39.624412 4776 scope.go:117] "RemoveContainer" containerID="abbfd694167843fa80feaa11dbb30b5a0931e128c7fec98484adaddf3c487af7" Nov 25 11:15:39 crc kubenswrapper[4776]: I1125 11:15:39.658649 4776 scope.go:117] "RemoveContainer" containerID="89f14821d2aaeda2739c1d10adaa4b87fb5a83414dc465a734f6b375da24fdee" Nov 25 11:15:39 crc kubenswrapper[4776]: I1125 11:15:39.959268 4776 scope.go:117] "RemoveContainer" containerID="048841f4110bcd3f5595becc35292ea8885ba21d1bde0f0d3e40f649c52ef3a8" Nov 25 11:15:39 crc kubenswrapper[4776]: I1125 11:15:39.990943 4776 scope.go:117] "RemoveContainer" containerID="15701fba63ade2b2868713c0edcf9c8ebd367f8392edeb5225004c085debdaf5" Nov 25 11:15:40 crc kubenswrapper[4776]: I1125 11:15:40.600798 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ddv8c"] Nov 25 11:15:40 crc kubenswrapper[4776]: I1125 11:15:40.602912 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ddv8c" Nov 25 11:15:40 crc kubenswrapper[4776]: I1125 11:15:40.612447 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ddv8c"] Nov 25 11:15:40 crc kubenswrapper[4776]: I1125 11:15:40.663966 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ns6rn\" (UniqueName: \"kubernetes.io/projected/91f5209c-ee2d-404f-a7b4-9305ec08f1ac-kube-api-access-ns6rn\") pod \"community-operators-ddv8c\" (UID: \"91f5209c-ee2d-404f-a7b4-9305ec08f1ac\") " pod="openshift-marketplace/community-operators-ddv8c" Nov 25 11:15:40 crc kubenswrapper[4776]: I1125 11:15:40.664312 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91f5209c-ee2d-404f-a7b4-9305ec08f1ac-catalog-content\") pod \"community-operators-ddv8c\" (UID: \"91f5209c-ee2d-404f-a7b4-9305ec08f1ac\") " pod="openshift-marketplace/community-operators-ddv8c" Nov 25 11:15:40 crc kubenswrapper[4776]: I1125 11:15:40.664403 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91f5209c-ee2d-404f-a7b4-9305ec08f1ac-utilities\") pod \"community-operators-ddv8c\" (UID: \"91f5209c-ee2d-404f-a7b4-9305ec08f1ac\") " pod="openshift-marketplace/community-operators-ddv8c" Nov 25 11:15:40 crc kubenswrapper[4776]: I1125 11:15:40.766014 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ns6rn\" (UniqueName: \"kubernetes.io/projected/91f5209c-ee2d-404f-a7b4-9305ec08f1ac-kube-api-access-ns6rn\") pod \"community-operators-ddv8c\" (UID: \"91f5209c-ee2d-404f-a7b4-9305ec08f1ac\") " pod="openshift-marketplace/community-operators-ddv8c" Nov 25 11:15:40 crc kubenswrapper[4776]: I1125 11:15:40.766463 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91f5209c-ee2d-404f-a7b4-9305ec08f1ac-catalog-content\") pod \"community-operators-ddv8c\" (UID: \"91f5209c-ee2d-404f-a7b4-9305ec08f1ac\") " pod="openshift-marketplace/community-operators-ddv8c" Nov 25 11:15:40 crc kubenswrapper[4776]: I1125 11:15:40.766493 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91f5209c-ee2d-404f-a7b4-9305ec08f1ac-utilities\") pod \"community-operators-ddv8c\" (UID: \"91f5209c-ee2d-404f-a7b4-9305ec08f1ac\") " pod="openshift-marketplace/community-operators-ddv8c" Nov 25 11:15:40 crc kubenswrapper[4776]: I1125 11:15:40.767234 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91f5209c-ee2d-404f-a7b4-9305ec08f1ac-catalog-content\") pod \"community-operators-ddv8c\" (UID: \"91f5209c-ee2d-404f-a7b4-9305ec08f1ac\") " pod="openshift-marketplace/community-operators-ddv8c" Nov 25 11:15:40 crc kubenswrapper[4776]: I1125 11:15:40.767356 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91f5209c-ee2d-404f-a7b4-9305ec08f1ac-utilities\") pod \"community-operators-ddv8c\" (UID: \"91f5209c-ee2d-404f-a7b4-9305ec08f1ac\") " pod="openshift-marketplace/community-operators-ddv8c" Nov 25 11:15:40 crc kubenswrapper[4776]: I1125 11:15:40.794416 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ns6rn\" (UniqueName: \"kubernetes.io/projected/91f5209c-ee2d-404f-a7b4-9305ec08f1ac-kube-api-access-ns6rn\") pod \"community-operators-ddv8c\" (UID: \"91f5209c-ee2d-404f-a7b4-9305ec08f1ac\") " pod="openshift-marketplace/community-operators-ddv8c" Nov 25 11:15:40 crc kubenswrapper[4776]: I1125 11:15:40.923236 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ddv8c" Nov 25 11:15:41 crc kubenswrapper[4776]: I1125 11:15:41.375113 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ddv8c"] Nov 25 11:15:42 crc kubenswrapper[4776]: I1125 11:15:42.134520 4776 generic.go:334] "Generic (PLEG): container finished" podID="91f5209c-ee2d-404f-a7b4-9305ec08f1ac" containerID="504fdba5e74691f9e3bcce97a71d92c5a0b3c5f1bc90b93dd4b520341ac54a77" exitCode=0 Nov 25 11:15:42 crc kubenswrapper[4776]: I1125 11:15:42.134607 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ddv8c" event={"ID":"91f5209c-ee2d-404f-a7b4-9305ec08f1ac","Type":"ContainerDied","Data":"504fdba5e74691f9e3bcce97a71d92c5a0b3c5f1bc90b93dd4b520341ac54a77"} Nov 25 11:15:42 crc kubenswrapper[4776]: I1125 11:15:42.134783 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ddv8c" event={"ID":"91f5209c-ee2d-404f-a7b4-9305ec08f1ac","Type":"ContainerStarted","Data":"92437fc98b78501215a5f4f27682e476b1c3fa26478fafea0ae04fcafc08e8a6"} Nov 25 11:15:44 crc kubenswrapper[4776]: I1125 11:15:44.160918 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ddv8c" event={"ID":"91f5209c-ee2d-404f-a7b4-9305ec08f1ac","Type":"ContainerStarted","Data":"d886e80e3395200c4f6418dc76667121dfbca749df4a39121194056d1c72a6c2"} Nov 25 11:15:47 crc kubenswrapper[4776]: I1125 11:15:47.819006 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:15:47 crc kubenswrapper[4776]: I1125 11:15:47.819663 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:15:49 crc kubenswrapper[4776]: I1125 11:15:49.214537 4776 generic.go:334] "Generic (PLEG): container finished" podID="91f5209c-ee2d-404f-a7b4-9305ec08f1ac" containerID="d886e80e3395200c4f6418dc76667121dfbca749df4a39121194056d1c72a6c2" exitCode=0 Nov 25 11:15:49 crc kubenswrapper[4776]: I1125 11:15:49.214850 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ddv8c" event={"ID":"91f5209c-ee2d-404f-a7b4-9305ec08f1ac","Type":"ContainerDied","Data":"d886e80e3395200c4f6418dc76667121dfbca749df4a39121194056d1c72a6c2"} Nov 25 11:15:51 crc kubenswrapper[4776]: I1125 11:15:51.235544 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ddv8c" event={"ID":"91f5209c-ee2d-404f-a7b4-9305ec08f1ac","Type":"ContainerStarted","Data":"24ba0e0e2b93eb754eec84a4e4014650436b31f0c6c9cdba3e865d8aecd8ae28"} Nov 25 11:15:51 crc kubenswrapper[4776]: I1125 11:15:51.270508 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ddv8c" podStartSLOduration=3.385063997 podStartE2EDuration="11.270484766s" podCreationTimestamp="2025-11-25 11:15:40 +0000 UTC" firstStartedPulling="2025-11-25 11:15:42.137614 +0000 UTC m=+6687.178673553" lastFinishedPulling="2025-11-25 11:15:50.023034759 +0000 UTC m=+6695.064094322" observedRunningTime="2025-11-25 11:15:51.254510835 +0000 UTC m=+6696.295570388" watchObservedRunningTime="2025-11-25 11:15:51.270484766 +0000 UTC m=+6696.311544319" Nov 25 11:16:00 crc kubenswrapper[4776]: I1125 11:16:00.924183 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ddv8c" Nov 25 11:16:00 crc kubenswrapper[4776]: I1125 11:16:00.924873 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ddv8c" Nov 25 11:16:01 crc kubenswrapper[4776]: I1125 11:16:01.971727 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-ddv8c" podUID="91f5209c-ee2d-404f-a7b4-9305ec08f1ac" containerName="registry-server" probeResult="failure" output=< Nov 25 11:16:01 crc kubenswrapper[4776]: timeout: failed to connect service ":50051" within 1s Nov 25 11:16:01 crc kubenswrapper[4776]: > Nov 25 11:16:10 crc kubenswrapper[4776]: I1125 11:16:10.977466 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ddv8c" Nov 25 11:16:11 crc kubenswrapper[4776]: I1125 11:16:11.032542 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ddv8c" Nov 25 11:16:11 crc kubenswrapper[4776]: I1125 11:16:11.941315 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ddv8c"] Nov 25 11:16:12 crc kubenswrapper[4776]: I1125 11:16:12.121295 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ddv8c" podUID="91f5209c-ee2d-404f-a7b4-9305ec08f1ac" containerName="registry-server" containerID="cri-o://24ba0e0e2b93eb754eec84a4e4014650436b31f0c6c9cdba3e865d8aecd8ae28" gracePeriod=2 Nov 25 11:16:12 crc kubenswrapper[4776]: I1125 11:16:12.713857 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ddv8c" Nov 25 11:16:12 crc kubenswrapper[4776]: I1125 11:16:12.860657 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91f5209c-ee2d-404f-a7b4-9305ec08f1ac-catalog-content\") pod \"91f5209c-ee2d-404f-a7b4-9305ec08f1ac\" (UID: \"91f5209c-ee2d-404f-a7b4-9305ec08f1ac\") " Nov 25 11:16:12 crc kubenswrapper[4776]: I1125 11:16:12.860889 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ns6rn\" (UniqueName: \"kubernetes.io/projected/91f5209c-ee2d-404f-a7b4-9305ec08f1ac-kube-api-access-ns6rn\") pod \"91f5209c-ee2d-404f-a7b4-9305ec08f1ac\" (UID: \"91f5209c-ee2d-404f-a7b4-9305ec08f1ac\") " Nov 25 11:16:12 crc kubenswrapper[4776]: I1125 11:16:12.861032 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91f5209c-ee2d-404f-a7b4-9305ec08f1ac-utilities\") pod \"91f5209c-ee2d-404f-a7b4-9305ec08f1ac\" (UID: \"91f5209c-ee2d-404f-a7b4-9305ec08f1ac\") " Nov 25 11:16:12 crc kubenswrapper[4776]: I1125 11:16:12.861854 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91f5209c-ee2d-404f-a7b4-9305ec08f1ac-utilities" (OuterVolumeSpecName: "utilities") pod "91f5209c-ee2d-404f-a7b4-9305ec08f1ac" (UID: "91f5209c-ee2d-404f-a7b4-9305ec08f1ac"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:16:12 crc kubenswrapper[4776]: I1125 11:16:12.867981 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91f5209c-ee2d-404f-a7b4-9305ec08f1ac-kube-api-access-ns6rn" (OuterVolumeSpecName: "kube-api-access-ns6rn") pod "91f5209c-ee2d-404f-a7b4-9305ec08f1ac" (UID: "91f5209c-ee2d-404f-a7b4-9305ec08f1ac"). InnerVolumeSpecName "kube-api-access-ns6rn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:16:12 crc kubenswrapper[4776]: I1125 11:16:12.916803 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91f5209c-ee2d-404f-a7b4-9305ec08f1ac-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91f5209c-ee2d-404f-a7b4-9305ec08f1ac" (UID: "91f5209c-ee2d-404f-a7b4-9305ec08f1ac"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:16:12 crc kubenswrapper[4776]: I1125 11:16:12.966728 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91f5209c-ee2d-404f-a7b4-9305ec08f1ac-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:16:12 crc kubenswrapper[4776]: I1125 11:16:12.966816 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91f5209c-ee2d-404f-a7b4-9305ec08f1ac-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:16:12 crc kubenswrapper[4776]: I1125 11:16:12.966831 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ns6rn\" (UniqueName: \"kubernetes.io/projected/91f5209c-ee2d-404f-a7b4-9305ec08f1ac-kube-api-access-ns6rn\") on node \"crc\" DevicePath \"\"" Nov 25 11:16:13 crc kubenswrapper[4776]: I1125 11:16:13.134488 4776 generic.go:334] "Generic (PLEG): container finished" podID="91f5209c-ee2d-404f-a7b4-9305ec08f1ac" containerID="24ba0e0e2b93eb754eec84a4e4014650436b31f0c6c9cdba3e865d8aecd8ae28" exitCode=0 Nov 25 11:16:13 crc kubenswrapper[4776]: I1125 11:16:13.134532 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ddv8c" event={"ID":"91f5209c-ee2d-404f-a7b4-9305ec08f1ac","Type":"ContainerDied","Data":"24ba0e0e2b93eb754eec84a4e4014650436b31f0c6c9cdba3e865d8aecd8ae28"} Nov 25 11:16:13 crc kubenswrapper[4776]: I1125 11:16:13.134559 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ddv8c" event={"ID":"91f5209c-ee2d-404f-a7b4-9305ec08f1ac","Type":"ContainerDied","Data":"92437fc98b78501215a5f4f27682e476b1c3fa26478fafea0ae04fcafc08e8a6"} Nov 25 11:16:13 crc kubenswrapper[4776]: I1125 11:16:13.134576 4776 scope.go:117] "RemoveContainer" containerID="24ba0e0e2b93eb754eec84a4e4014650436b31f0c6c9cdba3e865d8aecd8ae28" Nov 25 11:16:13 crc kubenswrapper[4776]: I1125 11:16:13.135845 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ddv8c" Nov 25 11:16:13 crc kubenswrapper[4776]: I1125 11:16:13.163384 4776 scope.go:117] "RemoveContainer" containerID="d886e80e3395200c4f6418dc76667121dfbca749df4a39121194056d1c72a6c2" Nov 25 11:16:13 crc kubenswrapper[4776]: I1125 11:16:13.168518 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ddv8c"] Nov 25 11:16:13 crc kubenswrapper[4776]: I1125 11:16:13.177114 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ddv8c"] Nov 25 11:16:13 crc kubenswrapper[4776]: I1125 11:16:13.193849 4776 scope.go:117] "RemoveContainer" containerID="504fdba5e74691f9e3bcce97a71d92c5a0b3c5f1bc90b93dd4b520341ac54a77" Nov 25 11:16:13 crc kubenswrapper[4776]: I1125 11:16:13.245728 4776 scope.go:117] "RemoveContainer" containerID="24ba0e0e2b93eb754eec84a4e4014650436b31f0c6c9cdba3e865d8aecd8ae28" Nov 25 11:16:13 crc kubenswrapper[4776]: E1125 11:16:13.246312 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24ba0e0e2b93eb754eec84a4e4014650436b31f0c6c9cdba3e865d8aecd8ae28\": container with ID starting with 24ba0e0e2b93eb754eec84a4e4014650436b31f0c6c9cdba3e865d8aecd8ae28 not found: ID does not exist" containerID="24ba0e0e2b93eb754eec84a4e4014650436b31f0c6c9cdba3e865d8aecd8ae28" Nov 25 11:16:13 crc kubenswrapper[4776]: I1125 11:16:13.246350 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24ba0e0e2b93eb754eec84a4e4014650436b31f0c6c9cdba3e865d8aecd8ae28"} err="failed to get container status \"24ba0e0e2b93eb754eec84a4e4014650436b31f0c6c9cdba3e865d8aecd8ae28\": rpc error: code = NotFound desc = could not find container \"24ba0e0e2b93eb754eec84a4e4014650436b31f0c6c9cdba3e865d8aecd8ae28\": container with ID starting with 24ba0e0e2b93eb754eec84a4e4014650436b31f0c6c9cdba3e865d8aecd8ae28 not found: ID does not exist" Nov 25 11:16:13 crc kubenswrapper[4776]: I1125 11:16:13.246379 4776 scope.go:117] "RemoveContainer" containerID="d886e80e3395200c4f6418dc76667121dfbca749df4a39121194056d1c72a6c2" Nov 25 11:16:13 crc kubenswrapper[4776]: E1125 11:16:13.246772 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d886e80e3395200c4f6418dc76667121dfbca749df4a39121194056d1c72a6c2\": container with ID starting with d886e80e3395200c4f6418dc76667121dfbca749df4a39121194056d1c72a6c2 not found: ID does not exist" containerID="d886e80e3395200c4f6418dc76667121dfbca749df4a39121194056d1c72a6c2" Nov 25 11:16:13 crc kubenswrapper[4776]: I1125 11:16:13.246880 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d886e80e3395200c4f6418dc76667121dfbca749df4a39121194056d1c72a6c2"} err="failed to get container status \"d886e80e3395200c4f6418dc76667121dfbca749df4a39121194056d1c72a6c2\": rpc error: code = NotFound desc = could not find container \"d886e80e3395200c4f6418dc76667121dfbca749df4a39121194056d1c72a6c2\": container with ID starting with d886e80e3395200c4f6418dc76667121dfbca749df4a39121194056d1c72a6c2 not found: ID does not exist" Nov 25 11:16:13 crc kubenswrapper[4776]: I1125 11:16:13.246967 4776 scope.go:117] "RemoveContainer" containerID="504fdba5e74691f9e3bcce97a71d92c5a0b3c5f1bc90b93dd4b520341ac54a77" Nov 25 11:16:13 crc kubenswrapper[4776]: E1125 11:16:13.247375 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"504fdba5e74691f9e3bcce97a71d92c5a0b3c5f1bc90b93dd4b520341ac54a77\": container with ID starting with 504fdba5e74691f9e3bcce97a71d92c5a0b3c5f1bc90b93dd4b520341ac54a77 not found: ID does not exist" containerID="504fdba5e74691f9e3bcce97a71d92c5a0b3c5f1bc90b93dd4b520341ac54a77" Nov 25 11:16:13 crc kubenswrapper[4776]: I1125 11:16:13.247403 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"504fdba5e74691f9e3bcce97a71d92c5a0b3c5f1bc90b93dd4b520341ac54a77"} err="failed to get container status \"504fdba5e74691f9e3bcce97a71d92c5a0b3c5f1bc90b93dd4b520341ac54a77\": rpc error: code = NotFound desc = could not find container \"504fdba5e74691f9e3bcce97a71d92c5a0b3c5f1bc90b93dd4b520341ac54a77\": container with ID starting with 504fdba5e74691f9e3bcce97a71d92c5a0b3c5f1bc90b93dd4b520341ac54a77 not found: ID does not exist" Nov 25 11:16:13 crc kubenswrapper[4776]: I1125 11:16:13.694376 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91f5209c-ee2d-404f-a7b4-9305ec08f1ac" path="/var/lib/kubelet/pods/91f5209c-ee2d-404f-a7b4-9305ec08f1ac/volumes" Nov 25 11:16:17 crc kubenswrapper[4776]: I1125 11:16:17.826243 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:16:17 crc kubenswrapper[4776]: I1125 11:16:17.826859 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:16:17 crc kubenswrapper[4776]: I1125 11:16:17.826916 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 11:16:17 crc kubenswrapper[4776]: I1125 11:16:17.827867 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 11:16:17 crc kubenswrapper[4776]: I1125 11:16:17.827961 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564" gracePeriod=600 Nov 25 11:16:18 crc kubenswrapper[4776]: I1125 11:16:18.182485 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564" exitCode=0 Nov 25 11:16:18 crc kubenswrapper[4776]: I1125 11:16:18.182553 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564"} Nov 25 11:16:18 crc kubenswrapper[4776]: I1125 11:16:18.182802 4776 scope.go:117] "RemoveContainer" containerID="d9823ef55a496477edc68cef40859ee8ddcd3f8a6fd6685be1f1eda0a305897e" Nov 25 11:16:18 crc kubenswrapper[4776]: E1125 11:16:18.526115 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:16:19 crc kubenswrapper[4776]: I1125 11:16:19.197614 4776 scope.go:117] "RemoveContainer" containerID="251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564" Nov 25 11:16:19 crc kubenswrapper[4776]: E1125 11:16:19.198497 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:16:31 crc kubenswrapper[4776]: I1125 11:16:31.664438 4776 scope.go:117] "RemoveContainer" containerID="251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564" Nov 25 11:16:31 crc kubenswrapper[4776]: E1125 11:16:31.665494 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:16:43 crc kubenswrapper[4776]: I1125 11:16:43.662464 4776 scope.go:117] "RemoveContainer" containerID="251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564" Nov 25 11:16:43 crc kubenswrapper[4776]: E1125 11:16:43.663549 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:16:56 crc kubenswrapper[4776]: I1125 11:16:56.664180 4776 scope.go:117] "RemoveContainer" containerID="251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564" Nov 25 11:16:56 crc kubenswrapper[4776]: E1125 11:16:56.664902 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:17:01 crc kubenswrapper[4776]: I1125 11:17:01.043724 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-sync-gwv2g"] Nov 25 11:17:01 crc kubenswrapper[4776]: I1125 11:17:01.054470 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-sync-gwv2g"] Nov 25 11:17:01 crc kubenswrapper[4776]: I1125 11:17:01.677003 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e" path="/var/lib/kubelet/pods/73f7c1dc-9edd-40f4-b8f4-16b99ea7ac7e/volumes" Nov 25 11:17:10 crc kubenswrapper[4776]: I1125 11:17:10.662971 4776 scope.go:117] "RemoveContainer" containerID="251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564" Nov 25 11:17:10 crc kubenswrapper[4776]: E1125 11:17:10.663706 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:17:22 crc kubenswrapper[4776]: I1125 11:17:22.663412 4776 scope.go:117] "RemoveContainer" containerID="251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564" Nov 25 11:17:22 crc kubenswrapper[4776]: E1125 11:17:22.666319 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:17:36 crc kubenswrapper[4776]: I1125 11:17:36.662875 4776 scope.go:117] "RemoveContainer" containerID="251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564" Nov 25 11:17:36 crc kubenswrapper[4776]: E1125 11:17:36.663662 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:17:40 crc kubenswrapper[4776]: I1125 11:17:40.372644 4776 scope.go:117] "RemoveContainer" containerID="0241f0320df1e6c1e5c6ca440d91ab0ec49adce12766987f286c37e32eb369e7" Nov 25 11:17:40 crc kubenswrapper[4776]: I1125 11:17:40.419796 4776 scope.go:117] "RemoveContainer" containerID="184170080ae095a5b221c8b587a3a4dfa3ecfbae323b183d5873a8c506cbc423" Nov 25 11:17:49 crc kubenswrapper[4776]: I1125 11:17:49.429177 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7lntt"] Nov 25 11:17:49 crc kubenswrapper[4776]: E1125 11:17:49.430161 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91f5209c-ee2d-404f-a7b4-9305ec08f1ac" containerName="registry-server" Nov 25 11:17:49 crc kubenswrapper[4776]: I1125 11:17:49.430172 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="91f5209c-ee2d-404f-a7b4-9305ec08f1ac" containerName="registry-server" Nov 25 11:17:49 crc kubenswrapper[4776]: E1125 11:17:49.430189 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91f5209c-ee2d-404f-a7b4-9305ec08f1ac" containerName="extract-utilities" Nov 25 11:17:49 crc kubenswrapper[4776]: I1125 11:17:49.430197 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="91f5209c-ee2d-404f-a7b4-9305ec08f1ac" containerName="extract-utilities" Nov 25 11:17:49 crc kubenswrapper[4776]: E1125 11:17:49.430233 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91f5209c-ee2d-404f-a7b4-9305ec08f1ac" containerName="extract-content" Nov 25 11:17:49 crc kubenswrapper[4776]: I1125 11:17:49.430246 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="91f5209c-ee2d-404f-a7b4-9305ec08f1ac" containerName="extract-content" Nov 25 11:17:49 crc kubenswrapper[4776]: I1125 11:17:49.430501 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="91f5209c-ee2d-404f-a7b4-9305ec08f1ac" containerName="registry-server" Nov 25 11:17:49 crc kubenswrapper[4776]: I1125 11:17:49.432034 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7lntt" Nov 25 11:17:49 crc kubenswrapper[4776]: I1125 11:17:49.457563 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7lntt"] Nov 25 11:17:49 crc kubenswrapper[4776]: I1125 11:17:49.585755 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12d5fcfc-9017-4a3b-8db1-27e48ccaf330-utilities\") pod \"redhat-operators-7lntt\" (UID: \"12d5fcfc-9017-4a3b-8db1-27e48ccaf330\") " pod="openshift-marketplace/redhat-operators-7lntt" Nov 25 11:17:49 crc kubenswrapper[4776]: I1125 11:17:49.586112 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgh6j\" (UniqueName: \"kubernetes.io/projected/12d5fcfc-9017-4a3b-8db1-27e48ccaf330-kube-api-access-vgh6j\") pod \"redhat-operators-7lntt\" (UID: \"12d5fcfc-9017-4a3b-8db1-27e48ccaf330\") " pod="openshift-marketplace/redhat-operators-7lntt" Nov 25 11:17:49 crc kubenswrapper[4776]: I1125 11:17:49.586968 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12d5fcfc-9017-4a3b-8db1-27e48ccaf330-catalog-content\") pod \"redhat-operators-7lntt\" (UID: \"12d5fcfc-9017-4a3b-8db1-27e48ccaf330\") " pod="openshift-marketplace/redhat-operators-7lntt" Nov 25 11:17:49 crc kubenswrapper[4776]: I1125 11:17:49.689457 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12d5fcfc-9017-4a3b-8db1-27e48ccaf330-utilities\") pod \"redhat-operators-7lntt\" (UID: \"12d5fcfc-9017-4a3b-8db1-27e48ccaf330\") " pod="openshift-marketplace/redhat-operators-7lntt" Nov 25 11:17:49 crc kubenswrapper[4776]: I1125 11:17:49.689551 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgh6j\" (UniqueName: \"kubernetes.io/projected/12d5fcfc-9017-4a3b-8db1-27e48ccaf330-kube-api-access-vgh6j\") pod \"redhat-operators-7lntt\" (UID: \"12d5fcfc-9017-4a3b-8db1-27e48ccaf330\") " pod="openshift-marketplace/redhat-operators-7lntt" Nov 25 11:17:49 crc kubenswrapper[4776]: I1125 11:17:49.689612 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12d5fcfc-9017-4a3b-8db1-27e48ccaf330-catalog-content\") pod \"redhat-operators-7lntt\" (UID: \"12d5fcfc-9017-4a3b-8db1-27e48ccaf330\") " pod="openshift-marketplace/redhat-operators-7lntt" Nov 25 11:17:49 crc kubenswrapper[4776]: I1125 11:17:49.690246 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12d5fcfc-9017-4a3b-8db1-27e48ccaf330-utilities\") pod \"redhat-operators-7lntt\" (UID: \"12d5fcfc-9017-4a3b-8db1-27e48ccaf330\") " pod="openshift-marketplace/redhat-operators-7lntt" Nov 25 11:17:49 crc kubenswrapper[4776]: I1125 11:17:49.690276 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12d5fcfc-9017-4a3b-8db1-27e48ccaf330-catalog-content\") pod \"redhat-operators-7lntt\" (UID: \"12d5fcfc-9017-4a3b-8db1-27e48ccaf330\") " pod="openshift-marketplace/redhat-operators-7lntt" Nov 25 11:17:49 crc kubenswrapper[4776]: I1125 11:17:49.707982 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgh6j\" (UniqueName: \"kubernetes.io/projected/12d5fcfc-9017-4a3b-8db1-27e48ccaf330-kube-api-access-vgh6j\") pod \"redhat-operators-7lntt\" (UID: \"12d5fcfc-9017-4a3b-8db1-27e48ccaf330\") " pod="openshift-marketplace/redhat-operators-7lntt" Nov 25 11:17:49 crc kubenswrapper[4776]: I1125 11:17:49.753711 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7lntt" Nov 25 11:17:50 crc kubenswrapper[4776]: I1125 11:17:50.219713 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7lntt"] Nov 25 11:17:50 crc kubenswrapper[4776]: I1125 11:17:50.662783 4776 scope.go:117] "RemoveContainer" containerID="251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564" Nov 25 11:17:50 crc kubenswrapper[4776]: E1125 11:17:50.663416 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:17:51 crc kubenswrapper[4776]: I1125 11:17:51.133211 4776 generic.go:334] "Generic (PLEG): container finished" podID="12d5fcfc-9017-4a3b-8db1-27e48ccaf330" containerID="a5ec4b5e69b7b1848c93b2c8d23cf47153e43a61e38b38c7c1474201e30c9411" exitCode=0 Nov 25 11:17:51 crc kubenswrapper[4776]: I1125 11:17:51.133314 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7lntt" event={"ID":"12d5fcfc-9017-4a3b-8db1-27e48ccaf330","Type":"ContainerDied","Data":"a5ec4b5e69b7b1848c93b2c8d23cf47153e43a61e38b38c7c1474201e30c9411"} Nov 25 11:17:51 crc kubenswrapper[4776]: I1125 11:17:51.133465 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7lntt" event={"ID":"12d5fcfc-9017-4a3b-8db1-27e48ccaf330","Type":"ContainerStarted","Data":"46b8b1bbcafe7699f2105ef641d2f466164741d2c6c2f245faf0043c00a25be2"} Nov 25 11:17:53 crc kubenswrapper[4776]: I1125 11:17:53.161343 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7lntt" event={"ID":"12d5fcfc-9017-4a3b-8db1-27e48ccaf330","Type":"ContainerStarted","Data":"b3eb77cc0b91c430aa4dbbbbd96267e058cf763786dec3e0a969c51846bf9f91"} Nov 25 11:18:03 crc kubenswrapper[4776]: I1125 11:18:03.921926 4776 generic.go:334] "Generic (PLEG): container finished" podID="12d5fcfc-9017-4a3b-8db1-27e48ccaf330" containerID="b3eb77cc0b91c430aa4dbbbbd96267e058cf763786dec3e0a969c51846bf9f91" exitCode=0 Nov 25 11:18:03 crc kubenswrapper[4776]: I1125 11:18:03.922025 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7lntt" event={"ID":"12d5fcfc-9017-4a3b-8db1-27e48ccaf330","Type":"ContainerDied","Data":"b3eb77cc0b91c430aa4dbbbbd96267e058cf763786dec3e0a969c51846bf9f91"} Nov 25 11:18:04 crc kubenswrapper[4776]: I1125 11:18:04.662617 4776 scope.go:117] "RemoveContainer" containerID="251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564" Nov 25 11:18:04 crc kubenswrapper[4776]: E1125 11:18:04.663112 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:18:04 crc kubenswrapper[4776]: I1125 11:18:04.935934 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7lntt" event={"ID":"12d5fcfc-9017-4a3b-8db1-27e48ccaf330","Type":"ContainerStarted","Data":"3007aacbc6e388fa32f6eb26137d5155066fbb4d8b7c26554bcc8d0d64d2b90d"} Nov 25 11:18:04 crc kubenswrapper[4776]: I1125 11:18:04.970926 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7lntt" podStartSLOduration=2.728461812 podStartE2EDuration="15.970905453s" podCreationTimestamp="2025-11-25 11:17:49 +0000 UTC" firstStartedPulling="2025-11-25 11:17:51.135545126 +0000 UTC m=+6816.176604679" lastFinishedPulling="2025-11-25 11:18:04.377988777 +0000 UTC m=+6829.419048320" observedRunningTime="2025-11-25 11:18:04.967557489 +0000 UTC m=+6830.008617042" watchObservedRunningTime="2025-11-25 11:18:04.970905453 +0000 UTC m=+6830.011965016" Nov 25 11:18:09 crc kubenswrapper[4776]: I1125 11:18:09.755915 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7lntt" Nov 25 11:18:09 crc kubenswrapper[4776]: I1125 11:18:09.756403 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7lntt" Nov 25 11:18:10 crc kubenswrapper[4776]: I1125 11:18:10.809324 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7lntt" podUID="12d5fcfc-9017-4a3b-8db1-27e48ccaf330" containerName="registry-server" probeResult="failure" output=< Nov 25 11:18:10 crc kubenswrapper[4776]: timeout: failed to connect service ":50051" within 1s Nov 25 11:18:10 crc kubenswrapper[4776]: > Nov 25 11:18:18 crc kubenswrapper[4776]: I1125 11:18:18.662794 4776 scope.go:117] "RemoveContainer" containerID="251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564" Nov 25 11:18:18 crc kubenswrapper[4776]: E1125 11:18:18.663853 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:18:19 crc kubenswrapper[4776]: I1125 11:18:19.822270 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7lntt" Nov 25 11:18:19 crc kubenswrapper[4776]: I1125 11:18:19.885358 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7lntt" Nov 25 11:18:20 crc kubenswrapper[4776]: I1125 11:18:20.635253 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7lntt"] Nov 25 11:18:21 crc kubenswrapper[4776]: I1125 11:18:21.115651 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7lntt" podUID="12d5fcfc-9017-4a3b-8db1-27e48ccaf330" containerName="registry-server" containerID="cri-o://3007aacbc6e388fa32f6eb26137d5155066fbb4d8b7c26554bcc8d0d64d2b90d" gracePeriod=2 Nov 25 11:18:22 crc kubenswrapper[4776]: I1125 11:18:22.128144 4776 generic.go:334] "Generic (PLEG): container finished" podID="12d5fcfc-9017-4a3b-8db1-27e48ccaf330" containerID="3007aacbc6e388fa32f6eb26137d5155066fbb4d8b7c26554bcc8d0d64d2b90d" exitCode=0 Nov 25 11:18:22 crc kubenswrapper[4776]: I1125 11:18:22.128200 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7lntt" event={"ID":"12d5fcfc-9017-4a3b-8db1-27e48ccaf330","Type":"ContainerDied","Data":"3007aacbc6e388fa32f6eb26137d5155066fbb4d8b7c26554bcc8d0d64d2b90d"} Nov 25 11:18:22 crc kubenswrapper[4776]: I1125 11:18:22.451710 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7lntt" Nov 25 11:18:22 crc kubenswrapper[4776]: I1125 11:18:22.541850 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12d5fcfc-9017-4a3b-8db1-27e48ccaf330-utilities\") pod \"12d5fcfc-9017-4a3b-8db1-27e48ccaf330\" (UID: \"12d5fcfc-9017-4a3b-8db1-27e48ccaf330\") " Nov 25 11:18:22 crc kubenswrapper[4776]: I1125 11:18:22.541936 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgh6j\" (UniqueName: \"kubernetes.io/projected/12d5fcfc-9017-4a3b-8db1-27e48ccaf330-kube-api-access-vgh6j\") pod \"12d5fcfc-9017-4a3b-8db1-27e48ccaf330\" (UID: \"12d5fcfc-9017-4a3b-8db1-27e48ccaf330\") " Nov 25 11:18:22 crc kubenswrapper[4776]: I1125 11:18:22.542129 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12d5fcfc-9017-4a3b-8db1-27e48ccaf330-catalog-content\") pod \"12d5fcfc-9017-4a3b-8db1-27e48ccaf330\" (UID: \"12d5fcfc-9017-4a3b-8db1-27e48ccaf330\") " Nov 25 11:18:22 crc kubenswrapper[4776]: I1125 11:18:22.542751 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12d5fcfc-9017-4a3b-8db1-27e48ccaf330-utilities" (OuterVolumeSpecName: "utilities") pod "12d5fcfc-9017-4a3b-8db1-27e48ccaf330" (UID: "12d5fcfc-9017-4a3b-8db1-27e48ccaf330"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:18:22 crc kubenswrapper[4776]: I1125 11:18:22.548371 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12d5fcfc-9017-4a3b-8db1-27e48ccaf330-kube-api-access-vgh6j" (OuterVolumeSpecName: "kube-api-access-vgh6j") pod "12d5fcfc-9017-4a3b-8db1-27e48ccaf330" (UID: "12d5fcfc-9017-4a3b-8db1-27e48ccaf330"). InnerVolumeSpecName "kube-api-access-vgh6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:18:22 crc kubenswrapper[4776]: I1125 11:18:22.640663 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12d5fcfc-9017-4a3b-8db1-27e48ccaf330-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "12d5fcfc-9017-4a3b-8db1-27e48ccaf330" (UID: "12d5fcfc-9017-4a3b-8db1-27e48ccaf330"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:18:22 crc kubenswrapper[4776]: I1125 11:18:22.649860 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12d5fcfc-9017-4a3b-8db1-27e48ccaf330-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:18:22 crc kubenswrapper[4776]: I1125 11:18:22.649898 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgh6j\" (UniqueName: \"kubernetes.io/projected/12d5fcfc-9017-4a3b-8db1-27e48ccaf330-kube-api-access-vgh6j\") on node \"crc\" DevicePath \"\"" Nov 25 11:18:22 crc kubenswrapper[4776]: I1125 11:18:22.649911 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12d5fcfc-9017-4a3b-8db1-27e48ccaf330-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:18:23 crc kubenswrapper[4776]: I1125 11:18:23.143578 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7lntt" event={"ID":"12d5fcfc-9017-4a3b-8db1-27e48ccaf330","Type":"ContainerDied","Data":"46b8b1bbcafe7699f2105ef641d2f466164741d2c6c2f245faf0043c00a25be2"} Nov 25 11:18:23 crc kubenswrapper[4776]: I1125 11:18:23.143635 4776 scope.go:117] "RemoveContainer" containerID="3007aacbc6e388fa32f6eb26137d5155066fbb4d8b7c26554bcc8d0d64d2b90d" Nov 25 11:18:23 crc kubenswrapper[4776]: I1125 11:18:23.143679 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7lntt" Nov 25 11:18:23 crc kubenswrapper[4776]: I1125 11:18:23.174522 4776 scope.go:117] "RemoveContainer" containerID="b3eb77cc0b91c430aa4dbbbbd96267e058cf763786dec3e0a969c51846bf9f91" Nov 25 11:18:23 crc kubenswrapper[4776]: I1125 11:18:23.183742 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7lntt"] Nov 25 11:18:23 crc kubenswrapper[4776]: I1125 11:18:23.195466 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7lntt"] Nov 25 11:18:23 crc kubenswrapper[4776]: I1125 11:18:23.205156 4776 scope.go:117] "RemoveContainer" containerID="a5ec4b5e69b7b1848c93b2c8d23cf47153e43a61e38b38c7c1474201e30c9411" Nov 25 11:18:23 crc kubenswrapper[4776]: I1125 11:18:23.691189 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12d5fcfc-9017-4a3b-8db1-27e48ccaf330" path="/var/lib/kubelet/pods/12d5fcfc-9017-4a3b-8db1-27e48ccaf330/volumes" Nov 25 11:18:31 crc kubenswrapper[4776]: I1125 11:18:31.663258 4776 scope.go:117] "RemoveContainer" containerID="251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564" Nov 25 11:18:31 crc kubenswrapper[4776]: E1125 11:18:31.664530 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:18:46 crc kubenswrapper[4776]: I1125 11:18:46.663398 4776 scope.go:117] "RemoveContainer" containerID="251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564" Nov 25 11:18:46 crc kubenswrapper[4776]: E1125 11:18:46.664708 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:19:00 crc kubenswrapper[4776]: I1125 11:19:00.662477 4776 scope.go:117] "RemoveContainer" containerID="251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564" Nov 25 11:19:00 crc kubenswrapper[4776]: E1125 11:19:00.663546 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:19:13 crc kubenswrapper[4776]: I1125 11:19:13.662382 4776 scope.go:117] "RemoveContainer" containerID="251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564" Nov 25 11:19:13 crc kubenswrapper[4776]: E1125 11:19:13.663117 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:19:26 crc kubenswrapper[4776]: I1125 11:19:26.662345 4776 scope.go:117] "RemoveContainer" containerID="251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564" Nov 25 11:19:26 crc kubenswrapper[4776]: E1125 11:19:26.663385 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:19:39 crc kubenswrapper[4776]: I1125 11:19:39.663322 4776 scope.go:117] "RemoveContainer" containerID="251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564" Nov 25 11:19:39 crc kubenswrapper[4776]: E1125 11:19:39.664145 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:19:40 crc kubenswrapper[4776]: I1125 11:19:40.548274 4776 scope.go:117] "RemoveContainer" containerID="fe3a993b7070572081a256a06201ec6d24e70d192df8371258444640c6b0b9d9" Nov 25 11:19:54 crc kubenswrapper[4776]: I1125 11:19:54.663359 4776 scope.go:117] "RemoveContainer" containerID="251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564" Nov 25 11:19:54 crc kubenswrapper[4776]: E1125 11:19:54.664282 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:20:01 crc kubenswrapper[4776]: I1125 11:20:01.052182 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-2jqfn"] Nov 25 11:20:01 crc kubenswrapper[4776]: I1125 11:20:01.061048 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-695d-account-create-fq4l2"] Nov 25 11:20:01 crc kubenswrapper[4776]: I1125 11:20:01.069513 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-2jqfn"] Nov 25 11:20:01 crc kubenswrapper[4776]: I1125 11:20:01.077059 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-695d-account-create-fq4l2"] Nov 25 11:20:01 crc kubenswrapper[4776]: I1125 11:20:01.673165 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="559c0e1c-d5ad-46f7-b9c5-3597858cc244" path="/var/lib/kubelet/pods/559c0e1c-d5ad-46f7-b9c5-3597858cc244/volumes" Nov 25 11:20:01 crc kubenswrapper[4776]: I1125 11:20:01.673825 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b650f074-0767-497f-8a4d-d7e477b759f6" path="/var/lib/kubelet/pods/b650f074-0767-497f-8a4d-d7e477b759f6/volumes" Nov 25 11:20:07 crc kubenswrapper[4776]: I1125 11:20:07.664409 4776 scope.go:117] "RemoveContainer" containerID="251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564" Nov 25 11:20:07 crc kubenswrapper[4776]: E1125 11:20:07.665249 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:20:15 crc kubenswrapper[4776]: I1125 11:20:15.357321 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bhqkb"] Nov 25 11:20:15 crc kubenswrapper[4776]: E1125 11:20:15.358257 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12d5fcfc-9017-4a3b-8db1-27e48ccaf330" containerName="extract-utilities" Nov 25 11:20:15 crc kubenswrapper[4776]: I1125 11:20:15.358269 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="12d5fcfc-9017-4a3b-8db1-27e48ccaf330" containerName="extract-utilities" Nov 25 11:20:15 crc kubenswrapper[4776]: E1125 11:20:15.358287 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12d5fcfc-9017-4a3b-8db1-27e48ccaf330" containerName="registry-server" Nov 25 11:20:15 crc kubenswrapper[4776]: I1125 11:20:15.358294 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="12d5fcfc-9017-4a3b-8db1-27e48ccaf330" containerName="registry-server" Nov 25 11:20:15 crc kubenswrapper[4776]: E1125 11:20:15.358311 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12d5fcfc-9017-4a3b-8db1-27e48ccaf330" containerName="extract-content" Nov 25 11:20:15 crc kubenswrapper[4776]: I1125 11:20:15.358317 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="12d5fcfc-9017-4a3b-8db1-27e48ccaf330" containerName="extract-content" Nov 25 11:20:15 crc kubenswrapper[4776]: I1125 11:20:15.358539 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="12d5fcfc-9017-4a3b-8db1-27e48ccaf330" containerName="registry-server" Nov 25 11:20:15 crc kubenswrapper[4776]: I1125 11:20:15.360162 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bhqkb" Nov 25 11:20:15 crc kubenswrapper[4776]: I1125 11:20:15.373478 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bhqkb"] Nov 25 11:20:15 crc kubenswrapper[4776]: I1125 11:20:15.506299 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3b4203d-ba68-4de4-aa64-87eb4d86be0d-catalog-content\") pod \"redhat-marketplace-bhqkb\" (UID: \"a3b4203d-ba68-4de4-aa64-87eb4d86be0d\") " pod="openshift-marketplace/redhat-marketplace-bhqkb" Nov 25 11:20:15 crc kubenswrapper[4776]: I1125 11:20:15.506738 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtx6f\" (UniqueName: \"kubernetes.io/projected/a3b4203d-ba68-4de4-aa64-87eb4d86be0d-kube-api-access-gtx6f\") pod \"redhat-marketplace-bhqkb\" (UID: \"a3b4203d-ba68-4de4-aa64-87eb4d86be0d\") " pod="openshift-marketplace/redhat-marketplace-bhqkb" Nov 25 11:20:15 crc kubenswrapper[4776]: I1125 11:20:15.506766 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3b4203d-ba68-4de4-aa64-87eb4d86be0d-utilities\") pod \"redhat-marketplace-bhqkb\" (UID: \"a3b4203d-ba68-4de4-aa64-87eb4d86be0d\") " pod="openshift-marketplace/redhat-marketplace-bhqkb" Nov 25 11:20:15 crc kubenswrapper[4776]: I1125 11:20:15.608744 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3b4203d-ba68-4de4-aa64-87eb4d86be0d-catalog-content\") pod \"redhat-marketplace-bhqkb\" (UID: \"a3b4203d-ba68-4de4-aa64-87eb4d86be0d\") " pod="openshift-marketplace/redhat-marketplace-bhqkb" Nov 25 11:20:15 crc kubenswrapper[4776]: I1125 11:20:15.608906 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtx6f\" (UniqueName: \"kubernetes.io/projected/a3b4203d-ba68-4de4-aa64-87eb4d86be0d-kube-api-access-gtx6f\") pod \"redhat-marketplace-bhqkb\" (UID: \"a3b4203d-ba68-4de4-aa64-87eb4d86be0d\") " pod="openshift-marketplace/redhat-marketplace-bhqkb" Nov 25 11:20:15 crc kubenswrapper[4776]: I1125 11:20:15.608930 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3b4203d-ba68-4de4-aa64-87eb4d86be0d-utilities\") pod \"redhat-marketplace-bhqkb\" (UID: \"a3b4203d-ba68-4de4-aa64-87eb4d86be0d\") " pod="openshift-marketplace/redhat-marketplace-bhqkb" Nov 25 11:20:15 crc kubenswrapper[4776]: I1125 11:20:15.609390 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3b4203d-ba68-4de4-aa64-87eb4d86be0d-utilities\") pod \"redhat-marketplace-bhqkb\" (UID: \"a3b4203d-ba68-4de4-aa64-87eb4d86be0d\") " pod="openshift-marketplace/redhat-marketplace-bhqkb" Nov 25 11:20:15 crc kubenswrapper[4776]: I1125 11:20:15.609563 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3b4203d-ba68-4de4-aa64-87eb4d86be0d-catalog-content\") pod \"redhat-marketplace-bhqkb\" (UID: \"a3b4203d-ba68-4de4-aa64-87eb4d86be0d\") " pod="openshift-marketplace/redhat-marketplace-bhqkb" Nov 25 11:20:15 crc kubenswrapper[4776]: I1125 11:20:15.631258 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtx6f\" (UniqueName: \"kubernetes.io/projected/a3b4203d-ba68-4de4-aa64-87eb4d86be0d-kube-api-access-gtx6f\") pod \"redhat-marketplace-bhqkb\" (UID: \"a3b4203d-ba68-4de4-aa64-87eb4d86be0d\") " pod="openshift-marketplace/redhat-marketplace-bhqkb" Nov 25 11:20:15 crc kubenswrapper[4776]: I1125 11:20:15.691807 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bhqkb" Nov 25 11:20:16 crc kubenswrapper[4776]: I1125 11:20:16.278526 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bhqkb"] Nov 25 11:20:16 crc kubenswrapper[4776]: I1125 11:20:16.296487 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bhqkb" event={"ID":"a3b4203d-ba68-4de4-aa64-87eb4d86be0d","Type":"ContainerStarted","Data":"998262973bed06726fedeb252f128b150dab80f2eb1b51c55a751c4c614daff0"} Nov 25 11:20:17 crc kubenswrapper[4776]: I1125 11:20:17.309171 4776 generic.go:334] "Generic (PLEG): container finished" podID="a3b4203d-ba68-4de4-aa64-87eb4d86be0d" containerID="1a266723e7a682fd7fddcb23cc8299c17da148579fb2f0f8aab069b7a77a8728" exitCode=0 Nov 25 11:20:17 crc kubenswrapper[4776]: I1125 11:20:17.309294 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bhqkb" event={"ID":"a3b4203d-ba68-4de4-aa64-87eb4d86be0d","Type":"ContainerDied","Data":"1a266723e7a682fd7fddcb23cc8299c17da148579fb2f0f8aab069b7a77a8728"} Nov 25 11:20:17 crc kubenswrapper[4776]: I1125 11:20:17.312518 4776 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 11:20:19 crc kubenswrapper[4776]: I1125 11:20:19.338347 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bhqkb" event={"ID":"a3b4203d-ba68-4de4-aa64-87eb4d86be0d","Type":"ContainerStarted","Data":"418d1322b8410f621c42aaf2536478fac4292725654722ed1244cd4351641740"} Nov 25 11:20:20 crc kubenswrapper[4776]: I1125 11:20:20.351900 4776 generic.go:334] "Generic (PLEG): container finished" podID="a3b4203d-ba68-4de4-aa64-87eb4d86be0d" containerID="418d1322b8410f621c42aaf2536478fac4292725654722ed1244cd4351641740" exitCode=0 Nov 25 11:20:20 crc kubenswrapper[4776]: I1125 11:20:20.351971 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bhqkb" event={"ID":"a3b4203d-ba68-4de4-aa64-87eb4d86be0d","Type":"ContainerDied","Data":"418d1322b8410f621c42aaf2536478fac4292725654722ed1244cd4351641740"} Nov 25 11:20:21 crc kubenswrapper[4776]: I1125 11:20:21.362278 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bhqkb" event={"ID":"a3b4203d-ba68-4de4-aa64-87eb4d86be0d","Type":"ContainerStarted","Data":"0a9404bfd567d931b9e5989092c2aa9c1a62f8261db040ff8ad486619d6e48a4"} Nov 25 11:20:21 crc kubenswrapper[4776]: I1125 11:20:21.663534 4776 scope.go:117] "RemoveContainer" containerID="251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564" Nov 25 11:20:21 crc kubenswrapper[4776]: E1125 11:20:21.664030 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:20:22 crc kubenswrapper[4776]: I1125 11:20:22.393311 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bhqkb" podStartSLOduration=3.6952281510000002 podStartE2EDuration="7.393291331s" podCreationTimestamp="2025-11-25 11:20:15 +0000 UTC" firstStartedPulling="2025-11-25 11:20:17.31226561 +0000 UTC m=+6962.353325153" lastFinishedPulling="2025-11-25 11:20:21.01032879 +0000 UTC m=+6966.051388333" observedRunningTime="2025-11-25 11:20:22.386711496 +0000 UTC m=+6967.427771049" watchObservedRunningTime="2025-11-25 11:20:22.393291331 +0000 UTC m=+6967.434350884" Nov 25 11:20:25 crc kubenswrapper[4776]: I1125 11:20:25.691910 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bhqkb" Nov 25 11:20:25 crc kubenswrapper[4776]: I1125 11:20:25.692174 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bhqkb" Nov 25 11:20:25 crc kubenswrapper[4776]: I1125 11:20:25.741656 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bhqkb" Nov 25 11:20:26 crc kubenswrapper[4776]: I1125 11:20:26.452644 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bhqkb" Nov 25 11:20:26 crc kubenswrapper[4776]: I1125 11:20:26.505399 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bhqkb"] Nov 25 11:20:28 crc kubenswrapper[4776]: I1125 11:20:28.432021 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bhqkb" podUID="a3b4203d-ba68-4de4-aa64-87eb4d86be0d" containerName="registry-server" containerID="cri-o://0a9404bfd567d931b9e5989092c2aa9c1a62f8261db040ff8ad486619d6e48a4" gracePeriod=2 Nov 25 11:20:29 crc kubenswrapper[4776]: I1125 11:20:29.444731 4776 generic.go:334] "Generic (PLEG): container finished" podID="a3b4203d-ba68-4de4-aa64-87eb4d86be0d" containerID="0a9404bfd567d931b9e5989092c2aa9c1a62f8261db040ff8ad486619d6e48a4" exitCode=0 Nov 25 11:20:29 crc kubenswrapper[4776]: I1125 11:20:29.444814 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bhqkb" event={"ID":"a3b4203d-ba68-4de4-aa64-87eb4d86be0d","Type":"ContainerDied","Data":"0a9404bfd567d931b9e5989092c2aa9c1a62f8261db040ff8ad486619d6e48a4"} Nov 25 11:20:29 crc kubenswrapper[4776]: I1125 11:20:29.551745 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bhqkb" Nov 25 11:20:29 crc kubenswrapper[4776]: I1125 11:20:29.642027 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtx6f\" (UniqueName: \"kubernetes.io/projected/a3b4203d-ba68-4de4-aa64-87eb4d86be0d-kube-api-access-gtx6f\") pod \"a3b4203d-ba68-4de4-aa64-87eb4d86be0d\" (UID: \"a3b4203d-ba68-4de4-aa64-87eb4d86be0d\") " Nov 25 11:20:29 crc kubenswrapper[4776]: I1125 11:20:29.642608 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3b4203d-ba68-4de4-aa64-87eb4d86be0d-utilities\") pod \"a3b4203d-ba68-4de4-aa64-87eb4d86be0d\" (UID: \"a3b4203d-ba68-4de4-aa64-87eb4d86be0d\") " Nov 25 11:20:29 crc kubenswrapper[4776]: I1125 11:20:29.642740 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3b4203d-ba68-4de4-aa64-87eb4d86be0d-catalog-content\") pod \"a3b4203d-ba68-4de4-aa64-87eb4d86be0d\" (UID: \"a3b4203d-ba68-4de4-aa64-87eb4d86be0d\") " Nov 25 11:20:29 crc kubenswrapper[4776]: I1125 11:20:29.643660 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3b4203d-ba68-4de4-aa64-87eb4d86be0d-utilities" (OuterVolumeSpecName: "utilities") pod "a3b4203d-ba68-4de4-aa64-87eb4d86be0d" (UID: "a3b4203d-ba68-4de4-aa64-87eb4d86be0d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:20:29 crc kubenswrapper[4776]: I1125 11:20:29.649636 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3b4203d-ba68-4de4-aa64-87eb4d86be0d-kube-api-access-gtx6f" (OuterVolumeSpecName: "kube-api-access-gtx6f") pod "a3b4203d-ba68-4de4-aa64-87eb4d86be0d" (UID: "a3b4203d-ba68-4de4-aa64-87eb4d86be0d"). InnerVolumeSpecName "kube-api-access-gtx6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:20:29 crc kubenswrapper[4776]: I1125 11:20:29.664293 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3b4203d-ba68-4de4-aa64-87eb4d86be0d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a3b4203d-ba68-4de4-aa64-87eb4d86be0d" (UID: "a3b4203d-ba68-4de4-aa64-87eb4d86be0d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:20:29 crc kubenswrapper[4776]: I1125 11:20:29.746022 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3b4203d-ba68-4de4-aa64-87eb4d86be0d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:20:29 crc kubenswrapper[4776]: I1125 11:20:29.746078 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtx6f\" (UniqueName: \"kubernetes.io/projected/a3b4203d-ba68-4de4-aa64-87eb4d86be0d-kube-api-access-gtx6f\") on node \"crc\" DevicePath \"\"" Nov 25 11:20:29 crc kubenswrapper[4776]: I1125 11:20:29.746094 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3b4203d-ba68-4de4-aa64-87eb4d86be0d-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:20:30 crc kubenswrapper[4776]: I1125 11:20:30.460232 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bhqkb" event={"ID":"a3b4203d-ba68-4de4-aa64-87eb4d86be0d","Type":"ContainerDied","Data":"998262973bed06726fedeb252f128b150dab80f2eb1b51c55a751c4c614daff0"} Nov 25 11:20:30 crc kubenswrapper[4776]: I1125 11:20:30.460296 4776 scope.go:117] "RemoveContainer" containerID="0a9404bfd567d931b9e5989092c2aa9c1a62f8261db040ff8ad486619d6e48a4" Nov 25 11:20:30 crc kubenswrapper[4776]: I1125 11:20:30.460337 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bhqkb" Nov 25 11:20:30 crc kubenswrapper[4776]: I1125 11:20:30.494837 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bhqkb"] Nov 25 11:20:30 crc kubenswrapper[4776]: I1125 11:20:30.504480 4776 scope.go:117] "RemoveContainer" containerID="418d1322b8410f621c42aaf2536478fac4292725654722ed1244cd4351641740" Nov 25 11:20:30 crc kubenswrapper[4776]: I1125 11:20:30.507254 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bhqkb"] Nov 25 11:20:30 crc kubenswrapper[4776]: I1125 11:20:30.526841 4776 scope.go:117] "RemoveContainer" containerID="1a266723e7a682fd7fddcb23cc8299c17da148579fb2f0f8aab069b7a77a8728" Nov 25 11:20:31 crc kubenswrapper[4776]: I1125 11:20:31.672236 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3b4203d-ba68-4de4-aa64-87eb4d86be0d" path="/var/lib/kubelet/pods/a3b4203d-ba68-4de4-aa64-87eb4d86be0d/volumes" Nov 25 11:20:35 crc kubenswrapper[4776]: I1125 11:20:35.679157 4776 scope.go:117] "RemoveContainer" containerID="251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564" Nov 25 11:20:35 crc kubenswrapper[4776]: E1125 11:20:35.679906 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:20:37 crc kubenswrapper[4776]: I1125 11:20:37.040860 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-vw6x4"] Nov 25 11:20:37 crc kubenswrapper[4776]: I1125 11:20:37.049480 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-vw6x4"] Nov 25 11:20:37 crc kubenswrapper[4776]: I1125 11:20:37.676857 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0" path="/var/lib/kubelet/pods/7ed9669a-44b6-4995-89a2-5b4bc6d7f4a0/volumes" Nov 25 11:20:40 crc kubenswrapper[4776]: I1125 11:20:40.604135 4776 scope.go:117] "RemoveContainer" containerID="ece40a72eb7c52a50691108104116e718253051f03f3be7484c2f5aed9864bfb" Nov 25 11:20:40 crc kubenswrapper[4776]: I1125 11:20:40.630159 4776 scope.go:117] "RemoveContainer" containerID="65f396e6b7d2e6b7884409da8eafde41099a0a8ec17fa3d9e48cf2c3303fde83" Nov 25 11:20:40 crc kubenswrapper[4776]: I1125 11:20:40.688887 4776 scope.go:117] "RemoveContainer" containerID="00d95e51686ebffab6c58c4e9ffdf4c917f7567955ab8d9a11b1fc46a5040455" Nov 25 11:20:40 crc kubenswrapper[4776]: I1125 11:20:40.715459 4776 scope.go:117] "RemoveContainer" containerID="bdb22909d55c3808601595581af91a9adf094fd07bfe6c30160ac0a116b2cb0a" Nov 25 11:20:40 crc kubenswrapper[4776]: I1125 11:20:40.738826 4776 scope.go:117] "RemoveContainer" containerID="c9adcfb517f34ad34e65da5b5067c75734c643d40cc5e3419dad41bf3caa54bd" Nov 25 11:20:40 crc kubenswrapper[4776]: I1125 11:20:40.763561 4776 scope.go:117] "RemoveContainer" containerID="c7a14088844565f332ea7726f4c85eed0320545074691122dd0f6f2040e3ed6d" Nov 25 11:20:47 crc kubenswrapper[4776]: I1125 11:20:47.663352 4776 scope.go:117] "RemoveContainer" containerID="251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564" Nov 25 11:20:47 crc kubenswrapper[4776]: E1125 11:20:47.664098 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:21:00 crc kubenswrapper[4776]: I1125 11:21:00.662554 4776 scope.go:117] "RemoveContainer" containerID="251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564" Nov 25 11:21:00 crc kubenswrapper[4776]: E1125 11:21:00.663628 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:21:15 crc kubenswrapper[4776]: I1125 11:21:15.675194 4776 scope.go:117] "RemoveContainer" containerID="251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564" Nov 25 11:21:15 crc kubenswrapper[4776]: E1125 11:21:15.676363 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:21:28 crc kubenswrapper[4776]: I1125 11:21:28.663043 4776 scope.go:117] "RemoveContainer" containerID="251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564" Nov 25 11:21:29 crc kubenswrapper[4776]: I1125 11:21:29.005300 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"147ff0aa7cd858a35205f61aa4a2bf176498cf75de472fc76e2ac38d941d5689"} Nov 25 11:23:16 crc kubenswrapper[4776]: I1125 11:23:16.039696 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-d479-account-create-6qg9g"] Nov 25 11:23:16 crc kubenswrapper[4776]: I1125 11:23:16.049093 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-d479-account-create-6qg9g"] Nov 25 11:23:17 crc kubenswrapper[4776]: I1125 11:23:17.024560 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-jh4vh"] Nov 25 11:23:17 crc kubenswrapper[4776]: I1125 11:23:17.034795 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-jh4vh"] Nov 25 11:23:17 crc kubenswrapper[4776]: I1125 11:23:17.698115 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93752b39-730c-470f-bbd9-2c9d39abbb03" path="/var/lib/kubelet/pods/93752b39-730c-470f-bbd9-2c9d39abbb03/volumes" Nov 25 11:23:17 crc kubenswrapper[4776]: I1125 11:23:17.700586 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2dfc4fe-0a7f-4670-9075-3712e8e9f775" path="/var/lib/kubelet/pods/c2dfc4fe-0a7f-4670-9075-3712e8e9f775/volumes" Nov 25 11:23:33 crc kubenswrapper[4776]: I1125 11:23:33.363956 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-plbg2"] Nov 25 11:23:33 crc kubenswrapper[4776]: E1125 11:23:33.365017 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3b4203d-ba68-4de4-aa64-87eb4d86be0d" containerName="extract-utilities" Nov 25 11:23:33 crc kubenswrapper[4776]: I1125 11:23:33.365037 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3b4203d-ba68-4de4-aa64-87eb4d86be0d" containerName="extract-utilities" Nov 25 11:23:33 crc kubenswrapper[4776]: E1125 11:23:33.365099 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3b4203d-ba68-4de4-aa64-87eb4d86be0d" containerName="extract-content" Nov 25 11:23:33 crc kubenswrapper[4776]: I1125 11:23:33.365110 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3b4203d-ba68-4de4-aa64-87eb4d86be0d" containerName="extract-content" Nov 25 11:23:33 crc kubenswrapper[4776]: E1125 11:23:33.365125 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3b4203d-ba68-4de4-aa64-87eb4d86be0d" containerName="registry-server" Nov 25 11:23:33 crc kubenswrapper[4776]: I1125 11:23:33.365133 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3b4203d-ba68-4de4-aa64-87eb4d86be0d" containerName="registry-server" Nov 25 11:23:33 crc kubenswrapper[4776]: I1125 11:23:33.365427 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3b4203d-ba68-4de4-aa64-87eb4d86be0d" containerName="registry-server" Nov 25 11:23:33 crc kubenswrapper[4776]: I1125 11:23:33.367470 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-plbg2" Nov 25 11:23:33 crc kubenswrapper[4776]: I1125 11:23:33.377149 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-plbg2"] Nov 25 11:23:33 crc kubenswrapper[4776]: I1125 11:23:33.479758 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qxms\" (UniqueName: \"kubernetes.io/projected/d29fc68c-9085-4cb6-900e-92a2ed2b49f1-kube-api-access-7qxms\") pod \"certified-operators-plbg2\" (UID: \"d29fc68c-9085-4cb6-900e-92a2ed2b49f1\") " pod="openshift-marketplace/certified-operators-plbg2" Nov 25 11:23:33 crc kubenswrapper[4776]: I1125 11:23:33.479947 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d29fc68c-9085-4cb6-900e-92a2ed2b49f1-utilities\") pod \"certified-operators-plbg2\" (UID: \"d29fc68c-9085-4cb6-900e-92a2ed2b49f1\") " pod="openshift-marketplace/certified-operators-plbg2" Nov 25 11:23:33 crc kubenswrapper[4776]: I1125 11:23:33.479973 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d29fc68c-9085-4cb6-900e-92a2ed2b49f1-catalog-content\") pod \"certified-operators-plbg2\" (UID: \"d29fc68c-9085-4cb6-900e-92a2ed2b49f1\") " pod="openshift-marketplace/certified-operators-plbg2" Nov 25 11:23:33 crc kubenswrapper[4776]: I1125 11:23:33.582850 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qxms\" (UniqueName: \"kubernetes.io/projected/d29fc68c-9085-4cb6-900e-92a2ed2b49f1-kube-api-access-7qxms\") pod \"certified-operators-plbg2\" (UID: \"d29fc68c-9085-4cb6-900e-92a2ed2b49f1\") " pod="openshift-marketplace/certified-operators-plbg2" Nov 25 11:23:33 crc kubenswrapper[4776]: I1125 11:23:33.583101 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d29fc68c-9085-4cb6-900e-92a2ed2b49f1-utilities\") pod \"certified-operators-plbg2\" (UID: \"d29fc68c-9085-4cb6-900e-92a2ed2b49f1\") " pod="openshift-marketplace/certified-operators-plbg2" Nov 25 11:23:33 crc kubenswrapper[4776]: I1125 11:23:33.583140 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d29fc68c-9085-4cb6-900e-92a2ed2b49f1-catalog-content\") pod \"certified-operators-plbg2\" (UID: \"d29fc68c-9085-4cb6-900e-92a2ed2b49f1\") " pod="openshift-marketplace/certified-operators-plbg2" Nov 25 11:23:33 crc kubenswrapper[4776]: I1125 11:23:33.583813 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d29fc68c-9085-4cb6-900e-92a2ed2b49f1-utilities\") pod \"certified-operators-plbg2\" (UID: \"d29fc68c-9085-4cb6-900e-92a2ed2b49f1\") " pod="openshift-marketplace/certified-operators-plbg2" Nov 25 11:23:33 crc kubenswrapper[4776]: I1125 11:23:33.583894 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d29fc68c-9085-4cb6-900e-92a2ed2b49f1-catalog-content\") pod \"certified-operators-plbg2\" (UID: \"d29fc68c-9085-4cb6-900e-92a2ed2b49f1\") " pod="openshift-marketplace/certified-operators-plbg2" Nov 25 11:23:33 crc kubenswrapper[4776]: I1125 11:23:33.606272 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qxms\" (UniqueName: \"kubernetes.io/projected/d29fc68c-9085-4cb6-900e-92a2ed2b49f1-kube-api-access-7qxms\") pod \"certified-operators-plbg2\" (UID: \"d29fc68c-9085-4cb6-900e-92a2ed2b49f1\") " pod="openshift-marketplace/certified-operators-plbg2" Nov 25 11:23:33 crc kubenswrapper[4776]: I1125 11:23:33.701876 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-plbg2" Nov 25 11:23:34 crc kubenswrapper[4776]: I1125 11:23:34.286880 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-plbg2"] Nov 25 11:23:34 crc kubenswrapper[4776]: I1125 11:23:34.340729 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-plbg2" event={"ID":"d29fc68c-9085-4cb6-900e-92a2ed2b49f1","Type":"ContainerStarted","Data":"f7c92c6e1699e3ef08839e6616a8ed58d30a026e919525be763289a1ff42f65c"} Nov 25 11:23:35 crc kubenswrapper[4776]: I1125 11:23:35.054274 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-zqlhx"] Nov 25 11:23:35 crc kubenswrapper[4776]: I1125 11:23:35.066173 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-zqlhx"] Nov 25 11:23:35 crc kubenswrapper[4776]: I1125 11:23:35.352896 4776 generic.go:334] "Generic (PLEG): container finished" podID="d29fc68c-9085-4cb6-900e-92a2ed2b49f1" containerID="936d9ee225c3f1b9535fe35ec24f3260da2b0ffdf6c7358774a866fcc9b55224" exitCode=0 Nov 25 11:23:35 crc kubenswrapper[4776]: I1125 11:23:35.353004 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-plbg2" event={"ID":"d29fc68c-9085-4cb6-900e-92a2ed2b49f1","Type":"ContainerDied","Data":"936d9ee225c3f1b9535fe35ec24f3260da2b0ffdf6c7358774a866fcc9b55224"} Nov 25 11:23:35 crc kubenswrapper[4776]: I1125 11:23:35.698534 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e759b1f-c2d3-4bf7-8e83-c18086f584b4" path="/var/lib/kubelet/pods/3e759b1f-c2d3-4bf7-8e83-c18086f584b4/volumes" Nov 25 11:23:40 crc kubenswrapper[4776]: I1125 11:23:40.977174 4776 scope.go:117] "RemoveContainer" containerID="55d975b1580e0afc94112b2d087d4f8efa5133c8f8f0c05a104079e8a370abd3" Nov 25 11:23:41 crc kubenswrapper[4776]: I1125 11:23:41.045281 4776 scope.go:117] "RemoveContainer" containerID="d47a1f096eb2f6c24dfa1f5885acd4fa5ce7a5ede02d8f2b3b92c9690e5297e7" Nov 25 11:23:41 crc kubenswrapper[4776]: I1125 11:23:41.192347 4776 scope.go:117] "RemoveContainer" containerID="bcd45529c80fb8ea89ea0f561d2248c2a779e17efa6e7528105e09d9cf4027aa" Nov 25 11:23:41 crc kubenswrapper[4776]: I1125 11:23:41.461609 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-plbg2" event={"ID":"d29fc68c-9085-4cb6-900e-92a2ed2b49f1","Type":"ContainerStarted","Data":"db43c08a7942c8f44c0cc7b79aeefcf1efa66e256ef3f324f34f52bfb8dad766"} Nov 25 11:23:42 crc kubenswrapper[4776]: I1125 11:23:42.472474 4776 generic.go:334] "Generic (PLEG): container finished" podID="d29fc68c-9085-4cb6-900e-92a2ed2b49f1" containerID="db43c08a7942c8f44c0cc7b79aeefcf1efa66e256ef3f324f34f52bfb8dad766" exitCode=0 Nov 25 11:23:42 crc kubenswrapper[4776]: I1125 11:23:42.472543 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-plbg2" event={"ID":"d29fc68c-9085-4cb6-900e-92a2ed2b49f1","Type":"ContainerDied","Data":"db43c08a7942c8f44c0cc7b79aeefcf1efa66e256ef3f324f34f52bfb8dad766"} Nov 25 11:23:44 crc kubenswrapper[4776]: I1125 11:23:44.491973 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-plbg2" event={"ID":"d29fc68c-9085-4cb6-900e-92a2ed2b49f1","Type":"ContainerStarted","Data":"ef3f0e66e3240da4fd8ab63db189412d5907d983dc71496718100d59f72e8040"} Nov 25 11:23:44 crc kubenswrapper[4776]: I1125 11:23:44.519216 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-plbg2" podStartSLOduration=2.657232239 podStartE2EDuration="11.519195467s" podCreationTimestamp="2025-11-25 11:23:33 +0000 UTC" firstStartedPulling="2025-11-25 11:23:35.355638681 +0000 UTC m=+7160.396698234" lastFinishedPulling="2025-11-25 11:23:44.217601889 +0000 UTC m=+7169.258661462" observedRunningTime="2025-11-25 11:23:44.51453902 +0000 UTC m=+7169.555598583" watchObservedRunningTime="2025-11-25 11:23:44.519195467 +0000 UTC m=+7169.560255020" Nov 25 11:23:47 crc kubenswrapper[4776]: I1125 11:23:47.818249 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:23:47 crc kubenswrapper[4776]: I1125 11:23:47.819269 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:23:53 crc kubenswrapper[4776]: I1125 11:23:53.703129 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-plbg2" Nov 25 11:23:53 crc kubenswrapper[4776]: I1125 11:23:53.703743 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-plbg2" Nov 25 11:23:53 crc kubenswrapper[4776]: I1125 11:23:53.768057 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-plbg2" Nov 25 11:23:54 crc kubenswrapper[4776]: I1125 11:23:54.648389 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-plbg2" Nov 25 11:23:54 crc kubenswrapper[4776]: I1125 11:23:54.708165 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-plbg2"] Nov 25 11:23:54 crc kubenswrapper[4776]: I1125 11:23:54.744722 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4mztr"] Nov 25 11:23:54 crc kubenswrapper[4776]: I1125 11:23:54.744975 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4mztr" podUID="3bd6a235-d6fa-4e0a-90fe-61c36fd645d4" containerName="registry-server" containerID="cri-o://9a54ea40141c6492c16647254890d7480212f504211cdd7182de33956d03007f" gracePeriod=2 Nov 25 11:23:55 crc kubenswrapper[4776]: I1125 11:23:55.624514 4776 generic.go:334] "Generic (PLEG): container finished" podID="3bd6a235-d6fa-4e0a-90fe-61c36fd645d4" containerID="9a54ea40141c6492c16647254890d7480212f504211cdd7182de33956d03007f" exitCode=0 Nov 25 11:23:55 crc kubenswrapper[4776]: I1125 11:23:55.624613 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4mztr" event={"ID":"3bd6a235-d6fa-4e0a-90fe-61c36fd645d4","Type":"ContainerDied","Data":"9a54ea40141c6492c16647254890d7480212f504211cdd7182de33956d03007f"} Nov 25 11:23:55 crc kubenswrapper[4776]: I1125 11:23:55.870524 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4mztr" Nov 25 11:23:55 crc kubenswrapper[4776]: I1125 11:23:55.903880 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mpz8\" (UniqueName: \"kubernetes.io/projected/3bd6a235-d6fa-4e0a-90fe-61c36fd645d4-kube-api-access-4mpz8\") pod \"3bd6a235-d6fa-4e0a-90fe-61c36fd645d4\" (UID: \"3bd6a235-d6fa-4e0a-90fe-61c36fd645d4\") " Nov 25 11:23:55 crc kubenswrapper[4776]: I1125 11:23:55.903944 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bd6a235-d6fa-4e0a-90fe-61c36fd645d4-utilities\") pod \"3bd6a235-d6fa-4e0a-90fe-61c36fd645d4\" (UID: \"3bd6a235-d6fa-4e0a-90fe-61c36fd645d4\") " Nov 25 11:23:55 crc kubenswrapper[4776]: I1125 11:23:55.904151 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bd6a235-d6fa-4e0a-90fe-61c36fd645d4-catalog-content\") pod \"3bd6a235-d6fa-4e0a-90fe-61c36fd645d4\" (UID: \"3bd6a235-d6fa-4e0a-90fe-61c36fd645d4\") " Nov 25 11:23:55 crc kubenswrapper[4776]: I1125 11:23:55.925133 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bd6a235-d6fa-4e0a-90fe-61c36fd645d4-kube-api-access-4mpz8" (OuterVolumeSpecName: "kube-api-access-4mpz8") pod "3bd6a235-d6fa-4e0a-90fe-61c36fd645d4" (UID: "3bd6a235-d6fa-4e0a-90fe-61c36fd645d4"). InnerVolumeSpecName "kube-api-access-4mpz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:23:55 crc kubenswrapper[4776]: I1125 11:23:55.935528 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bd6a235-d6fa-4e0a-90fe-61c36fd645d4-utilities" (OuterVolumeSpecName: "utilities") pod "3bd6a235-d6fa-4e0a-90fe-61c36fd645d4" (UID: "3bd6a235-d6fa-4e0a-90fe-61c36fd645d4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:23:56 crc kubenswrapper[4776]: I1125 11:23:56.006162 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mpz8\" (UniqueName: \"kubernetes.io/projected/3bd6a235-d6fa-4e0a-90fe-61c36fd645d4-kube-api-access-4mpz8\") on node \"crc\" DevicePath \"\"" Nov 25 11:23:56 crc kubenswrapper[4776]: I1125 11:23:56.006516 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3bd6a235-d6fa-4e0a-90fe-61c36fd645d4-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:23:56 crc kubenswrapper[4776]: I1125 11:23:56.070211 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bd6a235-d6fa-4e0a-90fe-61c36fd645d4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3bd6a235-d6fa-4e0a-90fe-61c36fd645d4" (UID: "3bd6a235-d6fa-4e0a-90fe-61c36fd645d4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:23:56 crc kubenswrapper[4776]: I1125 11:23:56.108871 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3bd6a235-d6fa-4e0a-90fe-61c36fd645d4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:23:56 crc kubenswrapper[4776]: I1125 11:23:56.639934 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4mztr" event={"ID":"3bd6a235-d6fa-4e0a-90fe-61c36fd645d4","Type":"ContainerDied","Data":"b563edd385fd045a89430f80cb581ced5bf54527abd1f29aec7af00dba156812"} Nov 25 11:23:56 crc kubenswrapper[4776]: I1125 11:23:56.640000 4776 scope.go:117] "RemoveContainer" containerID="9a54ea40141c6492c16647254890d7480212f504211cdd7182de33956d03007f" Nov 25 11:23:56 crc kubenswrapper[4776]: I1125 11:23:56.640005 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4mztr" Nov 25 11:23:56 crc kubenswrapper[4776]: I1125 11:23:56.678046 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4mztr"] Nov 25 11:23:56 crc kubenswrapper[4776]: I1125 11:23:56.678575 4776 scope.go:117] "RemoveContainer" containerID="76c02d64794adfd88244525abe367e225d920ff5adb6f7417c9a1713066a15e9" Nov 25 11:23:56 crc kubenswrapper[4776]: I1125 11:23:56.687785 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4mztr"] Nov 25 11:23:56 crc kubenswrapper[4776]: I1125 11:23:56.724798 4776 scope.go:117] "RemoveContainer" containerID="5686eecf274b670fdc58379788adcae83c5f637e042dc38f99d4d8f3efeafb91" Nov 25 11:23:57 crc kubenswrapper[4776]: I1125 11:23:57.694647 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bd6a235-d6fa-4e0a-90fe-61c36fd645d4" path="/var/lib/kubelet/pods/3bd6a235-d6fa-4e0a-90fe-61c36fd645d4/volumes" Nov 25 11:24:17 crc kubenswrapper[4776]: I1125 11:24:17.818377 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:24:17 crc kubenswrapper[4776]: I1125 11:24:17.819359 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:24:47 crc kubenswrapper[4776]: I1125 11:24:47.818234 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:24:47 crc kubenswrapper[4776]: I1125 11:24:47.818817 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:24:47 crc kubenswrapper[4776]: I1125 11:24:47.818871 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 11:24:47 crc kubenswrapper[4776]: I1125 11:24:47.819695 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"147ff0aa7cd858a35205f61aa4a2bf176498cf75de472fc76e2ac38d941d5689"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 11:24:47 crc kubenswrapper[4776]: I1125 11:24:47.819758 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://147ff0aa7cd858a35205f61aa4a2bf176498cf75de472fc76e2ac38d941d5689" gracePeriod=600 Nov 25 11:24:48 crc kubenswrapper[4776]: I1125 11:24:48.113575 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="147ff0aa7cd858a35205f61aa4a2bf176498cf75de472fc76e2ac38d941d5689" exitCode=0 Nov 25 11:24:48 crc kubenswrapper[4776]: I1125 11:24:48.113654 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"147ff0aa7cd858a35205f61aa4a2bf176498cf75de472fc76e2ac38d941d5689"} Nov 25 11:24:48 crc kubenswrapper[4776]: I1125 11:24:48.114012 4776 scope.go:117] "RemoveContainer" containerID="251a2aba0d6a9fa9760ed0ca963dd00ff3913b8c85170e410b0ab96351541564" Nov 25 11:24:49 crc kubenswrapper[4776]: I1125 11:24:49.142106 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c"} Nov 25 11:26:33 crc kubenswrapper[4776]: I1125 11:26:33.498808 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jdrw2"] Nov 25 11:26:33 crc kubenswrapper[4776]: E1125 11:26:33.500246 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bd6a235-d6fa-4e0a-90fe-61c36fd645d4" containerName="extract-content" Nov 25 11:26:33 crc kubenswrapper[4776]: I1125 11:26:33.500262 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bd6a235-d6fa-4e0a-90fe-61c36fd645d4" containerName="extract-content" Nov 25 11:26:33 crc kubenswrapper[4776]: E1125 11:26:33.500290 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bd6a235-d6fa-4e0a-90fe-61c36fd645d4" containerName="registry-server" Nov 25 11:26:33 crc kubenswrapper[4776]: I1125 11:26:33.500297 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bd6a235-d6fa-4e0a-90fe-61c36fd645d4" containerName="registry-server" Nov 25 11:26:33 crc kubenswrapper[4776]: E1125 11:26:33.500348 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bd6a235-d6fa-4e0a-90fe-61c36fd645d4" containerName="extract-utilities" Nov 25 11:26:33 crc kubenswrapper[4776]: I1125 11:26:33.500356 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bd6a235-d6fa-4e0a-90fe-61c36fd645d4" containerName="extract-utilities" Nov 25 11:26:33 crc kubenswrapper[4776]: I1125 11:26:33.500597 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bd6a235-d6fa-4e0a-90fe-61c36fd645d4" containerName="registry-server" Nov 25 11:26:33 crc kubenswrapper[4776]: I1125 11:26:33.502530 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jdrw2" Nov 25 11:26:33 crc kubenswrapper[4776]: I1125 11:26:33.519317 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jdrw2"] Nov 25 11:26:33 crc kubenswrapper[4776]: I1125 11:26:33.619974 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da4323b6-620c-4e4a-bcec-2cc0687f98bf-utilities\") pod \"community-operators-jdrw2\" (UID: \"da4323b6-620c-4e4a-bcec-2cc0687f98bf\") " pod="openshift-marketplace/community-operators-jdrw2" Nov 25 11:26:33 crc kubenswrapper[4776]: I1125 11:26:33.620125 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhkt5\" (UniqueName: \"kubernetes.io/projected/da4323b6-620c-4e4a-bcec-2cc0687f98bf-kube-api-access-qhkt5\") pod \"community-operators-jdrw2\" (UID: \"da4323b6-620c-4e4a-bcec-2cc0687f98bf\") " pod="openshift-marketplace/community-operators-jdrw2" Nov 25 11:26:33 crc kubenswrapper[4776]: I1125 11:26:33.620210 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da4323b6-620c-4e4a-bcec-2cc0687f98bf-catalog-content\") pod \"community-operators-jdrw2\" (UID: \"da4323b6-620c-4e4a-bcec-2cc0687f98bf\") " pod="openshift-marketplace/community-operators-jdrw2" Nov 25 11:26:33 crc kubenswrapper[4776]: I1125 11:26:33.723953 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhkt5\" (UniqueName: \"kubernetes.io/projected/da4323b6-620c-4e4a-bcec-2cc0687f98bf-kube-api-access-qhkt5\") pod \"community-operators-jdrw2\" (UID: \"da4323b6-620c-4e4a-bcec-2cc0687f98bf\") " pod="openshift-marketplace/community-operators-jdrw2" Nov 25 11:26:33 crc kubenswrapper[4776]: I1125 11:26:33.724162 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da4323b6-620c-4e4a-bcec-2cc0687f98bf-catalog-content\") pod \"community-operators-jdrw2\" (UID: \"da4323b6-620c-4e4a-bcec-2cc0687f98bf\") " pod="openshift-marketplace/community-operators-jdrw2" Nov 25 11:26:33 crc kubenswrapper[4776]: I1125 11:26:33.724352 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da4323b6-620c-4e4a-bcec-2cc0687f98bf-utilities\") pod \"community-operators-jdrw2\" (UID: \"da4323b6-620c-4e4a-bcec-2cc0687f98bf\") " pod="openshift-marketplace/community-operators-jdrw2" Nov 25 11:26:33 crc kubenswrapper[4776]: I1125 11:26:33.727015 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da4323b6-620c-4e4a-bcec-2cc0687f98bf-utilities\") pod \"community-operators-jdrw2\" (UID: \"da4323b6-620c-4e4a-bcec-2cc0687f98bf\") " pod="openshift-marketplace/community-operators-jdrw2" Nov 25 11:26:33 crc kubenswrapper[4776]: I1125 11:26:33.727228 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da4323b6-620c-4e4a-bcec-2cc0687f98bf-catalog-content\") pod \"community-operators-jdrw2\" (UID: \"da4323b6-620c-4e4a-bcec-2cc0687f98bf\") " pod="openshift-marketplace/community-operators-jdrw2" Nov 25 11:26:33 crc kubenswrapper[4776]: I1125 11:26:33.751587 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhkt5\" (UniqueName: \"kubernetes.io/projected/da4323b6-620c-4e4a-bcec-2cc0687f98bf-kube-api-access-qhkt5\") pod \"community-operators-jdrw2\" (UID: \"da4323b6-620c-4e4a-bcec-2cc0687f98bf\") " pod="openshift-marketplace/community-operators-jdrw2" Nov 25 11:26:33 crc kubenswrapper[4776]: I1125 11:26:33.831716 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jdrw2" Nov 25 11:26:34 crc kubenswrapper[4776]: I1125 11:26:34.449534 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jdrw2"] Nov 25 11:26:35 crc kubenswrapper[4776]: I1125 11:26:35.388950 4776 generic.go:334] "Generic (PLEG): container finished" podID="da4323b6-620c-4e4a-bcec-2cc0687f98bf" containerID="f144d521949d0f12492634763abf6ab68737243faade4d4ff39f3d4d3eb1ae01" exitCode=0 Nov 25 11:26:35 crc kubenswrapper[4776]: I1125 11:26:35.389018 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jdrw2" event={"ID":"da4323b6-620c-4e4a-bcec-2cc0687f98bf","Type":"ContainerDied","Data":"f144d521949d0f12492634763abf6ab68737243faade4d4ff39f3d4d3eb1ae01"} Nov 25 11:26:35 crc kubenswrapper[4776]: I1125 11:26:35.389932 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jdrw2" event={"ID":"da4323b6-620c-4e4a-bcec-2cc0687f98bf","Type":"ContainerStarted","Data":"c3345def2d43cc206895f5fc3d9130d9f27cf74d5c9fa1700d8a7a65f0ca19c8"} Nov 25 11:26:35 crc kubenswrapper[4776]: I1125 11:26:35.392029 4776 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 11:26:36 crc kubenswrapper[4776]: I1125 11:26:36.404404 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jdrw2" event={"ID":"da4323b6-620c-4e4a-bcec-2cc0687f98bf","Type":"ContainerStarted","Data":"dd95d7233dda1e11501e10cd30ec0d949663791e1fa5bff4d1fe943350549a99"} Nov 25 11:26:38 crc kubenswrapper[4776]: I1125 11:26:38.431918 4776 generic.go:334] "Generic (PLEG): container finished" podID="da4323b6-620c-4e4a-bcec-2cc0687f98bf" containerID="dd95d7233dda1e11501e10cd30ec0d949663791e1fa5bff4d1fe943350549a99" exitCode=0 Nov 25 11:26:38 crc kubenswrapper[4776]: I1125 11:26:38.432038 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jdrw2" event={"ID":"da4323b6-620c-4e4a-bcec-2cc0687f98bf","Type":"ContainerDied","Data":"dd95d7233dda1e11501e10cd30ec0d949663791e1fa5bff4d1fe943350549a99"} Nov 25 11:26:39 crc kubenswrapper[4776]: I1125 11:26:39.445484 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jdrw2" event={"ID":"da4323b6-620c-4e4a-bcec-2cc0687f98bf","Type":"ContainerStarted","Data":"5932338a1b1d99e4d30c2c53de85de74a70d6c40c41cbf97fc9c92eb6f0b99c0"} Nov 25 11:26:39 crc kubenswrapper[4776]: I1125 11:26:39.475905 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jdrw2" podStartSLOduration=2.966111041 podStartE2EDuration="6.475884196s" podCreationTimestamp="2025-11-25 11:26:33 +0000 UTC" firstStartedPulling="2025-11-25 11:26:35.391512113 +0000 UTC m=+7340.432571706" lastFinishedPulling="2025-11-25 11:26:38.901285308 +0000 UTC m=+7343.942344861" observedRunningTime="2025-11-25 11:26:39.465656029 +0000 UTC m=+7344.506715612" watchObservedRunningTime="2025-11-25 11:26:39.475884196 +0000 UTC m=+7344.516943749" Nov 25 11:26:40 crc kubenswrapper[4776]: I1125 11:26:40.460746 4776 generic.go:334] "Generic (PLEG): container finished" podID="4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8" containerID="12acfde4b5070a9f48760843816de0fe3aabf494f8eeb5c42ec02961dcc8dd30" exitCode=0 Nov 25 11:26:40 crc kubenswrapper[4776]: I1125 11:26:40.460812 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s" event={"ID":"4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8","Type":"ContainerDied","Data":"12acfde4b5070a9f48760843816de0fe3aabf494f8eeb5c42ec02961dcc8dd30"} Nov 25 11:26:41 crc kubenswrapper[4776]: I1125 11:26:41.982485 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s" Nov 25 11:26:42 crc kubenswrapper[4776]: I1125 11:26:42.037129 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8-inventory\") pod \"4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8\" (UID: \"4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8\") " Nov 25 11:26:42 crc kubenswrapper[4776]: I1125 11:26:42.037452 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kls22\" (UniqueName: \"kubernetes.io/projected/4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8-kube-api-access-kls22\") pod \"4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8\" (UID: \"4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8\") " Nov 25 11:26:42 crc kubenswrapper[4776]: I1125 11:26:42.037690 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8-ssh-key\") pod \"4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8\" (UID: \"4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8\") " Nov 25 11:26:42 crc kubenswrapper[4776]: I1125 11:26:42.037821 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8-tripleo-cleanup-combined-ca-bundle\") pod \"4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8\" (UID: \"4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8\") " Nov 25 11:26:42 crc kubenswrapper[4776]: I1125 11:26:42.042995 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8" (UID: "4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:26:42 crc kubenswrapper[4776]: I1125 11:26:42.043123 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8-kube-api-access-kls22" (OuterVolumeSpecName: "kube-api-access-kls22") pod "4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8" (UID: "4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8"). InnerVolumeSpecName "kube-api-access-kls22". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:26:42 crc kubenswrapper[4776]: I1125 11:26:42.068186 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8" (UID: "4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:26:42 crc kubenswrapper[4776]: I1125 11:26:42.075364 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8-inventory" (OuterVolumeSpecName: "inventory") pod "4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8" (UID: "4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:26:42 crc kubenswrapper[4776]: I1125 11:26:42.139909 4776 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:26:42 crc kubenswrapper[4776]: I1125 11:26:42.139946 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kls22\" (UniqueName: \"kubernetes.io/projected/4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8-kube-api-access-kls22\") on node \"crc\" DevicePath \"\"" Nov 25 11:26:42 crc kubenswrapper[4776]: I1125 11:26:42.139956 4776 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:26:42 crc kubenswrapper[4776]: I1125 11:26:42.139965 4776 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:26:42 crc kubenswrapper[4776]: I1125 11:26:42.480691 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s" event={"ID":"4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8","Type":"ContainerDied","Data":"fdd7374ddbc98c24e09ef70d8405488b1d5242071efb1094ec5c66f436a8e725"} Nov 25 11:26:42 crc kubenswrapper[4776]: I1125 11:26:42.480737 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s" Nov 25 11:26:42 crc kubenswrapper[4776]: I1125 11:26:42.480757 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdd7374ddbc98c24e09ef70d8405488b1d5242071efb1094ec5c66f436a8e725" Nov 25 11:26:43 crc kubenswrapper[4776]: I1125 11:26:43.832592 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jdrw2" Nov 25 11:26:43 crc kubenswrapper[4776]: I1125 11:26:43.833227 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jdrw2" Nov 25 11:26:43 crc kubenswrapper[4776]: I1125 11:26:43.891178 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jdrw2" Nov 25 11:26:44 crc kubenswrapper[4776]: I1125 11:26:44.549584 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jdrw2" Nov 25 11:26:44 crc kubenswrapper[4776]: I1125 11:26:44.600352 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jdrw2"] Nov 25 11:26:46 crc kubenswrapper[4776]: I1125 11:26:46.519887 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jdrw2" podUID="da4323b6-620c-4e4a-bcec-2cc0687f98bf" containerName="registry-server" containerID="cri-o://5932338a1b1d99e4d30c2c53de85de74a70d6c40c41cbf97fc9c92eb6f0b99c0" gracePeriod=2 Nov 25 11:26:47 crc kubenswrapper[4776]: I1125 11:26:47.081760 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jdrw2" Nov 25 11:26:47 crc kubenswrapper[4776]: I1125 11:26:47.149136 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhkt5\" (UniqueName: \"kubernetes.io/projected/da4323b6-620c-4e4a-bcec-2cc0687f98bf-kube-api-access-qhkt5\") pod \"da4323b6-620c-4e4a-bcec-2cc0687f98bf\" (UID: \"da4323b6-620c-4e4a-bcec-2cc0687f98bf\") " Nov 25 11:26:47 crc kubenswrapper[4776]: I1125 11:26:47.149218 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da4323b6-620c-4e4a-bcec-2cc0687f98bf-catalog-content\") pod \"da4323b6-620c-4e4a-bcec-2cc0687f98bf\" (UID: \"da4323b6-620c-4e4a-bcec-2cc0687f98bf\") " Nov 25 11:26:47 crc kubenswrapper[4776]: I1125 11:26:47.149254 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da4323b6-620c-4e4a-bcec-2cc0687f98bf-utilities\") pod \"da4323b6-620c-4e4a-bcec-2cc0687f98bf\" (UID: \"da4323b6-620c-4e4a-bcec-2cc0687f98bf\") " Nov 25 11:26:47 crc kubenswrapper[4776]: I1125 11:26:47.150325 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da4323b6-620c-4e4a-bcec-2cc0687f98bf-utilities" (OuterVolumeSpecName: "utilities") pod "da4323b6-620c-4e4a-bcec-2cc0687f98bf" (UID: "da4323b6-620c-4e4a-bcec-2cc0687f98bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:26:47 crc kubenswrapper[4776]: I1125 11:26:47.168430 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da4323b6-620c-4e4a-bcec-2cc0687f98bf-kube-api-access-qhkt5" (OuterVolumeSpecName: "kube-api-access-qhkt5") pod "da4323b6-620c-4e4a-bcec-2cc0687f98bf" (UID: "da4323b6-620c-4e4a-bcec-2cc0687f98bf"). InnerVolumeSpecName "kube-api-access-qhkt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:26:47 crc kubenswrapper[4776]: I1125 11:26:47.215214 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da4323b6-620c-4e4a-bcec-2cc0687f98bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "da4323b6-620c-4e4a-bcec-2cc0687f98bf" (UID: "da4323b6-620c-4e4a-bcec-2cc0687f98bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:26:47 crc kubenswrapper[4776]: I1125 11:26:47.251753 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhkt5\" (UniqueName: \"kubernetes.io/projected/da4323b6-620c-4e4a-bcec-2cc0687f98bf-kube-api-access-qhkt5\") on node \"crc\" DevicePath \"\"" Nov 25 11:26:47 crc kubenswrapper[4776]: I1125 11:26:47.251792 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da4323b6-620c-4e4a-bcec-2cc0687f98bf-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:26:47 crc kubenswrapper[4776]: I1125 11:26:47.251804 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da4323b6-620c-4e4a-bcec-2cc0687f98bf-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:26:47 crc kubenswrapper[4776]: I1125 11:26:47.534165 4776 generic.go:334] "Generic (PLEG): container finished" podID="da4323b6-620c-4e4a-bcec-2cc0687f98bf" containerID="5932338a1b1d99e4d30c2c53de85de74a70d6c40c41cbf97fc9c92eb6f0b99c0" exitCode=0 Nov 25 11:26:47 crc kubenswrapper[4776]: I1125 11:26:47.534264 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jdrw2" event={"ID":"da4323b6-620c-4e4a-bcec-2cc0687f98bf","Type":"ContainerDied","Data":"5932338a1b1d99e4d30c2c53de85de74a70d6c40c41cbf97fc9c92eb6f0b99c0"} Nov 25 11:26:47 crc kubenswrapper[4776]: I1125 11:26:47.534371 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jdrw2" event={"ID":"da4323b6-620c-4e4a-bcec-2cc0687f98bf","Type":"ContainerDied","Data":"c3345def2d43cc206895f5fc3d9130d9f27cf74d5c9fa1700d8a7a65f0ca19c8"} Nov 25 11:26:47 crc kubenswrapper[4776]: I1125 11:26:47.534355 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jdrw2" Nov 25 11:26:47 crc kubenswrapper[4776]: I1125 11:26:47.534402 4776 scope.go:117] "RemoveContainer" containerID="5932338a1b1d99e4d30c2c53de85de74a70d6c40c41cbf97fc9c92eb6f0b99c0" Nov 25 11:26:47 crc kubenswrapper[4776]: I1125 11:26:47.560135 4776 scope.go:117] "RemoveContainer" containerID="dd95d7233dda1e11501e10cd30ec0d949663791e1fa5bff4d1fe943350549a99" Nov 25 11:26:47 crc kubenswrapper[4776]: I1125 11:26:47.576458 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jdrw2"] Nov 25 11:26:47 crc kubenswrapper[4776]: I1125 11:26:47.588927 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jdrw2"] Nov 25 11:26:47 crc kubenswrapper[4776]: I1125 11:26:47.602924 4776 scope.go:117] "RemoveContainer" containerID="f144d521949d0f12492634763abf6ab68737243faade4d4ff39f3d4d3eb1ae01" Nov 25 11:26:47 crc kubenswrapper[4776]: I1125 11:26:47.658835 4776 scope.go:117] "RemoveContainer" containerID="5932338a1b1d99e4d30c2c53de85de74a70d6c40c41cbf97fc9c92eb6f0b99c0" Nov 25 11:26:47 crc kubenswrapper[4776]: E1125 11:26:47.659385 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5932338a1b1d99e4d30c2c53de85de74a70d6c40c41cbf97fc9c92eb6f0b99c0\": container with ID starting with 5932338a1b1d99e4d30c2c53de85de74a70d6c40c41cbf97fc9c92eb6f0b99c0 not found: ID does not exist" containerID="5932338a1b1d99e4d30c2c53de85de74a70d6c40c41cbf97fc9c92eb6f0b99c0" Nov 25 11:26:47 crc kubenswrapper[4776]: I1125 11:26:47.659420 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5932338a1b1d99e4d30c2c53de85de74a70d6c40c41cbf97fc9c92eb6f0b99c0"} err="failed to get container status \"5932338a1b1d99e4d30c2c53de85de74a70d6c40c41cbf97fc9c92eb6f0b99c0\": rpc error: code = NotFound desc = could not find container \"5932338a1b1d99e4d30c2c53de85de74a70d6c40c41cbf97fc9c92eb6f0b99c0\": container with ID starting with 5932338a1b1d99e4d30c2c53de85de74a70d6c40c41cbf97fc9c92eb6f0b99c0 not found: ID does not exist" Nov 25 11:26:47 crc kubenswrapper[4776]: I1125 11:26:47.659444 4776 scope.go:117] "RemoveContainer" containerID="dd95d7233dda1e11501e10cd30ec0d949663791e1fa5bff4d1fe943350549a99" Nov 25 11:26:47 crc kubenswrapper[4776]: E1125 11:26:47.659683 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd95d7233dda1e11501e10cd30ec0d949663791e1fa5bff4d1fe943350549a99\": container with ID starting with dd95d7233dda1e11501e10cd30ec0d949663791e1fa5bff4d1fe943350549a99 not found: ID does not exist" containerID="dd95d7233dda1e11501e10cd30ec0d949663791e1fa5bff4d1fe943350549a99" Nov 25 11:26:47 crc kubenswrapper[4776]: I1125 11:26:47.659706 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd95d7233dda1e11501e10cd30ec0d949663791e1fa5bff4d1fe943350549a99"} err="failed to get container status \"dd95d7233dda1e11501e10cd30ec0d949663791e1fa5bff4d1fe943350549a99\": rpc error: code = NotFound desc = could not find container \"dd95d7233dda1e11501e10cd30ec0d949663791e1fa5bff4d1fe943350549a99\": container with ID starting with dd95d7233dda1e11501e10cd30ec0d949663791e1fa5bff4d1fe943350549a99 not found: ID does not exist" Nov 25 11:26:47 crc kubenswrapper[4776]: I1125 11:26:47.659724 4776 scope.go:117] "RemoveContainer" containerID="f144d521949d0f12492634763abf6ab68737243faade4d4ff39f3d4d3eb1ae01" Nov 25 11:26:47 crc kubenswrapper[4776]: E1125 11:26:47.659990 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f144d521949d0f12492634763abf6ab68737243faade4d4ff39f3d4d3eb1ae01\": container with ID starting with f144d521949d0f12492634763abf6ab68737243faade4d4ff39f3d4d3eb1ae01 not found: ID does not exist" containerID="f144d521949d0f12492634763abf6ab68737243faade4d4ff39f3d4d3eb1ae01" Nov 25 11:26:47 crc kubenswrapper[4776]: I1125 11:26:47.660032 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f144d521949d0f12492634763abf6ab68737243faade4d4ff39f3d4d3eb1ae01"} err="failed to get container status \"f144d521949d0f12492634763abf6ab68737243faade4d4ff39f3d4d3eb1ae01\": rpc error: code = NotFound desc = could not find container \"f144d521949d0f12492634763abf6ab68737243faade4d4ff39f3d4d3eb1ae01\": container with ID starting with f144d521949d0f12492634763abf6ab68737243faade4d4ff39f3d4d3eb1ae01 not found: ID does not exist" Nov 25 11:26:47 crc kubenswrapper[4776]: I1125 11:26:47.717132 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da4323b6-620c-4e4a-bcec-2cc0687f98bf" path="/var/lib/kubelet/pods/da4323b6-620c-4e4a-bcec-2cc0687f98bf/volumes" Nov 25 11:26:49 crc kubenswrapper[4776]: I1125 11:26:49.810183 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-pzvxn"] Nov 25 11:26:49 crc kubenswrapper[4776]: E1125 11:26:49.810897 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Nov 25 11:26:49 crc kubenswrapper[4776]: I1125 11:26:49.810915 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Nov 25 11:26:49 crc kubenswrapper[4776]: E1125 11:26:49.810940 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da4323b6-620c-4e4a-bcec-2cc0687f98bf" containerName="registry-server" Nov 25 11:26:49 crc kubenswrapper[4776]: I1125 11:26:49.810947 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="da4323b6-620c-4e4a-bcec-2cc0687f98bf" containerName="registry-server" Nov 25 11:26:49 crc kubenswrapper[4776]: E1125 11:26:49.810968 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da4323b6-620c-4e4a-bcec-2cc0687f98bf" containerName="extract-content" Nov 25 11:26:49 crc kubenswrapper[4776]: I1125 11:26:49.810976 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="da4323b6-620c-4e4a-bcec-2cc0687f98bf" containerName="extract-content" Nov 25 11:26:49 crc kubenswrapper[4776]: E1125 11:26:49.810997 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da4323b6-620c-4e4a-bcec-2cc0687f98bf" containerName="extract-utilities" Nov 25 11:26:49 crc kubenswrapper[4776]: I1125 11:26:49.811003 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="da4323b6-620c-4e4a-bcec-2cc0687f98bf" containerName="extract-utilities" Nov 25 11:26:49 crc kubenswrapper[4776]: I1125 11:26:49.811215 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Nov 25 11:26:49 crc kubenswrapper[4776]: I1125 11:26:49.811235 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="da4323b6-620c-4e4a-bcec-2cc0687f98bf" containerName="registry-server" Nov 25 11:26:49 crc kubenswrapper[4776]: I1125 11:26:49.811922 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-pzvxn" Nov 25 11:26:49 crc kubenswrapper[4776]: I1125 11:26:49.816541 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 25 11:26:49 crc kubenswrapper[4776]: I1125 11:26:49.816658 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dknh7" Nov 25 11:26:49 crc kubenswrapper[4776]: I1125 11:26:49.816737 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 25 11:26:49 crc kubenswrapper[4776]: I1125 11:26:49.826400 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:26:49 crc kubenswrapper[4776]: I1125 11:26:49.827604 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-pzvxn"] Nov 25 11:26:49 crc kubenswrapper[4776]: I1125 11:26:49.907467 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4n2g\" (UniqueName: \"kubernetes.io/projected/b771a78c-c9b8-4c94-96b3-caa28c5cacc4-kube-api-access-p4n2g\") pod \"bootstrap-openstack-openstack-cell1-pzvxn\" (UID: \"b771a78c-c9b8-4c94-96b3-caa28c5cacc4\") " pod="openstack/bootstrap-openstack-openstack-cell1-pzvxn" Nov 25 11:26:49 crc kubenswrapper[4776]: I1125 11:26:49.907573 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b771a78c-c9b8-4c94-96b3-caa28c5cacc4-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-pzvxn\" (UID: \"b771a78c-c9b8-4c94-96b3-caa28c5cacc4\") " pod="openstack/bootstrap-openstack-openstack-cell1-pzvxn" Nov 25 11:26:49 crc kubenswrapper[4776]: I1125 11:26:49.907644 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b771a78c-c9b8-4c94-96b3-caa28c5cacc4-inventory\") pod \"bootstrap-openstack-openstack-cell1-pzvxn\" (UID: \"b771a78c-c9b8-4c94-96b3-caa28c5cacc4\") " pod="openstack/bootstrap-openstack-openstack-cell1-pzvxn" Nov 25 11:26:49 crc kubenswrapper[4776]: I1125 11:26:49.907713 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b771a78c-c9b8-4c94-96b3-caa28c5cacc4-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-pzvxn\" (UID: \"b771a78c-c9b8-4c94-96b3-caa28c5cacc4\") " pod="openstack/bootstrap-openstack-openstack-cell1-pzvxn" Nov 25 11:26:50 crc kubenswrapper[4776]: I1125 11:26:50.010871 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4n2g\" (UniqueName: \"kubernetes.io/projected/b771a78c-c9b8-4c94-96b3-caa28c5cacc4-kube-api-access-p4n2g\") pod \"bootstrap-openstack-openstack-cell1-pzvxn\" (UID: \"b771a78c-c9b8-4c94-96b3-caa28c5cacc4\") " pod="openstack/bootstrap-openstack-openstack-cell1-pzvxn" Nov 25 11:26:50 crc kubenswrapper[4776]: I1125 11:26:50.011415 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b771a78c-c9b8-4c94-96b3-caa28c5cacc4-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-pzvxn\" (UID: \"b771a78c-c9b8-4c94-96b3-caa28c5cacc4\") " pod="openstack/bootstrap-openstack-openstack-cell1-pzvxn" Nov 25 11:26:50 crc kubenswrapper[4776]: I1125 11:26:50.011472 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b771a78c-c9b8-4c94-96b3-caa28c5cacc4-inventory\") pod \"bootstrap-openstack-openstack-cell1-pzvxn\" (UID: \"b771a78c-c9b8-4c94-96b3-caa28c5cacc4\") " pod="openstack/bootstrap-openstack-openstack-cell1-pzvxn" Nov 25 11:26:50 crc kubenswrapper[4776]: I1125 11:26:50.011505 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b771a78c-c9b8-4c94-96b3-caa28c5cacc4-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-pzvxn\" (UID: \"b771a78c-c9b8-4c94-96b3-caa28c5cacc4\") " pod="openstack/bootstrap-openstack-openstack-cell1-pzvxn" Nov 25 11:26:50 crc kubenswrapper[4776]: I1125 11:26:50.020676 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b771a78c-c9b8-4c94-96b3-caa28c5cacc4-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-pzvxn\" (UID: \"b771a78c-c9b8-4c94-96b3-caa28c5cacc4\") " pod="openstack/bootstrap-openstack-openstack-cell1-pzvxn" Nov 25 11:26:50 crc kubenswrapper[4776]: I1125 11:26:50.021376 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b771a78c-c9b8-4c94-96b3-caa28c5cacc4-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-pzvxn\" (UID: \"b771a78c-c9b8-4c94-96b3-caa28c5cacc4\") " pod="openstack/bootstrap-openstack-openstack-cell1-pzvxn" Nov 25 11:26:50 crc kubenswrapper[4776]: I1125 11:26:50.022301 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b771a78c-c9b8-4c94-96b3-caa28c5cacc4-inventory\") pod \"bootstrap-openstack-openstack-cell1-pzvxn\" (UID: \"b771a78c-c9b8-4c94-96b3-caa28c5cacc4\") " pod="openstack/bootstrap-openstack-openstack-cell1-pzvxn" Nov 25 11:26:50 crc kubenswrapper[4776]: I1125 11:26:50.031694 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4n2g\" (UniqueName: \"kubernetes.io/projected/b771a78c-c9b8-4c94-96b3-caa28c5cacc4-kube-api-access-p4n2g\") pod \"bootstrap-openstack-openstack-cell1-pzvxn\" (UID: \"b771a78c-c9b8-4c94-96b3-caa28c5cacc4\") " pod="openstack/bootstrap-openstack-openstack-cell1-pzvxn" Nov 25 11:26:50 crc kubenswrapper[4776]: I1125 11:26:50.131515 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-pzvxn" Nov 25 11:26:50 crc kubenswrapper[4776]: I1125 11:26:50.727989 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-pzvxn"] Nov 25 11:26:51 crc kubenswrapper[4776]: I1125 11:26:51.572834 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-pzvxn" event={"ID":"b771a78c-c9b8-4c94-96b3-caa28c5cacc4","Type":"ContainerStarted","Data":"8613048b5f07f5801554d3223fa299c4ae56f76face4576ba4d8e2e09763b927"} Nov 25 11:26:52 crc kubenswrapper[4776]: I1125 11:26:52.589733 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-pzvxn" event={"ID":"b771a78c-c9b8-4c94-96b3-caa28c5cacc4","Type":"ContainerStarted","Data":"775718d39bc76e6362142c6103205cd1bac38e298422c399b2136563cd924fe2"} Nov 25 11:26:52 crc kubenswrapper[4776]: I1125 11:26:52.614022 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-cell1-pzvxn" podStartSLOduration=3.046989671 podStartE2EDuration="3.613976637s" podCreationTimestamp="2025-11-25 11:26:49 +0000 UTC" firstStartedPulling="2025-11-25 11:26:50.735389991 +0000 UTC m=+7355.776449544" lastFinishedPulling="2025-11-25 11:26:51.302376957 +0000 UTC m=+7356.343436510" observedRunningTime="2025-11-25 11:26:52.607620807 +0000 UTC m=+7357.648680360" watchObservedRunningTime="2025-11-25 11:26:52.613976637 +0000 UTC m=+7357.655036190" Nov 25 11:27:17 crc kubenswrapper[4776]: I1125 11:27:17.818472 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:27:17 crc kubenswrapper[4776]: I1125 11:27:17.819088 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:27:47 crc kubenswrapper[4776]: I1125 11:27:47.819044 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:27:47 crc kubenswrapper[4776]: I1125 11:27:47.819708 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:28:17 crc kubenswrapper[4776]: I1125 11:28:17.818128 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:28:17 crc kubenswrapper[4776]: I1125 11:28:17.818766 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:28:17 crc kubenswrapper[4776]: I1125 11:28:17.818826 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 11:28:17 crc kubenswrapper[4776]: I1125 11:28:17.819782 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 11:28:17 crc kubenswrapper[4776]: I1125 11:28:17.819860 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c" gracePeriod=600 Nov 25 11:28:17 crc kubenswrapper[4776]: E1125 11:28:17.944444 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:28:18 crc kubenswrapper[4776]: I1125 11:28:18.777039 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c" exitCode=0 Nov 25 11:28:18 crc kubenswrapper[4776]: I1125 11:28:18.777111 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c"} Nov 25 11:28:18 crc kubenswrapper[4776]: I1125 11:28:18.777199 4776 scope.go:117] "RemoveContainer" containerID="147ff0aa7cd858a35205f61aa4a2bf176498cf75de472fc76e2ac38d941d5689" Nov 25 11:28:18 crc kubenswrapper[4776]: I1125 11:28:18.777916 4776 scope.go:117] "RemoveContainer" containerID="1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c" Nov 25 11:28:18 crc kubenswrapper[4776]: E1125 11:28:18.778314 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:28:23 crc kubenswrapper[4776]: I1125 11:28:23.015018 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cm77x"] Nov 25 11:28:23 crc kubenswrapper[4776]: I1125 11:28:23.018016 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cm77x" Nov 25 11:28:23 crc kubenswrapper[4776]: I1125 11:28:23.025881 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cm77x"] Nov 25 11:28:23 crc kubenswrapper[4776]: I1125 11:28:23.124013 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/940d99ff-924c-4a98-93fc-00e862718c09-utilities\") pod \"redhat-operators-cm77x\" (UID: \"940d99ff-924c-4a98-93fc-00e862718c09\") " pod="openshift-marketplace/redhat-operators-cm77x" Nov 25 11:28:23 crc kubenswrapper[4776]: I1125 11:28:23.124928 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gw4l\" (UniqueName: \"kubernetes.io/projected/940d99ff-924c-4a98-93fc-00e862718c09-kube-api-access-4gw4l\") pod \"redhat-operators-cm77x\" (UID: \"940d99ff-924c-4a98-93fc-00e862718c09\") " pod="openshift-marketplace/redhat-operators-cm77x" Nov 25 11:28:23 crc kubenswrapper[4776]: I1125 11:28:23.125035 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/940d99ff-924c-4a98-93fc-00e862718c09-catalog-content\") pod \"redhat-operators-cm77x\" (UID: \"940d99ff-924c-4a98-93fc-00e862718c09\") " pod="openshift-marketplace/redhat-operators-cm77x" Nov 25 11:28:23 crc kubenswrapper[4776]: I1125 11:28:23.228759 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gw4l\" (UniqueName: \"kubernetes.io/projected/940d99ff-924c-4a98-93fc-00e862718c09-kube-api-access-4gw4l\") pod \"redhat-operators-cm77x\" (UID: \"940d99ff-924c-4a98-93fc-00e862718c09\") " pod="openshift-marketplace/redhat-operators-cm77x" Nov 25 11:28:23 crc kubenswrapper[4776]: I1125 11:28:23.228870 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/940d99ff-924c-4a98-93fc-00e862718c09-catalog-content\") pod \"redhat-operators-cm77x\" (UID: \"940d99ff-924c-4a98-93fc-00e862718c09\") " pod="openshift-marketplace/redhat-operators-cm77x" Nov 25 11:28:23 crc kubenswrapper[4776]: I1125 11:28:23.229115 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/940d99ff-924c-4a98-93fc-00e862718c09-utilities\") pod \"redhat-operators-cm77x\" (UID: \"940d99ff-924c-4a98-93fc-00e862718c09\") " pod="openshift-marketplace/redhat-operators-cm77x" Nov 25 11:28:23 crc kubenswrapper[4776]: I1125 11:28:23.230105 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/940d99ff-924c-4a98-93fc-00e862718c09-utilities\") pod \"redhat-operators-cm77x\" (UID: \"940d99ff-924c-4a98-93fc-00e862718c09\") " pod="openshift-marketplace/redhat-operators-cm77x" Nov 25 11:28:23 crc kubenswrapper[4776]: I1125 11:28:23.230223 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/940d99ff-924c-4a98-93fc-00e862718c09-catalog-content\") pod \"redhat-operators-cm77x\" (UID: \"940d99ff-924c-4a98-93fc-00e862718c09\") " pod="openshift-marketplace/redhat-operators-cm77x" Nov 25 11:28:23 crc kubenswrapper[4776]: I1125 11:28:23.249687 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gw4l\" (UniqueName: \"kubernetes.io/projected/940d99ff-924c-4a98-93fc-00e862718c09-kube-api-access-4gw4l\") pod \"redhat-operators-cm77x\" (UID: \"940d99ff-924c-4a98-93fc-00e862718c09\") " pod="openshift-marketplace/redhat-operators-cm77x" Nov 25 11:28:23 crc kubenswrapper[4776]: I1125 11:28:23.351932 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cm77x" Nov 25 11:28:23 crc kubenswrapper[4776]: I1125 11:28:23.841138 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cm77x"] Nov 25 11:28:24 crc kubenswrapper[4776]: I1125 11:28:24.845062 4776 generic.go:334] "Generic (PLEG): container finished" podID="940d99ff-924c-4a98-93fc-00e862718c09" containerID="5af1f4ef92f8150c53411b709fb515a981f06ae613a5d0fe1fbcb27a01b639c5" exitCode=0 Nov 25 11:28:24 crc kubenswrapper[4776]: I1125 11:28:24.845156 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cm77x" event={"ID":"940d99ff-924c-4a98-93fc-00e862718c09","Type":"ContainerDied","Data":"5af1f4ef92f8150c53411b709fb515a981f06ae613a5d0fe1fbcb27a01b639c5"} Nov 25 11:28:24 crc kubenswrapper[4776]: I1125 11:28:24.845404 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cm77x" event={"ID":"940d99ff-924c-4a98-93fc-00e862718c09","Type":"ContainerStarted","Data":"b5753c075c49779cf9120e3fa62c608e62f1fd5188243c89e6fb071ac7ec2747"} Nov 25 11:28:27 crc kubenswrapper[4776]: I1125 11:28:27.882152 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cm77x" event={"ID":"940d99ff-924c-4a98-93fc-00e862718c09","Type":"ContainerStarted","Data":"a54b1d4e864c9f85b01a23be4442500c9101729c42eca5cd756a37c10edd6ed4"} Nov 25 11:28:30 crc kubenswrapper[4776]: I1125 11:28:30.663231 4776 scope.go:117] "RemoveContainer" containerID="1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c" Nov 25 11:28:30 crc kubenswrapper[4776]: E1125 11:28:30.663926 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:28:34 crc kubenswrapper[4776]: I1125 11:28:34.942515 4776 generic.go:334] "Generic (PLEG): container finished" podID="940d99ff-924c-4a98-93fc-00e862718c09" containerID="a54b1d4e864c9f85b01a23be4442500c9101729c42eca5cd756a37c10edd6ed4" exitCode=0 Nov 25 11:28:34 crc kubenswrapper[4776]: I1125 11:28:34.942656 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cm77x" event={"ID":"940d99ff-924c-4a98-93fc-00e862718c09","Type":"ContainerDied","Data":"a54b1d4e864c9f85b01a23be4442500c9101729c42eca5cd756a37c10edd6ed4"} Nov 25 11:28:35 crc kubenswrapper[4776]: I1125 11:28:35.953140 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cm77x" event={"ID":"940d99ff-924c-4a98-93fc-00e862718c09","Type":"ContainerStarted","Data":"81435967d29ea22171548fdfb660ac3cada120e831ce070b6339d444d3ff1d6e"} Nov 25 11:28:35 crc kubenswrapper[4776]: I1125 11:28:35.978174 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cm77x" podStartSLOduration=3.322603987 podStartE2EDuration="13.978156159s" podCreationTimestamp="2025-11-25 11:28:22 +0000 UTC" firstStartedPulling="2025-11-25 11:28:24.846922882 +0000 UTC m=+7449.887982435" lastFinishedPulling="2025-11-25 11:28:35.502475054 +0000 UTC m=+7460.543534607" observedRunningTime="2025-11-25 11:28:35.971043321 +0000 UTC m=+7461.012102884" watchObservedRunningTime="2025-11-25 11:28:35.978156159 +0000 UTC m=+7461.019215712" Nov 25 11:28:43 crc kubenswrapper[4776]: I1125 11:28:43.352987 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cm77x" Nov 25 11:28:43 crc kubenswrapper[4776]: I1125 11:28:43.353747 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cm77x" Nov 25 11:28:43 crc kubenswrapper[4776]: I1125 11:28:43.662391 4776 scope.go:117] "RemoveContainer" containerID="1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c" Nov 25 11:28:43 crc kubenswrapper[4776]: E1125 11:28:43.662684 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:28:44 crc kubenswrapper[4776]: I1125 11:28:44.427265 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cm77x" podUID="940d99ff-924c-4a98-93fc-00e862718c09" containerName="registry-server" probeResult="failure" output=< Nov 25 11:28:44 crc kubenswrapper[4776]: timeout: failed to connect service ":50051" within 1s Nov 25 11:28:44 crc kubenswrapper[4776]: > Nov 25 11:28:53 crc kubenswrapper[4776]: I1125 11:28:53.406423 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cm77x" Nov 25 11:28:53 crc kubenswrapper[4776]: I1125 11:28:53.458229 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cm77x" Nov 25 11:28:54 crc kubenswrapper[4776]: I1125 11:28:54.214779 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cm77x"] Nov 25 11:28:54 crc kubenswrapper[4776]: I1125 11:28:54.663345 4776 scope.go:117] "RemoveContainer" containerID="1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c" Nov 25 11:28:54 crc kubenswrapper[4776]: E1125 11:28:54.663685 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:28:55 crc kubenswrapper[4776]: I1125 11:28:55.136272 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cm77x" podUID="940d99ff-924c-4a98-93fc-00e862718c09" containerName="registry-server" containerID="cri-o://81435967d29ea22171548fdfb660ac3cada120e831ce070b6339d444d3ff1d6e" gracePeriod=2 Nov 25 11:28:55 crc kubenswrapper[4776]: I1125 11:28:55.630282 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cm77x" Nov 25 11:28:55 crc kubenswrapper[4776]: I1125 11:28:55.751987 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/940d99ff-924c-4a98-93fc-00e862718c09-utilities\") pod \"940d99ff-924c-4a98-93fc-00e862718c09\" (UID: \"940d99ff-924c-4a98-93fc-00e862718c09\") " Nov 25 11:28:55 crc kubenswrapper[4776]: I1125 11:28:55.752199 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/940d99ff-924c-4a98-93fc-00e862718c09-catalog-content\") pod \"940d99ff-924c-4a98-93fc-00e862718c09\" (UID: \"940d99ff-924c-4a98-93fc-00e862718c09\") " Nov 25 11:28:55 crc kubenswrapper[4776]: I1125 11:28:55.752421 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gw4l\" (UniqueName: \"kubernetes.io/projected/940d99ff-924c-4a98-93fc-00e862718c09-kube-api-access-4gw4l\") pod \"940d99ff-924c-4a98-93fc-00e862718c09\" (UID: \"940d99ff-924c-4a98-93fc-00e862718c09\") " Nov 25 11:28:55 crc kubenswrapper[4776]: I1125 11:28:55.753804 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/940d99ff-924c-4a98-93fc-00e862718c09-utilities" (OuterVolumeSpecName: "utilities") pod "940d99ff-924c-4a98-93fc-00e862718c09" (UID: "940d99ff-924c-4a98-93fc-00e862718c09"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:28:55 crc kubenswrapper[4776]: I1125 11:28:55.761360 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/940d99ff-924c-4a98-93fc-00e862718c09-kube-api-access-4gw4l" (OuterVolumeSpecName: "kube-api-access-4gw4l") pod "940d99ff-924c-4a98-93fc-00e862718c09" (UID: "940d99ff-924c-4a98-93fc-00e862718c09"). InnerVolumeSpecName "kube-api-access-4gw4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:28:55 crc kubenswrapper[4776]: I1125 11:28:55.855652 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gw4l\" (UniqueName: \"kubernetes.io/projected/940d99ff-924c-4a98-93fc-00e862718c09-kube-api-access-4gw4l\") on node \"crc\" DevicePath \"\"" Nov 25 11:28:55 crc kubenswrapper[4776]: I1125 11:28:55.855704 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/940d99ff-924c-4a98-93fc-00e862718c09-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:28:55 crc kubenswrapper[4776]: I1125 11:28:55.856015 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/940d99ff-924c-4a98-93fc-00e862718c09-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "940d99ff-924c-4a98-93fc-00e862718c09" (UID: "940d99ff-924c-4a98-93fc-00e862718c09"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:28:55 crc kubenswrapper[4776]: I1125 11:28:55.958353 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/940d99ff-924c-4a98-93fc-00e862718c09-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:28:56 crc kubenswrapper[4776]: I1125 11:28:56.152118 4776 generic.go:334] "Generic (PLEG): container finished" podID="940d99ff-924c-4a98-93fc-00e862718c09" containerID="81435967d29ea22171548fdfb660ac3cada120e831ce070b6339d444d3ff1d6e" exitCode=0 Nov 25 11:28:56 crc kubenswrapper[4776]: I1125 11:28:56.152174 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cm77x" event={"ID":"940d99ff-924c-4a98-93fc-00e862718c09","Type":"ContainerDied","Data":"81435967d29ea22171548fdfb660ac3cada120e831ce070b6339d444d3ff1d6e"} Nov 25 11:28:56 crc kubenswrapper[4776]: I1125 11:28:56.152213 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cm77x" event={"ID":"940d99ff-924c-4a98-93fc-00e862718c09","Type":"ContainerDied","Data":"b5753c075c49779cf9120e3fa62c608e62f1fd5188243c89e6fb071ac7ec2747"} Nov 25 11:28:56 crc kubenswrapper[4776]: I1125 11:28:56.152234 4776 scope.go:117] "RemoveContainer" containerID="81435967d29ea22171548fdfb660ac3cada120e831ce070b6339d444d3ff1d6e" Nov 25 11:28:56 crc kubenswrapper[4776]: I1125 11:28:56.152262 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cm77x" Nov 25 11:28:56 crc kubenswrapper[4776]: I1125 11:28:56.181355 4776 scope.go:117] "RemoveContainer" containerID="a54b1d4e864c9f85b01a23be4442500c9101729c42eca5cd756a37c10edd6ed4" Nov 25 11:28:56 crc kubenswrapper[4776]: I1125 11:28:56.188546 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cm77x"] Nov 25 11:28:56 crc kubenswrapper[4776]: I1125 11:28:56.201214 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cm77x"] Nov 25 11:28:56 crc kubenswrapper[4776]: I1125 11:28:56.211663 4776 scope.go:117] "RemoveContainer" containerID="5af1f4ef92f8150c53411b709fb515a981f06ae613a5d0fe1fbcb27a01b639c5" Nov 25 11:28:56 crc kubenswrapper[4776]: I1125 11:28:56.275329 4776 scope.go:117] "RemoveContainer" containerID="81435967d29ea22171548fdfb660ac3cada120e831ce070b6339d444d3ff1d6e" Nov 25 11:28:56 crc kubenswrapper[4776]: E1125 11:28:56.276025 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81435967d29ea22171548fdfb660ac3cada120e831ce070b6339d444d3ff1d6e\": container with ID starting with 81435967d29ea22171548fdfb660ac3cada120e831ce070b6339d444d3ff1d6e not found: ID does not exist" containerID="81435967d29ea22171548fdfb660ac3cada120e831ce070b6339d444d3ff1d6e" Nov 25 11:28:56 crc kubenswrapper[4776]: I1125 11:28:56.276077 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81435967d29ea22171548fdfb660ac3cada120e831ce070b6339d444d3ff1d6e"} err="failed to get container status \"81435967d29ea22171548fdfb660ac3cada120e831ce070b6339d444d3ff1d6e\": rpc error: code = NotFound desc = could not find container \"81435967d29ea22171548fdfb660ac3cada120e831ce070b6339d444d3ff1d6e\": container with ID starting with 81435967d29ea22171548fdfb660ac3cada120e831ce070b6339d444d3ff1d6e not found: ID does not exist" Nov 25 11:28:56 crc kubenswrapper[4776]: I1125 11:28:56.276116 4776 scope.go:117] "RemoveContainer" containerID="a54b1d4e864c9f85b01a23be4442500c9101729c42eca5cd756a37c10edd6ed4" Nov 25 11:28:56 crc kubenswrapper[4776]: E1125 11:28:56.276895 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a54b1d4e864c9f85b01a23be4442500c9101729c42eca5cd756a37c10edd6ed4\": container with ID starting with a54b1d4e864c9f85b01a23be4442500c9101729c42eca5cd756a37c10edd6ed4 not found: ID does not exist" containerID="a54b1d4e864c9f85b01a23be4442500c9101729c42eca5cd756a37c10edd6ed4" Nov 25 11:28:56 crc kubenswrapper[4776]: I1125 11:28:56.276940 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a54b1d4e864c9f85b01a23be4442500c9101729c42eca5cd756a37c10edd6ed4"} err="failed to get container status \"a54b1d4e864c9f85b01a23be4442500c9101729c42eca5cd756a37c10edd6ed4\": rpc error: code = NotFound desc = could not find container \"a54b1d4e864c9f85b01a23be4442500c9101729c42eca5cd756a37c10edd6ed4\": container with ID starting with a54b1d4e864c9f85b01a23be4442500c9101729c42eca5cd756a37c10edd6ed4 not found: ID does not exist" Nov 25 11:28:56 crc kubenswrapper[4776]: I1125 11:28:56.276970 4776 scope.go:117] "RemoveContainer" containerID="5af1f4ef92f8150c53411b709fb515a981f06ae613a5d0fe1fbcb27a01b639c5" Nov 25 11:28:56 crc kubenswrapper[4776]: E1125 11:28:56.277533 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5af1f4ef92f8150c53411b709fb515a981f06ae613a5d0fe1fbcb27a01b639c5\": container with ID starting with 5af1f4ef92f8150c53411b709fb515a981f06ae613a5d0fe1fbcb27a01b639c5 not found: ID does not exist" containerID="5af1f4ef92f8150c53411b709fb515a981f06ae613a5d0fe1fbcb27a01b639c5" Nov 25 11:28:56 crc kubenswrapper[4776]: I1125 11:28:56.277596 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5af1f4ef92f8150c53411b709fb515a981f06ae613a5d0fe1fbcb27a01b639c5"} err="failed to get container status \"5af1f4ef92f8150c53411b709fb515a981f06ae613a5d0fe1fbcb27a01b639c5\": rpc error: code = NotFound desc = could not find container \"5af1f4ef92f8150c53411b709fb515a981f06ae613a5d0fe1fbcb27a01b639c5\": container with ID starting with 5af1f4ef92f8150c53411b709fb515a981f06ae613a5d0fe1fbcb27a01b639c5 not found: ID does not exist" Nov 25 11:28:57 crc kubenswrapper[4776]: I1125 11:28:57.682206 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="940d99ff-924c-4a98-93fc-00e862718c09" path="/var/lib/kubelet/pods/940d99ff-924c-4a98-93fc-00e862718c09/volumes" Nov 25 11:29:09 crc kubenswrapper[4776]: I1125 11:29:09.662896 4776 scope.go:117] "RemoveContainer" containerID="1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c" Nov 25 11:29:09 crc kubenswrapper[4776]: E1125 11:29:09.663774 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:29:21 crc kubenswrapper[4776]: I1125 11:29:21.663389 4776 scope.go:117] "RemoveContainer" containerID="1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c" Nov 25 11:29:21 crc kubenswrapper[4776]: E1125 11:29:21.664360 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:29:35 crc kubenswrapper[4776]: I1125 11:29:35.670666 4776 scope.go:117] "RemoveContainer" containerID="1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c" Nov 25 11:29:35 crc kubenswrapper[4776]: E1125 11:29:35.671814 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:29:48 crc kubenswrapper[4776]: I1125 11:29:48.662148 4776 scope.go:117] "RemoveContainer" containerID="1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c" Nov 25 11:29:48 crc kubenswrapper[4776]: E1125 11:29:48.662959 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:29:57 crc kubenswrapper[4776]: I1125 11:29:57.744083 4776 generic.go:334] "Generic (PLEG): container finished" podID="b771a78c-c9b8-4c94-96b3-caa28c5cacc4" containerID="775718d39bc76e6362142c6103205cd1bac38e298422c399b2136563cd924fe2" exitCode=0 Nov 25 11:29:57 crc kubenswrapper[4776]: I1125 11:29:57.744190 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-pzvxn" event={"ID":"b771a78c-c9b8-4c94-96b3-caa28c5cacc4","Type":"ContainerDied","Data":"775718d39bc76e6362142c6103205cd1bac38e298422c399b2136563cd924fe2"} Nov 25 11:29:59 crc kubenswrapper[4776]: I1125 11:29:59.279054 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-pzvxn" Nov 25 11:29:59 crc kubenswrapper[4776]: I1125 11:29:59.447593 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b771a78c-c9b8-4c94-96b3-caa28c5cacc4-inventory\") pod \"b771a78c-c9b8-4c94-96b3-caa28c5cacc4\" (UID: \"b771a78c-c9b8-4c94-96b3-caa28c5cacc4\") " Nov 25 11:29:59 crc kubenswrapper[4776]: I1125 11:29:59.447919 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b771a78c-c9b8-4c94-96b3-caa28c5cacc4-ssh-key\") pod \"b771a78c-c9b8-4c94-96b3-caa28c5cacc4\" (UID: \"b771a78c-c9b8-4c94-96b3-caa28c5cacc4\") " Nov 25 11:29:59 crc kubenswrapper[4776]: I1125 11:29:59.448045 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4n2g\" (UniqueName: \"kubernetes.io/projected/b771a78c-c9b8-4c94-96b3-caa28c5cacc4-kube-api-access-p4n2g\") pod \"b771a78c-c9b8-4c94-96b3-caa28c5cacc4\" (UID: \"b771a78c-c9b8-4c94-96b3-caa28c5cacc4\") " Nov 25 11:29:59 crc kubenswrapper[4776]: I1125 11:29:59.448153 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b771a78c-c9b8-4c94-96b3-caa28c5cacc4-bootstrap-combined-ca-bundle\") pod \"b771a78c-c9b8-4c94-96b3-caa28c5cacc4\" (UID: \"b771a78c-c9b8-4c94-96b3-caa28c5cacc4\") " Nov 25 11:29:59 crc kubenswrapper[4776]: I1125 11:29:59.453312 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b771a78c-c9b8-4c94-96b3-caa28c5cacc4-kube-api-access-p4n2g" (OuterVolumeSpecName: "kube-api-access-p4n2g") pod "b771a78c-c9b8-4c94-96b3-caa28c5cacc4" (UID: "b771a78c-c9b8-4c94-96b3-caa28c5cacc4"). InnerVolumeSpecName "kube-api-access-p4n2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:29:59 crc kubenswrapper[4776]: I1125 11:29:59.457720 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b771a78c-c9b8-4c94-96b3-caa28c5cacc4-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "b771a78c-c9b8-4c94-96b3-caa28c5cacc4" (UID: "b771a78c-c9b8-4c94-96b3-caa28c5cacc4"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:29:59 crc kubenswrapper[4776]: I1125 11:29:59.476580 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b771a78c-c9b8-4c94-96b3-caa28c5cacc4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b771a78c-c9b8-4c94-96b3-caa28c5cacc4" (UID: "b771a78c-c9b8-4c94-96b3-caa28c5cacc4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:29:59 crc kubenswrapper[4776]: I1125 11:29:59.482548 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b771a78c-c9b8-4c94-96b3-caa28c5cacc4-inventory" (OuterVolumeSpecName: "inventory") pod "b771a78c-c9b8-4c94-96b3-caa28c5cacc4" (UID: "b771a78c-c9b8-4c94-96b3-caa28c5cacc4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:29:59 crc kubenswrapper[4776]: I1125 11:29:59.552663 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4n2g\" (UniqueName: \"kubernetes.io/projected/b771a78c-c9b8-4c94-96b3-caa28c5cacc4-kube-api-access-p4n2g\") on node \"crc\" DevicePath \"\"" Nov 25 11:29:59 crc kubenswrapper[4776]: I1125 11:29:59.552932 4776 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b771a78c-c9b8-4c94-96b3-caa28c5cacc4-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:29:59 crc kubenswrapper[4776]: I1125 11:29:59.552971 4776 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b771a78c-c9b8-4c94-96b3-caa28c5cacc4-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:29:59 crc kubenswrapper[4776]: I1125 11:29:59.552983 4776 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b771a78c-c9b8-4c94-96b3-caa28c5cacc4-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:29:59 crc kubenswrapper[4776]: I1125 11:29:59.767913 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-pzvxn" event={"ID":"b771a78c-c9b8-4c94-96b3-caa28c5cacc4","Type":"ContainerDied","Data":"8613048b5f07f5801554d3223fa299c4ae56f76face4576ba4d8e2e09763b927"} Nov 25 11:29:59 crc kubenswrapper[4776]: I1125 11:29:59.768207 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8613048b5f07f5801554d3223fa299c4ae56f76face4576ba4d8e2e09763b927" Nov 25 11:29:59 crc kubenswrapper[4776]: I1125 11:29:59.767961 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-pzvxn" Nov 25 11:29:59 crc kubenswrapper[4776]: I1125 11:29:59.854021 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-nhxsr"] Nov 25 11:29:59 crc kubenswrapper[4776]: E1125 11:29:59.854536 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="940d99ff-924c-4a98-93fc-00e862718c09" containerName="extract-utilities" Nov 25 11:29:59 crc kubenswrapper[4776]: I1125 11:29:59.854561 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="940d99ff-924c-4a98-93fc-00e862718c09" containerName="extract-utilities" Nov 25 11:29:59 crc kubenswrapper[4776]: E1125 11:29:59.854580 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="940d99ff-924c-4a98-93fc-00e862718c09" containerName="registry-server" Nov 25 11:29:59 crc kubenswrapper[4776]: I1125 11:29:59.854586 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="940d99ff-924c-4a98-93fc-00e862718c09" containerName="registry-server" Nov 25 11:29:59 crc kubenswrapper[4776]: E1125 11:29:59.854608 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b771a78c-c9b8-4c94-96b3-caa28c5cacc4" containerName="bootstrap-openstack-openstack-cell1" Nov 25 11:29:59 crc kubenswrapper[4776]: I1125 11:29:59.854615 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="b771a78c-c9b8-4c94-96b3-caa28c5cacc4" containerName="bootstrap-openstack-openstack-cell1" Nov 25 11:29:59 crc kubenswrapper[4776]: E1125 11:29:59.854645 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="940d99ff-924c-4a98-93fc-00e862718c09" containerName="extract-content" Nov 25 11:29:59 crc kubenswrapper[4776]: I1125 11:29:59.854650 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="940d99ff-924c-4a98-93fc-00e862718c09" containerName="extract-content" Nov 25 11:29:59 crc kubenswrapper[4776]: I1125 11:29:59.854893 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="940d99ff-924c-4a98-93fc-00e862718c09" containerName="registry-server" Nov 25 11:29:59 crc kubenswrapper[4776]: I1125 11:29:59.854936 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="b771a78c-c9b8-4c94-96b3-caa28c5cacc4" containerName="bootstrap-openstack-openstack-cell1" Nov 25 11:29:59 crc kubenswrapper[4776]: I1125 11:29:59.855783 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-nhxsr" Nov 25 11:29:59 crc kubenswrapper[4776]: I1125 11:29:59.858307 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dknh7" Nov 25 11:29:59 crc kubenswrapper[4776]: I1125 11:29:59.858489 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:29:59 crc kubenswrapper[4776]: I1125 11:29:59.858845 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 25 11:29:59 crc kubenswrapper[4776]: I1125 11:29:59.858995 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 25 11:30:00 crc kubenswrapper[4776]: I1125 11:29:59.867627 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-nhxsr"] Nov 25 11:30:00 crc kubenswrapper[4776]: I1125 11:29:59.961400 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9eccbe5b-80cd-4633-829c-0f79f228bf48-ssh-key\") pod \"download-cache-openstack-openstack-cell1-nhxsr\" (UID: \"9eccbe5b-80cd-4633-829c-0f79f228bf48\") " pod="openstack/download-cache-openstack-openstack-cell1-nhxsr" Nov 25 11:30:00 crc kubenswrapper[4776]: I1125 11:29:59.961602 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pb7bz\" (UniqueName: \"kubernetes.io/projected/9eccbe5b-80cd-4633-829c-0f79f228bf48-kube-api-access-pb7bz\") pod \"download-cache-openstack-openstack-cell1-nhxsr\" (UID: \"9eccbe5b-80cd-4633-829c-0f79f228bf48\") " pod="openstack/download-cache-openstack-openstack-cell1-nhxsr" Nov 25 11:30:00 crc kubenswrapper[4776]: I1125 11:29:59.961657 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9eccbe5b-80cd-4633-829c-0f79f228bf48-inventory\") pod \"download-cache-openstack-openstack-cell1-nhxsr\" (UID: \"9eccbe5b-80cd-4633-829c-0f79f228bf48\") " pod="openstack/download-cache-openstack-openstack-cell1-nhxsr" Nov 25 11:30:00 crc kubenswrapper[4776]: I1125 11:30:00.063575 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9eccbe5b-80cd-4633-829c-0f79f228bf48-ssh-key\") pod \"download-cache-openstack-openstack-cell1-nhxsr\" (UID: \"9eccbe5b-80cd-4633-829c-0f79f228bf48\") " pod="openstack/download-cache-openstack-openstack-cell1-nhxsr" Nov 25 11:30:00 crc kubenswrapper[4776]: I1125 11:30:00.063681 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pb7bz\" (UniqueName: \"kubernetes.io/projected/9eccbe5b-80cd-4633-829c-0f79f228bf48-kube-api-access-pb7bz\") pod \"download-cache-openstack-openstack-cell1-nhxsr\" (UID: \"9eccbe5b-80cd-4633-829c-0f79f228bf48\") " pod="openstack/download-cache-openstack-openstack-cell1-nhxsr" Nov 25 11:30:00 crc kubenswrapper[4776]: I1125 11:30:00.063721 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9eccbe5b-80cd-4633-829c-0f79f228bf48-inventory\") pod \"download-cache-openstack-openstack-cell1-nhxsr\" (UID: \"9eccbe5b-80cd-4633-829c-0f79f228bf48\") " pod="openstack/download-cache-openstack-openstack-cell1-nhxsr" Nov 25 11:30:00 crc kubenswrapper[4776]: I1125 11:30:00.067270 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9eccbe5b-80cd-4633-829c-0f79f228bf48-inventory\") pod \"download-cache-openstack-openstack-cell1-nhxsr\" (UID: \"9eccbe5b-80cd-4633-829c-0f79f228bf48\") " pod="openstack/download-cache-openstack-openstack-cell1-nhxsr" Nov 25 11:30:00 crc kubenswrapper[4776]: I1125 11:30:00.067943 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9eccbe5b-80cd-4633-829c-0f79f228bf48-ssh-key\") pod \"download-cache-openstack-openstack-cell1-nhxsr\" (UID: \"9eccbe5b-80cd-4633-829c-0f79f228bf48\") " pod="openstack/download-cache-openstack-openstack-cell1-nhxsr" Nov 25 11:30:00 crc kubenswrapper[4776]: I1125 11:30:00.083787 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pb7bz\" (UniqueName: \"kubernetes.io/projected/9eccbe5b-80cd-4633-829c-0f79f228bf48-kube-api-access-pb7bz\") pod \"download-cache-openstack-openstack-cell1-nhxsr\" (UID: \"9eccbe5b-80cd-4633-829c-0f79f228bf48\") " pod="openstack/download-cache-openstack-openstack-cell1-nhxsr" Nov 25 11:30:00 crc kubenswrapper[4776]: I1125 11:30:00.136505 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401170-t572v"] Nov 25 11:30:00 crc kubenswrapper[4776]: I1125 11:30:00.138502 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-t572v" Nov 25 11:30:00 crc kubenswrapper[4776]: I1125 11:30:00.141771 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 11:30:00 crc kubenswrapper[4776]: I1125 11:30:00.142023 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 11:30:00 crc kubenswrapper[4776]: I1125 11:30:00.147468 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401170-t572v"] Nov 25 11:30:00 crc kubenswrapper[4776]: I1125 11:30:00.185664 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-nhxsr" Nov 25 11:30:00 crc kubenswrapper[4776]: I1125 11:30:00.274605 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe-config-volume\") pod \"collect-profiles-29401170-t572v\" (UID: \"a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-t572v" Nov 25 11:30:00 crc kubenswrapper[4776]: I1125 11:30:00.274873 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vd55z\" (UniqueName: \"kubernetes.io/projected/a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe-kube-api-access-vd55z\") pod \"collect-profiles-29401170-t572v\" (UID: \"a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-t572v" Nov 25 11:30:00 crc kubenswrapper[4776]: I1125 11:30:00.274932 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe-secret-volume\") pod \"collect-profiles-29401170-t572v\" (UID: \"a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-t572v" Nov 25 11:30:00 crc kubenswrapper[4776]: I1125 11:30:00.377564 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe-config-volume\") pod \"collect-profiles-29401170-t572v\" (UID: \"a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-t572v" Nov 25 11:30:00 crc kubenswrapper[4776]: I1125 11:30:00.378015 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vd55z\" (UniqueName: \"kubernetes.io/projected/a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe-kube-api-access-vd55z\") pod \"collect-profiles-29401170-t572v\" (UID: \"a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-t572v" Nov 25 11:30:00 crc kubenswrapper[4776]: I1125 11:30:00.378099 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe-secret-volume\") pod \"collect-profiles-29401170-t572v\" (UID: \"a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-t572v" Nov 25 11:30:00 crc kubenswrapper[4776]: I1125 11:30:00.378650 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe-config-volume\") pod \"collect-profiles-29401170-t572v\" (UID: \"a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-t572v" Nov 25 11:30:00 crc kubenswrapper[4776]: I1125 11:30:00.382781 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe-secret-volume\") pod \"collect-profiles-29401170-t572v\" (UID: \"a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-t572v" Nov 25 11:30:00 crc kubenswrapper[4776]: I1125 11:30:00.397731 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vd55z\" (UniqueName: \"kubernetes.io/projected/a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe-kube-api-access-vd55z\") pod \"collect-profiles-29401170-t572v\" (UID: \"a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-t572v" Nov 25 11:30:00 crc kubenswrapper[4776]: I1125 11:30:00.612272 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-t572v" Nov 25 11:30:00 crc kubenswrapper[4776]: I1125 11:30:00.790540 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-nhxsr"] Nov 25 11:30:00 crc kubenswrapper[4776]: W1125 11:30:00.795656 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9eccbe5b_80cd_4633_829c_0f79f228bf48.slice/crio-95b92d82dc77c0166a24847174ad728a7819d2b4756e2d24b9eb5d8908e7055a WatchSource:0}: Error finding container 95b92d82dc77c0166a24847174ad728a7819d2b4756e2d24b9eb5d8908e7055a: Status 404 returned error can't find the container with id 95b92d82dc77c0166a24847174ad728a7819d2b4756e2d24b9eb5d8908e7055a Nov 25 11:30:01 crc kubenswrapper[4776]: I1125 11:30:01.061209 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401170-t572v"] Nov 25 11:30:01 crc kubenswrapper[4776]: W1125 11:30:01.063850 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda38ad9f5_2ea1_4da3_93ce_d1c0033a3fbe.slice/crio-afd041feebe3d8da353d8d1bed1596839eff5a6c82be88e497acbacc36abaef5 WatchSource:0}: Error finding container afd041feebe3d8da353d8d1bed1596839eff5a6c82be88e497acbacc36abaef5: Status 404 returned error can't find the container with id afd041feebe3d8da353d8d1bed1596839eff5a6c82be88e497acbacc36abaef5 Nov 25 11:30:01 crc kubenswrapper[4776]: I1125 11:30:01.662502 4776 scope.go:117] "RemoveContainer" containerID="1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c" Nov 25 11:30:01 crc kubenswrapper[4776]: E1125 11:30:01.663137 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:30:01 crc kubenswrapper[4776]: I1125 11:30:01.793356 4776 generic.go:334] "Generic (PLEG): container finished" podID="a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe" containerID="7b381a199dd0fbf4c2df793c6a00aa1f2b429bf2974992eb737c590f24e9c3be" exitCode=0 Nov 25 11:30:01 crc kubenswrapper[4776]: I1125 11:30:01.793405 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-t572v" event={"ID":"a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe","Type":"ContainerDied","Data":"7b381a199dd0fbf4c2df793c6a00aa1f2b429bf2974992eb737c590f24e9c3be"} Nov 25 11:30:01 crc kubenswrapper[4776]: I1125 11:30:01.793445 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-t572v" event={"ID":"a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe","Type":"ContainerStarted","Data":"afd041feebe3d8da353d8d1bed1596839eff5a6c82be88e497acbacc36abaef5"} Nov 25 11:30:01 crc kubenswrapper[4776]: I1125 11:30:01.796005 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-nhxsr" event={"ID":"9eccbe5b-80cd-4633-829c-0f79f228bf48","Type":"ContainerStarted","Data":"e68f8d4efa09d0d226e07429f8ed7f104d2334624333ddb1a4e744b910a978ca"} Nov 25 11:30:01 crc kubenswrapper[4776]: I1125 11:30:01.796092 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-nhxsr" event={"ID":"9eccbe5b-80cd-4633-829c-0f79f228bf48","Type":"ContainerStarted","Data":"95b92d82dc77c0166a24847174ad728a7819d2b4756e2d24b9eb5d8908e7055a"} Nov 25 11:30:01 crc kubenswrapper[4776]: I1125 11:30:01.825863 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-cell1-nhxsr" podStartSLOduration=2.38028937 podStartE2EDuration="2.825846229s" podCreationTimestamp="2025-11-25 11:29:59 +0000 UTC" firstStartedPulling="2025-11-25 11:30:00.799405734 +0000 UTC m=+7545.840465287" lastFinishedPulling="2025-11-25 11:30:01.244962593 +0000 UTC m=+7546.286022146" observedRunningTime="2025-11-25 11:30:01.822746211 +0000 UTC m=+7546.863805764" watchObservedRunningTime="2025-11-25 11:30:01.825846229 +0000 UTC m=+7546.866905772" Nov 25 11:30:03 crc kubenswrapper[4776]: I1125 11:30:03.170236 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-t572v" Nov 25 11:30:03 crc kubenswrapper[4776]: I1125 11:30:03.349424 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe-config-volume\") pod \"a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe\" (UID: \"a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe\") " Nov 25 11:30:03 crc kubenswrapper[4776]: I1125 11:30:03.349587 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vd55z\" (UniqueName: \"kubernetes.io/projected/a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe-kube-api-access-vd55z\") pod \"a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe\" (UID: \"a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe\") " Nov 25 11:30:03 crc kubenswrapper[4776]: I1125 11:30:03.349727 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe-secret-volume\") pod \"a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe\" (UID: \"a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe\") " Nov 25 11:30:03 crc kubenswrapper[4776]: I1125 11:30:03.350734 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe-config-volume" (OuterVolumeSpecName: "config-volume") pod "a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe" (UID: "a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:30:03 crc kubenswrapper[4776]: I1125 11:30:03.354850 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe-kube-api-access-vd55z" (OuterVolumeSpecName: "kube-api-access-vd55z") pod "a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe" (UID: "a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe"). InnerVolumeSpecName "kube-api-access-vd55z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:30:03 crc kubenswrapper[4776]: I1125 11:30:03.355178 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe" (UID: "a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:30:03 crc kubenswrapper[4776]: I1125 11:30:03.452474 4776 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 11:30:03 crc kubenswrapper[4776]: I1125 11:30:03.452519 4776 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 11:30:03 crc kubenswrapper[4776]: I1125 11:30:03.452531 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vd55z\" (UniqueName: \"kubernetes.io/projected/a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe-kube-api-access-vd55z\") on node \"crc\" DevicePath \"\"" Nov 25 11:30:03 crc kubenswrapper[4776]: I1125 11:30:03.818090 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-t572v" event={"ID":"a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe","Type":"ContainerDied","Data":"afd041feebe3d8da353d8d1bed1596839eff5a6c82be88e497acbacc36abaef5"} Nov 25 11:30:03 crc kubenswrapper[4776]: I1125 11:30:03.818143 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afd041feebe3d8da353d8d1bed1596839eff5a6c82be88e497acbacc36abaef5" Nov 25 11:30:03 crc kubenswrapper[4776]: I1125 11:30:03.818171 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401170-t572v" Nov 25 11:30:04 crc kubenswrapper[4776]: I1125 11:30:04.256092 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401125-hhgzx"] Nov 25 11:30:04 crc kubenswrapper[4776]: I1125 11:30:04.267781 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401125-hhgzx"] Nov 25 11:30:05 crc kubenswrapper[4776]: I1125 11:30:05.677057 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d6f9eeb-c801-495d-97e5-7c83d46e754a" path="/var/lib/kubelet/pods/4d6f9eeb-c801-495d-97e5-7c83d46e754a/volumes" Nov 25 11:30:15 crc kubenswrapper[4776]: I1125 11:30:15.669494 4776 scope.go:117] "RemoveContainer" containerID="1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c" Nov 25 11:30:15 crc kubenswrapper[4776]: E1125 11:30:15.670369 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:30:29 crc kubenswrapper[4776]: I1125 11:30:29.663013 4776 scope.go:117] "RemoveContainer" containerID="1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c" Nov 25 11:30:29 crc kubenswrapper[4776]: E1125 11:30:29.663857 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:30:41 crc kubenswrapper[4776]: I1125 11:30:41.462763 4776 scope.go:117] "RemoveContainer" containerID="ef36b1748f5b0da835396b431acf96950ff611d61b980b26afd69b0e39ccff7a" Nov 25 11:30:43 crc kubenswrapper[4776]: I1125 11:30:43.663746 4776 scope.go:117] "RemoveContainer" containerID="1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c" Nov 25 11:30:43 crc kubenswrapper[4776]: E1125 11:30:43.664359 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:30:56 crc kubenswrapper[4776]: I1125 11:30:56.662846 4776 scope.go:117] "RemoveContainer" containerID="1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c" Nov 25 11:30:56 crc kubenswrapper[4776]: E1125 11:30:56.663670 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:31:10 crc kubenswrapper[4776]: I1125 11:31:10.663631 4776 scope.go:117] "RemoveContainer" containerID="1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c" Nov 25 11:31:10 crc kubenswrapper[4776]: E1125 11:31:10.664471 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:31:24 crc kubenswrapper[4776]: I1125 11:31:24.663850 4776 scope.go:117] "RemoveContainer" containerID="1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c" Nov 25 11:31:24 crc kubenswrapper[4776]: E1125 11:31:24.666183 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:31:26 crc kubenswrapper[4776]: I1125 11:31:26.581024 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-28hck"] Nov 25 11:31:26 crc kubenswrapper[4776]: E1125 11:31:26.581949 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe" containerName="collect-profiles" Nov 25 11:31:26 crc kubenswrapper[4776]: I1125 11:31:26.581964 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe" containerName="collect-profiles" Nov 25 11:31:26 crc kubenswrapper[4776]: I1125 11:31:26.582216 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe" containerName="collect-profiles" Nov 25 11:31:26 crc kubenswrapper[4776]: I1125 11:31:26.584050 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-28hck" Nov 25 11:31:26 crc kubenswrapper[4776]: I1125 11:31:26.607172 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-28hck"] Nov 25 11:31:26 crc kubenswrapper[4776]: I1125 11:31:26.707636 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d6c7357-e7ce-437e-98a5-eeea39389273-utilities\") pod \"redhat-marketplace-28hck\" (UID: \"8d6c7357-e7ce-437e-98a5-eeea39389273\") " pod="openshift-marketplace/redhat-marketplace-28hck" Nov 25 11:31:26 crc kubenswrapper[4776]: I1125 11:31:26.708243 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d6c7357-e7ce-437e-98a5-eeea39389273-catalog-content\") pod \"redhat-marketplace-28hck\" (UID: \"8d6c7357-e7ce-437e-98a5-eeea39389273\") " pod="openshift-marketplace/redhat-marketplace-28hck" Nov 25 11:31:26 crc kubenswrapper[4776]: I1125 11:31:26.708365 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8ppx\" (UniqueName: \"kubernetes.io/projected/8d6c7357-e7ce-437e-98a5-eeea39389273-kube-api-access-n8ppx\") pod \"redhat-marketplace-28hck\" (UID: \"8d6c7357-e7ce-437e-98a5-eeea39389273\") " pod="openshift-marketplace/redhat-marketplace-28hck" Nov 25 11:31:26 crc kubenswrapper[4776]: I1125 11:31:26.810980 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d6c7357-e7ce-437e-98a5-eeea39389273-utilities\") pod \"redhat-marketplace-28hck\" (UID: \"8d6c7357-e7ce-437e-98a5-eeea39389273\") " pod="openshift-marketplace/redhat-marketplace-28hck" Nov 25 11:31:26 crc kubenswrapper[4776]: I1125 11:31:26.811417 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d6c7357-e7ce-437e-98a5-eeea39389273-utilities\") pod \"redhat-marketplace-28hck\" (UID: \"8d6c7357-e7ce-437e-98a5-eeea39389273\") " pod="openshift-marketplace/redhat-marketplace-28hck" Nov 25 11:31:26 crc kubenswrapper[4776]: I1125 11:31:26.811876 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d6c7357-e7ce-437e-98a5-eeea39389273-catalog-content\") pod \"redhat-marketplace-28hck\" (UID: \"8d6c7357-e7ce-437e-98a5-eeea39389273\") " pod="openshift-marketplace/redhat-marketplace-28hck" Nov 25 11:31:26 crc kubenswrapper[4776]: I1125 11:31:26.812019 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8ppx\" (UniqueName: \"kubernetes.io/projected/8d6c7357-e7ce-437e-98a5-eeea39389273-kube-api-access-n8ppx\") pod \"redhat-marketplace-28hck\" (UID: \"8d6c7357-e7ce-437e-98a5-eeea39389273\") " pod="openshift-marketplace/redhat-marketplace-28hck" Nov 25 11:31:26 crc kubenswrapper[4776]: I1125 11:31:26.812537 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d6c7357-e7ce-437e-98a5-eeea39389273-catalog-content\") pod \"redhat-marketplace-28hck\" (UID: \"8d6c7357-e7ce-437e-98a5-eeea39389273\") " pod="openshift-marketplace/redhat-marketplace-28hck" Nov 25 11:31:26 crc kubenswrapper[4776]: I1125 11:31:26.832980 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8ppx\" (UniqueName: \"kubernetes.io/projected/8d6c7357-e7ce-437e-98a5-eeea39389273-kube-api-access-n8ppx\") pod \"redhat-marketplace-28hck\" (UID: \"8d6c7357-e7ce-437e-98a5-eeea39389273\") " pod="openshift-marketplace/redhat-marketplace-28hck" Nov 25 11:31:26 crc kubenswrapper[4776]: I1125 11:31:26.924542 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-28hck" Nov 25 11:31:27 crc kubenswrapper[4776]: I1125 11:31:27.383717 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-28hck"] Nov 25 11:31:27 crc kubenswrapper[4776]: I1125 11:31:27.687836 4776 generic.go:334] "Generic (PLEG): container finished" podID="8d6c7357-e7ce-437e-98a5-eeea39389273" containerID="fc69f99d410394b64ac2a4045a0841e1e1a299dddf91b78a2287a2c5a926cb5f" exitCode=0 Nov 25 11:31:27 crc kubenswrapper[4776]: I1125 11:31:27.688001 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-28hck" event={"ID":"8d6c7357-e7ce-437e-98a5-eeea39389273","Type":"ContainerDied","Data":"fc69f99d410394b64ac2a4045a0841e1e1a299dddf91b78a2287a2c5a926cb5f"} Nov 25 11:31:27 crc kubenswrapper[4776]: I1125 11:31:27.689056 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-28hck" event={"ID":"8d6c7357-e7ce-437e-98a5-eeea39389273","Type":"ContainerStarted","Data":"332bae89292c564308e07e939fd3976d6aaff6b2a8c3f8030433932cbf2d3276"} Nov 25 11:31:28 crc kubenswrapper[4776]: I1125 11:31:28.702951 4776 generic.go:334] "Generic (PLEG): container finished" podID="9eccbe5b-80cd-4633-829c-0f79f228bf48" containerID="e68f8d4efa09d0d226e07429f8ed7f104d2334624333ddb1a4e744b910a978ca" exitCode=0 Nov 25 11:31:28 crc kubenswrapper[4776]: I1125 11:31:28.703036 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-nhxsr" event={"ID":"9eccbe5b-80cd-4633-829c-0f79f228bf48","Type":"ContainerDied","Data":"e68f8d4efa09d0d226e07429f8ed7f104d2334624333ddb1a4e744b910a978ca"} Nov 25 11:31:29 crc kubenswrapper[4776]: I1125 11:31:29.714238 4776 generic.go:334] "Generic (PLEG): container finished" podID="8d6c7357-e7ce-437e-98a5-eeea39389273" containerID="3d95376268539d86f09b788a17a70a0291db7addb7ac64321d169453ca3a2d01" exitCode=0 Nov 25 11:31:29 crc kubenswrapper[4776]: I1125 11:31:29.714294 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-28hck" event={"ID":"8d6c7357-e7ce-437e-98a5-eeea39389273","Type":"ContainerDied","Data":"3d95376268539d86f09b788a17a70a0291db7addb7ac64321d169453ca3a2d01"} Nov 25 11:31:30 crc kubenswrapper[4776]: I1125 11:31:30.135618 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-nhxsr" Nov 25 11:31:30 crc kubenswrapper[4776]: I1125 11:31:30.193249 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9eccbe5b-80cd-4633-829c-0f79f228bf48-ssh-key\") pod \"9eccbe5b-80cd-4633-829c-0f79f228bf48\" (UID: \"9eccbe5b-80cd-4633-829c-0f79f228bf48\") " Nov 25 11:31:30 crc kubenswrapper[4776]: I1125 11:31:30.262796 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9eccbe5b-80cd-4633-829c-0f79f228bf48-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9eccbe5b-80cd-4633-829c-0f79f228bf48" (UID: "9eccbe5b-80cd-4633-829c-0f79f228bf48"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:31:30 crc kubenswrapper[4776]: I1125 11:31:30.294756 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pb7bz\" (UniqueName: \"kubernetes.io/projected/9eccbe5b-80cd-4633-829c-0f79f228bf48-kube-api-access-pb7bz\") pod \"9eccbe5b-80cd-4633-829c-0f79f228bf48\" (UID: \"9eccbe5b-80cd-4633-829c-0f79f228bf48\") " Nov 25 11:31:30 crc kubenswrapper[4776]: I1125 11:31:30.295242 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9eccbe5b-80cd-4633-829c-0f79f228bf48-inventory\") pod \"9eccbe5b-80cd-4633-829c-0f79f228bf48\" (UID: \"9eccbe5b-80cd-4633-829c-0f79f228bf48\") " Nov 25 11:31:30 crc kubenswrapper[4776]: I1125 11:31:30.296165 4776 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9eccbe5b-80cd-4633-829c-0f79f228bf48-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:31:30 crc kubenswrapper[4776]: I1125 11:31:30.301803 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9eccbe5b-80cd-4633-829c-0f79f228bf48-kube-api-access-pb7bz" (OuterVolumeSpecName: "kube-api-access-pb7bz") pod "9eccbe5b-80cd-4633-829c-0f79f228bf48" (UID: "9eccbe5b-80cd-4633-829c-0f79f228bf48"). InnerVolumeSpecName "kube-api-access-pb7bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:31:30 crc kubenswrapper[4776]: I1125 11:31:30.325904 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9eccbe5b-80cd-4633-829c-0f79f228bf48-inventory" (OuterVolumeSpecName: "inventory") pod "9eccbe5b-80cd-4633-829c-0f79f228bf48" (UID: "9eccbe5b-80cd-4633-829c-0f79f228bf48"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:31:30 crc kubenswrapper[4776]: I1125 11:31:30.398050 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pb7bz\" (UniqueName: \"kubernetes.io/projected/9eccbe5b-80cd-4633-829c-0f79f228bf48-kube-api-access-pb7bz\") on node \"crc\" DevicePath \"\"" Nov 25 11:31:30 crc kubenswrapper[4776]: I1125 11:31:30.398129 4776 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9eccbe5b-80cd-4633-829c-0f79f228bf48-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:31:30 crc kubenswrapper[4776]: I1125 11:31:30.725957 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-nhxsr" Nov 25 11:31:30 crc kubenswrapper[4776]: I1125 11:31:30.725949 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-nhxsr" event={"ID":"9eccbe5b-80cd-4633-829c-0f79f228bf48","Type":"ContainerDied","Data":"95b92d82dc77c0166a24847174ad728a7819d2b4756e2d24b9eb5d8908e7055a"} Nov 25 11:31:30 crc kubenswrapper[4776]: I1125 11:31:30.726101 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95b92d82dc77c0166a24847174ad728a7819d2b4756e2d24b9eb5d8908e7055a" Nov 25 11:31:30 crc kubenswrapper[4776]: I1125 11:31:30.729636 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-28hck" event={"ID":"8d6c7357-e7ce-437e-98a5-eeea39389273","Type":"ContainerStarted","Data":"5e0abd68e32c20f77809fcabcef24ef3050ce337148d6f434b568cbc2bff4ca3"} Nov 25 11:31:30 crc kubenswrapper[4776]: I1125 11:31:30.766902 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-28hck" podStartSLOduration=2.329668306 podStartE2EDuration="4.76688246s" podCreationTimestamp="2025-11-25 11:31:26 +0000 UTC" firstStartedPulling="2025-11-25 11:31:27.6901294 +0000 UTC m=+7632.731188963" lastFinishedPulling="2025-11-25 11:31:30.127343564 +0000 UTC m=+7635.168403117" observedRunningTime="2025-11-25 11:31:30.743470812 +0000 UTC m=+7635.784530375" watchObservedRunningTime="2025-11-25 11:31:30.76688246 +0000 UTC m=+7635.807942013" Nov 25 11:31:30 crc kubenswrapper[4776]: I1125 11:31:30.815515 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-75kbb"] Nov 25 11:31:30 crc kubenswrapper[4776]: E1125 11:31:30.816374 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9eccbe5b-80cd-4633-829c-0f79f228bf48" containerName="download-cache-openstack-openstack-cell1" Nov 25 11:31:30 crc kubenswrapper[4776]: I1125 11:31:30.816396 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="9eccbe5b-80cd-4633-829c-0f79f228bf48" containerName="download-cache-openstack-openstack-cell1" Nov 25 11:31:30 crc kubenswrapper[4776]: I1125 11:31:30.816663 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="9eccbe5b-80cd-4633-829c-0f79f228bf48" containerName="download-cache-openstack-openstack-cell1" Nov 25 11:31:30 crc kubenswrapper[4776]: I1125 11:31:30.817640 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-75kbb" Nov 25 11:31:30 crc kubenswrapper[4776]: I1125 11:31:30.820183 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dknh7" Nov 25 11:31:30 crc kubenswrapper[4776]: I1125 11:31:30.820473 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:31:30 crc kubenswrapper[4776]: I1125 11:31:30.820500 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 25 11:31:30 crc kubenswrapper[4776]: I1125 11:31:30.821682 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 25 11:31:30 crc kubenswrapper[4776]: I1125 11:31:30.842233 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-75kbb"] Nov 25 11:31:31 crc kubenswrapper[4776]: I1125 11:31:31.012778 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5cb01ee8-be4e-4cb2-9d08-0ee2509afac6-inventory\") pod \"configure-network-openstack-openstack-cell1-75kbb\" (UID: \"5cb01ee8-be4e-4cb2-9d08-0ee2509afac6\") " pod="openstack/configure-network-openstack-openstack-cell1-75kbb" Nov 25 11:31:31 crc kubenswrapper[4776]: I1125 11:31:31.012866 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5cb01ee8-be4e-4cb2-9d08-0ee2509afac6-ssh-key\") pod \"configure-network-openstack-openstack-cell1-75kbb\" (UID: \"5cb01ee8-be4e-4cb2-9d08-0ee2509afac6\") " pod="openstack/configure-network-openstack-openstack-cell1-75kbb" Nov 25 11:31:31 crc kubenswrapper[4776]: I1125 11:31:31.013321 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74n8x\" (UniqueName: \"kubernetes.io/projected/5cb01ee8-be4e-4cb2-9d08-0ee2509afac6-kube-api-access-74n8x\") pod \"configure-network-openstack-openstack-cell1-75kbb\" (UID: \"5cb01ee8-be4e-4cb2-9d08-0ee2509afac6\") " pod="openstack/configure-network-openstack-openstack-cell1-75kbb" Nov 25 11:31:31 crc kubenswrapper[4776]: I1125 11:31:31.115882 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5cb01ee8-be4e-4cb2-9d08-0ee2509afac6-inventory\") pod \"configure-network-openstack-openstack-cell1-75kbb\" (UID: \"5cb01ee8-be4e-4cb2-9d08-0ee2509afac6\") " pod="openstack/configure-network-openstack-openstack-cell1-75kbb" Nov 25 11:31:31 crc kubenswrapper[4776]: I1125 11:31:31.115980 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5cb01ee8-be4e-4cb2-9d08-0ee2509afac6-ssh-key\") pod \"configure-network-openstack-openstack-cell1-75kbb\" (UID: \"5cb01ee8-be4e-4cb2-9d08-0ee2509afac6\") " pod="openstack/configure-network-openstack-openstack-cell1-75kbb" Nov 25 11:31:31 crc kubenswrapper[4776]: I1125 11:31:31.116126 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74n8x\" (UniqueName: \"kubernetes.io/projected/5cb01ee8-be4e-4cb2-9d08-0ee2509afac6-kube-api-access-74n8x\") pod \"configure-network-openstack-openstack-cell1-75kbb\" (UID: \"5cb01ee8-be4e-4cb2-9d08-0ee2509afac6\") " pod="openstack/configure-network-openstack-openstack-cell1-75kbb" Nov 25 11:31:31 crc kubenswrapper[4776]: I1125 11:31:31.121071 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5cb01ee8-be4e-4cb2-9d08-0ee2509afac6-ssh-key\") pod \"configure-network-openstack-openstack-cell1-75kbb\" (UID: \"5cb01ee8-be4e-4cb2-9d08-0ee2509afac6\") " pod="openstack/configure-network-openstack-openstack-cell1-75kbb" Nov 25 11:31:31 crc kubenswrapper[4776]: I1125 11:31:31.121319 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5cb01ee8-be4e-4cb2-9d08-0ee2509afac6-inventory\") pod \"configure-network-openstack-openstack-cell1-75kbb\" (UID: \"5cb01ee8-be4e-4cb2-9d08-0ee2509afac6\") " pod="openstack/configure-network-openstack-openstack-cell1-75kbb" Nov 25 11:31:31 crc kubenswrapper[4776]: I1125 11:31:31.135619 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74n8x\" (UniqueName: \"kubernetes.io/projected/5cb01ee8-be4e-4cb2-9d08-0ee2509afac6-kube-api-access-74n8x\") pod \"configure-network-openstack-openstack-cell1-75kbb\" (UID: \"5cb01ee8-be4e-4cb2-9d08-0ee2509afac6\") " pod="openstack/configure-network-openstack-openstack-cell1-75kbb" Nov 25 11:31:31 crc kubenswrapper[4776]: I1125 11:31:31.435370 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-75kbb" Nov 25 11:31:31 crc kubenswrapper[4776]: I1125 11:31:31.990723 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-75kbb"] Nov 25 11:31:31 crc kubenswrapper[4776]: W1125 11:31:31.992455 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5cb01ee8_be4e_4cb2_9d08_0ee2509afac6.slice/crio-b9d396ee4920c68a0c44f08a85279fd60ecddf83bddf93413a1813a51cc9412c WatchSource:0}: Error finding container b9d396ee4920c68a0c44f08a85279fd60ecddf83bddf93413a1813a51cc9412c: Status 404 returned error can't find the container with id b9d396ee4920c68a0c44f08a85279fd60ecddf83bddf93413a1813a51cc9412c Nov 25 11:31:32 crc kubenswrapper[4776]: I1125 11:31:32.746946 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-75kbb" event={"ID":"5cb01ee8-be4e-4cb2-9d08-0ee2509afac6","Type":"ContainerStarted","Data":"4cc77e27bb6bbe123eac1da46a87f70c7fa44f1f28ff4c3bdddfff8387117f0c"} Nov 25 11:31:32 crc kubenswrapper[4776]: I1125 11:31:32.746991 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-75kbb" event={"ID":"5cb01ee8-be4e-4cb2-9d08-0ee2509afac6","Type":"ContainerStarted","Data":"b9d396ee4920c68a0c44f08a85279fd60ecddf83bddf93413a1813a51cc9412c"} Nov 25 11:31:32 crc kubenswrapper[4776]: I1125 11:31:32.774833 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-cell1-75kbb" podStartSLOduration=2.332319419 podStartE2EDuration="2.774812612s" podCreationTimestamp="2025-11-25 11:31:30 +0000 UTC" firstStartedPulling="2025-11-25 11:31:31.995624631 +0000 UTC m=+7637.036684184" lastFinishedPulling="2025-11-25 11:31:32.438117814 +0000 UTC m=+7637.479177377" observedRunningTime="2025-11-25 11:31:32.765452657 +0000 UTC m=+7637.806512210" watchObservedRunningTime="2025-11-25 11:31:32.774812612 +0000 UTC m=+7637.815872165" Nov 25 11:31:36 crc kubenswrapper[4776]: I1125 11:31:36.925194 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-28hck" Nov 25 11:31:36 crc kubenswrapper[4776]: I1125 11:31:36.925933 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-28hck" Nov 25 11:31:36 crc kubenswrapper[4776]: I1125 11:31:36.985047 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-28hck" Nov 25 11:31:37 crc kubenswrapper[4776]: I1125 11:31:37.664435 4776 scope.go:117] "RemoveContainer" containerID="1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c" Nov 25 11:31:37 crc kubenswrapper[4776]: E1125 11:31:37.664716 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:31:37 crc kubenswrapper[4776]: I1125 11:31:37.843929 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-28hck" Nov 25 11:31:37 crc kubenswrapper[4776]: I1125 11:31:37.907880 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-28hck"] Nov 25 11:31:39 crc kubenswrapper[4776]: I1125 11:31:39.831045 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-28hck" podUID="8d6c7357-e7ce-437e-98a5-eeea39389273" containerName="registry-server" containerID="cri-o://5e0abd68e32c20f77809fcabcef24ef3050ce337148d6f434b568cbc2bff4ca3" gracePeriod=2 Nov 25 11:31:40 crc kubenswrapper[4776]: I1125 11:31:40.331987 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-28hck" Nov 25 11:31:40 crc kubenswrapper[4776]: I1125 11:31:40.527774 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8ppx\" (UniqueName: \"kubernetes.io/projected/8d6c7357-e7ce-437e-98a5-eeea39389273-kube-api-access-n8ppx\") pod \"8d6c7357-e7ce-437e-98a5-eeea39389273\" (UID: \"8d6c7357-e7ce-437e-98a5-eeea39389273\") " Nov 25 11:31:40 crc kubenswrapper[4776]: I1125 11:31:40.528091 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d6c7357-e7ce-437e-98a5-eeea39389273-utilities\") pod \"8d6c7357-e7ce-437e-98a5-eeea39389273\" (UID: \"8d6c7357-e7ce-437e-98a5-eeea39389273\") " Nov 25 11:31:40 crc kubenswrapper[4776]: I1125 11:31:40.528330 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d6c7357-e7ce-437e-98a5-eeea39389273-catalog-content\") pod \"8d6c7357-e7ce-437e-98a5-eeea39389273\" (UID: \"8d6c7357-e7ce-437e-98a5-eeea39389273\") " Nov 25 11:31:40 crc kubenswrapper[4776]: I1125 11:31:40.528878 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d6c7357-e7ce-437e-98a5-eeea39389273-utilities" (OuterVolumeSpecName: "utilities") pod "8d6c7357-e7ce-437e-98a5-eeea39389273" (UID: "8d6c7357-e7ce-437e-98a5-eeea39389273"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:31:40 crc kubenswrapper[4776]: I1125 11:31:40.533681 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d6c7357-e7ce-437e-98a5-eeea39389273-kube-api-access-n8ppx" (OuterVolumeSpecName: "kube-api-access-n8ppx") pod "8d6c7357-e7ce-437e-98a5-eeea39389273" (UID: "8d6c7357-e7ce-437e-98a5-eeea39389273"). InnerVolumeSpecName "kube-api-access-n8ppx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:31:40 crc kubenswrapper[4776]: I1125 11:31:40.547524 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d6c7357-e7ce-437e-98a5-eeea39389273-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d6c7357-e7ce-437e-98a5-eeea39389273" (UID: "8d6c7357-e7ce-437e-98a5-eeea39389273"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:31:40 crc kubenswrapper[4776]: I1125 11:31:40.630891 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8ppx\" (UniqueName: \"kubernetes.io/projected/8d6c7357-e7ce-437e-98a5-eeea39389273-kube-api-access-n8ppx\") on node \"crc\" DevicePath \"\"" Nov 25 11:31:40 crc kubenswrapper[4776]: I1125 11:31:40.630937 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d6c7357-e7ce-437e-98a5-eeea39389273-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:31:40 crc kubenswrapper[4776]: I1125 11:31:40.630955 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d6c7357-e7ce-437e-98a5-eeea39389273-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:31:40 crc kubenswrapper[4776]: I1125 11:31:40.842776 4776 generic.go:334] "Generic (PLEG): container finished" podID="8d6c7357-e7ce-437e-98a5-eeea39389273" containerID="5e0abd68e32c20f77809fcabcef24ef3050ce337148d6f434b568cbc2bff4ca3" exitCode=0 Nov 25 11:31:40 crc kubenswrapper[4776]: I1125 11:31:40.842845 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-28hck" event={"ID":"8d6c7357-e7ce-437e-98a5-eeea39389273","Type":"ContainerDied","Data":"5e0abd68e32c20f77809fcabcef24ef3050ce337148d6f434b568cbc2bff4ca3"} Nov 25 11:31:40 crc kubenswrapper[4776]: I1125 11:31:40.842860 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-28hck" Nov 25 11:31:40 crc kubenswrapper[4776]: I1125 11:31:40.842877 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-28hck" event={"ID":"8d6c7357-e7ce-437e-98a5-eeea39389273","Type":"ContainerDied","Data":"332bae89292c564308e07e939fd3976d6aaff6b2a8c3f8030433932cbf2d3276"} Nov 25 11:31:40 crc kubenswrapper[4776]: I1125 11:31:40.842897 4776 scope.go:117] "RemoveContainer" containerID="5e0abd68e32c20f77809fcabcef24ef3050ce337148d6f434b568cbc2bff4ca3" Nov 25 11:31:40 crc kubenswrapper[4776]: I1125 11:31:40.869881 4776 scope.go:117] "RemoveContainer" containerID="3d95376268539d86f09b788a17a70a0291db7addb7ac64321d169453ca3a2d01" Nov 25 11:31:40 crc kubenswrapper[4776]: I1125 11:31:40.876315 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-28hck"] Nov 25 11:31:40 crc kubenswrapper[4776]: I1125 11:31:40.885708 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-28hck"] Nov 25 11:31:40 crc kubenswrapper[4776]: I1125 11:31:40.902576 4776 scope.go:117] "RemoveContainer" containerID="fc69f99d410394b64ac2a4045a0841e1e1a299dddf91b78a2287a2c5a926cb5f" Nov 25 11:31:40 crc kubenswrapper[4776]: I1125 11:31:40.952262 4776 scope.go:117] "RemoveContainer" containerID="5e0abd68e32c20f77809fcabcef24ef3050ce337148d6f434b568cbc2bff4ca3" Nov 25 11:31:40 crc kubenswrapper[4776]: E1125 11:31:40.952727 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e0abd68e32c20f77809fcabcef24ef3050ce337148d6f434b568cbc2bff4ca3\": container with ID starting with 5e0abd68e32c20f77809fcabcef24ef3050ce337148d6f434b568cbc2bff4ca3 not found: ID does not exist" containerID="5e0abd68e32c20f77809fcabcef24ef3050ce337148d6f434b568cbc2bff4ca3" Nov 25 11:31:40 crc kubenswrapper[4776]: I1125 11:31:40.952771 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e0abd68e32c20f77809fcabcef24ef3050ce337148d6f434b568cbc2bff4ca3"} err="failed to get container status \"5e0abd68e32c20f77809fcabcef24ef3050ce337148d6f434b568cbc2bff4ca3\": rpc error: code = NotFound desc = could not find container \"5e0abd68e32c20f77809fcabcef24ef3050ce337148d6f434b568cbc2bff4ca3\": container with ID starting with 5e0abd68e32c20f77809fcabcef24ef3050ce337148d6f434b568cbc2bff4ca3 not found: ID does not exist" Nov 25 11:31:40 crc kubenswrapper[4776]: I1125 11:31:40.952805 4776 scope.go:117] "RemoveContainer" containerID="3d95376268539d86f09b788a17a70a0291db7addb7ac64321d169453ca3a2d01" Nov 25 11:31:40 crc kubenswrapper[4776]: E1125 11:31:40.953229 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d95376268539d86f09b788a17a70a0291db7addb7ac64321d169453ca3a2d01\": container with ID starting with 3d95376268539d86f09b788a17a70a0291db7addb7ac64321d169453ca3a2d01 not found: ID does not exist" containerID="3d95376268539d86f09b788a17a70a0291db7addb7ac64321d169453ca3a2d01" Nov 25 11:31:40 crc kubenswrapper[4776]: I1125 11:31:40.953283 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d95376268539d86f09b788a17a70a0291db7addb7ac64321d169453ca3a2d01"} err="failed to get container status \"3d95376268539d86f09b788a17a70a0291db7addb7ac64321d169453ca3a2d01\": rpc error: code = NotFound desc = could not find container \"3d95376268539d86f09b788a17a70a0291db7addb7ac64321d169453ca3a2d01\": container with ID starting with 3d95376268539d86f09b788a17a70a0291db7addb7ac64321d169453ca3a2d01 not found: ID does not exist" Nov 25 11:31:40 crc kubenswrapper[4776]: I1125 11:31:40.953325 4776 scope.go:117] "RemoveContainer" containerID="fc69f99d410394b64ac2a4045a0841e1e1a299dddf91b78a2287a2c5a926cb5f" Nov 25 11:31:40 crc kubenswrapper[4776]: E1125 11:31:40.953703 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc69f99d410394b64ac2a4045a0841e1e1a299dddf91b78a2287a2c5a926cb5f\": container with ID starting with fc69f99d410394b64ac2a4045a0841e1e1a299dddf91b78a2287a2c5a926cb5f not found: ID does not exist" containerID="fc69f99d410394b64ac2a4045a0841e1e1a299dddf91b78a2287a2c5a926cb5f" Nov 25 11:31:40 crc kubenswrapper[4776]: I1125 11:31:40.953730 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc69f99d410394b64ac2a4045a0841e1e1a299dddf91b78a2287a2c5a926cb5f"} err="failed to get container status \"fc69f99d410394b64ac2a4045a0841e1e1a299dddf91b78a2287a2c5a926cb5f\": rpc error: code = NotFound desc = could not find container \"fc69f99d410394b64ac2a4045a0841e1e1a299dddf91b78a2287a2c5a926cb5f\": container with ID starting with fc69f99d410394b64ac2a4045a0841e1e1a299dddf91b78a2287a2c5a926cb5f not found: ID does not exist" Nov 25 11:31:41 crc kubenswrapper[4776]: I1125 11:31:41.673036 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d6c7357-e7ce-437e-98a5-eeea39389273" path="/var/lib/kubelet/pods/8d6c7357-e7ce-437e-98a5-eeea39389273/volumes" Nov 25 11:31:48 crc kubenswrapper[4776]: I1125 11:31:48.663852 4776 scope.go:117] "RemoveContainer" containerID="1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c" Nov 25 11:31:48 crc kubenswrapper[4776]: E1125 11:31:48.664626 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:32:02 crc kubenswrapper[4776]: I1125 11:32:02.662327 4776 scope.go:117] "RemoveContainer" containerID="1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c" Nov 25 11:32:02 crc kubenswrapper[4776]: E1125 11:32:02.664583 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:32:15 crc kubenswrapper[4776]: I1125 11:32:15.672475 4776 scope.go:117] "RemoveContainer" containerID="1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c" Nov 25 11:32:15 crc kubenswrapper[4776]: E1125 11:32:15.673662 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:32:28 crc kubenswrapper[4776]: I1125 11:32:28.662567 4776 scope.go:117] "RemoveContainer" containerID="1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c" Nov 25 11:32:28 crc kubenswrapper[4776]: E1125 11:32:28.663623 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:32:42 crc kubenswrapper[4776]: I1125 11:32:42.664560 4776 scope.go:117] "RemoveContainer" containerID="1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c" Nov 25 11:32:42 crc kubenswrapper[4776]: E1125 11:32:42.666687 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:32:51 crc kubenswrapper[4776]: I1125 11:32:51.532565 4776 generic.go:334] "Generic (PLEG): container finished" podID="5cb01ee8-be4e-4cb2-9d08-0ee2509afac6" containerID="4cc77e27bb6bbe123eac1da46a87f70c7fa44f1f28ff4c3bdddfff8387117f0c" exitCode=0 Nov 25 11:32:51 crc kubenswrapper[4776]: I1125 11:32:51.532665 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-75kbb" event={"ID":"5cb01ee8-be4e-4cb2-9d08-0ee2509afac6","Type":"ContainerDied","Data":"4cc77e27bb6bbe123eac1da46a87f70c7fa44f1f28ff4c3bdddfff8387117f0c"} Nov 25 11:32:52 crc kubenswrapper[4776]: I1125 11:32:52.963027 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-75kbb" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.051538 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74n8x\" (UniqueName: \"kubernetes.io/projected/5cb01ee8-be4e-4cb2-9d08-0ee2509afac6-kube-api-access-74n8x\") pod \"5cb01ee8-be4e-4cb2-9d08-0ee2509afac6\" (UID: \"5cb01ee8-be4e-4cb2-9d08-0ee2509afac6\") " Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.051643 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5cb01ee8-be4e-4cb2-9d08-0ee2509afac6-ssh-key\") pod \"5cb01ee8-be4e-4cb2-9d08-0ee2509afac6\" (UID: \"5cb01ee8-be4e-4cb2-9d08-0ee2509afac6\") " Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.051825 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5cb01ee8-be4e-4cb2-9d08-0ee2509afac6-inventory\") pod \"5cb01ee8-be4e-4cb2-9d08-0ee2509afac6\" (UID: \"5cb01ee8-be4e-4cb2-9d08-0ee2509afac6\") " Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.059266 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cb01ee8-be4e-4cb2-9d08-0ee2509afac6-kube-api-access-74n8x" (OuterVolumeSpecName: "kube-api-access-74n8x") pod "5cb01ee8-be4e-4cb2-9d08-0ee2509afac6" (UID: "5cb01ee8-be4e-4cb2-9d08-0ee2509afac6"). InnerVolumeSpecName "kube-api-access-74n8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.080812 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cb01ee8-be4e-4cb2-9d08-0ee2509afac6-inventory" (OuterVolumeSpecName: "inventory") pod "5cb01ee8-be4e-4cb2-9d08-0ee2509afac6" (UID: "5cb01ee8-be4e-4cb2-9d08-0ee2509afac6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.084976 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cb01ee8-be4e-4cb2-9d08-0ee2509afac6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5cb01ee8-be4e-4cb2-9d08-0ee2509afac6" (UID: "5cb01ee8-be4e-4cb2-9d08-0ee2509afac6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.153861 4776 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5cb01ee8-be4e-4cb2-9d08-0ee2509afac6-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.153893 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74n8x\" (UniqueName: \"kubernetes.io/projected/5cb01ee8-be4e-4cb2-9d08-0ee2509afac6-kube-api-access-74n8x\") on node \"crc\" DevicePath \"\"" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.153903 4776 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5cb01ee8-be4e-4cb2-9d08-0ee2509afac6-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.558146 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-75kbb" event={"ID":"5cb01ee8-be4e-4cb2-9d08-0ee2509afac6","Type":"ContainerDied","Data":"b9d396ee4920c68a0c44f08a85279fd60ecddf83bddf93413a1813a51cc9412c"} Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.558715 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9d396ee4920c68a0c44f08a85279fd60ecddf83bddf93413a1813a51cc9412c" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.558219 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-75kbb" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.678883 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-k5b8t"] Nov 25 11:32:53 crc kubenswrapper[4776]: E1125 11:32:53.679698 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d6c7357-e7ce-437e-98a5-eeea39389273" containerName="extract-content" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.679751 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d6c7357-e7ce-437e-98a5-eeea39389273" containerName="extract-content" Nov 25 11:32:53 crc kubenswrapper[4776]: E1125 11:32:53.679826 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d6c7357-e7ce-437e-98a5-eeea39389273" containerName="extract-utilities" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.679846 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d6c7357-e7ce-437e-98a5-eeea39389273" containerName="extract-utilities" Nov 25 11:32:53 crc kubenswrapper[4776]: E1125 11:32:53.679886 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cb01ee8-be4e-4cb2-9d08-0ee2509afac6" containerName="configure-network-openstack-openstack-cell1" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.679906 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cb01ee8-be4e-4cb2-9d08-0ee2509afac6" containerName="configure-network-openstack-openstack-cell1" Nov 25 11:32:53 crc kubenswrapper[4776]: E1125 11:32:53.679927 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d6c7357-e7ce-437e-98a5-eeea39389273" containerName="registry-server" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.679944 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d6c7357-e7ce-437e-98a5-eeea39389273" containerName="registry-server" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.680526 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cb01ee8-be4e-4cb2-9d08-0ee2509afac6" containerName="configure-network-openstack-openstack-cell1" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.680625 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d6c7357-e7ce-437e-98a5-eeea39389273" containerName="registry-server" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.682293 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-k5b8t" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.685783 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.685963 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dknh7" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.686231 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.689530 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-k5b8t"] Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.696276 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.867175 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ec16c9c-ee71-4edb-b4e8-6570328756d8-ssh-key\") pod \"validate-network-openstack-openstack-cell1-k5b8t\" (UID: \"0ec16c9c-ee71-4edb-b4e8-6570328756d8\") " pod="openstack/validate-network-openstack-openstack-cell1-k5b8t" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.867575 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ec16c9c-ee71-4edb-b4e8-6570328756d8-inventory\") pod \"validate-network-openstack-openstack-cell1-k5b8t\" (UID: \"0ec16c9c-ee71-4edb-b4e8-6570328756d8\") " pod="openstack/validate-network-openstack-openstack-cell1-k5b8t" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.867628 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtsfw\" (UniqueName: \"kubernetes.io/projected/0ec16c9c-ee71-4edb-b4e8-6570328756d8-kube-api-access-qtsfw\") pod \"validate-network-openstack-openstack-cell1-k5b8t\" (UID: \"0ec16c9c-ee71-4edb-b4e8-6570328756d8\") " pod="openstack/validate-network-openstack-openstack-cell1-k5b8t" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.969160 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ec16c9c-ee71-4edb-b4e8-6570328756d8-ssh-key\") pod \"validate-network-openstack-openstack-cell1-k5b8t\" (UID: \"0ec16c9c-ee71-4edb-b4e8-6570328756d8\") " pod="openstack/validate-network-openstack-openstack-cell1-k5b8t" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.969521 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ec16c9c-ee71-4edb-b4e8-6570328756d8-inventory\") pod \"validate-network-openstack-openstack-cell1-k5b8t\" (UID: \"0ec16c9c-ee71-4edb-b4e8-6570328756d8\") " pod="openstack/validate-network-openstack-openstack-cell1-k5b8t" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.969547 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtsfw\" (UniqueName: \"kubernetes.io/projected/0ec16c9c-ee71-4edb-b4e8-6570328756d8-kube-api-access-qtsfw\") pod \"validate-network-openstack-openstack-cell1-k5b8t\" (UID: \"0ec16c9c-ee71-4edb-b4e8-6570328756d8\") " pod="openstack/validate-network-openstack-openstack-cell1-k5b8t" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.976012 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ec16c9c-ee71-4edb-b4e8-6570328756d8-ssh-key\") pod \"validate-network-openstack-openstack-cell1-k5b8t\" (UID: \"0ec16c9c-ee71-4edb-b4e8-6570328756d8\") " pod="openstack/validate-network-openstack-openstack-cell1-k5b8t" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.978518 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ec16c9c-ee71-4edb-b4e8-6570328756d8-inventory\") pod \"validate-network-openstack-openstack-cell1-k5b8t\" (UID: \"0ec16c9c-ee71-4edb-b4e8-6570328756d8\") " pod="openstack/validate-network-openstack-openstack-cell1-k5b8t" Nov 25 11:32:53 crc kubenswrapper[4776]: I1125 11:32:53.990085 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtsfw\" (UniqueName: \"kubernetes.io/projected/0ec16c9c-ee71-4edb-b4e8-6570328756d8-kube-api-access-qtsfw\") pod \"validate-network-openstack-openstack-cell1-k5b8t\" (UID: \"0ec16c9c-ee71-4edb-b4e8-6570328756d8\") " pod="openstack/validate-network-openstack-openstack-cell1-k5b8t" Nov 25 11:32:54 crc kubenswrapper[4776]: I1125 11:32:54.015865 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-k5b8t" Nov 25 11:32:54 crc kubenswrapper[4776]: I1125 11:32:54.561432 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-k5b8t"] Nov 25 11:32:54 crc kubenswrapper[4776]: I1125 11:32:54.570942 4776 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 11:32:54 crc kubenswrapper[4776]: I1125 11:32:54.662809 4776 scope.go:117] "RemoveContainer" containerID="1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c" Nov 25 11:32:54 crc kubenswrapper[4776]: E1125 11:32:54.663451 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:32:55 crc kubenswrapper[4776]: I1125 11:32:55.581038 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-k5b8t" event={"ID":"0ec16c9c-ee71-4edb-b4e8-6570328756d8","Type":"ContainerStarted","Data":"d25d5708206cc2199ca517b47b278077d31c19d5cc1d07005972a01ae4ea7912"} Nov 25 11:32:55 crc kubenswrapper[4776]: I1125 11:32:55.581394 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-k5b8t" event={"ID":"0ec16c9c-ee71-4edb-b4e8-6570328756d8","Type":"ContainerStarted","Data":"45b450f9fc3773c82ba39cf8d114736fa28a245a54b447b502a8008362305ec2"} Nov 25 11:32:55 crc kubenswrapper[4776]: I1125 11:32:55.603982 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-cell1-k5b8t" podStartSLOduration=2.005418106 podStartE2EDuration="2.603962164s" podCreationTimestamp="2025-11-25 11:32:53 +0000 UTC" firstStartedPulling="2025-11-25 11:32:54.570701898 +0000 UTC m=+7719.611761451" lastFinishedPulling="2025-11-25 11:32:55.169245956 +0000 UTC m=+7720.210305509" observedRunningTime="2025-11-25 11:32:55.597344408 +0000 UTC m=+7720.638403961" watchObservedRunningTime="2025-11-25 11:32:55.603962164 +0000 UTC m=+7720.645021717" Nov 25 11:33:00 crc kubenswrapper[4776]: I1125 11:33:00.634733 4776 generic.go:334] "Generic (PLEG): container finished" podID="0ec16c9c-ee71-4edb-b4e8-6570328756d8" containerID="d25d5708206cc2199ca517b47b278077d31c19d5cc1d07005972a01ae4ea7912" exitCode=0 Nov 25 11:33:00 crc kubenswrapper[4776]: I1125 11:33:00.634824 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-k5b8t" event={"ID":"0ec16c9c-ee71-4edb-b4e8-6570328756d8","Type":"ContainerDied","Data":"d25d5708206cc2199ca517b47b278077d31c19d5cc1d07005972a01ae4ea7912"} Nov 25 11:33:02 crc kubenswrapper[4776]: I1125 11:33:02.733170 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-k5b8t" Nov 25 11:33:02 crc kubenswrapper[4776]: I1125 11:33:02.865746 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ec16c9c-ee71-4edb-b4e8-6570328756d8-ssh-key\") pod \"0ec16c9c-ee71-4edb-b4e8-6570328756d8\" (UID: \"0ec16c9c-ee71-4edb-b4e8-6570328756d8\") " Nov 25 11:33:02 crc kubenswrapper[4776]: I1125 11:33:02.866211 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtsfw\" (UniqueName: \"kubernetes.io/projected/0ec16c9c-ee71-4edb-b4e8-6570328756d8-kube-api-access-qtsfw\") pod \"0ec16c9c-ee71-4edb-b4e8-6570328756d8\" (UID: \"0ec16c9c-ee71-4edb-b4e8-6570328756d8\") " Nov 25 11:33:02 crc kubenswrapper[4776]: I1125 11:33:02.866308 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ec16c9c-ee71-4edb-b4e8-6570328756d8-inventory\") pod \"0ec16c9c-ee71-4edb-b4e8-6570328756d8\" (UID: \"0ec16c9c-ee71-4edb-b4e8-6570328756d8\") " Nov 25 11:33:02 crc kubenswrapper[4776]: I1125 11:33:02.873318 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ec16c9c-ee71-4edb-b4e8-6570328756d8-kube-api-access-qtsfw" (OuterVolumeSpecName: "kube-api-access-qtsfw") pod "0ec16c9c-ee71-4edb-b4e8-6570328756d8" (UID: "0ec16c9c-ee71-4edb-b4e8-6570328756d8"). InnerVolumeSpecName "kube-api-access-qtsfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:33:02 crc kubenswrapper[4776]: I1125 11:33:02.927349 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ec16c9c-ee71-4edb-b4e8-6570328756d8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0ec16c9c-ee71-4edb-b4e8-6570328756d8" (UID: "0ec16c9c-ee71-4edb-b4e8-6570328756d8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:33:02 crc kubenswrapper[4776]: I1125 11:33:02.927391 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ec16c9c-ee71-4edb-b4e8-6570328756d8-inventory" (OuterVolumeSpecName: "inventory") pod "0ec16c9c-ee71-4edb-b4e8-6570328756d8" (UID: "0ec16c9c-ee71-4edb-b4e8-6570328756d8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:33:02 crc kubenswrapper[4776]: I1125 11:33:02.969933 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtsfw\" (UniqueName: \"kubernetes.io/projected/0ec16c9c-ee71-4edb-b4e8-6570328756d8-kube-api-access-qtsfw\") on node \"crc\" DevicePath \"\"" Nov 25 11:33:02 crc kubenswrapper[4776]: I1125 11:33:02.969980 4776 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ec16c9c-ee71-4edb-b4e8-6570328756d8-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:33:02 crc kubenswrapper[4776]: I1125 11:33:02.969996 4776 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0ec16c9c-ee71-4edb-b4e8-6570328756d8-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:33:03 crc kubenswrapper[4776]: I1125 11:33:03.297852 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-k5b8t" event={"ID":"0ec16c9c-ee71-4edb-b4e8-6570328756d8","Type":"ContainerDied","Data":"45b450f9fc3773c82ba39cf8d114736fa28a245a54b447b502a8008362305ec2"} Nov 25 11:33:03 crc kubenswrapper[4776]: I1125 11:33:03.297896 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45b450f9fc3773c82ba39cf8d114736fa28a245a54b447b502a8008362305ec2" Nov 25 11:33:03 crc kubenswrapper[4776]: I1125 11:33:03.297960 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-k5b8t" Nov 25 11:33:03 crc kubenswrapper[4776]: I1125 11:33:03.850520 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-cell1-lxhrp"] Nov 25 11:33:03 crc kubenswrapper[4776]: E1125 11:33:03.851308 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ec16c9c-ee71-4edb-b4e8-6570328756d8" containerName="validate-network-openstack-openstack-cell1" Nov 25 11:33:03 crc kubenswrapper[4776]: I1125 11:33:03.851325 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ec16c9c-ee71-4edb-b4e8-6570328756d8" containerName="validate-network-openstack-openstack-cell1" Nov 25 11:33:03 crc kubenswrapper[4776]: I1125 11:33:03.851614 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ec16c9c-ee71-4edb-b4e8-6570328756d8" containerName="validate-network-openstack-openstack-cell1" Nov 25 11:33:03 crc kubenswrapper[4776]: I1125 11:33:03.852614 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-lxhrp" Nov 25 11:33:03 crc kubenswrapper[4776]: I1125 11:33:03.856825 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:33:03 crc kubenswrapper[4776]: I1125 11:33:03.857274 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 25 11:33:03 crc kubenswrapper[4776]: I1125 11:33:03.857496 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dknh7" Nov 25 11:33:03 crc kubenswrapper[4776]: I1125 11:33:03.857644 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 25 11:33:03 crc kubenswrapper[4776]: I1125 11:33:03.863771 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-lxhrp"] Nov 25 11:33:03 crc kubenswrapper[4776]: I1125 11:33:03.999436 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcp95\" (UniqueName: \"kubernetes.io/projected/8ac69e36-fa36-4a86-843d-4deffacc0f1f-kube-api-access-xcp95\") pod \"install-os-openstack-openstack-cell1-lxhrp\" (UID: \"8ac69e36-fa36-4a86-843d-4deffacc0f1f\") " pod="openstack/install-os-openstack-openstack-cell1-lxhrp" Nov 25 11:33:03 crc kubenswrapper[4776]: I1125 11:33:03.999527 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ac69e36-fa36-4a86-843d-4deffacc0f1f-ssh-key\") pod \"install-os-openstack-openstack-cell1-lxhrp\" (UID: \"8ac69e36-fa36-4a86-843d-4deffacc0f1f\") " pod="openstack/install-os-openstack-openstack-cell1-lxhrp" Nov 25 11:33:03 crc kubenswrapper[4776]: I1125 11:33:03.999623 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ac69e36-fa36-4a86-843d-4deffacc0f1f-inventory\") pod \"install-os-openstack-openstack-cell1-lxhrp\" (UID: \"8ac69e36-fa36-4a86-843d-4deffacc0f1f\") " pod="openstack/install-os-openstack-openstack-cell1-lxhrp" Nov 25 11:33:04 crc kubenswrapper[4776]: I1125 11:33:04.101257 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcp95\" (UniqueName: \"kubernetes.io/projected/8ac69e36-fa36-4a86-843d-4deffacc0f1f-kube-api-access-xcp95\") pod \"install-os-openstack-openstack-cell1-lxhrp\" (UID: \"8ac69e36-fa36-4a86-843d-4deffacc0f1f\") " pod="openstack/install-os-openstack-openstack-cell1-lxhrp" Nov 25 11:33:04 crc kubenswrapper[4776]: I1125 11:33:04.101364 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ac69e36-fa36-4a86-843d-4deffacc0f1f-ssh-key\") pod \"install-os-openstack-openstack-cell1-lxhrp\" (UID: \"8ac69e36-fa36-4a86-843d-4deffacc0f1f\") " pod="openstack/install-os-openstack-openstack-cell1-lxhrp" Nov 25 11:33:04 crc kubenswrapper[4776]: I1125 11:33:04.101456 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ac69e36-fa36-4a86-843d-4deffacc0f1f-inventory\") pod \"install-os-openstack-openstack-cell1-lxhrp\" (UID: \"8ac69e36-fa36-4a86-843d-4deffacc0f1f\") " pod="openstack/install-os-openstack-openstack-cell1-lxhrp" Nov 25 11:33:04 crc kubenswrapper[4776]: I1125 11:33:04.106316 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ac69e36-fa36-4a86-843d-4deffacc0f1f-inventory\") pod \"install-os-openstack-openstack-cell1-lxhrp\" (UID: \"8ac69e36-fa36-4a86-843d-4deffacc0f1f\") " pod="openstack/install-os-openstack-openstack-cell1-lxhrp" Nov 25 11:33:04 crc kubenswrapper[4776]: I1125 11:33:04.106672 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ac69e36-fa36-4a86-843d-4deffacc0f1f-ssh-key\") pod \"install-os-openstack-openstack-cell1-lxhrp\" (UID: \"8ac69e36-fa36-4a86-843d-4deffacc0f1f\") " pod="openstack/install-os-openstack-openstack-cell1-lxhrp" Nov 25 11:33:04 crc kubenswrapper[4776]: I1125 11:33:04.125899 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcp95\" (UniqueName: \"kubernetes.io/projected/8ac69e36-fa36-4a86-843d-4deffacc0f1f-kube-api-access-xcp95\") pod \"install-os-openstack-openstack-cell1-lxhrp\" (UID: \"8ac69e36-fa36-4a86-843d-4deffacc0f1f\") " pod="openstack/install-os-openstack-openstack-cell1-lxhrp" Nov 25 11:33:04 crc kubenswrapper[4776]: I1125 11:33:04.175393 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-lxhrp" Nov 25 11:33:04 crc kubenswrapper[4776]: I1125 11:33:04.740779 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-lxhrp"] Nov 25 11:33:05 crc kubenswrapper[4776]: I1125 11:33:05.325976 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-lxhrp" event={"ID":"8ac69e36-fa36-4a86-843d-4deffacc0f1f","Type":"ContainerStarted","Data":"f5b14801cc9fe6df4c6889c6e887a05160a3a1478bd34490075feb6cd0fdf673"} Nov 25 11:33:06 crc kubenswrapper[4776]: I1125 11:33:06.335721 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-lxhrp" event={"ID":"8ac69e36-fa36-4a86-843d-4deffacc0f1f","Type":"ContainerStarted","Data":"f81992852fddf7c3440d27f145f11c5b83a7d73f0188acf32b33c8d41f75d553"} Nov 25 11:33:06 crc kubenswrapper[4776]: I1125 11:33:06.368285 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-cell1-lxhrp" podStartSLOduration=2.8405120520000002 podStartE2EDuration="3.368262065s" podCreationTimestamp="2025-11-25 11:33:03 +0000 UTC" firstStartedPulling="2025-11-25 11:33:04.746268806 +0000 UTC m=+7729.787328359" lastFinishedPulling="2025-11-25 11:33:05.274018799 +0000 UTC m=+7730.315078372" observedRunningTime="2025-11-25 11:33:06.35173178 +0000 UTC m=+7731.392791343" watchObservedRunningTime="2025-11-25 11:33:06.368262065 +0000 UTC m=+7731.409321618" Nov 25 11:33:06 crc kubenswrapper[4776]: I1125 11:33:06.663229 4776 scope.go:117] "RemoveContainer" containerID="1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c" Nov 25 11:33:06 crc kubenswrapper[4776]: E1125 11:33:06.663587 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:33:21 crc kubenswrapper[4776]: I1125 11:33:21.663504 4776 scope.go:117] "RemoveContainer" containerID="1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c" Nov 25 11:33:22 crc kubenswrapper[4776]: I1125 11:33:22.499870 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"cced90c452d27e661d744bc42ceb0710b870fc7135d4b27a5d442a29ab283935"} Nov 25 11:33:50 crc kubenswrapper[4776]: I1125 11:33:50.795318 4776 generic.go:334] "Generic (PLEG): container finished" podID="8ac69e36-fa36-4a86-843d-4deffacc0f1f" containerID="f81992852fddf7c3440d27f145f11c5b83a7d73f0188acf32b33c8d41f75d553" exitCode=0 Nov 25 11:33:50 crc kubenswrapper[4776]: I1125 11:33:50.795423 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-lxhrp" event={"ID":"8ac69e36-fa36-4a86-843d-4deffacc0f1f","Type":"ContainerDied","Data":"f81992852fddf7c3440d27f145f11c5b83a7d73f0188acf32b33c8d41f75d553"} Nov 25 11:33:52 crc kubenswrapper[4776]: I1125 11:33:52.311176 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-lxhrp" Nov 25 11:33:52 crc kubenswrapper[4776]: I1125 11:33:52.388274 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ac69e36-fa36-4a86-843d-4deffacc0f1f-inventory\") pod \"8ac69e36-fa36-4a86-843d-4deffacc0f1f\" (UID: \"8ac69e36-fa36-4a86-843d-4deffacc0f1f\") " Nov 25 11:33:52 crc kubenswrapper[4776]: I1125 11:33:52.388316 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ac69e36-fa36-4a86-843d-4deffacc0f1f-ssh-key\") pod \"8ac69e36-fa36-4a86-843d-4deffacc0f1f\" (UID: \"8ac69e36-fa36-4a86-843d-4deffacc0f1f\") " Nov 25 11:33:52 crc kubenswrapper[4776]: I1125 11:33:52.388432 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcp95\" (UniqueName: \"kubernetes.io/projected/8ac69e36-fa36-4a86-843d-4deffacc0f1f-kube-api-access-xcp95\") pod \"8ac69e36-fa36-4a86-843d-4deffacc0f1f\" (UID: \"8ac69e36-fa36-4a86-843d-4deffacc0f1f\") " Nov 25 11:33:52 crc kubenswrapper[4776]: I1125 11:33:52.396574 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ac69e36-fa36-4a86-843d-4deffacc0f1f-kube-api-access-xcp95" (OuterVolumeSpecName: "kube-api-access-xcp95") pod "8ac69e36-fa36-4a86-843d-4deffacc0f1f" (UID: "8ac69e36-fa36-4a86-843d-4deffacc0f1f"). InnerVolumeSpecName "kube-api-access-xcp95". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:33:52 crc kubenswrapper[4776]: I1125 11:33:52.428694 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ac69e36-fa36-4a86-843d-4deffacc0f1f-inventory" (OuterVolumeSpecName: "inventory") pod "8ac69e36-fa36-4a86-843d-4deffacc0f1f" (UID: "8ac69e36-fa36-4a86-843d-4deffacc0f1f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:33:52 crc kubenswrapper[4776]: I1125 11:33:52.441986 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ac69e36-fa36-4a86-843d-4deffacc0f1f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8ac69e36-fa36-4a86-843d-4deffacc0f1f" (UID: "8ac69e36-fa36-4a86-843d-4deffacc0f1f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:33:52 crc kubenswrapper[4776]: I1125 11:33:52.491844 4776 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8ac69e36-fa36-4a86-843d-4deffacc0f1f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:33:52 crc kubenswrapper[4776]: I1125 11:33:52.491896 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcp95\" (UniqueName: \"kubernetes.io/projected/8ac69e36-fa36-4a86-843d-4deffacc0f1f-kube-api-access-xcp95\") on node \"crc\" DevicePath \"\"" Nov 25 11:33:52 crc kubenswrapper[4776]: I1125 11:33:52.491915 4776 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8ac69e36-fa36-4a86-843d-4deffacc0f1f-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:33:52 crc kubenswrapper[4776]: I1125 11:33:52.820796 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-lxhrp" event={"ID":"8ac69e36-fa36-4a86-843d-4deffacc0f1f","Type":"ContainerDied","Data":"f5b14801cc9fe6df4c6889c6e887a05160a3a1478bd34490075feb6cd0fdf673"} Nov 25 11:33:52 crc kubenswrapper[4776]: I1125 11:33:52.820838 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5b14801cc9fe6df4c6889c6e887a05160a3a1478bd34490075feb6cd0fdf673" Nov 25 11:33:52 crc kubenswrapper[4776]: I1125 11:33:52.820858 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-lxhrp" Nov 25 11:33:52 crc kubenswrapper[4776]: I1125 11:33:52.927628 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-lwvwk"] Nov 25 11:33:52 crc kubenswrapper[4776]: E1125 11:33:52.928257 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ac69e36-fa36-4a86-843d-4deffacc0f1f" containerName="install-os-openstack-openstack-cell1" Nov 25 11:33:52 crc kubenswrapper[4776]: I1125 11:33:52.928318 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ac69e36-fa36-4a86-843d-4deffacc0f1f" containerName="install-os-openstack-openstack-cell1" Nov 25 11:33:52 crc kubenswrapper[4776]: I1125 11:33:52.928579 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ac69e36-fa36-4a86-843d-4deffacc0f1f" containerName="install-os-openstack-openstack-cell1" Nov 25 11:33:52 crc kubenswrapper[4776]: I1125 11:33:52.929406 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-lwvwk" Nov 25 11:33:52 crc kubenswrapper[4776]: I1125 11:33:52.934755 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 25 11:33:52 crc kubenswrapper[4776]: I1125 11:33:52.934883 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dknh7" Nov 25 11:33:52 crc kubenswrapper[4776]: I1125 11:33:52.934970 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:33:52 crc kubenswrapper[4776]: I1125 11:33:52.937401 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 25 11:33:52 crc kubenswrapper[4776]: I1125 11:33:52.939042 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-lwvwk"] Nov 25 11:33:53 crc kubenswrapper[4776]: I1125 11:33:53.002544 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ed85cc7-157f-4a26-85aa-5266b68246f7-ssh-key\") pod \"configure-os-openstack-openstack-cell1-lwvwk\" (UID: \"7ed85cc7-157f-4a26-85aa-5266b68246f7\") " pod="openstack/configure-os-openstack-openstack-cell1-lwvwk" Nov 25 11:33:53 crc kubenswrapper[4776]: I1125 11:33:53.002617 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jfgc\" (UniqueName: \"kubernetes.io/projected/7ed85cc7-157f-4a26-85aa-5266b68246f7-kube-api-access-6jfgc\") pod \"configure-os-openstack-openstack-cell1-lwvwk\" (UID: \"7ed85cc7-157f-4a26-85aa-5266b68246f7\") " pod="openstack/configure-os-openstack-openstack-cell1-lwvwk" Nov 25 11:33:53 crc kubenswrapper[4776]: I1125 11:33:53.002651 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ed85cc7-157f-4a26-85aa-5266b68246f7-inventory\") pod \"configure-os-openstack-openstack-cell1-lwvwk\" (UID: \"7ed85cc7-157f-4a26-85aa-5266b68246f7\") " pod="openstack/configure-os-openstack-openstack-cell1-lwvwk" Nov 25 11:33:53 crc kubenswrapper[4776]: I1125 11:33:53.105270 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ed85cc7-157f-4a26-85aa-5266b68246f7-ssh-key\") pod \"configure-os-openstack-openstack-cell1-lwvwk\" (UID: \"7ed85cc7-157f-4a26-85aa-5266b68246f7\") " pod="openstack/configure-os-openstack-openstack-cell1-lwvwk" Nov 25 11:33:53 crc kubenswrapper[4776]: I1125 11:33:53.105312 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jfgc\" (UniqueName: \"kubernetes.io/projected/7ed85cc7-157f-4a26-85aa-5266b68246f7-kube-api-access-6jfgc\") pod \"configure-os-openstack-openstack-cell1-lwvwk\" (UID: \"7ed85cc7-157f-4a26-85aa-5266b68246f7\") " pod="openstack/configure-os-openstack-openstack-cell1-lwvwk" Nov 25 11:33:53 crc kubenswrapper[4776]: I1125 11:33:53.105337 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ed85cc7-157f-4a26-85aa-5266b68246f7-inventory\") pod \"configure-os-openstack-openstack-cell1-lwvwk\" (UID: \"7ed85cc7-157f-4a26-85aa-5266b68246f7\") " pod="openstack/configure-os-openstack-openstack-cell1-lwvwk" Nov 25 11:33:53 crc kubenswrapper[4776]: I1125 11:33:53.110717 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ed85cc7-157f-4a26-85aa-5266b68246f7-ssh-key\") pod \"configure-os-openstack-openstack-cell1-lwvwk\" (UID: \"7ed85cc7-157f-4a26-85aa-5266b68246f7\") " pod="openstack/configure-os-openstack-openstack-cell1-lwvwk" Nov 25 11:33:53 crc kubenswrapper[4776]: I1125 11:33:53.112336 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ed85cc7-157f-4a26-85aa-5266b68246f7-inventory\") pod \"configure-os-openstack-openstack-cell1-lwvwk\" (UID: \"7ed85cc7-157f-4a26-85aa-5266b68246f7\") " pod="openstack/configure-os-openstack-openstack-cell1-lwvwk" Nov 25 11:33:53 crc kubenswrapper[4776]: I1125 11:33:53.127761 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jfgc\" (UniqueName: \"kubernetes.io/projected/7ed85cc7-157f-4a26-85aa-5266b68246f7-kube-api-access-6jfgc\") pod \"configure-os-openstack-openstack-cell1-lwvwk\" (UID: \"7ed85cc7-157f-4a26-85aa-5266b68246f7\") " pod="openstack/configure-os-openstack-openstack-cell1-lwvwk" Nov 25 11:33:53 crc kubenswrapper[4776]: I1125 11:33:53.255556 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-lwvwk" Nov 25 11:33:53 crc kubenswrapper[4776]: I1125 11:33:53.783967 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-lwvwk"] Nov 25 11:33:53 crc kubenswrapper[4776]: I1125 11:33:53.838405 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-lwvwk" event={"ID":"7ed85cc7-157f-4a26-85aa-5266b68246f7","Type":"ContainerStarted","Data":"758073d4c2ed1ae58d1ace11b21188efbfb396890d7029ec986436df856a6e79"} Nov 25 11:33:54 crc kubenswrapper[4776]: I1125 11:33:54.851713 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-lwvwk" event={"ID":"7ed85cc7-157f-4a26-85aa-5266b68246f7","Type":"ContainerStarted","Data":"cc6277bf6142a44a051775ab7c7f1bf8d4e7cea2951265d29ec339e4f025254e"} Nov 25 11:33:54 crc kubenswrapper[4776]: I1125 11:33:54.884029 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-lwvwk" podStartSLOduration=2.460616733 podStartE2EDuration="2.884008137s" podCreationTimestamp="2025-11-25 11:33:52 +0000 UTC" firstStartedPulling="2025-11-25 11:33:53.787162655 +0000 UTC m=+7778.828222208" lastFinishedPulling="2025-11-25 11:33:54.210554059 +0000 UTC m=+7779.251613612" observedRunningTime="2025-11-25 11:33:54.873543784 +0000 UTC m=+7779.914603337" watchObservedRunningTime="2025-11-25 11:33:54.884008137 +0000 UTC m=+7779.925067690" Nov 25 11:34:38 crc kubenswrapper[4776]: I1125 11:34:38.349168 4776 generic.go:334] "Generic (PLEG): container finished" podID="7ed85cc7-157f-4a26-85aa-5266b68246f7" containerID="cc6277bf6142a44a051775ab7c7f1bf8d4e7cea2951265d29ec339e4f025254e" exitCode=0 Nov 25 11:34:38 crc kubenswrapper[4776]: I1125 11:34:38.349265 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-lwvwk" event={"ID":"7ed85cc7-157f-4a26-85aa-5266b68246f7","Type":"ContainerDied","Data":"cc6277bf6142a44a051775ab7c7f1bf8d4e7cea2951265d29ec339e4f025254e"} Nov 25 11:34:39 crc kubenswrapper[4776]: I1125 11:34:39.812029 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-lwvwk" Nov 25 11:34:39 crc kubenswrapper[4776]: I1125 11:34:39.829678 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ed85cc7-157f-4a26-85aa-5266b68246f7-ssh-key\") pod \"7ed85cc7-157f-4a26-85aa-5266b68246f7\" (UID: \"7ed85cc7-157f-4a26-85aa-5266b68246f7\") " Nov 25 11:34:39 crc kubenswrapper[4776]: I1125 11:34:39.829919 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jfgc\" (UniqueName: \"kubernetes.io/projected/7ed85cc7-157f-4a26-85aa-5266b68246f7-kube-api-access-6jfgc\") pod \"7ed85cc7-157f-4a26-85aa-5266b68246f7\" (UID: \"7ed85cc7-157f-4a26-85aa-5266b68246f7\") " Nov 25 11:34:39 crc kubenswrapper[4776]: I1125 11:34:39.830174 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ed85cc7-157f-4a26-85aa-5266b68246f7-inventory\") pod \"7ed85cc7-157f-4a26-85aa-5266b68246f7\" (UID: \"7ed85cc7-157f-4a26-85aa-5266b68246f7\") " Nov 25 11:34:39 crc kubenswrapper[4776]: I1125 11:34:39.837829 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ed85cc7-157f-4a26-85aa-5266b68246f7-kube-api-access-6jfgc" (OuterVolumeSpecName: "kube-api-access-6jfgc") pod "7ed85cc7-157f-4a26-85aa-5266b68246f7" (UID: "7ed85cc7-157f-4a26-85aa-5266b68246f7"). InnerVolumeSpecName "kube-api-access-6jfgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:34:39 crc kubenswrapper[4776]: I1125 11:34:39.862674 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ed85cc7-157f-4a26-85aa-5266b68246f7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7ed85cc7-157f-4a26-85aa-5266b68246f7" (UID: "7ed85cc7-157f-4a26-85aa-5266b68246f7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:34:39 crc kubenswrapper[4776]: I1125 11:34:39.864897 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ed85cc7-157f-4a26-85aa-5266b68246f7-inventory" (OuterVolumeSpecName: "inventory") pod "7ed85cc7-157f-4a26-85aa-5266b68246f7" (UID: "7ed85cc7-157f-4a26-85aa-5266b68246f7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:34:39 crc kubenswrapper[4776]: I1125 11:34:39.932827 4776 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ed85cc7-157f-4a26-85aa-5266b68246f7-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:34:39 crc kubenswrapper[4776]: I1125 11:34:39.932868 4776 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ed85cc7-157f-4a26-85aa-5266b68246f7-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:34:39 crc kubenswrapper[4776]: I1125 11:34:39.932879 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jfgc\" (UniqueName: \"kubernetes.io/projected/7ed85cc7-157f-4a26-85aa-5266b68246f7-kube-api-access-6jfgc\") on node \"crc\" DevicePath \"\"" Nov 25 11:34:40 crc kubenswrapper[4776]: I1125 11:34:40.373918 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-lwvwk" event={"ID":"7ed85cc7-157f-4a26-85aa-5266b68246f7","Type":"ContainerDied","Data":"758073d4c2ed1ae58d1ace11b21188efbfb396890d7029ec986436df856a6e79"} Nov 25 11:34:40 crc kubenswrapper[4776]: I1125 11:34:40.373970 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="758073d4c2ed1ae58d1ace11b21188efbfb396890d7029ec986436df856a6e79" Nov 25 11:34:40 crc kubenswrapper[4776]: I1125 11:34:40.374016 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-lwvwk" Nov 25 11:34:40 crc kubenswrapper[4776]: I1125 11:34:40.478272 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-openstack-h7v7g"] Nov 25 11:34:40 crc kubenswrapper[4776]: E1125 11:34:40.478927 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ed85cc7-157f-4a26-85aa-5266b68246f7" containerName="configure-os-openstack-openstack-cell1" Nov 25 11:34:40 crc kubenswrapper[4776]: I1125 11:34:40.478956 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ed85cc7-157f-4a26-85aa-5266b68246f7" containerName="configure-os-openstack-openstack-cell1" Nov 25 11:34:40 crc kubenswrapper[4776]: I1125 11:34:40.479278 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ed85cc7-157f-4a26-85aa-5266b68246f7" containerName="configure-os-openstack-openstack-cell1" Nov 25 11:34:40 crc kubenswrapper[4776]: I1125 11:34:40.480211 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-h7v7g" Nov 25 11:34:40 crc kubenswrapper[4776]: I1125 11:34:40.482562 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 25 11:34:40 crc kubenswrapper[4776]: I1125 11:34:40.483252 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:34:40 crc kubenswrapper[4776]: I1125 11:34:40.483577 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 25 11:34:40 crc kubenswrapper[4776]: I1125 11:34:40.488505 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-h7v7g"] Nov 25 11:34:40 crc kubenswrapper[4776]: I1125 11:34:40.493898 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dknh7" Nov 25 11:34:40 crc kubenswrapper[4776]: I1125 11:34:40.572355 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mz94\" (UniqueName: \"kubernetes.io/projected/087b7d2d-0326-4d89-9262-902367375a92-kube-api-access-9mz94\") pod \"ssh-known-hosts-openstack-h7v7g\" (UID: \"087b7d2d-0326-4d89-9262-902367375a92\") " pod="openstack/ssh-known-hosts-openstack-h7v7g" Nov 25 11:34:40 crc kubenswrapper[4776]: I1125 11:34:40.572480 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/087b7d2d-0326-4d89-9262-902367375a92-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-h7v7g\" (UID: \"087b7d2d-0326-4d89-9262-902367375a92\") " pod="openstack/ssh-known-hosts-openstack-h7v7g" Nov 25 11:34:40 crc kubenswrapper[4776]: I1125 11:34:40.572680 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/087b7d2d-0326-4d89-9262-902367375a92-inventory-0\") pod \"ssh-known-hosts-openstack-h7v7g\" (UID: \"087b7d2d-0326-4d89-9262-902367375a92\") " pod="openstack/ssh-known-hosts-openstack-h7v7g" Nov 25 11:34:40 crc kubenswrapper[4776]: I1125 11:34:40.674509 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mz94\" (UniqueName: \"kubernetes.io/projected/087b7d2d-0326-4d89-9262-902367375a92-kube-api-access-9mz94\") pod \"ssh-known-hosts-openstack-h7v7g\" (UID: \"087b7d2d-0326-4d89-9262-902367375a92\") " pod="openstack/ssh-known-hosts-openstack-h7v7g" Nov 25 11:34:40 crc kubenswrapper[4776]: I1125 11:34:40.675008 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/087b7d2d-0326-4d89-9262-902367375a92-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-h7v7g\" (UID: \"087b7d2d-0326-4d89-9262-902367375a92\") " pod="openstack/ssh-known-hosts-openstack-h7v7g" Nov 25 11:34:40 crc kubenswrapper[4776]: I1125 11:34:40.675145 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/087b7d2d-0326-4d89-9262-902367375a92-inventory-0\") pod \"ssh-known-hosts-openstack-h7v7g\" (UID: \"087b7d2d-0326-4d89-9262-902367375a92\") " pod="openstack/ssh-known-hosts-openstack-h7v7g" Nov 25 11:34:40 crc kubenswrapper[4776]: I1125 11:34:40.680768 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/087b7d2d-0326-4d89-9262-902367375a92-inventory-0\") pod \"ssh-known-hosts-openstack-h7v7g\" (UID: \"087b7d2d-0326-4d89-9262-902367375a92\") " pod="openstack/ssh-known-hosts-openstack-h7v7g" Nov 25 11:34:40 crc kubenswrapper[4776]: I1125 11:34:40.680969 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/087b7d2d-0326-4d89-9262-902367375a92-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-h7v7g\" (UID: \"087b7d2d-0326-4d89-9262-902367375a92\") " pod="openstack/ssh-known-hosts-openstack-h7v7g" Nov 25 11:34:40 crc kubenswrapper[4776]: I1125 11:34:40.696728 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mz94\" (UniqueName: \"kubernetes.io/projected/087b7d2d-0326-4d89-9262-902367375a92-kube-api-access-9mz94\") pod \"ssh-known-hosts-openstack-h7v7g\" (UID: \"087b7d2d-0326-4d89-9262-902367375a92\") " pod="openstack/ssh-known-hosts-openstack-h7v7g" Nov 25 11:34:40 crc kubenswrapper[4776]: I1125 11:34:40.808040 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-h7v7g" Nov 25 11:34:41 crc kubenswrapper[4776]: I1125 11:34:41.401819 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-h7v7g"] Nov 25 11:34:42 crc kubenswrapper[4776]: I1125 11:34:42.397462 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-h7v7g" event={"ID":"087b7d2d-0326-4d89-9262-902367375a92","Type":"ContainerStarted","Data":"34c3958a12c9fa7fb265f873f879c2346cd9245469bd1af16f26b960c41ffab8"} Nov 25 11:34:42 crc kubenswrapper[4776]: I1125 11:34:42.397511 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-h7v7g" event={"ID":"087b7d2d-0326-4d89-9262-902367375a92","Type":"ContainerStarted","Data":"0f6197d8003f1adfa6c265552d9d4f7ba2e4db6830477ff4d6fe140566e1cde7"} Nov 25 11:34:42 crc kubenswrapper[4776]: I1125 11:34:42.421612 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-openstack-h7v7g" podStartSLOduration=1.991407554 podStartE2EDuration="2.421595279s" podCreationTimestamp="2025-11-25 11:34:40 +0000 UTC" firstStartedPulling="2025-11-25 11:34:41.390457606 +0000 UTC m=+7826.431517149" lastFinishedPulling="2025-11-25 11:34:41.820645321 +0000 UTC m=+7826.861704874" observedRunningTime="2025-11-25 11:34:42.414240765 +0000 UTC m=+7827.455300318" watchObservedRunningTime="2025-11-25 11:34:42.421595279 +0000 UTC m=+7827.462654832" Nov 25 11:34:51 crc kubenswrapper[4776]: I1125 11:34:51.488252 4776 generic.go:334] "Generic (PLEG): container finished" podID="087b7d2d-0326-4d89-9262-902367375a92" containerID="34c3958a12c9fa7fb265f873f879c2346cd9245469bd1af16f26b960c41ffab8" exitCode=0 Nov 25 11:34:51 crc kubenswrapper[4776]: I1125 11:34:51.488409 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-h7v7g" event={"ID":"087b7d2d-0326-4d89-9262-902367375a92","Type":"ContainerDied","Data":"34c3958a12c9fa7fb265f873f879c2346cd9245469bd1af16f26b960c41ffab8"} Nov 25 11:34:52 crc kubenswrapper[4776]: I1125 11:34:52.967130 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-h7v7g" Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.075138 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/087b7d2d-0326-4d89-9262-902367375a92-ssh-key-openstack-cell1\") pod \"087b7d2d-0326-4d89-9262-902367375a92\" (UID: \"087b7d2d-0326-4d89-9262-902367375a92\") " Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.075244 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mz94\" (UniqueName: \"kubernetes.io/projected/087b7d2d-0326-4d89-9262-902367375a92-kube-api-access-9mz94\") pod \"087b7d2d-0326-4d89-9262-902367375a92\" (UID: \"087b7d2d-0326-4d89-9262-902367375a92\") " Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.075584 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/087b7d2d-0326-4d89-9262-902367375a92-inventory-0\") pod \"087b7d2d-0326-4d89-9262-902367375a92\" (UID: \"087b7d2d-0326-4d89-9262-902367375a92\") " Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.083280 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/087b7d2d-0326-4d89-9262-902367375a92-kube-api-access-9mz94" (OuterVolumeSpecName: "kube-api-access-9mz94") pod "087b7d2d-0326-4d89-9262-902367375a92" (UID: "087b7d2d-0326-4d89-9262-902367375a92"). InnerVolumeSpecName "kube-api-access-9mz94". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.106838 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/087b7d2d-0326-4d89-9262-902367375a92-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "087b7d2d-0326-4d89-9262-902367375a92" (UID: "087b7d2d-0326-4d89-9262-902367375a92"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.119532 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/087b7d2d-0326-4d89-9262-902367375a92-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "087b7d2d-0326-4d89-9262-902367375a92" (UID: "087b7d2d-0326-4d89-9262-902367375a92"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.178785 4776 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/087b7d2d-0326-4d89-9262-902367375a92-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.178840 4776 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/087b7d2d-0326-4d89-9262-902367375a92-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.178856 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mz94\" (UniqueName: \"kubernetes.io/projected/087b7d2d-0326-4d89-9262-902367375a92-kube-api-access-9mz94\") on node \"crc\" DevicePath \"\"" Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.508319 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-h7v7g" event={"ID":"087b7d2d-0326-4d89-9262-902367375a92","Type":"ContainerDied","Data":"0f6197d8003f1adfa6c265552d9d4f7ba2e4db6830477ff4d6fe140566e1cde7"} Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.508360 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f6197d8003f1adfa6c265552d9d4f7ba2e4db6830477ff4d6fe140566e1cde7" Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.508408 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-h7v7g" Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.581638 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-cell1-q69dh"] Nov 25 11:34:53 crc kubenswrapper[4776]: E1125 11:34:53.582199 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="087b7d2d-0326-4d89-9262-902367375a92" containerName="ssh-known-hosts-openstack" Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.582220 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="087b7d2d-0326-4d89-9262-902367375a92" containerName="ssh-known-hosts-openstack" Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.582500 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="087b7d2d-0326-4d89-9262-902367375a92" containerName="ssh-known-hosts-openstack" Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.586045 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-q69dh" Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.588340 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dknh7" Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.588390 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.588411 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.590440 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.590833 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-q69dh"] Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.691561 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857-inventory\") pod \"run-os-openstack-openstack-cell1-q69dh\" (UID: \"d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857\") " pod="openstack/run-os-openstack-openstack-cell1-q69dh" Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.691917 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857-ssh-key\") pod \"run-os-openstack-openstack-cell1-q69dh\" (UID: \"d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857\") " pod="openstack/run-os-openstack-openstack-cell1-q69dh" Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.692405 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6f5x\" (UniqueName: \"kubernetes.io/projected/d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857-kube-api-access-c6f5x\") pod \"run-os-openstack-openstack-cell1-q69dh\" (UID: \"d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857\") " pod="openstack/run-os-openstack-openstack-cell1-q69dh" Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.795011 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6f5x\" (UniqueName: \"kubernetes.io/projected/d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857-kube-api-access-c6f5x\") pod \"run-os-openstack-openstack-cell1-q69dh\" (UID: \"d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857\") " pod="openstack/run-os-openstack-openstack-cell1-q69dh" Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.795368 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857-inventory\") pod \"run-os-openstack-openstack-cell1-q69dh\" (UID: \"d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857\") " pod="openstack/run-os-openstack-openstack-cell1-q69dh" Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.795449 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857-ssh-key\") pod \"run-os-openstack-openstack-cell1-q69dh\" (UID: \"d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857\") " pod="openstack/run-os-openstack-openstack-cell1-q69dh" Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.799634 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857-ssh-key\") pod \"run-os-openstack-openstack-cell1-q69dh\" (UID: \"d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857\") " pod="openstack/run-os-openstack-openstack-cell1-q69dh" Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.799677 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857-inventory\") pod \"run-os-openstack-openstack-cell1-q69dh\" (UID: \"d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857\") " pod="openstack/run-os-openstack-openstack-cell1-q69dh" Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.822878 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6f5x\" (UniqueName: \"kubernetes.io/projected/d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857-kube-api-access-c6f5x\") pod \"run-os-openstack-openstack-cell1-q69dh\" (UID: \"d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857\") " pod="openstack/run-os-openstack-openstack-cell1-q69dh" Nov 25 11:34:53 crc kubenswrapper[4776]: I1125 11:34:53.901110 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-q69dh" Nov 25 11:34:54 crc kubenswrapper[4776]: I1125 11:34:54.446419 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-q69dh"] Nov 25 11:34:54 crc kubenswrapper[4776]: I1125 11:34:54.516730 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-q69dh" event={"ID":"d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857","Type":"ContainerStarted","Data":"48e6b550f6e2857f97054fc7540dc6937a448dfd3031f49c945479c805a1a6b0"} Nov 25 11:34:55 crc kubenswrapper[4776]: I1125 11:34:55.528232 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-q69dh" event={"ID":"d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857","Type":"ContainerStarted","Data":"95fbb2b4ccdc8d137b2482025653d24cc76894a53c007d5b1cceb60a1c41787f"} Nov 25 11:34:55 crc kubenswrapper[4776]: I1125 11:34:55.548403 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-cell1-q69dh" podStartSLOduration=2.10553047 podStartE2EDuration="2.548384871s" podCreationTimestamp="2025-11-25 11:34:53 +0000 UTC" firstStartedPulling="2025-11-25 11:34:54.452196337 +0000 UTC m=+7839.493255890" lastFinishedPulling="2025-11-25 11:34:54.895050738 +0000 UTC m=+7839.936110291" observedRunningTime="2025-11-25 11:34:55.546195146 +0000 UTC m=+7840.587254719" watchObservedRunningTime="2025-11-25 11:34:55.548384871 +0000 UTC m=+7840.589444424" Nov 25 11:35:03 crc kubenswrapper[4776]: I1125 11:35:03.626304 4776 generic.go:334] "Generic (PLEG): container finished" podID="d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857" containerID="95fbb2b4ccdc8d137b2482025653d24cc76894a53c007d5b1cceb60a1c41787f" exitCode=0 Nov 25 11:35:03 crc kubenswrapper[4776]: I1125 11:35:03.626847 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-q69dh" event={"ID":"d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857","Type":"ContainerDied","Data":"95fbb2b4ccdc8d137b2482025653d24cc76894a53c007d5b1cceb60a1c41787f"} Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.133425 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-q69dh" Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.158346 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6f5x\" (UniqueName: \"kubernetes.io/projected/d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857-kube-api-access-c6f5x\") pod \"d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857\" (UID: \"d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857\") " Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.158563 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857-ssh-key\") pod \"d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857\" (UID: \"d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857\") " Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.158728 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857-inventory\") pod \"d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857\" (UID: \"d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857\") " Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.164857 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857-kube-api-access-c6f5x" (OuterVolumeSpecName: "kube-api-access-c6f5x") pod "d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857" (UID: "d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857"). InnerVolumeSpecName "kube-api-access-c6f5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.196574 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857-inventory" (OuterVolumeSpecName: "inventory") pod "d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857" (UID: "d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.201714 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857" (UID: "d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.262172 4776 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.262230 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6f5x\" (UniqueName: \"kubernetes.io/projected/d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857-kube-api-access-c6f5x\") on node \"crc\" DevicePath \"\"" Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.262251 4776 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.648161 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-q69dh" event={"ID":"d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857","Type":"ContainerDied","Data":"48e6b550f6e2857f97054fc7540dc6937a448dfd3031f49c945479c805a1a6b0"} Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.648201 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48e6b550f6e2857f97054fc7540dc6937a448dfd3031f49c945479c805a1a6b0" Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.648272 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-q69dh" Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.728451 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-jrtwp"] Nov 25 11:35:05 crc kubenswrapper[4776]: E1125 11:35:05.728854 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857" containerName="run-os-openstack-openstack-cell1" Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.728869 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857" containerName="run-os-openstack-openstack-cell1" Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.729117 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857" containerName="run-os-openstack-openstack-cell1" Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.729809 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-jrtwp" Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.732050 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.732975 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.733250 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dknh7" Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.733509 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.746890 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-jrtwp"] Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.772945 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjcjt\" (UniqueName: \"kubernetes.io/projected/4404e0f3-9f58-4639-9045-afdcc964e742-kube-api-access-jjcjt\") pod \"reboot-os-openstack-openstack-cell1-jrtwp\" (UID: \"4404e0f3-9f58-4639-9045-afdcc964e742\") " pod="openstack/reboot-os-openstack-openstack-cell1-jrtwp" Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.773045 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4404e0f3-9f58-4639-9045-afdcc964e742-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-jrtwp\" (UID: \"4404e0f3-9f58-4639-9045-afdcc964e742\") " pod="openstack/reboot-os-openstack-openstack-cell1-jrtwp" Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.773801 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4404e0f3-9f58-4639-9045-afdcc964e742-inventory\") pod \"reboot-os-openstack-openstack-cell1-jrtwp\" (UID: \"4404e0f3-9f58-4639-9045-afdcc964e742\") " pod="openstack/reboot-os-openstack-openstack-cell1-jrtwp" Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.875646 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjcjt\" (UniqueName: \"kubernetes.io/projected/4404e0f3-9f58-4639-9045-afdcc964e742-kube-api-access-jjcjt\") pod \"reboot-os-openstack-openstack-cell1-jrtwp\" (UID: \"4404e0f3-9f58-4639-9045-afdcc964e742\") " pod="openstack/reboot-os-openstack-openstack-cell1-jrtwp" Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.875771 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4404e0f3-9f58-4639-9045-afdcc964e742-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-jrtwp\" (UID: \"4404e0f3-9f58-4639-9045-afdcc964e742\") " pod="openstack/reboot-os-openstack-openstack-cell1-jrtwp" Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.875950 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4404e0f3-9f58-4639-9045-afdcc964e742-inventory\") pod \"reboot-os-openstack-openstack-cell1-jrtwp\" (UID: \"4404e0f3-9f58-4639-9045-afdcc964e742\") " pod="openstack/reboot-os-openstack-openstack-cell1-jrtwp" Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.880092 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4404e0f3-9f58-4639-9045-afdcc964e742-inventory\") pod \"reboot-os-openstack-openstack-cell1-jrtwp\" (UID: \"4404e0f3-9f58-4639-9045-afdcc964e742\") " pod="openstack/reboot-os-openstack-openstack-cell1-jrtwp" Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.888670 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4404e0f3-9f58-4639-9045-afdcc964e742-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-jrtwp\" (UID: \"4404e0f3-9f58-4639-9045-afdcc964e742\") " pod="openstack/reboot-os-openstack-openstack-cell1-jrtwp" Nov 25 11:35:05 crc kubenswrapper[4776]: I1125 11:35:05.893367 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjcjt\" (UniqueName: \"kubernetes.io/projected/4404e0f3-9f58-4639-9045-afdcc964e742-kube-api-access-jjcjt\") pod \"reboot-os-openstack-openstack-cell1-jrtwp\" (UID: \"4404e0f3-9f58-4639-9045-afdcc964e742\") " pod="openstack/reboot-os-openstack-openstack-cell1-jrtwp" Nov 25 11:35:06 crc kubenswrapper[4776]: I1125 11:35:06.054302 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-jrtwp" Nov 25 11:35:06 crc kubenswrapper[4776]: I1125 11:35:06.605128 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-jrtwp"] Nov 25 11:35:06 crc kubenswrapper[4776]: W1125 11:35:06.630404 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4404e0f3_9f58_4639_9045_afdcc964e742.slice/crio-1009895c9222cf9919c6e04b85cdd919bdf888fcd41d7511c6901e3f83379806 WatchSource:0}: Error finding container 1009895c9222cf9919c6e04b85cdd919bdf888fcd41d7511c6901e3f83379806: Status 404 returned error can't find the container with id 1009895c9222cf9919c6e04b85cdd919bdf888fcd41d7511c6901e3f83379806 Nov 25 11:35:06 crc kubenswrapper[4776]: I1125 11:35:06.658457 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-jrtwp" event={"ID":"4404e0f3-9f58-4639-9045-afdcc964e742","Type":"ContainerStarted","Data":"1009895c9222cf9919c6e04b85cdd919bdf888fcd41d7511c6901e3f83379806"} Nov 25 11:35:07 crc kubenswrapper[4776]: I1125 11:35:07.690209 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-jrtwp" event={"ID":"4404e0f3-9f58-4639-9045-afdcc964e742","Type":"ContainerStarted","Data":"e26eeeba1fe0a8232a7875138e89eca0717c3e3ecfc4bab8758ac65ccd34b483"} Nov 25 11:35:07 crc kubenswrapper[4776]: I1125 11:35:07.695411 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-cell1-jrtwp" podStartSLOduration=2.069672075 podStartE2EDuration="2.695391384s" podCreationTimestamp="2025-11-25 11:35:05 +0000 UTC" firstStartedPulling="2025-11-25 11:35:06.636857355 +0000 UTC m=+7851.677916908" lastFinishedPulling="2025-11-25 11:35:07.262576664 +0000 UTC m=+7852.303636217" observedRunningTime="2025-11-25 11:35:07.694003929 +0000 UTC m=+7852.735063482" watchObservedRunningTime="2025-11-25 11:35:07.695391384 +0000 UTC m=+7852.736450937" Nov 25 11:35:23 crc kubenswrapper[4776]: I1125 11:35:23.852961 4776 generic.go:334] "Generic (PLEG): container finished" podID="4404e0f3-9f58-4639-9045-afdcc964e742" containerID="e26eeeba1fe0a8232a7875138e89eca0717c3e3ecfc4bab8758ac65ccd34b483" exitCode=0 Nov 25 11:35:23 crc kubenswrapper[4776]: I1125 11:35:23.853038 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-jrtwp" event={"ID":"4404e0f3-9f58-4639-9045-afdcc964e742","Type":"ContainerDied","Data":"e26eeeba1fe0a8232a7875138e89eca0717c3e3ecfc4bab8758ac65ccd34b483"} Nov 25 11:35:25 crc kubenswrapper[4776]: I1125 11:35:25.337212 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-jrtwp" Nov 25 11:35:25 crc kubenswrapper[4776]: I1125 11:35:25.454431 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjcjt\" (UniqueName: \"kubernetes.io/projected/4404e0f3-9f58-4639-9045-afdcc964e742-kube-api-access-jjcjt\") pod \"4404e0f3-9f58-4639-9045-afdcc964e742\" (UID: \"4404e0f3-9f58-4639-9045-afdcc964e742\") " Nov 25 11:35:25 crc kubenswrapper[4776]: I1125 11:35:25.454531 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4404e0f3-9f58-4639-9045-afdcc964e742-ssh-key\") pod \"4404e0f3-9f58-4639-9045-afdcc964e742\" (UID: \"4404e0f3-9f58-4639-9045-afdcc964e742\") " Nov 25 11:35:25 crc kubenswrapper[4776]: I1125 11:35:25.454661 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4404e0f3-9f58-4639-9045-afdcc964e742-inventory\") pod \"4404e0f3-9f58-4639-9045-afdcc964e742\" (UID: \"4404e0f3-9f58-4639-9045-afdcc964e742\") " Nov 25 11:35:25 crc kubenswrapper[4776]: I1125 11:35:25.465448 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4404e0f3-9f58-4639-9045-afdcc964e742-kube-api-access-jjcjt" (OuterVolumeSpecName: "kube-api-access-jjcjt") pod "4404e0f3-9f58-4639-9045-afdcc964e742" (UID: "4404e0f3-9f58-4639-9045-afdcc964e742"). InnerVolumeSpecName "kube-api-access-jjcjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:35:25 crc kubenswrapper[4776]: I1125 11:35:25.493666 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4404e0f3-9f58-4639-9045-afdcc964e742-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4404e0f3-9f58-4639-9045-afdcc964e742" (UID: "4404e0f3-9f58-4639-9045-afdcc964e742"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:35:25 crc kubenswrapper[4776]: I1125 11:35:25.503183 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4404e0f3-9f58-4639-9045-afdcc964e742-inventory" (OuterVolumeSpecName: "inventory") pod "4404e0f3-9f58-4639-9045-afdcc964e742" (UID: "4404e0f3-9f58-4639-9045-afdcc964e742"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:35:25 crc kubenswrapper[4776]: I1125 11:35:25.558223 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjcjt\" (UniqueName: \"kubernetes.io/projected/4404e0f3-9f58-4639-9045-afdcc964e742-kube-api-access-jjcjt\") on node \"crc\" DevicePath \"\"" Nov 25 11:35:25 crc kubenswrapper[4776]: I1125 11:35:25.558262 4776 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4404e0f3-9f58-4639-9045-afdcc964e742-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:35:25 crc kubenswrapper[4776]: I1125 11:35:25.558274 4776 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4404e0f3-9f58-4639-9045-afdcc964e742-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:35:25 crc kubenswrapper[4776]: I1125 11:35:25.875991 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-jrtwp" event={"ID":"4404e0f3-9f58-4639-9045-afdcc964e742","Type":"ContainerDied","Data":"1009895c9222cf9919c6e04b85cdd919bdf888fcd41d7511c6901e3f83379806"} Nov 25 11:35:25 crc kubenswrapper[4776]: I1125 11:35:25.876385 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1009895c9222cf9919c6e04b85cdd919bdf888fcd41d7511c6901e3f83379806" Nov 25 11:35:25 crc kubenswrapper[4776]: I1125 11:35:25.876082 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-jrtwp" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.013553 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-bhxjr"] Nov 25 11:35:26 crc kubenswrapper[4776]: E1125 11:35:26.013983 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4404e0f3-9f58-4639-9045-afdcc964e742" containerName="reboot-os-openstack-openstack-cell1" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.013999 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="4404e0f3-9f58-4639-9045-afdcc964e742" containerName="reboot-os-openstack-openstack-cell1" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.014224 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="4404e0f3-9f58-4639-9045-afdcc964e742" containerName="reboot-os-openstack-openstack-cell1" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.015019 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.018674 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-ovn-default-certs-0" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.019001 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-libvirt-default-certs-0" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.019003 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.019047 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-neutron-metadata-default-certs-0" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.019188 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dknh7" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.019113 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.019330 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.020602 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-telemetry-default-certs-0" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.027866 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-bhxjr"] Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.067270 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6x8d\" (UniqueName: \"kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-kube-api-access-v6x8d\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.067840 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.067894 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.067929 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.067969 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-openstack-cell1-libvirt-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.068198 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-openstack-cell1-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.068269 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-inventory\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.068443 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.068592 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-openstack-cell1-telemetry-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.068765 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.068927 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.068985 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.069016 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.069109 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-openstack-cell1-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.069217 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-ssh-key\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.171680 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.171750 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.171786 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.171809 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.171882 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-openstack-cell1-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.171948 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-ssh-key\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.172028 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6x8d\" (UniqueName: \"kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-kube-api-access-v6x8d\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.172112 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.172168 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.172200 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.172242 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-openstack-cell1-libvirt-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.172290 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-openstack-cell1-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.172316 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-inventory\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.172370 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.172430 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-openstack-cell1-telemetry-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.177003 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-ssh-key\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.178200 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.179413 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.182349 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-inventory\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.183672 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.188200 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-openstack-cell1-libvirt-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.188871 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-openstack-cell1-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.189061 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.189338 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.189688 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.189698 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.190700 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-openstack-cell1-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.191377 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.192027 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-openstack-cell1-telemetry-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.193003 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6x8d\" (UniqueName: \"kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-kube-api-access-v6x8d\") pod \"install-certs-openstack-openstack-cell1-bhxjr\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.369138 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:35:26 crc kubenswrapper[4776]: I1125 11:35:26.996220 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-bhxjr"] Nov 25 11:35:27 crc kubenswrapper[4776]: I1125 11:35:27.897623 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" event={"ID":"4af7b0e8-aca0-4803-9d2c-0502406152e0","Type":"ContainerStarted","Data":"a3efe91bb85b81da95c808e83ec3e4ed3417f23588f3374064d04df2ae56dd70"} Nov 25 11:35:27 crc kubenswrapper[4776]: I1125 11:35:27.897983 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" event={"ID":"4af7b0e8-aca0-4803-9d2c-0502406152e0","Type":"ContainerStarted","Data":"ea557c4d15a0d18738bb2f7d345b0a4a4285eee4888994c9c919cb1869449e58"} Nov 25 11:35:27 crc kubenswrapper[4776]: I1125 11:35:27.952300 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" podStartSLOduration=2.437376127 podStartE2EDuration="2.952276526s" podCreationTimestamp="2025-11-25 11:35:25 +0000 UTC" firstStartedPulling="2025-11-25 11:35:27.021371629 +0000 UTC m=+7872.062431172" lastFinishedPulling="2025-11-25 11:35:27.536272008 +0000 UTC m=+7872.577331571" observedRunningTime="2025-11-25 11:35:27.92454386 +0000 UTC m=+7872.965603413" watchObservedRunningTime="2025-11-25 11:35:27.952276526 +0000 UTC m=+7872.993336079" Nov 25 11:35:47 crc kubenswrapper[4776]: I1125 11:35:47.817916 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:35:47 crc kubenswrapper[4776]: I1125 11:35:47.818502 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:36:03 crc kubenswrapper[4776]: I1125 11:36:03.430245 4776 generic.go:334] "Generic (PLEG): container finished" podID="4af7b0e8-aca0-4803-9d2c-0502406152e0" containerID="a3efe91bb85b81da95c808e83ec3e4ed3417f23588f3374064d04df2ae56dd70" exitCode=0 Nov 25 11:36:03 crc kubenswrapper[4776]: I1125 11:36:03.430326 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" event={"ID":"4af7b0e8-aca0-4803-9d2c-0502406152e0","Type":"ContainerDied","Data":"a3efe91bb85b81da95c808e83ec3e4ed3417f23588f3374064d04df2ae56dd70"} Nov 25 11:36:04 crc kubenswrapper[4776]: I1125 11:36:04.902138 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.012217 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6x8d\" (UniqueName: \"kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-kube-api-access-v6x8d\") pod \"4af7b0e8-aca0-4803-9d2c-0502406152e0\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.012594 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-nova-combined-ca-bundle\") pod \"4af7b0e8-aca0-4803-9d2c-0502406152e0\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.012626 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-openstack-cell1-telemetry-default-certs-0\") pod \"4af7b0e8-aca0-4803-9d2c-0502406152e0\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.012653 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-neutron-metadata-combined-ca-bundle\") pod \"4af7b0e8-aca0-4803-9d2c-0502406152e0\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.012730 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-openstack-cell1-neutron-metadata-default-certs-0\") pod \"4af7b0e8-aca0-4803-9d2c-0502406152e0\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.012777 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-ovn-combined-ca-bundle\") pod \"4af7b0e8-aca0-4803-9d2c-0502406152e0\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.012804 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-ssh-key\") pod \"4af7b0e8-aca0-4803-9d2c-0502406152e0\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.012837 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-libvirt-combined-ca-bundle\") pod \"4af7b0e8-aca0-4803-9d2c-0502406152e0\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.012887 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-openstack-cell1-ovn-default-certs-0\") pod \"4af7b0e8-aca0-4803-9d2c-0502406152e0\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.012969 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-inventory\") pod \"4af7b0e8-aca0-4803-9d2c-0502406152e0\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.013036 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-telemetry-combined-ca-bundle\") pod \"4af7b0e8-aca0-4803-9d2c-0502406152e0\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.013082 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-openstack-cell1-libvirt-default-certs-0\") pod \"4af7b0e8-aca0-4803-9d2c-0502406152e0\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.013134 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-bootstrap-combined-ca-bundle\") pod \"4af7b0e8-aca0-4803-9d2c-0502406152e0\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.013164 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-neutron-dhcp-combined-ca-bundle\") pod \"4af7b0e8-aca0-4803-9d2c-0502406152e0\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.013186 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-neutron-sriov-combined-ca-bundle\") pod \"4af7b0e8-aca0-4803-9d2c-0502406152e0\" (UID: \"4af7b0e8-aca0-4803-9d2c-0502406152e0\") " Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.021161 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-openstack-cell1-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-libvirt-default-certs-0") pod "4af7b0e8-aca0-4803-9d2c-0502406152e0" (UID: "4af7b0e8-aca0-4803-9d2c-0502406152e0"). InnerVolumeSpecName "openstack-cell1-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.022642 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "4af7b0e8-aca0-4803-9d2c-0502406152e0" (UID: "4af7b0e8-aca0-4803-9d2c-0502406152e0"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.022904 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-openstack-cell1-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-telemetry-default-certs-0") pod "4af7b0e8-aca0-4803-9d2c-0502406152e0" (UID: "4af7b0e8-aca0-4803-9d2c-0502406152e0"). InnerVolumeSpecName "openstack-cell1-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.022938 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "4af7b0e8-aca0-4803-9d2c-0502406152e0" (UID: "4af7b0e8-aca0-4803-9d2c-0502406152e0"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.023003 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-openstack-cell1-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-neutron-metadata-default-certs-0") pod "4af7b0e8-aca0-4803-9d2c-0502406152e0" (UID: "4af7b0e8-aca0-4803-9d2c-0502406152e0"). InnerVolumeSpecName "openstack-cell1-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.024365 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "4af7b0e8-aca0-4803-9d2c-0502406152e0" (UID: "4af7b0e8-aca0-4803-9d2c-0502406152e0"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.024458 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "4af7b0e8-aca0-4803-9d2c-0502406152e0" (UID: "4af7b0e8-aca0-4803-9d2c-0502406152e0"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.025027 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "4af7b0e8-aca0-4803-9d2c-0502406152e0" (UID: "4af7b0e8-aca0-4803-9d2c-0502406152e0"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.025830 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "4af7b0e8-aca0-4803-9d2c-0502406152e0" (UID: "4af7b0e8-aca0-4803-9d2c-0502406152e0"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.032764 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-openstack-cell1-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-ovn-default-certs-0") pod "4af7b0e8-aca0-4803-9d2c-0502406152e0" (UID: "4af7b0e8-aca0-4803-9d2c-0502406152e0"). InnerVolumeSpecName "openstack-cell1-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.042236 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "4af7b0e8-aca0-4803-9d2c-0502406152e0" (UID: "4af7b0e8-aca0-4803-9d2c-0502406152e0"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.042432 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-kube-api-access-v6x8d" (OuterVolumeSpecName: "kube-api-access-v6x8d") pod "4af7b0e8-aca0-4803-9d2c-0502406152e0" (UID: "4af7b0e8-aca0-4803-9d2c-0502406152e0"). InnerVolumeSpecName "kube-api-access-v6x8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.044748 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "4af7b0e8-aca0-4803-9d2c-0502406152e0" (UID: "4af7b0e8-aca0-4803-9d2c-0502406152e0"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.055789 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4af7b0e8-aca0-4803-9d2c-0502406152e0" (UID: "4af7b0e8-aca0-4803-9d2c-0502406152e0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.058945 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-inventory" (OuterVolumeSpecName: "inventory") pod "4af7b0e8-aca0-4803-9d2c-0502406152e0" (UID: "4af7b0e8-aca0-4803-9d2c-0502406152e0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.115445 4776 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.115479 4776 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.115490 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6x8d\" (UniqueName: \"kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-kube-api-access-v6x8d\") on node \"crc\" DevicePath \"\"" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.115499 4776 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.115509 4776 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-openstack-cell1-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.115520 4776 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.115532 4776 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-openstack-cell1-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.115541 4776 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.115551 4776 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.115558 4776 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.115569 4776 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-openstack-cell1-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.115577 4776 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.115586 4776 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.115597 4776 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4af7b0e8-aca0-4803-9d2c-0502406152e0-openstack-cell1-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.115605 4776 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4af7b0e8-aca0-4803-9d2c-0502406152e0-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.451693 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" event={"ID":"4af7b0e8-aca0-4803-9d2c-0502406152e0","Type":"ContainerDied","Data":"ea557c4d15a0d18738bb2f7d345b0a4a4285eee4888994c9c919cb1869449e58"} Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.451968 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea557c4d15a0d18738bb2f7d345b0a4a4285eee4888994c9c919cb1869449e58" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.451796 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-bhxjr" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.608717 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-cell1-jppkj"] Nov 25 11:36:05 crc kubenswrapper[4776]: E1125 11:36:05.609210 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4af7b0e8-aca0-4803-9d2c-0502406152e0" containerName="install-certs-openstack-openstack-cell1" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.609229 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="4af7b0e8-aca0-4803-9d2c-0502406152e0" containerName="install-certs-openstack-openstack-cell1" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.609438 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="4af7b0e8-aca0-4803-9d2c-0502406152e0" containerName="install-certs-openstack-openstack-cell1" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.610284 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-jppkj" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.617347 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.617373 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.617705 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dknh7" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.617844 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.617851 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.622531 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-jppkj"] Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.624933 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6lzr\" (UniqueName: \"kubernetes.io/projected/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-kube-api-access-b6lzr\") pod \"ovn-openstack-openstack-cell1-jppkj\" (UID: \"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4\") " pod="openstack/ovn-openstack-openstack-cell1-jppkj" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.625028 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-ssh-key\") pod \"ovn-openstack-openstack-cell1-jppkj\" (UID: \"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4\") " pod="openstack/ovn-openstack-openstack-cell1-jppkj" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.625062 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-jppkj\" (UID: \"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4\") " pod="openstack/ovn-openstack-openstack-cell1-jppkj" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.625131 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-inventory\") pod \"ovn-openstack-openstack-cell1-jppkj\" (UID: \"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4\") " pod="openstack/ovn-openstack-openstack-cell1-jppkj" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.625477 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-jppkj\" (UID: \"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4\") " pod="openstack/ovn-openstack-openstack-cell1-jppkj" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.728851 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-jppkj\" (UID: \"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4\") " pod="openstack/ovn-openstack-openstack-cell1-jppkj" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.729050 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6lzr\" (UniqueName: \"kubernetes.io/projected/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-kube-api-access-b6lzr\") pod \"ovn-openstack-openstack-cell1-jppkj\" (UID: \"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4\") " pod="openstack/ovn-openstack-openstack-cell1-jppkj" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.729139 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-ssh-key\") pod \"ovn-openstack-openstack-cell1-jppkj\" (UID: \"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4\") " pod="openstack/ovn-openstack-openstack-cell1-jppkj" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.729174 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-jppkj\" (UID: \"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4\") " pod="openstack/ovn-openstack-openstack-cell1-jppkj" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.729223 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-inventory\") pod \"ovn-openstack-openstack-cell1-jppkj\" (UID: \"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4\") " pod="openstack/ovn-openstack-openstack-cell1-jppkj" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.733789 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-jppkj\" (UID: \"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4\") " pod="openstack/ovn-openstack-openstack-cell1-jppkj" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.734521 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-ssh-key\") pod \"ovn-openstack-openstack-cell1-jppkj\" (UID: \"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4\") " pod="openstack/ovn-openstack-openstack-cell1-jppkj" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.736830 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-inventory\") pod \"ovn-openstack-openstack-cell1-jppkj\" (UID: \"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4\") " pod="openstack/ovn-openstack-openstack-cell1-jppkj" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.745720 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-jppkj\" (UID: \"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4\") " pod="openstack/ovn-openstack-openstack-cell1-jppkj" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.748476 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6lzr\" (UniqueName: \"kubernetes.io/projected/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-kube-api-access-b6lzr\") pod \"ovn-openstack-openstack-cell1-jppkj\" (UID: \"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4\") " pod="openstack/ovn-openstack-openstack-cell1-jppkj" Nov 25 11:36:05 crc kubenswrapper[4776]: I1125 11:36:05.947587 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-jppkj" Nov 25 11:36:06 crc kubenswrapper[4776]: I1125 11:36:06.534728 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-jppkj"] Nov 25 11:36:07 crc kubenswrapper[4776]: I1125 11:36:07.475308 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-jppkj" event={"ID":"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4","Type":"ContainerStarted","Data":"3fa9f9b0a1adedc2ecb6d375cce19369b18fe6c81eb6b7e65c6b5e5e25058e51"} Nov 25 11:36:07 crc kubenswrapper[4776]: I1125 11:36:07.475596 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-jppkj" event={"ID":"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4","Type":"ContainerStarted","Data":"e01a45921e7432dff991969c2304f62d33b329c7c6c46cde5343170244d41ab4"} Nov 25 11:36:07 crc kubenswrapper[4776]: I1125 11:36:07.494147 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-cell1-jppkj" podStartSLOduration=2.005313949 podStartE2EDuration="2.494128714s" podCreationTimestamp="2025-11-25 11:36:05 +0000 UTC" firstStartedPulling="2025-11-25 11:36:06.536048145 +0000 UTC m=+7911.577107708" lastFinishedPulling="2025-11-25 11:36:07.02486291 +0000 UTC m=+7912.065922473" observedRunningTime="2025-11-25 11:36:07.490045682 +0000 UTC m=+7912.531105245" watchObservedRunningTime="2025-11-25 11:36:07.494128714 +0000 UTC m=+7912.535188267" Nov 25 11:36:17 crc kubenswrapper[4776]: I1125 11:36:17.818302 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:36:17 crc kubenswrapper[4776]: I1125 11:36:17.819120 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:36:47 crc kubenswrapper[4776]: I1125 11:36:47.818267 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:36:47 crc kubenswrapper[4776]: I1125 11:36:47.818888 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:36:47 crc kubenswrapper[4776]: I1125 11:36:47.818941 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 11:36:47 crc kubenswrapper[4776]: I1125 11:36:47.819963 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cced90c452d27e661d744bc42ceb0710b870fc7135d4b27a5d442a29ab283935"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 11:36:47 crc kubenswrapper[4776]: I1125 11:36:47.820021 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://cced90c452d27e661d744bc42ceb0710b870fc7135d4b27a5d442a29ab283935" gracePeriod=600 Nov 25 11:36:48 crc kubenswrapper[4776]: I1125 11:36:48.895343 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="cced90c452d27e661d744bc42ceb0710b870fc7135d4b27a5d442a29ab283935" exitCode=0 Nov 25 11:36:48 crc kubenswrapper[4776]: I1125 11:36:48.895443 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"cced90c452d27e661d744bc42ceb0710b870fc7135d4b27a5d442a29ab283935"} Nov 25 11:36:48 crc kubenswrapper[4776]: I1125 11:36:48.895875 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b"} Nov 25 11:36:48 crc kubenswrapper[4776]: I1125 11:36:48.895911 4776 scope.go:117] "RemoveContainer" containerID="1d708a2faa34d54c477db2f5dabe6e1a26af6f350f603015d32accb30fc3515c" Nov 25 11:37:10 crc kubenswrapper[4776]: I1125 11:37:10.157985 4776 generic.go:334] "Generic (PLEG): container finished" podID="458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4" containerID="3fa9f9b0a1adedc2ecb6d375cce19369b18fe6c81eb6b7e65c6b5e5e25058e51" exitCode=0 Nov 25 11:37:10 crc kubenswrapper[4776]: I1125 11:37:10.158174 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-jppkj" event={"ID":"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4","Type":"ContainerDied","Data":"3fa9f9b0a1adedc2ecb6d375cce19369b18fe6c81eb6b7e65c6b5e5e25058e51"} Nov 25 11:37:11 crc kubenswrapper[4776]: I1125 11:37:11.734825 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-jppkj" Nov 25 11:37:11 crc kubenswrapper[4776]: I1125 11:37:11.844306 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6lzr\" (UniqueName: \"kubernetes.io/projected/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-kube-api-access-b6lzr\") pod \"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4\" (UID: \"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4\") " Nov 25 11:37:11 crc kubenswrapper[4776]: I1125 11:37:11.844378 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-ovncontroller-config-0\") pod \"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4\" (UID: \"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4\") " Nov 25 11:37:11 crc kubenswrapper[4776]: I1125 11:37:11.844528 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-inventory\") pod \"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4\" (UID: \"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4\") " Nov 25 11:37:11 crc kubenswrapper[4776]: I1125 11:37:11.844555 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-ssh-key\") pod \"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4\" (UID: \"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4\") " Nov 25 11:37:11 crc kubenswrapper[4776]: I1125 11:37:11.844738 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-ovn-combined-ca-bundle\") pod \"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4\" (UID: \"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4\") " Nov 25 11:37:11 crc kubenswrapper[4776]: I1125 11:37:11.850848 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4" (UID: "458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:37:11 crc kubenswrapper[4776]: I1125 11:37:11.851358 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-kube-api-access-b6lzr" (OuterVolumeSpecName: "kube-api-access-b6lzr") pod "458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4" (UID: "458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4"). InnerVolumeSpecName "kube-api-access-b6lzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:37:11 crc kubenswrapper[4776]: I1125 11:37:11.873038 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4" (UID: "458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:37:11 crc kubenswrapper[4776]: I1125 11:37:11.877690 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4" (UID: "458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:37:11 crc kubenswrapper[4776]: I1125 11:37:11.878048 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-inventory" (OuterVolumeSpecName: "inventory") pod "458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4" (UID: "458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:37:11 crc kubenswrapper[4776]: I1125 11:37:11.948020 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6lzr\" (UniqueName: \"kubernetes.io/projected/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-kube-api-access-b6lzr\") on node \"crc\" DevicePath \"\"" Nov 25 11:37:11 crc kubenswrapper[4776]: I1125 11:37:11.948116 4776 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:37:11 crc kubenswrapper[4776]: I1125 11:37:11.948128 4776 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:37:11 crc kubenswrapper[4776]: I1125 11:37:11.948137 4776 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:37:11 crc kubenswrapper[4776]: I1125 11:37:11.948146 4776 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.178877 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-jppkj" event={"ID":"458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4","Type":"ContainerDied","Data":"e01a45921e7432dff991969c2304f62d33b329c7c6c46cde5343170244d41ab4"} Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.178929 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e01a45921e7432dff991969c2304f62d33b329c7c6c46cde5343170244d41ab4" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.178951 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-jppkj" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.306528 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-nzsns"] Nov 25 11:37:12 crc kubenswrapper[4776]: E1125 11:37:12.307469 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4" containerName="ovn-openstack-openstack-cell1" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.307494 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4" containerName="ovn-openstack-openstack-cell1" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.307762 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4" containerName="ovn-openstack-openstack-cell1" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.308720 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-nzsns" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.311840 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.312039 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.312244 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.312483 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.312879 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.312829 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dknh7" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.319115 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-nzsns"] Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.457884 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-nzsns\" (UID: \"795a3ef2-55b2-4550-a681-12b17cb26dad\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-nzsns" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.457967 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-nzsns\" (UID: \"795a3ef2-55b2-4550-a681-12b17cb26dad\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-nzsns" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.457991 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdqmx\" (UniqueName: \"kubernetes.io/projected/795a3ef2-55b2-4550-a681-12b17cb26dad-kube-api-access-fdqmx\") pod \"neutron-metadata-openstack-openstack-cell1-nzsns\" (UID: \"795a3ef2-55b2-4550-a681-12b17cb26dad\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-nzsns" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.458122 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-nzsns\" (UID: \"795a3ef2-55b2-4550-a681-12b17cb26dad\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-nzsns" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.458156 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-nzsns\" (UID: \"795a3ef2-55b2-4550-a681-12b17cb26dad\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-nzsns" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.458225 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-nzsns\" (UID: \"795a3ef2-55b2-4550-a681-12b17cb26dad\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-nzsns" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.561988 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-nzsns\" (UID: \"795a3ef2-55b2-4550-a681-12b17cb26dad\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-nzsns" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.562053 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdqmx\" (UniqueName: \"kubernetes.io/projected/795a3ef2-55b2-4550-a681-12b17cb26dad-kube-api-access-fdqmx\") pod \"neutron-metadata-openstack-openstack-cell1-nzsns\" (UID: \"795a3ef2-55b2-4550-a681-12b17cb26dad\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-nzsns" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.562655 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-nzsns\" (UID: \"795a3ef2-55b2-4550-a681-12b17cb26dad\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-nzsns" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.562709 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-nzsns\" (UID: \"795a3ef2-55b2-4550-a681-12b17cb26dad\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-nzsns" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.562811 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-nzsns\" (UID: \"795a3ef2-55b2-4550-a681-12b17cb26dad\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-nzsns" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.562928 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-nzsns\" (UID: \"795a3ef2-55b2-4550-a681-12b17cb26dad\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-nzsns" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.566372 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-nzsns\" (UID: \"795a3ef2-55b2-4550-a681-12b17cb26dad\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-nzsns" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.567763 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-nzsns\" (UID: \"795a3ef2-55b2-4550-a681-12b17cb26dad\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-nzsns" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.568556 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-nzsns\" (UID: \"795a3ef2-55b2-4550-a681-12b17cb26dad\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-nzsns" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.570307 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-nzsns\" (UID: \"795a3ef2-55b2-4550-a681-12b17cb26dad\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-nzsns" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.577378 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-nzsns\" (UID: \"795a3ef2-55b2-4550-a681-12b17cb26dad\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-nzsns" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.585876 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdqmx\" (UniqueName: \"kubernetes.io/projected/795a3ef2-55b2-4550-a681-12b17cb26dad-kube-api-access-fdqmx\") pod \"neutron-metadata-openstack-openstack-cell1-nzsns\" (UID: \"795a3ef2-55b2-4550-a681-12b17cb26dad\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-nzsns" Nov 25 11:37:12 crc kubenswrapper[4776]: I1125 11:37:12.636446 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-nzsns" Nov 25 11:37:13 crc kubenswrapper[4776]: I1125 11:37:13.227396 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-nzsns"] Nov 25 11:37:14 crc kubenswrapper[4776]: I1125 11:37:14.202096 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-nzsns" event={"ID":"795a3ef2-55b2-4550-a681-12b17cb26dad","Type":"ContainerStarted","Data":"c042df3d79039fa3456ab0d3d332093c33b3b8424ded22c69c5f4240bb733433"} Nov 25 11:37:14 crc kubenswrapper[4776]: I1125 11:37:14.202861 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-nzsns" event={"ID":"795a3ef2-55b2-4550-a681-12b17cb26dad","Type":"ContainerStarted","Data":"41afbbad82e3e675e5a7657a61454b743d51d14fd4a064fb53fc00fb10d45096"} Nov 25 11:38:04 crc kubenswrapper[4776]: I1125 11:38:04.699862 4776 generic.go:334] "Generic (PLEG): container finished" podID="795a3ef2-55b2-4550-a681-12b17cb26dad" containerID="c042df3d79039fa3456ab0d3d332093c33b3b8424ded22c69c5f4240bb733433" exitCode=0 Nov 25 11:38:04 crc kubenswrapper[4776]: I1125 11:38:04.699959 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-nzsns" event={"ID":"795a3ef2-55b2-4550-a681-12b17cb26dad","Type":"ContainerDied","Data":"c042df3d79039fa3456ab0d3d332093c33b3b8424ded22c69c5f4240bb733433"} Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.220638 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-nzsns" Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.276033 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-inventory\") pod \"795a3ef2-55b2-4550-a681-12b17cb26dad\" (UID: \"795a3ef2-55b2-4550-a681-12b17cb26dad\") " Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.309608 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-inventory" (OuterVolumeSpecName: "inventory") pod "795a3ef2-55b2-4550-a681-12b17cb26dad" (UID: "795a3ef2-55b2-4550-a681-12b17cb26dad"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.378654 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-nova-metadata-neutron-config-0\") pod \"795a3ef2-55b2-4550-a681-12b17cb26dad\" (UID: \"795a3ef2-55b2-4550-a681-12b17cb26dad\") " Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.379011 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-ssh-key\") pod \"795a3ef2-55b2-4550-a681-12b17cb26dad\" (UID: \"795a3ef2-55b2-4550-a681-12b17cb26dad\") " Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.379412 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdqmx\" (UniqueName: \"kubernetes.io/projected/795a3ef2-55b2-4550-a681-12b17cb26dad-kube-api-access-fdqmx\") pod \"795a3ef2-55b2-4550-a681-12b17cb26dad\" (UID: \"795a3ef2-55b2-4550-a681-12b17cb26dad\") " Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.379935 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-neutron-metadata-combined-ca-bundle\") pod \"795a3ef2-55b2-4550-a681-12b17cb26dad\" (UID: \"795a3ef2-55b2-4550-a681-12b17cb26dad\") " Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.379993 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-neutron-ovn-metadata-agent-neutron-config-0\") pod \"795a3ef2-55b2-4550-a681-12b17cb26dad\" (UID: \"795a3ef2-55b2-4550-a681-12b17cb26dad\") " Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.381181 4776 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.384140 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "795a3ef2-55b2-4550-a681-12b17cb26dad" (UID: "795a3ef2-55b2-4550-a681-12b17cb26dad"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.384696 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/795a3ef2-55b2-4550-a681-12b17cb26dad-kube-api-access-fdqmx" (OuterVolumeSpecName: "kube-api-access-fdqmx") pod "795a3ef2-55b2-4550-a681-12b17cb26dad" (UID: "795a3ef2-55b2-4550-a681-12b17cb26dad"). InnerVolumeSpecName "kube-api-access-fdqmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.412270 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "795a3ef2-55b2-4550-a681-12b17cb26dad" (UID: "795a3ef2-55b2-4550-a681-12b17cb26dad"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.415306 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "795a3ef2-55b2-4550-a681-12b17cb26dad" (UID: "795a3ef2-55b2-4550-a681-12b17cb26dad"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.418000 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "795a3ef2-55b2-4550-a681-12b17cb26dad" (UID: "795a3ef2-55b2-4550-a681-12b17cb26dad"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.482267 4776 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.482305 4776 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.482320 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdqmx\" (UniqueName: \"kubernetes.io/projected/795a3ef2-55b2-4550-a681-12b17cb26dad-kube-api-access-fdqmx\") on node \"crc\" DevicePath \"\"" Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.482333 4776 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.482346 4776 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/795a3ef2-55b2-4550-a681-12b17cb26dad-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.724900 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-nzsns" event={"ID":"795a3ef2-55b2-4550-a681-12b17cb26dad","Type":"ContainerDied","Data":"41afbbad82e3e675e5a7657a61454b743d51d14fd4a064fb53fc00fb10d45096"} Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.724943 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41afbbad82e3e675e5a7657a61454b743d51d14fd4a064fb53fc00fb10d45096" Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.724999 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-nzsns" Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.821618 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-c2plb"] Nov 25 11:38:06 crc kubenswrapper[4776]: E1125 11:38:06.822269 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="795a3ef2-55b2-4550-a681-12b17cb26dad" containerName="neutron-metadata-openstack-openstack-cell1" Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.822292 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="795a3ef2-55b2-4550-a681-12b17cb26dad" containerName="neutron-metadata-openstack-openstack-cell1" Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.822602 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="795a3ef2-55b2-4550-a681-12b17cb26dad" containerName="neutron-metadata-openstack-openstack-cell1" Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.823612 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-c2plb" Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.825957 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.826328 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dknh7" Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.828012 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.829250 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.829615 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.831469 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-c2plb"] Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.994931 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-ssh-key\") pod \"libvirt-openstack-openstack-cell1-c2plb\" (UID: \"bc20c5f3-b0e5-47e2-8878-015b66dad6f0\") " pod="openstack/libvirt-openstack-openstack-cell1-c2plb" Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.995475 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-inventory\") pod \"libvirt-openstack-openstack-cell1-c2plb\" (UID: \"bc20c5f3-b0e5-47e2-8878-015b66dad6f0\") " pod="openstack/libvirt-openstack-openstack-cell1-c2plb" Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.995515 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4fkw\" (UniqueName: \"kubernetes.io/projected/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-kube-api-access-m4fkw\") pod \"libvirt-openstack-openstack-cell1-c2plb\" (UID: \"bc20c5f3-b0e5-47e2-8878-015b66dad6f0\") " pod="openstack/libvirt-openstack-openstack-cell1-c2plb" Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.995599 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-c2plb\" (UID: \"bc20c5f3-b0e5-47e2-8878-015b66dad6f0\") " pod="openstack/libvirt-openstack-openstack-cell1-c2plb" Nov 25 11:38:06 crc kubenswrapper[4776]: I1125 11:38:06.995692 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-c2plb\" (UID: \"bc20c5f3-b0e5-47e2-8878-015b66dad6f0\") " pod="openstack/libvirt-openstack-openstack-cell1-c2plb" Nov 25 11:38:07 crc kubenswrapper[4776]: I1125 11:38:07.097351 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-c2plb\" (UID: \"bc20c5f3-b0e5-47e2-8878-015b66dad6f0\") " pod="openstack/libvirt-openstack-openstack-cell1-c2plb" Nov 25 11:38:07 crc kubenswrapper[4776]: I1125 11:38:07.097504 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-ssh-key\") pod \"libvirt-openstack-openstack-cell1-c2plb\" (UID: \"bc20c5f3-b0e5-47e2-8878-015b66dad6f0\") " pod="openstack/libvirt-openstack-openstack-cell1-c2plb" Nov 25 11:38:07 crc kubenswrapper[4776]: I1125 11:38:07.097570 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-inventory\") pod \"libvirt-openstack-openstack-cell1-c2plb\" (UID: \"bc20c5f3-b0e5-47e2-8878-015b66dad6f0\") " pod="openstack/libvirt-openstack-openstack-cell1-c2plb" Nov 25 11:38:07 crc kubenswrapper[4776]: I1125 11:38:07.097591 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4fkw\" (UniqueName: \"kubernetes.io/projected/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-kube-api-access-m4fkw\") pod \"libvirt-openstack-openstack-cell1-c2plb\" (UID: \"bc20c5f3-b0e5-47e2-8878-015b66dad6f0\") " pod="openstack/libvirt-openstack-openstack-cell1-c2plb" Nov 25 11:38:07 crc kubenswrapper[4776]: I1125 11:38:07.097624 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-c2plb\" (UID: \"bc20c5f3-b0e5-47e2-8878-015b66dad6f0\") " pod="openstack/libvirt-openstack-openstack-cell1-c2plb" Nov 25 11:38:07 crc kubenswrapper[4776]: I1125 11:38:07.102232 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-ssh-key\") pod \"libvirt-openstack-openstack-cell1-c2plb\" (UID: \"bc20c5f3-b0e5-47e2-8878-015b66dad6f0\") " pod="openstack/libvirt-openstack-openstack-cell1-c2plb" Nov 25 11:38:07 crc kubenswrapper[4776]: I1125 11:38:07.102276 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-c2plb\" (UID: \"bc20c5f3-b0e5-47e2-8878-015b66dad6f0\") " pod="openstack/libvirt-openstack-openstack-cell1-c2plb" Nov 25 11:38:07 crc kubenswrapper[4776]: I1125 11:38:07.111300 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-c2plb\" (UID: \"bc20c5f3-b0e5-47e2-8878-015b66dad6f0\") " pod="openstack/libvirt-openstack-openstack-cell1-c2plb" Nov 25 11:38:07 crc kubenswrapper[4776]: I1125 11:38:07.111502 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-inventory\") pod \"libvirt-openstack-openstack-cell1-c2plb\" (UID: \"bc20c5f3-b0e5-47e2-8878-015b66dad6f0\") " pod="openstack/libvirt-openstack-openstack-cell1-c2plb" Nov 25 11:38:07 crc kubenswrapper[4776]: I1125 11:38:07.119410 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4fkw\" (UniqueName: \"kubernetes.io/projected/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-kube-api-access-m4fkw\") pod \"libvirt-openstack-openstack-cell1-c2plb\" (UID: \"bc20c5f3-b0e5-47e2-8878-015b66dad6f0\") " pod="openstack/libvirt-openstack-openstack-cell1-c2plb" Nov 25 11:38:07 crc kubenswrapper[4776]: I1125 11:38:07.142581 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-c2plb" Nov 25 11:38:07 crc kubenswrapper[4776]: I1125 11:38:07.732391 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-c2plb"] Nov 25 11:38:07 crc kubenswrapper[4776]: I1125 11:38:07.753443 4776 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 11:38:08 crc kubenswrapper[4776]: I1125 11:38:08.756456 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-c2plb" event={"ID":"bc20c5f3-b0e5-47e2-8878-015b66dad6f0","Type":"ContainerStarted","Data":"6460bf3c8dc1b7263aeb3b614a79758110284fc4c65c837f6eb42db240b11502"} Nov 25 11:38:08 crc kubenswrapper[4776]: I1125 11:38:08.756856 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-c2plb" event={"ID":"bc20c5f3-b0e5-47e2-8878-015b66dad6f0","Type":"ContainerStarted","Data":"6daaa3738cf087a97cf513fe5e4b64fcf6898f1714b30c590d3d991a2ed48098"} Nov 25 11:38:08 crc kubenswrapper[4776]: I1125 11:38:08.787373 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-openstack-openstack-cell1-c2plb" podStartSLOduration=2.337108 podStartE2EDuration="2.787354387s" podCreationTimestamp="2025-11-25 11:38:06 +0000 UTC" firstStartedPulling="2025-11-25 11:38:07.752796659 +0000 UTC m=+8032.793856212" lastFinishedPulling="2025-11-25 11:38:08.203043046 +0000 UTC m=+8033.244102599" observedRunningTime="2025-11-25 11:38:08.778772751 +0000 UTC m=+8033.819832304" watchObservedRunningTime="2025-11-25 11:38:08.787354387 +0000 UTC m=+8033.828413940" Nov 25 11:38:37 crc kubenswrapper[4776]: I1125 11:38:37.897885 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-54gpf"] Nov 25 11:38:37 crc kubenswrapper[4776]: I1125 11:38:37.900653 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-54gpf" Nov 25 11:38:37 crc kubenswrapper[4776]: I1125 11:38:37.910582 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-54gpf"] Nov 25 11:38:38 crc kubenswrapper[4776]: I1125 11:38:38.049013 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7066eebe-a597-40d1-b9f1-f6cde8df2666-catalog-content\") pod \"redhat-operators-54gpf\" (UID: \"7066eebe-a597-40d1-b9f1-f6cde8df2666\") " pod="openshift-marketplace/redhat-operators-54gpf" Nov 25 11:38:38 crc kubenswrapper[4776]: I1125 11:38:38.049102 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7066eebe-a597-40d1-b9f1-f6cde8df2666-utilities\") pod \"redhat-operators-54gpf\" (UID: \"7066eebe-a597-40d1-b9f1-f6cde8df2666\") " pod="openshift-marketplace/redhat-operators-54gpf" Nov 25 11:38:38 crc kubenswrapper[4776]: I1125 11:38:38.049275 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxt6g\" (UniqueName: \"kubernetes.io/projected/7066eebe-a597-40d1-b9f1-f6cde8df2666-kube-api-access-lxt6g\") pod \"redhat-operators-54gpf\" (UID: \"7066eebe-a597-40d1-b9f1-f6cde8df2666\") " pod="openshift-marketplace/redhat-operators-54gpf" Nov 25 11:38:38 crc kubenswrapper[4776]: I1125 11:38:38.095978 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rzzlh"] Nov 25 11:38:38 crc kubenswrapper[4776]: I1125 11:38:38.099795 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rzzlh" Nov 25 11:38:38 crc kubenswrapper[4776]: I1125 11:38:38.110544 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rzzlh"] Nov 25 11:38:38 crc kubenswrapper[4776]: I1125 11:38:38.151640 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7066eebe-a597-40d1-b9f1-f6cde8df2666-catalog-content\") pod \"redhat-operators-54gpf\" (UID: \"7066eebe-a597-40d1-b9f1-f6cde8df2666\") " pod="openshift-marketplace/redhat-operators-54gpf" Nov 25 11:38:38 crc kubenswrapper[4776]: I1125 11:38:38.151704 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7066eebe-a597-40d1-b9f1-f6cde8df2666-utilities\") pod \"redhat-operators-54gpf\" (UID: \"7066eebe-a597-40d1-b9f1-f6cde8df2666\") " pod="openshift-marketplace/redhat-operators-54gpf" Nov 25 11:38:38 crc kubenswrapper[4776]: I1125 11:38:38.151860 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxt6g\" (UniqueName: \"kubernetes.io/projected/7066eebe-a597-40d1-b9f1-f6cde8df2666-kube-api-access-lxt6g\") pod \"redhat-operators-54gpf\" (UID: \"7066eebe-a597-40d1-b9f1-f6cde8df2666\") " pod="openshift-marketplace/redhat-operators-54gpf" Nov 25 11:38:38 crc kubenswrapper[4776]: I1125 11:38:38.152249 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7066eebe-a597-40d1-b9f1-f6cde8df2666-catalog-content\") pod \"redhat-operators-54gpf\" (UID: \"7066eebe-a597-40d1-b9f1-f6cde8df2666\") " pod="openshift-marketplace/redhat-operators-54gpf" Nov 25 11:38:38 crc kubenswrapper[4776]: I1125 11:38:38.152414 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7066eebe-a597-40d1-b9f1-f6cde8df2666-utilities\") pod \"redhat-operators-54gpf\" (UID: \"7066eebe-a597-40d1-b9f1-f6cde8df2666\") " pod="openshift-marketplace/redhat-operators-54gpf" Nov 25 11:38:38 crc kubenswrapper[4776]: I1125 11:38:38.178377 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxt6g\" (UniqueName: \"kubernetes.io/projected/7066eebe-a597-40d1-b9f1-f6cde8df2666-kube-api-access-lxt6g\") pod \"redhat-operators-54gpf\" (UID: \"7066eebe-a597-40d1-b9f1-f6cde8df2666\") " pod="openshift-marketplace/redhat-operators-54gpf" Nov 25 11:38:38 crc kubenswrapper[4776]: I1125 11:38:38.235940 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-54gpf" Nov 25 11:38:38 crc kubenswrapper[4776]: I1125 11:38:38.254131 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de4c5588-c756-4f79-b6a3-3d533fce4008-utilities\") pod \"certified-operators-rzzlh\" (UID: \"de4c5588-c756-4f79-b6a3-3d533fce4008\") " pod="openshift-marketplace/certified-operators-rzzlh" Nov 25 11:38:38 crc kubenswrapper[4776]: I1125 11:38:38.254186 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de4c5588-c756-4f79-b6a3-3d533fce4008-catalog-content\") pod \"certified-operators-rzzlh\" (UID: \"de4c5588-c756-4f79-b6a3-3d533fce4008\") " pod="openshift-marketplace/certified-operators-rzzlh" Nov 25 11:38:38 crc kubenswrapper[4776]: I1125 11:38:38.254308 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjms4\" (UniqueName: \"kubernetes.io/projected/de4c5588-c756-4f79-b6a3-3d533fce4008-kube-api-access-bjms4\") pod \"certified-operators-rzzlh\" (UID: \"de4c5588-c756-4f79-b6a3-3d533fce4008\") " pod="openshift-marketplace/certified-operators-rzzlh" Nov 25 11:38:38 crc kubenswrapper[4776]: I1125 11:38:38.357536 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de4c5588-c756-4f79-b6a3-3d533fce4008-utilities\") pod \"certified-operators-rzzlh\" (UID: \"de4c5588-c756-4f79-b6a3-3d533fce4008\") " pod="openshift-marketplace/certified-operators-rzzlh" Nov 25 11:38:38 crc kubenswrapper[4776]: I1125 11:38:38.357925 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de4c5588-c756-4f79-b6a3-3d533fce4008-catalog-content\") pod \"certified-operators-rzzlh\" (UID: \"de4c5588-c756-4f79-b6a3-3d533fce4008\") " pod="openshift-marketplace/certified-operators-rzzlh" Nov 25 11:38:38 crc kubenswrapper[4776]: I1125 11:38:38.358194 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjms4\" (UniqueName: \"kubernetes.io/projected/de4c5588-c756-4f79-b6a3-3d533fce4008-kube-api-access-bjms4\") pod \"certified-operators-rzzlh\" (UID: \"de4c5588-c756-4f79-b6a3-3d533fce4008\") " pod="openshift-marketplace/certified-operators-rzzlh" Nov 25 11:38:38 crc kubenswrapper[4776]: I1125 11:38:38.358653 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de4c5588-c756-4f79-b6a3-3d533fce4008-utilities\") pod \"certified-operators-rzzlh\" (UID: \"de4c5588-c756-4f79-b6a3-3d533fce4008\") " pod="openshift-marketplace/certified-operators-rzzlh" Nov 25 11:38:38 crc kubenswrapper[4776]: I1125 11:38:38.359160 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de4c5588-c756-4f79-b6a3-3d533fce4008-catalog-content\") pod \"certified-operators-rzzlh\" (UID: \"de4c5588-c756-4f79-b6a3-3d533fce4008\") " pod="openshift-marketplace/certified-operators-rzzlh" Nov 25 11:38:38 crc kubenswrapper[4776]: I1125 11:38:38.380451 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjms4\" (UniqueName: \"kubernetes.io/projected/de4c5588-c756-4f79-b6a3-3d533fce4008-kube-api-access-bjms4\") pod \"certified-operators-rzzlh\" (UID: \"de4c5588-c756-4f79-b6a3-3d533fce4008\") " pod="openshift-marketplace/certified-operators-rzzlh" Nov 25 11:38:38 crc kubenswrapper[4776]: I1125 11:38:38.423770 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rzzlh" Nov 25 11:38:38 crc kubenswrapper[4776]: I1125 11:38:38.806538 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-54gpf"] Nov 25 11:38:38 crc kubenswrapper[4776]: I1125 11:38:38.977822 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rzzlh"] Nov 25 11:38:39 crc kubenswrapper[4776]: I1125 11:38:39.071644 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-54gpf" event={"ID":"7066eebe-a597-40d1-b9f1-f6cde8df2666","Type":"ContainerStarted","Data":"e92483d2c7f3f642b9a4a96360cf677f08ee178e32e95ef42619c4836c943020"} Nov 25 11:38:39 crc kubenswrapper[4776]: I1125 11:38:39.073295 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rzzlh" event={"ID":"de4c5588-c756-4f79-b6a3-3d533fce4008","Type":"ContainerStarted","Data":"8191655786ea8980a5afaa86aa921f7541e389aa3791992106aa9cb3a67dc43f"} Nov 25 11:38:40 crc kubenswrapper[4776]: I1125 11:38:40.085648 4776 generic.go:334] "Generic (PLEG): container finished" podID="de4c5588-c756-4f79-b6a3-3d533fce4008" containerID="f29fc92a3fbf4d441302e4f7c622d34788208eb9a4e4122e23e6b37b2093da52" exitCode=0 Nov 25 11:38:40 crc kubenswrapper[4776]: I1125 11:38:40.085786 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rzzlh" event={"ID":"de4c5588-c756-4f79-b6a3-3d533fce4008","Type":"ContainerDied","Data":"f29fc92a3fbf4d441302e4f7c622d34788208eb9a4e4122e23e6b37b2093da52"} Nov 25 11:38:40 crc kubenswrapper[4776]: I1125 11:38:40.088851 4776 generic.go:334] "Generic (PLEG): container finished" podID="7066eebe-a597-40d1-b9f1-f6cde8df2666" containerID="5687b5e1d8cdedc8c37fe871b0e6a016cb89ff745077d03b02410dfd40ff560b" exitCode=0 Nov 25 11:38:40 crc kubenswrapper[4776]: I1125 11:38:40.088874 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-54gpf" event={"ID":"7066eebe-a597-40d1-b9f1-f6cde8df2666","Type":"ContainerDied","Data":"5687b5e1d8cdedc8c37fe871b0e6a016cb89ff745077d03b02410dfd40ff560b"} Nov 25 11:38:40 crc kubenswrapper[4776]: I1125 11:38:40.290105 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qhwlh"] Nov 25 11:38:40 crc kubenswrapper[4776]: I1125 11:38:40.292570 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qhwlh" Nov 25 11:38:40 crc kubenswrapper[4776]: I1125 11:38:40.300669 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qhwlh"] Nov 25 11:38:40 crc kubenswrapper[4776]: I1125 11:38:40.353279 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3110db76-1723-48ad-a077-d5550aaa64d7-catalog-content\") pod \"community-operators-qhwlh\" (UID: \"3110db76-1723-48ad-a077-d5550aaa64d7\") " pod="openshift-marketplace/community-operators-qhwlh" Nov 25 11:38:40 crc kubenswrapper[4776]: I1125 11:38:40.353347 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcjqj\" (UniqueName: \"kubernetes.io/projected/3110db76-1723-48ad-a077-d5550aaa64d7-kube-api-access-dcjqj\") pod \"community-operators-qhwlh\" (UID: \"3110db76-1723-48ad-a077-d5550aaa64d7\") " pod="openshift-marketplace/community-operators-qhwlh" Nov 25 11:38:40 crc kubenswrapper[4776]: I1125 11:38:40.353426 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3110db76-1723-48ad-a077-d5550aaa64d7-utilities\") pod \"community-operators-qhwlh\" (UID: \"3110db76-1723-48ad-a077-d5550aaa64d7\") " pod="openshift-marketplace/community-operators-qhwlh" Nov 25 11:38:40 crc kubenswrapper[4776]: I1125 11:38:40.455554 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3110db76-1723-48ad-a077-d5550aaa64d7-catalog-content\") pod \"community-operators-qhwlh\" (UID: \"3110db76-1723-48ad-a077-d5550aaa64d7\") " pod="openshift-marketplace/community-operators-qhwlh" Nov 25 11:38:40 crc kubenswrapper[4776]: I1125 11:38:40.455620 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcjqj\" (UniqueName: \"kubernetes.io/projected/3110db76-1723-48ad-a077-d5550aaa64d7-kube-api-access-dcjqj\") pod \"community-operators-qhwlh\" (UID: \"3110db76-1723-48ad-a077-d5550aaa64d7\") " pod="openshift-marketplace/community-operators-qhwlh" Nov 25 11:38:40 crc kubenswrapper[4776]: I1125 11:38:40.455665 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3110db76-1723-48ad-a077-d5550aaa64d7-utilities\") pod \"community-operators-qhwlh\" (UID: \"3110db76-1723-48ad-a077-d5550aaa64d7\") " pod="openshift-marketplace/community-operators-qhwlh" Nov 25 11:38:40 crc kubenswrapper[4776]: I1125 11:38:40.456640 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3110db76-1723-48ad-a077-d5550aaa64d7-catalog-content\") pod \"community-operators-qhwlh\" (UID: \"3110db76-1723-48ad-a077-d5550aaa64d7\") " pod="openshift-marketplace/community-operators-qhwlh" Nov 25 11:38:40 crc kubenswrapper[4776]: I1125 11:38:40.456728 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3110db76-1723-48ad-a077-d5550aaa64d7-utilities\") pod \"community-operators-qhwlh\" (UID: \"3110db76-1723-48ad-a077-d5550aaa64d7\") " pod="openshift-marketplace/community-operators-qhwlh" Nov 25 11:38:40 crc kubenswrapper[4776]: I1125 11:38:40.476976 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcjqj\" (UniqueName: \"kubernetes.io/projected/3110db76-1723-48ad-a077-d5550aaa64d7-kube-api-access-dcjqj\") pod \"community-operators-qhwlh\" (UID: \"3110db76-1723-48ad-a077-d5550aaa64d7\") " pod="openshift-marketplace/community-operators-qhwlh" Nov 25 11:38:40 crc kubenswrapper[4776]: I1125 11:38:40.616241 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qhwlh" Nov 25 11:38:41 crc kubenswrapper[4776]: I1125 11:38:41.185876 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qhwlh"] Nov 25 11:38:42 crc kubenswrapper[4776]: I1125 11:38:42.113101 4776 generic.go:334] "Generic (PLEG): container finished" podID="3110db76-1723-48ad-a077-d5550aaa64d7" containerID="7f0a4ae427be20118b865f3530314cfec5efa10cc9f7ef53db9ef4aa5007dc8f" exitCode=0 Nov 25 11:38:42 crc kubenswrapper[4776]: I1125 11:38:42.113182 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qhwlh" event={"ID":"3110db76-1723-48ad-a077-d5550aaa64d7","Type":"ContainerDied","Data":"7f0a4ae427be20118b865f3530314cfec5efa10cc9f7ef53db9ef4aa5007dc8f"} Nov 25 11:38:42 crc kubenswrapper[4776]: I1125 11:38:42.113469 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qhwlh" event={"ID":"3110db76-1723-48ad-a077-d5550aaa64d7","Type":"ContainerStarted","Data":"7faeff04e3091ac3fc9be84b0a6ef596486710141c74f4b44b8f36d4537308c9"} Nov 25 11:38:52 crc kubenswrapper[4776]: I1125 11:38:52.222196 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-54gpf" event={"ID":"7066eebe-a597-40d1-b9f1-f6cde8df2666","Type":"ContainerStarted","Data":"a62c20d9b3b7300c478a5443dd50d616686b5715fcec5645c01d6e2c237f34f1"} Nov 25 11:38:52 crc kubenswrapper[4776]: I1125 11:38:52.224490 4776 generic.go:334] "Generic (PLEG): container finished" podID="de4c5588-c756-4f79-b6a3-3d533fce4008" containerID="9de5691a087cc3d71702e65277f840c7b6efc5328e9df261576f98b9f71a2690" exitCode=0 Nov 25 11:38:52 crc kubenswrapper[4776]: I1125 11:38:52.224569 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rzzlh" event={"ID":"de4c5588-c756-4f79-b6a3-3d533fce4008","Type":"ContainerDied","Data":"9de5691a087cc3d71702e65277f840c7b6efc5328e9df261576f98b9f71a2690"} Nov 25 11:38:52 crc kubenswrapper[4776]: I1125 11:38:52.235687 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qhwlh" event={"ID":"3110db76-1723-48ad-a077-d5550aaa64d7","Type":"ContainerStarted","Data":"5df8358a9ed5531405f37fc8a2b0b86ac4d1cdbe18618c865101997d611d0d57"} Nov 25 11:38:53 crc kubenswrapper[4776]: I1125 11:38:53.250041 4776 generic.go:334] "Generic (PLEG): container finished" podID="7066eebe-a597-40d1-b9f1-f6cde8df2666" containerID="a62c20d9b3b7300c478a5443dd50d616686b5715fcec5645c01d6e2c237f34f1" exitCode=0 Nov 25 11:38:53 crc kubenswrapper[4776]: I1125 11:38:53.250129 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-54gpf" event={"ID":"7066eebe-a597-40d1-b9f1-f6cde8df2666","Type":"ContainerDied","Data":"a62c20d9b3b7300c478a5443dd50d616686b5715fcec5645c01d6e2c237f34f1"} Nov 25 11:38:56 crc kubenswrapper[4776]: I1125 11:38:56.289441 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-54gpf" event={"ID":"7066eebe-a597-40d1-b9f1-f6cde8df2666","Type":"ContainerStarted","Data":"bffa44dbf4705824967a378be71a548630361ec0d949a1793858a16bda1292c8"} Nov 25 11:38:56 crc kubenswrapper[4776]: I1125 11:38:56.295193 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rzzlh" event={"ID":"de4c5588-c756-4f79-b6a3-3d533fce4008","Type":"ContainerStarted","Data":"7410820d010f3e990e45cf3fe734039a4b2c891940e41e4f93463cbc1e758821"} Nov 25 11:38:56 crc kubenswrapper[4776]: I1125 11:38:56.299262 4776 generic.go:334] "Generic (PLEG): container finished" podID="3110db76-1723-48ad-a077-d5550aaa64d7" containerID="5df8358a9ed5531405f37fc8a2b0b86ac4d1cdbe18618c865101997d611d0d57" exitCode=0 Nov 25 11:38:56 crc kubenswrapper[4776]: I1125 11:38:56.299322 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qhwlh" event={"ID":"3110db76-1723-48ad-a077-d5550aaa64d7","Type":"ContainerDied","Data":"5df8358a9ed5531405f37fc8a2b0b86ac4d1cdbe18618c865101997d611d0d57"} Nov 25 11:38:56 crc kubenswrapper[4776]: I1125 11:38:56.322677 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-54gpf" podStartSLOduration=4.192531594 podStartE2EDuration="19.322647794s" podCreationTimestamp="2025-11-25 11:38:37 +0000 UTC" firstStartedPulling="2025-11-25 11:38:40.091480547 +0000 UTC m=+8065.132540100" lastFinishedPulling="2025-11-25 11:38:55.221596747 +0000 UTC m=+8080.262656300" observedRunningTime="2025-11-25 11:38:56.313752831 +0000 UTC m=+8081.354812404" watchObservedRunningTime="2025-11-25 11:38:56.322647794 +0000 UTC m=+8081.363707357" Nov 25 11:38:56 crc kubenswrapper[4776]: I1125 11:38:56.333736 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rzzlh" podStartSLOduration=3.138165209 podStartE2EDuration="18.333702021s" podCreationTimestamp="2025-11-25 11:38:38 +0000 UTC" firstStartedPulling="2025-11-25 11:38:40.088666526 +0000 UTC m=+8065.129726079" lastFinishedPulling="2025-11-25 11:38:55.284203338 +0000 UTC m=+8080.325262891" observedRunningTime="2025-11-25 11:38:56.332922132 +0000 UTC m=+8081.373981685" watchObservedRunningTime="2025-11-25 11:38:56.333702021 +0000 UTC m=+8081.374761574" Nov 25 11:38:57 crc kubenswrapper[4776]: I1125 11:38:57.325300 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qhwlh" event={"ID":"3110db76-1723-48ad-a077-d5550aaa64d7","Type":"ContainerStarted","Data":"82ba30d51c491c89ebf98ac494484d799d86a8f0184d492204ea4d3586926042"} Nov 25 11:38:57 crc kubenswrapper[4776]: I1125 11:38:57.355957 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qhwlh" podStartSLOduration=2.64997242 podStartE2EDuration="17.35593765s" podCreationTimestamp="2025-11-25 11:38:40 +0000 UTC" firstStartedPulling="2025-11-25 11:38:42.115252614 +0000 UTC m=+8067.156312167" lastFinishedPulling="2025-11-25 11:38:56.821217844 +0000 UTC m=+8081.862277397" observedRunningTime="2025-11-25 11:38:57.345248992 +0000 UTC m=+8082.386308565" watchObservedRunningTime="2025-11-25 11:38:57.35593765 +0000 UTC m=+8082.396997203" Nov 25 11:38:58 crc kubenswrapper[4776]: I1125 11:38:58.236888 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-54gpf" Nov 25 11:38:58 crc kubenswrapper[4776]: I1125 11:38:58.236935 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-54gpf" Nov 25 11:38:58 crc kubenswrapper[4776]: I1125 11:38:58.425113 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rzzlh" Nov 25 11:38:58 crc kubenswrapper[4776]: I1125 11:38:58.425176 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rzzlh" Nov 25 11:38:58 crc kubenswrapper[4776]: I1125 11:38:58.478792 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rzzlh" Nov 25 11:38:59 crc kubenswrapper[4776]: I1125 11:38:59.292669 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-54gpf" podUID="7066eebe-a597-40d1-b9f1-f6cde8df2666" containerName="registry-server" probeResult="failure" output=< Nov 25 11:38:59 crc kubenswrapper[4776]: timeout: failed to connect service ":50051" within 1s Nov 25 11:38:59 crc kubenswrapper[4776]: > Nov 25 11:39:00 crc kubenswrapper[4776]: I1125 11:39:00.616890 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qhwlh" Nov 25 11:39:00 crc kubenswrapper[4776]: I1125 11:39:00.617327 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qhwlh" Nov 25 11:39:01 crc kubenswrapper[4776]: I1125 11:39:01.683131 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-qhwlh" podUID="3110db76-1723-48ad-a077-d5550aaa64d7" containerName="registry-server" probeResult="failure" output=< Nov 25 11:39:01 crc kubenswrapper[4776]: timeout: failed to connect service ":50051" within 1s Nov 25 11:39:01 crc kubenswrapper[4776]: > Nov 25 11:39:08 crc kubenswrapper[4776]: I1125 11:39:08.292756 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-54gpf" Nov 25 11:39:08 crc kubenswrapper[4776]: I1125 11:39:08.348713 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-54gpf" Nov 25 11:39:08 crc kubenswrapper[4776]: I1125 11:39:08.470526 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rzzlh" Nov 25 11:39:09 crc kubenswrapper[4776]: I1125 11:39:09.116712 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-54gpf"] Nov 25 11:39:09 crc kubenswrapper[4776]: I1125 11:39:09.498047 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tqb6t"] Nov 25 11:39:09 crc kubenswrapper[4776]: I1125 11:39:09.499143 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tqb6t" podUID="dbcd062b-3c33-49ea-9f28-30d316e43273" containerName="registry-server" containerID="cri-o://e77b469ae67457d49b6659b5386c8d363e8819ee62a34d206e15cfff06a5230e" gracePeriod=2 Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.117028 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tqb6t" Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.275985 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmvbc\" (UniqueName: \"kubernetes.io/projected/dbcd062b-3c33-49ea-9f28-30d316e43273-kube-api-access-lmvbc\") pod \"dbcd062b-3c33-49ea-9f28-30d316e43273\" (UID: \"dbcd062b-3c33-49ea-9f28-30d316e43273\") " Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.276578 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbcd062b-3c33-49ea-9f28-30d316e43273-catalog-content\") pod \"dbcd062b-3c33-49ea-9f28-30d316e43273\" (UID: \"dbcd062b-3c33-49ea-9f28-30d316e43273\") " Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.276830 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbcd062b-3c33-49ea-9f28-30d316e43273-utilities\") pod \"dbcd062b-3c33-49ea-9f28-30d316e43273\" (UID: \"dbcd062b-3c33-49ea-9f28-30d316e43273\") " Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.281058 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbcd062b-3c33-49ea-9f28-30d316e43273-utilities" (OuterVolumeSpecName: "utilities") pod "dbcd062b-3c33-49ea-9f28-30d316e43273" (UID: "dbcd062b-3c33-49ea-9f28-30d316e43273"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.291295 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbcd062b-3c33-49ea-9f28-30d316e43273-kube-api-access-lmvbc" (OuterVolumeSpecName: "kube-api-access-lmvbc") pod "dbcd062b-3c33-49ea-9f28-30d316e43273" (UID: "dbcd062b-3c33-49ea-9f28-30d316e43273"). InnerVolumeSpecName "kube-api-access-lmvbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.380334 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmvbc\" (UniqueName: \"kubernetes.io/projected/dbcd062b-3c33-49ea-9f28-30d316e43273-kube-api-access-lmvbc\") on node \"crc\" DevicePath \"\"" Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.380373 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbcd062b-3c33-49ea-9f28-30d316e43273-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.454879 4776 generic.go:334] "Generic (PLEG): container finished" podID="dbcd062b-3c33-49ea-9f28-30d316e43273" containerID="e77b469ae67457d49b6659b5386c8d363e8819ee62a34d206e15cfff06a5230e" exitCode=0 Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.454933 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqb6t" event={"ID":"dbcd062b-3c33-49ea-9f28-30d316e43273","Type":"ContainerDied","Data":"e77b469ae67457d49b6659b5386c8d363e8819ee62a34d206e15cfff06a5230e"} Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.454973 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqb6t" event={"ID":"dbcd062b-3c33-49ea-9f28-30d316e43273","Type":"ContainerDied","Data":"1a5552c4b6ec0ee5470a0cfdd0382881f74f2e210a78442eeefaa39482ecd8d7"} Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.454989 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tqb6t" Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.454996 4776 scope.go:117] "RemoveContainer" containerID="e77b469ae67457d49b6659b5386c8d363e8819ee62a34d206e15cfff06a5230e" Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.461788 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbcd062b-3c33-49ea-9f28-30d316e43273-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dbcd062b-3c33-49ea-9f28-30d316e43273" (UID: "dbcd062b-3c33-49ea-9f28-30d316e43273"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.482973 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbcd062b-3c33-49ea-9f28-30d316e43273-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.491793 4776 scope.go:117] "RemoveContainer" containerID="40365c4b18b60e9ee19374b72e0fd9a8838770606898037ee7485253ad172e8c" Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.531252 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rzzlh"] Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.576295 4776 scope.go:117] "RemoveContainer" containerID="0315c666685e64377a951b2145043bb0de8fec43d33e20e4b2577e423a97f240" Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.651926 4776 scope.go:117] "RemoveContainer" containerID="e77b469ae67457d49b6659b5386c8d363e8819ee62a34d206e15cfff06a5230e" Nov 25 11:39:10 crc kubenswrapper[4776]: E1125 11:39:10.652513 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e77b469ae67457d49b6659b5386c8d363e8819ee62a34d206e15cfff06a5230e\": container with ID starting with e77b469ae67457d49b6659b5386c8d363e8819ee62a34d206e15cfff06a5230e not found: ID does not exist" containerID="e77b469ae67457d49b6659b5386c8d363e8819ee62a34d206e15cfff06a5230e" Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.652553 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e77b469ae67457d49b6659b5386c8d363e8819ee62a34d206e15cfff06a5230e"} err="failed to get container status \"e77b469ae67457d49b6659b5386c8d363e8819ee62a34d206e15cfff06a5230e\": rpc error: code = NotFound desc = could not find container \"e77b469ae67457d49b6659b5386c8d363e8819ee62a34d206e15cfff06a5230e\": container with ID starting with e77b469ae67457d49b6659b5386c8d363e8819ee62a34d206e15cfff06a5230e not found: ID does not exist" Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.652581 4776 scope.go:117] "RemoveContainer" containerID="40365c4b18b60e9ee19374b72e0fd9a8838770606898037ee7485253ad172e8c" Nov 25 11:39:10 crc kubenswrapper[4776]: E1125 11:39:10.652894 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40365c4b18b60e9ee19374b72e0fd9a8838770606898037ee7485253ad172e8c\": container with ID starting with 40365c4b18b60e9ee19374b72e0fd9a8838770606898037ee7485253ad172e8c not found: ID does not exist" containerID="40365c4b18b60e9ee19374b72e0fd9a8838770606898037ee7485253ad172e8c" Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.652923 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40365c4b18b60e9ee19374b72e0fd9a8838770606898037ee7485253ad172e8c"} err="failed to get container status \"40365c4b18b60e9ee19374b72e0fd9a8838770606898037ee7485253ad172e8c\": rpc error: code = NotFound desc = could not find container \"40365c4b18b60e9ee19374b72e0fd9a8838770606898037ee7485253ad172e8c\": container with ID starting with 40365c4b18b60e9ee19374b72e0fd9a8838770606898037ee7485253ad172e8c not found: ID does not exist" Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.652940 4776 scope.go:117] "RemoveContainer" containerID="0315c666685e64377a951b2145043bb0de8fec43d33e20e4b2577e423a97f240" Nov 25 11:39:10 crc kubenswrapper[4776]: E1125 11:39:10.654465 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0315c666685e64377a951b2145043bb0de8fec43d33e20e4b2577e423a97f240\": container with ID starting with 0315c666685e64377a951b2145043bb0de8fec43d33e20e4b2577e423a97f240 not found: ID does not exist" containerID="0315c666685e64377a951b2145043bb0de8fec43d33e20e4b2577e423a97f240" Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.654504 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0315c666685e64377a951b2145043bb0de8fec43d33e20e4b2577e423a97f240"} err="failed to get container status \"0315c666685e64377a951b2145043bb0de8fec43d33e20e4b2577e423a97f240\": rpc error: code = NotFound desc = could not find container \"0315c666685e64377a951b2145043bb0de8fec43d33e20e4b2577e423a97f240\": container with ID starting with 0315c666685e64377a951b2145043bb0de8fec43d33e20e4b2577e423a97f240 not found: ID does not exist" Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.685875 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qhwlh" Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.796270 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qhwlh" Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.797769 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tqb6t"] Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.806700 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tqb6t"] Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.901030 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-plbg2"] Nov 25 11:39:10 crc kubenswrapper[4776]: I1125 11:39:10.901326 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-plbg2" podUID="d29fc68c-9085-4cb6-900e-92a2ed2b49f1" containerName="registry-server" containerID="cri-o://ef3f0e66e3240da4fd8ab63db189412d5907d983dc71496718100d59f72e8040" gracePeriod=2 Nov 25 11:39:11 crc kubenswrapper[4776]: I1125 11:39:11.465240 4776 generic.go:334] "Generic (PLEG): container finished" podID="d29fc68c-9085-4cb6-900e-92a2ed2b49f1" containerID="ef3f0e66e3240da4fd8ab63db189412d5907d983dc71496718100d59f72e8040" exitCode=0 Nov 25 11:39:11 crc kubenswrapper[4776]: I1125 11:39:11.465322 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-plbg2" event={"ID":"d29fc68c-9085-4cb6-900e-92a2ed2b49f1","Type":"ContainerDied","Data":"ef3f0e66e3240da4fd8ab63db189412d5907d983dc71496718100d59f72e8040"} Nov 25 11:39:11 crc kubenswrapper[4776]: I1125 11:39:11.675044 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbcd062b-3c33-49ea-9f28-30d316e43273" path="/var/lib/kubelet/pods/dbcd062b-3c33-49ea-9f28-30d316e43273/volumes" Nov 25 11:39:11 crc kubenswrapper[4776]: I1125 11:39:11.983335 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-plbg2" Nov 25 11:39:12 crc kubenswrapper[4776]: I1125 11:39:12.018984 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qxms\" (UniqueName: \"kubernetes.io/projected/d29fc68c-9085-4cb6-900e-92a2ed2b49f1-kube-api-access-7qxms\") pod \"d29fc68c-9085-4cb6-900e-92a2ed2b49f1\" (UID: \"d29fc68c-9085-4cb6-900e-92a2ed2b49f1\") " Nov 25 11:39:12 crc kubenswrapper[4776]: I1125 11:39:12.019085 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d29fc68c-9085-4cb6-900e-92a2ed2b49f1-catalog-content\") pod \"d29fc68c-9085-4cb6-900e-92a2ed2b49f1\" (UID: \"d29fc68c-9085-4cb6-900e-92a2ed2b49f1\") " Nov 25 11:39:12 crc kubenswrapper[4776]: I1125 11:39:12.019266 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d29fc68c-9085-4cb6-900e-92a2ed2b49f1-utilities\") pod \"d29fc68c-9085-4cb6-900e-92a2ed2b49f1\" (UID: \"d29fc68c-9085-4cb6-900e-92a2ed2b49f1\") " Nov 25 11:39:12 crc kubenswrapper[4776]: I1125 11:39:12.019778 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d29fc68c-9085-4cb6-900e-92a2ed2b49f1-utilities" (OuterVolumeSpecName: "utilities") pod "d29fc68c-9085-4cb6-900e-92a2ed2b49f1" (UID: "d29fc68c-9085-4cb6-900e-92a2ed2b49f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:39:12 crc kubenswrapper[4776]: I1125 11:39:12.020752 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d29fc68c-9085-4cb6-900e-92a2ed2b49f1-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:39:12 crc kubenswrapper[4776]: I1125 11:39:12.038281 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d29fc68c-9085-4cb6-900e-92a2ed2b49f1-kube-api-access-7qxms" (OuterVolumeSpecName: "kube-api-access-7qxms") pod "d29fc68c-9085-4cb6-900e-92a2ed2b49f1" (UID: "d29fc68c-9085-4cb6-900e-92a2ed2b49f1"). InnerVolumeSpecName "kube-api-access-7qxms". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:39:12 crc kubenswrapper[4776]: I1125 11:39:12.081595 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d29fc68c-9085-4cb6-900e-92a2ed2b49f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d29fc68c-9085-4cb6-900e-92a2ed2b49f1" (UID: "d29fc68c-9085-4cb6-900e-92a2ed2b49f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:39:12 crc kubenswrapper[4776]: I1125 11:39:12.122221 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d29fc68c-9085-4cb6-900e-92a2ed2b49f1-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:39:12 crc kubenswrapper[4776]: I1125 11:39:12.122273 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qxms\" (UniqueName: \"kubernetes.io/projected/d29fc68c-9085-4cb6-900e-92a2ed2b49f1-kube-api-access-7qxms\") on node \"crc\" DevicePath \"\"" Nov 25 11:39:12 crc kubenswrapper[4776]: I1125 11:39:12.484028 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-plbg2" event={"ID":"d29fc68c-9085-4cb6-900e-92a2ed2b49f1","Type":"ContainerDied","Data":"f7c92c6e1699e3ef08839e6616a8ed58d30a026e919525be763289a1ff42f65c"} Nov 25 11:39:12 crc kubenswrapper[4776]: I1125 11:39:12.484433 4776 scope.go:117] "RemoveContainer" containerID="ef3f0e66e3240da4fd8ab63db189412d5907d983dc71496718100d59f72e8040" Nov 25 11:39:12 crc kubenswrapper[4776]: I1125 11:39:12.484136 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-plbg2" Nov 25 11:39:12 crc kubenswrapper[4776]: I1125 11:39:12.519242 4776 scope.go:117] "RemoveContainer" containerID="db43c08a7942c8f44c0cc7b79aeefcf1efa66e256ef3f324f34f52bfb8dad766" Nov 25 11:39:12 crc kubenswrapper[4776]: I1125 11:39:12.538060 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-plbg2"] Nov 25 11:39:12 crc kubenswrapper[4776]: I1125 11:39:12.555524 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-plbg2"] Nov 25 11:39:12 crc kubenswrapper[4776]: I1125 11:39:12.598201 4776 scope.go:117] "RemoveContainer" containerID="936d9ee225c3f1b9535fe35ec24f3260da2b0ffdf6c7358774a866fcc9b55224" Nov 25 11:39:12 crc kubenswrapper[4776]: I1125 11:39:12.908651 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qhwlh"] Nov 25 11:39:12 crc kubenswrapper[4776]: I1125 11:39:12.909423 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qhwlh" podUID="3110db76-1723-48ad-a077-d5550aaa64d7" containerName="registry-server" containerID="cri-o://82ba30d51c491c89ebf98ac494484d799d86a8f0184d492204ea4d3586926042" gracePeriod=2 Nov 25 11:39:13 crc kubenswrapper[4776]: I1125 11:39:13.498636 4776 generic.go:334] "Generic (PLEG): container finished" podID="3110db76-1723-48ad-a077-d5550aaa64d7" containerID="82ba30d51c491c89ebf98ac494484d799d86a8f0184d492204ea4d3586926042" exitCode=0 Nov 25 11:39:13 crc kubenswrapper[4776]: I1125 11:39:13.498679 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qhwlh" event={"ID":"3110db76-1723-48ad-a077-d5550aaa64d7","Type":"ContainerDied","Data":"82ba30d51c491c89ebf98ac494484d799d86a8f0184d492204ea4d3586926042"} Nov 25 11:39:13 crc kubenswrapper[4776]: I1125 11:39:13.676983 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d29fc68c-9085-4cb6-900e-92a2ed2b49f1" path="/var/lib/kubelet/pods/d29fc68c-9085-4cb6-900e-92a2ed2b49f1/volumes" Nov 25 11:39:14 crc kubenswrapper[4776]: I1125 11:39:14.157367 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qhwlh" Nov 25 11:39:14 crc kubenswrapper[4776]: I1125 11:39:14.195895 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3110db76-1723-48ad-a077-d5550aaa64d7-utilities\") pod \"3110db76-1723-48ad-a077-d5550aaa64d7\" (UID: \"3110db76-1723-48ad-a077-d5550aaa64d7\") " Nov 25 11:39:14 crc kubenswrapper[4776]: I1125 11:39:14.196216 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcjqj\" (UniqueName: \"kubernetes.io/projected/3110db76-1723-48ad-a077-d5550aaa64d7-kube-api-access-dcjqj\") pod \"3110db76-1723-48ad-a077-d5550aaa64d7\" (UID: \"3110db76-1723-48ad-a077-d5550aaa64d7\") " Nov 25 11:39:14 crc kubenswrapper[4776]: I1125 11:39:14.196290 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3110db76-1723-48ad-a077-d5550aaa64d7-catalog-content\") pod \"3110db76-1723-48ad-a077-d5550aaa64d7\" (UID: \"3110db76-1723-48ad-a077-d5550aaa64d7\") " Nov 25 11:39:14 crc kubenswrapper[4776]: I1125 11:39:14.196828 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3110db76-1723-48ad-a077-d5550aaa64d7-utilities" (OuterVolumeSpecName: "utilities") pod "3110db76-1723-48ad-a077-d5550aaa64d7" (UID: "3110db76-1723-48ad-a077-d5550aaa64d7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:39:14 crc kubenswrapper[4776]: I1125 11:39:14.210339 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3110db76-1723-48ad-a077-d5550aaa64d7-kube-api-access-dcjqj" (OuterVolumeSpecName: "kube-api-access-dcjqj") pod "3110db76-1723-48ad-a077-d5550aaa64d7" (UID: "3110db76-1723-48ad-a077-d5550aaa64d7"). InnerVolumeSpecName "kube-api-access-dcjqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:39:14 crc kubenswrapper[4776]: I1125 11:39:14.275089 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3110db76-1723-48ad-a077-d5550aaa64d7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3110db76-1723-48ad-a077-d5550aaa64d7" (UID: "3110db76-1723-48ad-a077-d5550aaa64d7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:39:14 crc kubenswrapper[4776]: I1125 11:39:14.299257 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcjqj\" (UniqueName: \"kubernetes.io/projected/3110db76-1723-48ad-a077-d5550aaa64d7-kube-api-access-dcjqj\") on node \"crc\" DevicePath \"\"" Nov 25 11:39:14 crc kubenswrapper[4776]: I1125 11:39:14.299298 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3110db76-1723-48ad-a077-d5550aaa64d7-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:39:14 crc kubenswrapper[4776]: I1125 11:39:14.299307 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3110db76-1723-48ad-a077-d5550aaa64d7-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:39:14 crc kubenswrapper[4776]: I1125 11:39:14.510606 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qhwlh" event={"ID":"3110db76-1723-48ad-a077-d5550aaa64d7","Type":"ContainerDied","Data":"7faeff04e3091ac3fc9be84b0a6ef596486710141c74f4b44b8f36d4537308c9"} Nov 25 11:39:14 crc kubenswrapper[4776]: I1125 11:39:14.510671 4776 scope.go:117] "RemoveContainer" containerID="82ba30d51c491c89ebf98ac494484d799d86a8f0184d492204ea4d3586926042" Nov 25 11:39:14 crc kubenswrapper[4776]: I1125 11:39:14.510691 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qhwlh" Nov 25 11:39:14 crc kubenswrapper[4776]: I1125 11:39:14.557883 4776 scope.go:117] "RemoveContainer" containerID="5df8358a9ed5531405f37fc8a2b0b86ac4d1cdbe18618c865101997d611d0d57" Nov 25 11:39:14 crc kubenswrapper[4776]: I1125 11:39:14.570144 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qhwlh"] Nov 25 11:39:14 crc kubenswrapper[4776]: I1125 11:39:14.580578 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qhwlh"] Nov 25 11:39:14 crc kubenswrapper[4776]: I1125 11:39:14.606283 4776 scope.go:117] "RemoveContainer" containerID="7f0a4ae427be20118b865f3530314cfec5efa10cc9f7ef53db9ef4aa5007dc8f" Nov 25 11:39:15 crc kubenswrapper[4776]: I1125 11:39:15.679460 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3110db76-1723-48ad-a077-d5550aaa64d7" path="/var/lib/kubelet/pods/3110db76-1723-48ad-a077-d5550aaa64d7/volumes" Nov 25 11:39:17 crc kubenswrapper[4776]: I1125 11:39:17.817889 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:39:17 crc kubenswrapper[4776]: I1125 11:39:17.818270 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:39:47 crc kubenswrapper[4776]: I1125 11:39:47.819173 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:39:47 crc kubenswrapper[4776]: I1125 11:39:47.819727 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:40:17 crc kubenswrapper[4776]: I1125 11:40:17.818414 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:40:17 crc kubenswrapper[4776]: I1125 11:40:17.819036 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:40:17 crc kubenswrapper[4776]: I1125 11:40:17.819319 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 11:40:17 crc kubenswrapper[4776]: I1125 11:40:17.820131 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 11:40:17 crc kubenswrapper[4776]: I1125 11:40:17.820191 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b" gracePeriod=600 Nov 25 11:40:18 crc kubenswrapper[4776]: E1125 11:40:18.000221 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:40:18 crc kubenswrapper[4776]: I1125 11:40:18.137986 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b" exitCode=0 Nov 25 11:40:18 crc kubenswrapper[4776]: I1125 11:40:18.138030 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b"} Nov 25 11:40:18 crc kubenswrapper[4776]: I1125 11:40:18.138075 4776 scope.go:117] "RemoveContainer" containerID="cced90c452d27e661d744bc42ceb0710b870fc7135d4b27a5d442a29ab283935" Nov 25 11:40:18 crc kubenswrapper[4776]: I1125 11:40:18.138860 4776 scope.go:117] "RemoveContainer" containerID="3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b" Nov 25 11:40:18 crc kubenswrapper[4776]: E1125 11:40:18.139218 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:40:32 crc kubenswrapper[4776]: I1125 11:40:32.662605 4776 scope.go:117] "RemoveContainer" containerID="3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b" Nov 25 11:40:32 crc kubenswrapper[4776]: E1125 11:40:32.664464 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:40:43 crc kubenswrapper[4776]: I1125 11:40:43.661885 4776 scope.go:117] "RemoveContainer" containerID="3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b" Nov 25 11:40:43 crc kubenswrapper[4776]: E1125 11:40:43.662796 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:40:56 crc kubenswrapper[4776]: I1125 11:40:56.662451 4776 scope.go:117] "RemoveContainer" containerID="3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b" Nov 25 11:40:56 crc kubenswrapper[4776]: E1125 11:40:56.663295 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:41:07 crc kubenswrapper[4776]: I1125 11:41:07.662868 4776 scope.go:117] "RemoveContainer" containerID="3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b" Nov 25 11:41:07 crc kubenswrapper[4776]: E1125 11:41:07.663909 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:41:19 crc kubenswrapper[4776]: I1125 11:41:19.663406 4776 scope.go:117] "RemoveContainer" containerID="3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b" Nov 25 11:41:19 crc kubenswrapper[4776]: E1125 11:41:19.664210 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:41:34 crc kubenswrapper[4776]: I1125 11:41:34.663136 4776 scope.go:117] "RemoveContainer" containerID="3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b" Nov 25 11:41:34 crc kubenswrapper[4776]: E1125 11:41:34.664123 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:41:45 crc kubenswrapper[4776]: I1125 11:41:45.680189 4776 scope.go:117] "RemoveContainer" containerID="3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b" Nov 25 11:41:45 crc kubenswrapper[4776]: E1125 11:41:45.681163 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:41:49 crc kubenswrapper[4776]: I1125 11:41:49.775780 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8f8hl"] Nov 25 11:41:49 crc kubenswrapper[4776]: E1125 11:41:49.776794 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3110db76-1723-48ad-a077-d5550aaa64d7" containerName="extract-content" Nov 25 11:41:49 crc kubenswrapper[4776]: I1125 11:41:49.776808 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="3110db76-1723-48ad-a077-d5550aaa64d7" containerName="extract-content" Nov 25 11:41:49 crc kubenswrapper[4776]: E1125 11:41:49.776820 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbcd062b-3c33-49ea-9f28-30d316e43273" containerName="extract-content" Nov 25 11:41:49 crc kubenswrapper[4776]: I1125 11:41:49.776827 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbcd062b-3c33-49ea-9f28-30d316e43273" containerName="extract-content" Nov 25 11:41:49 crc kubenswrapper[4776]: E1125 11:41:49.776846 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbcd062b-3c33-49ea-9f28-30d316e43273" containerName="extract-utilities" Nov 25 11:41:49 crc kubenswrapper[4776]: I1125 11:41:49.776854 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbcd062b-3c33-49ea-9f28-30d316e43273" containerName="extract-utilities" Nov 25 11:41:49 crc kubenswrapper[4776]: E1125 11:41:49.776867 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d29fc68c-9085-4cb6-900e-92a2ed2b49f1" containerName="extract-content" Nov 25 11:41:49 crc kubenswrapper[4776]: I1125 11:41:49.776875 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="d29fc68c-9085-4cb6-900e-92a2ed2b49f1" containerName="extract-content" Nov 25 11:41:49 crc kubenswrapper[4776]: E1125 11:41:49.776890 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbcd062b-3c33-49ea-9f28-30d316e43273" containerName="registry-server" Nov 25 11:41:49 crc kubenswrapper[4776]: I1125 11:41:49.776896 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbcd062b-3c33-49ea-9f28-30d316e43273" containerName="registry-server" Nov 25 11:41:49 crc kubenswrapper[4776]: E1125 11:41:49.776912 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3110db76-1723-48ad-a077-d5550aaa64d7" containerName="extract-utilities" Nov 25 11:41:49 crc kubenswrapper[4776]: I1125 11:41:49.776918 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="3110db76-1723-48ad-a077-d5550aaa64d7" containerName="extract-utilities" Nov 25 11:41:49 crc kubenswrapper[4776]: E1125 11:41:49.776929 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3110db76-1723-48ad-a077-d5550aaa64d7" containerName="registry-server" Nov 25 11:41:49 crc kubenswrapper[4776]: I1125 11:41:49.776936 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="3110db76-1723-48ad-a077-d5550aaa64d7" containerName="registry-server" Nov 25 11:41:49 crc kubenswrapper[4776]: E1125 11:41:49.776949 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d29fc68c-9085-4cb6-900e-92a2ed2b49f1" containerName="registry-server" Nov 25 11:41:49 crc kubenswrapper[4776]: I1125 11:41:49.776955 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="d29fc68c-9085-4cb6-900e-92a2ed2b49f1" containerName="registry-server" Nov 25 11:41:49 crc kubenswrapper[4776]: E1125 11:41:49.776985 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d29fc68c-9085-4cb6-900e-92a2ed2b49f1" containerName="extract-utilities" Nov 25 11:41:49 crc kubenswrapper[4776]: I1125 11:41:49.776992 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="d29fc68c-9085-4cb6-900e-92a2ed2b49f1" containerName="extract-utilities" Nov 25 11:41:49 crc kubenswrapper[4776]: I1125 11:41:49.777215 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbcd062b-3c33-49ea-9f28-30d316e43273" containerName="registry-server" Nov 25 11:41:49 crc kubenswrapper[4776]: I1125 11:41:49.777226 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="d29fc68c-9085-4cb6-900e-92a2ed2b49f1" containerName="registry-server" Nov 25 11:41:49 crc kubenswrapper[4776]: I1125 11:41:49.777241 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="3110db76-1723-48ad-a077-d5550aaa64d7" containerName="registry-server" Nov 25 11:41:49 crc kubenswrapper[4776]: I1125 11:41:49.778965 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8f8hl" Nov 25 11:41:49 crc kubenswrapper[4776]: I1125 11:41:49.797208 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8f8hl"] Nov 25 11:41:49 crc kubenswrapper[4776]: I1125 11:41:49.936054 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84nq8\" (UniqueName: \"kubernetes.io/projected/d2510036-0cf5-471f-8166-71f71f8b6bb9-kube-api-access-84nq8\") pod \"redhat-marketplace-8f8hl\" (UID: \"d2510036-0cf5-471f-8166-71f71f8b6bb9\") " pod="openshift-marketplace/redhat-marketplace-8f8hl" Nov 25 11:41:49 crc kubenswrapper[4776]: I1125 11:41:49.936266 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2510036-0cf5-471f-8166-71f71f8b6bb9-utilities\") pod \"redhat-marketplace-8f8hl\" (UID: \"d2510036-0cf5-471f-8166-71f71f8b6bb9\") " pod="openshift-marketplace/redhat-marketplace-8f8hl" Nov 25 11:41:49 crc kubenswrapper[4776]: I1125 11:41:49.936742 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2510036-0cf5-471f-8166-71f71f8b6bb9-catalog-content\") pod \"redhat-marketplace-8f8hl\" (UID: \"d2510036-0cf5-471f-8166-71f71f8b6bb9\") " pod="openshift-marketplace/redhat-marketplace-8f8hl" Nov 25 11:41:50 crc kubenswrapper[4776]: I1125 11:41:50.039449 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2510036-0cf5-471f-8166-71f71f8b6bb9-utilities\") pod \"redhat-marketplace-8f8hl\" (UID: \"d2510036-0cf5-471f-8166-71f71f8b6bb9\") " pod="openshift-marketplace/redhat-marketplace-8f8hl" Nov 25 11:41:50 crc kubenswrapper[4776]: I1125 11:41:50.039640 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2510036-0cf5-471f-8166-71f71f8b6bb9-catalog-content\") pod \"redhat-marketplace-8f8hl\" (UID: \"d2510036-0cf5-471f-8166-71f71f8b6bb9\") " pod="openshift-marketplace/redhat-marketplace-8f8hl" Nov 25 11:41:50 crc kubenswrapper[4776]: I1125 11:41:50.039696 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84nq8\" (UniqueName: \"kubernetes.io/projected/d2510036-0cf5-471f-8166-71f71f8b6bb9-kube-api-access-84nq8\") pod \"redhat-marketplace-8f8hl\" (UID: \"d2510036-0cf5-471f-8166-71f71f8b6bb9\") " pod="openshift-marketplace/redhat-marketplace-8f8hl" Nov 25 11:41:50 crc kubenswrapper[4776]: I1125 11:41:50.040101 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2510036-0cf5-471f-8166-71f71f8b6bb9-catalog-content\") pod \"redhat-marketplace-8f8hl\" (UID: \"d2510036-0cf5-471f-8166-71f71f8b6bb9\") " pod="openshift-marketplace/redhat-marketplace-8f8hl" Nov 25 11:41:50 crc kubenswrapper[4776]: I1125 11:41:50.040101 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2510036-0cf5-471f-8166-71f71f8b6bb9-utilities\") pod \"redhat-marketplace-8f8hl\" (UID: \"d2510036-0cf5-471f-8166-71f71f8b6bb9\") " pod="openshift-marketplace/redhat-marketplace-8f8hl" Nov 25 11:41:50 crc kubenswrapper[4776]: I1125 11:41:50.061515 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84nq8\" (UniqueName: \"kubernetes.io/projected/d2510036-0cf5-471f-8166-71f71f8b6bb9-kube-api-access-84nq8\") pod \"redhat-marketplace-8f8hl\" (UID: \"d2510036-0cf5-471f-8166-71f71f8b6bb9\") " pod="openshift-marketplace/redhat-marketplace-8f8hl" Nov 25 11:41:50 crc kubenswrapper[4776]: I1125 11:41:50.109543 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8f8hl" Nov 25 11:41:50 crc kubenswrapper[4776]: I1125 11:41:50.626350 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8f8hl"] Nov 25 11:41:51 crc kubenswrapper[4776]: I1125 11:41:51.107313 4776 generic.go:334] "Generic (PLEG): container finished" podID="d2510036-0cf5-471f-8166-71f71f8b6bb9" containerID="2447ef287d45c09842233c4490dc2cbe81085f1838dfef7f8d3b766a37caa6c7" exitCode=0 Nov 25 11:41:51 crc kubenswrapper[4776]: I1125 11:41:51.107385 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8f8hl" event={"ID":"d2510036-0cf5-471f-8166-71f71f8b6bb9","Type":"ContainerDied","Data":"2447ef287d45c09842233c4490dc2cbe81085f1838dfef7f8d3b766a37caa6c7"} Nov 25 11:41:51 crc kubenswrapper[4776]: I1125 11:41:51.107634 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8f8hl" event={"ID":"d2510036-0cf5-471f-8166-71f71f8b6bb9","Type":"ContainerStarted","Data":"aa87855e1b94bcb98b6accba379318f004c46f8d1f57bae55d7697f280a3a0a3"} Nov 25 11:41:53 crc kubenswrapper[4776]: I1125 11:41:53.128158 4776 generic.go:334] "Generic (PLEG): container finished" podID="d2510036-0cf5-471f-8166-71f71f8b6bb9" containerID="b5ca75d8ceb14fc9dc6bac478668d5b252b42b096b338473fd00151c2bec9db9" exitCode=0 Nov 25 11:41:53 crc kubenswrapper[4776]: I1125 11:41:53.128285 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8f8hl" event={"ID":"d2510036-0cf5-471f-8166-71f71f8b6bb9","Type":"ContainerDied","Data":"b5ca75d8ceb14fc9dc6bac478668d5b252b42b096b338473fd00151c2bec9db9"} Nov 25 11:41:54 crc kubenswrapper[4776]: I1125 11:41:54.140305 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8f8hl" event={"ID":"d2510036-0cf5-471f-8166-71f71f8b6bb9","Type":"ContainerStarted","Data":"f98179abffccb2b78b4884915c646e3d1a75237c5d13c6d9a0137cc0120e16ef"} Nov 25 11:41:54 crc kubenswrapper[4776]: I1125 11:41:54.163036 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8f8hl" podStartSLOduration=2.5408032560000002 podStartE2EDuration="5.163019608s" podCreationTimestamp="2025-11-25 11:41:49 +0000 UTC" firstStartedPulling="2025-11-25 11:41:51.109198157 +0000 UTC m=+8256.150257710" lastFinishedPulling="2025-11-25 11:41:53.731414509 +0000 UTC m=+8258.772474062" observedRunningTime="2025-11-25 11:41:54.156821922 +0000 UTC m=+8259.197881495" watchObservedRunningTime="2025-11-25 11:41:54.163019608 +0000 UTC m=+8259.204079161" Nov 25 11:42:00 crc kubenswrapper[4776]: I1125 11:42:00.111518 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8f8hl" Nov 25 11:42:00 crc kubenswrapper[4776]: I1125 11:42:00.112212 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8f8hl" Nov 25 11:42:00 crc kubenswrapper[4776]: I1125 11:42:00.160185 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8f8hl" Nov 25 11:42:00 crc kubenswrapper[4776]: I1125 11:42:00.249422 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8f8hl" Nov 25 11:42:00 crc kubenswrapper[4776]: I1125 11:42:00.400187 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8f8hl"] Nov 25 11:42:00 crc kubenswrapper[4776]: I1125 11:42:00.662618 4776 scope.go:117] "RemoveContainer" containerID="3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b" Nov 25 11:42:00 crc kubenswrapper[4776]: E1125 11:42:00.663027 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:42:02 crc kubenswrapper[4776]: I1125 11:42:02.215823 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8f8hl" podUID="d2510036-0cf5-471f-8166-71f71f8b6bb9" containerName="registry-server" containerID="cri-o://f98179abffccb2b78b4884915c646e3d1a75237c5d13c6d9a0137cc0120e16ef" gracePeriod=2 Nov 25 11:42:02 crc kubenswrapper[4776]: I1125 11:42:02.717882 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8f8hl" Nov 25 11:42:02 crc kubenswrapper[4776]: I1125 11:42:02.909788 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2510036-0cf5-471f-8166-71f71f8b6bb9-utilities\") pod \"d2510036-0cf5-471f-8166-71f71f8b6bb9\" (UID: \"d2510036-0cf5-471f-8166-71f71f8b6bb9\") " Nov 25 11:42:02 crc kubenswrapper[4776]: I1125 11:42:02.909966 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2510036-0cf5-471f-8166-71f71f8b6bb9-catalog-content\") pod \"d2510036-0cf5-471f-8166-71f71f8b6bb9\" (UID: \"d2510036-0cf5-471f-8166-71f71f8b6bb9\") " Nov 25 11:42:02 crc kubenswrapper[4776]: I1125 11:42:02.910296 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84nq8\" (UniqueName: \"kubernetes.io/projected/d2510036-0cf5-471f-8166-71f71f8b6bb9-kube-api-access-84nq8\") pod \"d2510036-0cf5-471f-8166-71f71f8b6bb9\" (UID: \"d2510036-0cf5-471f-8166-71f71f8b6bb9\") " Nov 25 11:42:02 crc kubenswrapper[4776]: I1125 11:42:02.910965 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2510036-0cf5-471f-8166-71f71f8b6bb9-utilities" (OuterVolumeSpecName: "utilities") pod "d2510036-0cf5-471f-8166-71f71f8b6bb9" (UID: "d2510036-0cf5-471f-8166-71f71f8b6bb9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:42:02 crc kubenswrapper[4776]: I1125 11:42:02.925969 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2510036-0cf5-471f-8166-71f71f8b6bb9-kube-api-access-84nq8" (OuterVolumeSpecName: "kube-api-access-84nq8") pod "d2510036-0cf5-471f-8166-71f71f8b6bb9" (UID: "d2510036-0cf5-471f-8166-71f71f8b6bb9"). InnerVolumeSpecName "kube-api-access-84nq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:42:02 crc kubenswrapper[4776]: I1125 11:42:02.929650 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2510036-0cf5-471f-8166-71f71f8b6bb9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d2510036-0cf5-471f-8166-71f71f8b6bb9" (UID: "d2510036-0cf5-471f-8166-71f71f8b6bb9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:42:03 crc kubenswrapper[4776]: I1125 11:42:03.012935 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2510036-0cf5-471f-8166-71f71f8b6bb9-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:42:03 crc kubenswrapper[4776]: I1125 11:42:03.012993 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2510036-0cf5-471f-8166-71f71f8b6bb9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:42:03 crc kubenswrapper[4776]: I1125 11:42:03.013005 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84nq8\" (UniqueName: \"kubernetes.io/projected/d2510036-0cf5-471f-8166-71f71f8b6bb9-kube-api-access-84nq8\") on node \"crc\" DevicePath \"\"" Nov 25 11:42:03 crc kubenswrapper[4776]: I1125 11:42:03.227960 4776 generic.go:334] "Generic (PLEG): container finished" podID="d2510036-0cf5-471f-8166-71f71f8b6bb9" containerID="f98179abffccb2b78b4884915c646e3d1a75237c5d13c6d9a0137cc0120e16ef" exitCode=0 Nov 25 11:42:03 crc kubenswrapper[4776]: I1125 11:42:03.228243 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8f8hl" event={"ID":"d2510036-0cf5-471f-8166-71f71f8b6bb9","Type":"ContainerDied","Data":"f98179abffccb2b78b4884915c646e3d1a75237c5d13c6d9a0137cc0120e16ef"} Nov 25 11:42:03 crc kubenswrapper[4776]: I1125 11:42:03.228517 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8f8hl" event={"ID":"d2510036-0cf5-471f-8166-71f71f8b6bb9","Type":"ContainerDied","Data":"aa87855e1b94bcb98b6accba379318f004c46f8d1f57bae55d7697f280a3a0a3"} Nov 25 11:42:03 crc kubenswrapper[4776]: I1125 11:42:03.228542 4776 scope.go:117] "RemoveContainer" containerID="f98179abffccb2b78b4884915c646e3d1a75237c5d13c6d9a0137cc0120e16ef" Nov 25 11:42:03 crc kubenswrapper[4776]: I1125 11:42:03.228258 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8f8hl" Nov 25 11:42:03 crc kubenswrapper[4776]: I1125 11:42:03.264347 4776 scope.go:117] "RemoveContainer" containerID="b5ca75d8ceb14fc9dc6bac478668d5b252b42b096b338473fd00151c2bec9db9" Nov 25 11:42:03 crc kubenswrapper[4776]: I1125 11:42:03.276212 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8f8hl"] Nov 25 11:42:03 crc kubenswrapper[4776]: I1125 11:42:03.286814 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8f8hl"] Nov 25 11:42:03 crc kubenswrapper[4776]: I1125 11:42:03.299825 4776 scope.go:117] "RemoveContainer" containerID="2447ef287d45c09842233c4490dc2cbe81085f1838dfef7f8d3b766a37caa6c7" Nov 25 11:42:03 crc kubenswrapper[4776]: I1125 11:42:03.350590 4776 scope.go:117] "RemoveContainer" containerID="f98179abffccb2b78b4884915c646e3d1a75237c5d13c6d9a0137cc0120e16ef" Nov 25 11:42:03 crc kubenswrapper[4776]: E1125 11:42:03.351045 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f98179abffccb2b78b4884915c646e3d1a75237c5d13c6d9a0137cc0120e16ef\": container with ID starting with f98179abffccb2b78b4884915c646e3d1a75237c5d13c6d9a0137cc0120e16ef not found: ID does not exist" containerID="f98179abffccb2b78b4884915c646e3d1a75237c5d13c6d9a0137cc0120e16ef" Nov 25 11:42:03 crc kubenswrapper[4776]: I1125 11:42:03.351102 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f98179abffccb2b78b4884915c646e3d1a75237c5d13c6d9a0137cc0120e16ef"} err="failed to get container status \"f98179abffccb2b78b4884915c646e3d1a75237c5d13c6d9a0137cc0120e16ef\": rpc error: code = NotFound desc = could not find container \"f98179abffccb2b78b4884915c646e3d1a75237c5d13c6d9a0137cc0120e16ef\": container with ID starting with f98179abffccb2b78b4884915c646e3d1a75237c5d13c6d9a0137cc0120e16ef not found: ID does not exist" Nov 25 11:42:03 crc kubenswrapper[4776]: I1125 11:42:03.351128 4776 scope.go:117] "RemoveContainer" containerID="b5ca75d8ceb14fc9dc6bac478668d5b252b42b096b338473fd00151c2bec9db9" Nov 25 11:42:03 crc kubenswrapper[4776]: E1125 11:42:03.351704 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5ca75d8ceb14fc9dc6bac478668d5b252b42b096b338473fd00151c2bec9db9\": container with ID starting with b5ca75d8ceb14fc9dc6bac478668d5b252b42b096b338473fd00151c2bec9db9 not found: ID does not exist" containerID="b5ca75d8ceb14fc9dc6bac478668d5b252b42b096b338473fd00151c2bec9db9" Nov 25 11:42:03 crc kubenswrapper[4776]: I1125 11:42:03.351748 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5ca75d8ceb14fc9dc6bac478668d5b252b42b096b338473fd00151c2bec9db9"} err="failed to get container status \"b5ca75d8ceb14fc9dc6bac478668d5b252b42b096b338473fd00151c2bec9db9\": rpc error: code = NotFound desc = could not find container \"b5ca75d8ceb14fc9dc6bac478668d5b252b42b096b338473fd00151c2bec9db9\": container with ID starting with b5ca75d8ceb14fc9dc6bac478668d5b252b42b096b338473fd00151c2bec9db9 not found: ID does not exist" Nov 25 11:42:03 crc kubenswrapper[4776]: I1125 11:42:03.351778 4776 scope.go:117] "RemoveContainer" containerID="2447ef287d45c09842233c4490dc2cbe81085f1838dfef7f8d3b766a37caa6c7" Nov 25 11:42:03 crc kubenswrapper[4776]: E1125 11:42:03.352185 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2447ef287d45c09842233c4490dc2cbe81085f1838dfef7f8d3b766a37caa6c7\": container with ID starting with 2447ef287d45c09842233c4490dc2cbe81085f1838dfef7f8d3b766a37caa6c7 not found: ID does not exist" containerID="2447ef287d45c09842233c4490dc2cbe81085f1838dfef7f8d3b766a37caa6c7" Nov 25 11:42:03 crc kubenswrapper[4776]: I1125 11:42:03.352222 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2447ef287d45c09842233c4490dc2cbe81085f1838dfef7f8d3b766a37caa6c7"} err="failed to get container status \"2447ef287d45c09842233c4490dc2cbe81085f1838dfef7f8d3b766a37caa6c7\": rpc error: code = NotFound desc = could not find container \"2447ef287d45c09842233c4490dc2cbe81085f1838dfef7f8d3b766a37caa6c7\": container with ID starting with 2447ef287d45c09842233c4490dc2cbe81085f1838dfef7f8d3b766a37caa6c7 not found: ID does not exist" Nov 25 11:42:03 crc kubenswrapper[4776]: I1125 11:42:03.674380 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2510036-0cf5-471f-8166-71f71f8b6bb9" path="/var/lib/kubelet/pods/d2510036-0cf5-471f-8166-71f71f8b6bb9/volumes" Nov 25 11:42:11 crc kubenswrapper[4776]: I1125 11:42:11.663235 4776 scope.go:117] "RemoveContainer" containerID="3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b" Nov 25 11:42:11 crc kubenswrapper[4776]: E1125 11:42:11.664267 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:42:26 crc kubenswrapper[4776]: I1125 11:42:26.662597 4776 scope.go:117] "RemoveContainer" containerID="3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b" Nov 25 11:42:26 crc kubenswrapper[4776]: E1125 11:42:26.663425 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:42:38 crc kubenswrapper[4776]: I1125 11:42:38.601236 4776 generic.go:334] "Generic (PLEG): container finished" podID="bc20c5f3-b0e5-47e2-8878-015b66dad6f0" containerID="6460bf3c8dc1b7263aeb3b614a79758110284fc4c65c837f6eb42db240b11502" exitCode=0 Nov 25 11:42:38 crc kubenswrapper[4776]: I1125 11:42:38.601352 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-c2plb" event={"ID":"bc20c5f3-b0e5-47e2-8878-015b66dad6f0","Type":"ContainerDied","Data":"6460bf3c8dc1b7263aeb3b614a79758110284fc4c65c837f6eb42db240b11502"} Nov 25 11:42:38 crc kubenswrapper[4776]: I1125 11:42:38.663659 4776 scope.go:117] "RemoveContainer" containerID="3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b" Nov 25 11:42:38 crc kubenswrapper[4776]: E1125 11:42:38.664128 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.119846 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-c2plb" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.209429 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-libvirt-combined-ca-bundle\") pod \"bc20c5f3-b0e5-47e2-8878-015b66dad6f0\" (UID: \"bc20c5f3-b0e5-47e2-8878-015b66dad6f0\") " Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.209550 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-ssh-key\") pod \"bc20c5f3-b0e5-47e2-8878-015b66dad6f0\" (UID: \"bc20c5f3-b0e5-47e2-8878-015b66dad6f0\") " Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.209579 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4fkw\" (UniqueName: \"kubernetes.io/projected/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-kube-api-access-m4fkw\") pod \"bc20c5f3-b0e5-47e2-8878-015b66dad6f0\" (UID: \"bc20c5f3-b0e5-47e2-8878-015b66dad6f0\") " Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.209829 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-libvirt-secret-0\") pod \"bc20c5f3-b0e5-47e2-8878-015b66dad6f0\" (UID: \"bc20c5f3-b0e5-47e2-8878-015b66dad6f0\") " Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.209947 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-inventory\") pod \"bc20c5f3-b0e5-47e2-8878-015b66dad6f0\" (UID: \"bc20c5f3-b0e5-47e2-8878-015b66dad6f0\") " Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.215671 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "bc20c5f3-b0e5-47e2-8878-015b66dad6f0" (UID: "bc20c5f3-b0e5-47e2-8878-015b66dad6f0"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.215898 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-kube-api-access-m4fkw" (OuterVolumeSpecName: "kube-api-access-m4fkw") pod "bc20c5f3-b0e5-47e2-8878-015b66dad6f0" (UID: "bc20c5f3-b0e5-47e2-8878-015b66dad6f0"). InnerVolumeSpecName "kube-api-access-m4fkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.240252 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bc20c5f3-b0e5-47e2-8878-015b66dad6f0" (UID: "bc20c5f3-b0e5-47e2-8878-015b66dad6f0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.246002 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "bc20c5f3-b0e5-47e2-8878-015b66dad6f0" (UID: "bc20c5f3-b0e5-47e2-8878-015b66dad6f0"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.248398 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-inventory" (OuterVolumeSpecName: "inventory") pod "bc20c5f3-b0e5-47e2-8878-015b66dad6f0" (UID: "bc20c5f3-b0e5-47e2-8878-015b66dad6f0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.313415 4776 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.313456 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4fkw\" (UniqueName: \"kubernetes.io/projected/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-kube-api-access-m4fkw\") on node \"crc\" DevicePath \"\"" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.313475 4776 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.313486 4776 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.313499 4776 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc20c5f3-b0e5-47e2-8878-015b66dad6f0-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.627334 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-c2plb" event={"ID":"bc20c5f3-b0e5-47e2-8878-015b66dad6f0","Type":"ContainerDied","Data":"6daaa3738cf087a97cf513fe5e4b64fcf6898f1714b30c590d3d991a2ed48098"} Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.627389 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6daaa3738cf087a97cf513fe5e4b64fcf6898f1714b30c590d3d991a2ed48098" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.627414 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-c2plb" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.731211 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-bh42p"] Nov 25 11:42:40 crc kubenswrapper[4776]: E1125 11:42:40.731721 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2510036-0cf5-471f-8166-71f71f8b6bb9" containerName="extract-utilities" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.731744 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2510036-0cf5-471f-8166-71f71f8b6bb9" containerName="extract-utilities" Nov 25 11:42:40 crc kubenswrapper[4776]: E1125 11:42:40.731765 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2510036-0cf5-471f-8166-71f71f8b6bb9" containerName="extract-content" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.731774 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2510036-0cf5-471f-8166-71f71f8b6bb9" containerName="extract-content" Nov 25 11:42:40 crc kubenswrapper[4776]: E1125 11:42:40.731806 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2510036-0cf5-471f-8166-71f71f8b6bb9" containerName="registry-server" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.731813 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2510036-0cf5-471f-8166-71f71f8b6bb9" containerName="registry-server" Nov 25 11:42:40 crc kubenswrapper[4776]: E1125 11:42:40.731835 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc20c5f3-b0e5-47e2-8878-015b66dad6f0" containerName="libvirt-openstack-openstack-cell1" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.731843 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc20c5f3-b0e5-47e2-8878-015b66dad6f0" containerName="libvirt-openstack-openstack-cell1" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.732108 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2510036-0cf5-471f-8166-71f71f8b6bb9" containerName="registry-server" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.732133 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc20c5f3-b0e5-47e2-8878-015b66dad6f0" containerName="libvirt-openstack-openstack-cell1" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.734403 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.758499 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.758851 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.759149 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.759390 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.759587 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.759828 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dknh7" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.760014 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.772555 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-bh42p"] Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.825790 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-bh42p\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.825854 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-bh42p\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.825937 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-bh42p\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.825966 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24dh6\" (UniqueName: \"kubernetes.io/projected/2aba01e8-085c-4bef-b6bc-1ae855ced88f-kube-api-access-24dh6\") pod \"nova-cell1-openstack-openstack-cell1-bh42p\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.826090 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-bh42p\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.826128 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-inventory\") pod \"nova-cell1-openstack-openstack-cell1-bh42p\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.826173 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-bh42p\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.826273 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-bh42p\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.826358 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-bh42p\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.927989 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-bh42p\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.928458 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24dh6\" (UniqueName: \"kubernetes.io/projected/2aba01e8-085c-4bef-b6bc-1ae855ced88f-kube-api-access-24dh6\") pod \"nova-cell1-openstack-openstack-cell1-bh42p\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.928632 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-bh42p\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.928727 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-inventory\") pod \"nova-cell1-openstack-openstack-cell1-bh42p\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.928880 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-bh42p\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.929034 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-bh42p\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.929247 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-bh42p\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.929394 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-bh42p\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.929496 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-bh42p\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.930226 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-bh42p\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.935649 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-bh42p\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.936468 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-bh42p\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.937376 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-bh42p\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.938857 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-inventory\") pod \"nova-cell1-openstack-openstack-cell1-bh42p\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.939226 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-bh42p\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.939371 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-bh42p\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.945732 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-bh42p\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:40 crc kubenswrapper[4776]: I1125 11:42:40.949780 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24dh6\" (UniqueName: \"kubernetes.io/projected/2aba01e8-085c-4bef-b6bc-1ae855ced88f-kube-api-access-24dh6\") pod \"nova-cell1-openstack-openstack-cell1-bh42p\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:41 crc kubenswrapper[4776]: I1125 11:42:41.073044 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:42:41 crc kubenswrapper[4776]: I1125 11:42:41.678758 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-bh42p"] Nov 25 11:42:42 crc kubenswrapper[4776]: I1125 11:42:42.652683 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" event={"ID":"2aba01e8-085c-4bef-b6bc-1ae855ced88f","Type":"ContainerStarted","Data":"ca041aef605bdafb3c79e72e6233888fb7b8cf69c8e11109a0ab953235b0254c"} Nov 25 11:42:42 crc kubenswrapper[4776]: I1125 11:42:42.653365 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" event={"ID":"2aba01e8-085c-4bef-b6bc-1ae855ced88f","Type":"ContainerStarted","Data":"9706d8a8ad261d49f0ff4cb351ba83a70c78b9b1f162b5e524669577ff3882aa"} Nov 25 11:42:42 crc kubenswrapper[4776]: I1125 11:42:42.682751 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" podStartSLOduration=2.187038087 podStartE2EDuration="2.682733115s" podCreationTimestamp="2025-11-25 11:42:40 +0000 UTC" firstStartedPulling="2025-11-25 11:42:41.669462821 +0000 UTC m=+8306.710522374" lastFinishedPulling="2025-11-25 11:42:42.165157849 +0000 UTC m=+8307.206217402" observedRunningTime="2025-11-25 11:42:42.676207271 +0000 UTC m=+8307.717266834" watchObservedRunningTime="2025-11-25 11:42:42.682733115 +0000 UTC m=+8307.723792668" Nov 25 11:42:49 crc kubenswrapper[4776]: I1125 11:42:49.663146 4776 scope.go:117] "RemoveContainer" containerID="3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b" Nov 25 11:42:49 crc kubenswrapper[4776]: E1125 11:42:49.664453 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:43:01 crc kubenswrapper[4776]: I1125 11:43:01.663746 4776 scope.go:117] "RemoveContainer" containerID="3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b" Nov 25 11:43:01 crc kubenswrapper[4776]: E1125 11:43:01.664506 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:43:12 crc kubenswrapper[4776]: I1125 11:43:12.662142 4776 scope.go:117] "RemoveContainer" containerID="3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b" Nov 25 11:43:12 crc kubenswrapper[4776]: E1125 11:43:12.662890 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:43:25 crc kubenswrapper[4776]: I1125 11:43:25.669132 4776 scope.go:117] "RemoveContainer" containerID="3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b" Nov 25 11:43:25 crc kubenswrapper[4776]: E1125 11:43:25.669914 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:43:39 crc kubenswrapper[4776]: I1125 11:43:39.662826 4776 scope.go:117] "RemoveContainer" containerID="3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b" Nov 25 11:43:39 crc kubenswrapper[4776]: E1125 11:43:39.663749 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:43:54 crc kubenswrapper[4776]: I1125 11:43:54.663585 4776 scope.go:117] "RemoveContainer" containerID="3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b" Nov 25 11:43:54 crc kubenswrapper[4776]: E1125 11:43:54.664221 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:44:06 crc kubenswrapper[4776]: I1125 11:44:06.663227 4776 scope.go:117] "RemoveContainer" containerID="3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b" Nov 25 11:44:06 crc kubenswrapper[4776]: E1125 11:44:06.664685 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:44:21 crc kubenswrapper[4776]: I1125 11:44:21.662326 4776 scope.go:117] "RemoveContainer" containerID="3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b" Nov 25 11:44:21 crc kubenswrapper[4776]: E1125 11:44:21.663121 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:44:34 crc kubenswrapper[4776]: I1125 11:44:34.663358 4776 scope.go:117] "RemoveContainer" containerID="3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b" Nov 25 11:44:34 crc kubenswrapper[4776]: E1125 11:44:34.664731 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:44:48 crc kubenswrapper[4776]: I1125 11:44:48.662766 4776 scope.go:117] "RemoveContainer" containerID="3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b" Nov 25 11:44:48 crc kubenswrapper[4776]: E1125 11:44:48.663823 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:45:00 crc kubenswrapper[4776]: I1125 11:45:00.148728 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401185-rwjj5"] Nov 25 11:45:00 crc kubenswrapper[4776]: I1125 11:45:00.151006 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-rwjj5" Nov 25 11:45:00 crc kubenswrapper[4776]: I1125 11:45:00.153674 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 11:45:00 crc kubenswrapper[4776]: I1125 11:45:00.153674 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 11:45:00 crc kubenswrapper[4776]: I1125 11:45:00.160616 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401185-rwjj5"] Nov 25 11:45:00 crc kubenswrapper[4776]: I1125 11:45:00.185841 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8dd6c817-f320-44c4-b3d9-46551476961a-config-volume\") pod \"collect-profiles-29401185-rwjj5\" (UID: \"8dd6c817-f320-44c4-b3d9-46551476961a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-rwjj5" Nov 25 11:45:00 crc kubenswrapper[4776]: I1125 11:45:00.185902 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slgpw\" (UniqueName: \"kubernetes.io/projected/8dd6c817-f320-44c4-b3d9-46551476961a-kube-api-access-slgpw\") pod \"collect-profiles-29401185-rwjj5\" (UID: \"8dd6c817-f320-44c4-b3d9-46551476961a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-rwjj5" Nov 25 11:45:00 crc kubenswrapper[4776]: I1125 11:45:00.186193 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8dd6c817-f320-44c4-b3d9-46551476961a-secret-volume\") pod \"collect-profiles-29401185-rwjj5\" (UID: \"8dd6c817-f320-44c4-b3d9-46551476961a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-rwjj5" Nov 25 11:45:00 crc kubenswrapper[4776]: I1125 11:45:00.288658 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8dd6c817-f320-44c4-b3d9-46551476961a-config-volume\") pod \"collect-profiles-29401185-rwjj5\" (UID: \"8dd6c817-f320-44c4-b3d9-46551476961a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-rwjj5" Nov 25 11:45:00 crc kubenswrapper[4776]: I1125 11:45:00.288726 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slgpw\" (UniqueName: \"kubernetes.io/projected/8dd6c817-f320-44c4-b3d9-46551476961a-kube-api-access-slgpw\") pod \"collect-profiles-29401185-rwjj5\" (UID: \"8dd6c817-f320-44c4-b3d9-46551476961a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-rwjj5" Nov 25 11:45:00 crc kubenswrapper[4776]: I1125 11:45:00.288828 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8dd6c817-f320-44c4-b3d9-46551476961a-secret-volume\") pod \"collect-profiles-29401185-rwjj5\" (UID: \"8dd6c817-f320-44c4-b3d9-46551476961a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-rwjj5" Nov 25 11:45:00 crc kubenswrapper[4776]: I1125 11:45:00.289588 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8dd6c817-f320-44c4-b3d9-46551476961a-config-volume\") pod \"collect-profiles-29401185-rwjj5\" (UID: \"8dd6c817-f320-44c4-b3d9-46551476961a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-rwjj5" Nov 25 11:45:00 crc kubenswrapper[4776]: I1125 11:45:00.296504 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8dd6c817-f320-44c4-b3d9-46551476961a-secret-volume\") pod \"collect-profiles-29401185-rwjj5\" (UID: \"8dd6c817-f320-44c4-b3d9-46551476961a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-rwjj5" Nov 25 11:45:00 crc kubenswrapper[4776]: I1125 11:45:00.304900 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slgpw\" (UniqueName: \"kubernetes.io/projected/8dd6c817-f320-44c4-b3d9-46551476961a-kube-api-access-slgpw\") pod \"collect-profiles-29401185-rwjj5\" (UID: \"8dd6c817-f320-44c4-b3d9-46551476961a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-rwjj5" Nov 25 11:45:00 crc kubenswrapper[4776]: I1125 11:45:00.480359 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-rwjj5" Nov 25 11:45:00 crc kubenswrapper[4776]: I1125 11:45:00.970489 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401185-rwjj5"] Nov 25 11:45:01 crc kubenswrapper[4776]: I1125 11:45:01.130482 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-rwjj5" event={"ID":"8dd6c817-f320-44c4-b3d9-46551476961a","Type":"ContainerStarted","Data":"82f7580b72dc3f8635239a56ac3ae3d3665e75e61cb046446322c5f5a0dd6f4b"} Nov 25 11:45:02 crc kubenswrapper[4776]: I1125 11:45:02.141319 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-rwjj5" event={"ID":"8dd6c817-f320-44c4-b3d9-46551476961a","Type":"ContainerStarted","Data":"e1781e134e2e83d48777366bb15706ac4c749c760d67ba16cbd8db4315ee74e4"} Nov 25 11:45:02 crc kubenswrapper[4776]: I1125 11:45:02.163109 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-rwjj5" podStartSLOduration=2.163088007 podStartE2EDuration="2.163088007s" podCreationTimestamp="2025-11-25 11:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:45:02.157905497 +0000 UTC m=+8447.198965060" watchObservedRunningTime="2025-11-25 11:45:02.163088007 +0000 UTC m=+8447.204147560" Nov 25 11:45:03 crc kubenswrapper[4776]: I1125 11:45:03.152413 4776 generic.go:334] "Generic (PLEG): container finished" podID="8dd6c817-f320-44c4-b3d9-46551476961a" containerID="e1781e134e2e83d48777366bb15706ac4c749c760d67ba16cbd8db4315ee74e4" exitCode=0 Nov 25 11:45:03 crc kubenswrapper[4776]: I1125 11:45:03.152459 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-rwjj5" event={"ID":"8dd6c817-f320-44c4-b3d9-46551476961a","Type":"ContainerDied","Data":"e1781e134e2e83d48777366bb15706ac4c749c760d67ba16cbd8db4315ee74e4"} Nov 25 11:45:03 crc kubenswrapper[4776]: I1125 11:45:03.662435 4776 scope.go:117] "RemoveContainer" containerID="3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b" Nov 25 11:45:03 crc kubenswrapper[4776]: E1125 11:45:03.662752 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:45:04 crc kubenswrapper[4776]: I1125 11:45:04.677370 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-rwjj5" Nov 25 11:45:04 crc kubenswrapper[4776]: I1125 11:45:04.793577 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8dd6c817-f320-44c4-b3d9-46551476961a-secret-volume\") pod \"8dd6c817-f320-44c4-b3d9-46551476961a\" (UID: \"8dd6c817-f320-44c4-b3d9-46551476961a\") " Nov 25 11:45:04 crc kubenswrapper[4776]: I1125 11:45:04.793707 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slgpw\" (UniqueName: \"kubernetes.io/projected/8dd6c817-f320-44c4-b3d9-46551476961a-kube-api-access-slgpw\") pod \"8dd6c817-f320-44c4-b3d9-46551476961a\" (UID: \"8dd6c817-f320-44c4-b3d9-46551476961a\") " Nov 25 11:45:04 crc kubenswrapper[4776]: I1125 11:45:04.793739 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8dd6c817-f320-44c4-b3d9-46551476961a-config-volume\") pod \"8dd6c817-f320-44c4-b3d9-46551476961a\" (UID: \"8dd6c817-f320-44c4-b3d9-46551476961a\") " Nov 25 11:45:04 crc kubenswrapper[4776]: I1125 11:45:04.794991 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8dd6c817-f320-44c4-b3d9-46551476961a-config-volume" (OuterVolumeSpecName: "config-volume") pod "8dd6c817-f320-44c4-b3d9-46551476961a" (UID: "8dd6c817-f320-44c4-b3d9-46551476961a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:45:04 crc kubenswrapper[4776]: I1125 11:45:04.799526 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dd6c817-f320-44c4-b3d9-46551476961a-kube-api-access-slgpw" (OuterVolumeSpecName: "kube-api-access-slgpw") pod "8dd6c817-f320-44c4-b3d9-46551476961a" (UID: "8dd6c817-f320-44c4-b3d9-46551476961a"). InnerVolumeSpecName "kube-api-access-slgpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:45:04 crc kubenswrapper[4776]: I1125 11:45:04.800795 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8dd6c817-f320-44c4-b3d9-46551476961a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8dd6c817-f320-44c4-b3d9-46551476961a" (UID: "8dd6c817-f320-44c4-b3d9-46551476961a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:45:04 crc kubenswrapper[4776]: I1125 11:45:04.898649 4776 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8dd6c817-f320-44c4-b3d9-46551476961a-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 11:45:04 crc kubenswrapper[4776]: I1125 11:45:04.898687 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slgpw\" (UniqueName: \"kubernetes.io/projected/8dd6c817-f320-44c4-b3d9-46551476961a-kube-api-access-slgpw\") on node \"crc\" DevicePath \"\"" Nov 25 11:45:04 crc kubenswrapper[4776]: I1125 11:45:04.898696 4776 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8dd6c817-f320-44c4-b3d9-46551476961a-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 11:45:05 crc kubenswrapper[4776]: I1125 11:45:05.170842 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-rwjj5" event={"ID":"8dd6c817-f320-44c4-b3d9-46551476961a","Type":"ContainerDied","Data":"82f7580b72dc3f8635239a56ac3ae3d3665e75e61cb046446322c5f5a0dd6f4b"} Nov 25 11:45:05 crc kubenswrapper[4776]: I1125 11:45:05.170883 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82f7580b72dc3f8635239a56ac3ae3d3665e75e61cb046446322c5f5a0dd6f4b" Nov 25 11:45:05 crc kubenswrapper[4776]: I1125 11:45:05.170921 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401185-rwjj5" Nov 25 11:45:05 crc kubenswrapper[4776]: I1125 11:45:05.256218 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401140-xc252"] Nov 25 11:45:05 crc kubenswrapper[4776]: I1125 11:45:05.274131 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401140-xc252"] Nov 25 11:45:05 crc kubenswrapper[4776]: I1125 11:45:05.675881 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51d095bb-1d12-4939-aeb6-104e88e0163f" path="/var/lib/kubelet/pods/51d095bb-1d12-4939-aeb6-104e88e0163f/volumes" Nov 25 11:45:15 crc kubenswrapper[4776]: I1125 11:45:15.669890 4776 scope.go:117] "RemoveContainer" containerID="3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b" Nov 25 11:45:15 crc kubenswrapper[4776]: E1125 11:45:15.670767 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:45:29 crc kubenswrapper[4776]: I1125 11:45:29.663239 4776 scope.go:117] "RemoveContainer" containerID="3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b" Nov 25 11:45:30 crc kubenswrapper[4776]: I1125 11:45:30.401878 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"b3060280eff383f881b2e41e47023c8d8874f894afc80e5917b1ff4623cd5d74"} Nov 25 11:45:41 crc kubenswrapper[4776]: I1125 11:45:41.936144 4776 scope.go:117] "RemoveContainer" containerID="ba23c2ddb4c24f894d2222b9501f72d438e8d6a63f91eba6b3c8c1c692bd9bdf" Nov 25 11:45:43 crc kubenswrapper[4776]: I1125 11:45:43.555723 4776 generic.go:334] "Generic (PLEG): container finished" podID="2aba01e8-085c-4bef-b6bc-1ae855ced88f" containerID="ca041aef605bdafb3c79e72e6233888fb7b8cf69c8e11109a0ab953235b0254c" exitCode=0 Nov 25 11:45:43 crc kubenswrapper[4776]: I1125 11:45:43.555831 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" event={"ID":"2aba01e8-085c-4bef-b6bc-1ae855ced88f","Type":"ContainerDied","Data":"ca041aef605bdafb3c79e72e6233888fb7b8cf69c8e11109a0ab953235b0254c"} Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.014642 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.106690 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-migration-ssh-key-1\") pod \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.106996 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-ssh-key\") pod \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.107147 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-cells-global-config-0\") pod \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.107179 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-cell1-combined-ca-bundle\") pod \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.107440 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-cell1-compute-config-0\") pod \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.107656 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-cell1-compute-config-1\") pod \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.107734 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-migration-ssh-key-0\") pod \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.107768 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24dh6\" (UniqueName: \"kubernetes.io/projected/2aba01e8-085c-4bef-b6bc-1ae855ced88f-kube-api-access-24dh6\") pod \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.107794 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-inventory\") pod \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\" (UID: \"2aba01e8-085c-4bef-b6bc-1ae855ced88f\") " Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.116821 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "2aba01e8-085c-4bef-b6bc-1ae855ced88f" (UID: "2aba01e8-085c-4bef-b6bc-1ae855ced88f"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.116851 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aba01e8-085c-4bef-b6bc-1ae855ced88f-kube-api-access-24dh6" (OuterVolumeSpecName: "kube-api-access-24dh6") pod "2aba01e8-085c-4bef-b6bc-1ae855ced88f" (UID: "2aba01e8-085c-4bef-b6bc-1ae855ced88f"). InnerVolumeSpecName "kube-api-access-24dh6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.140076 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "2aba01e8-085c-4bef-b6bc-1ae855ced88f" (UID: "2aba01e8-085c-4bef-b6bc-1ae855ced88f"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.145321 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "2aba01e8-085c-4bef-b6bc-1ae855ced88f" (UID: "2aba01e8-085c-4bef-b6bc-1ae855ced88f"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.146222 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-inventory" (OuterVolumeSpecName: "inventory") pod "2aba01e8-085c-4bef-b6bc-1ae855ced88f" (UID: "2aba01e8-085c-4bef-b6bc-1ae855ced88f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.147972 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2aba01e8-085c-4bef-b6bc-1ae855ced88f" (UID: "2aba01e8-085c-4bef-b6bc-1ae855ced88f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.153537 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "2aba01e8-085c-4bef-b6bc-1ae855ced88f" (UID: "2aba01e8-085c-4bef-b6bc-1ae855ced88f"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.161523 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "2aba01e8-085c-4bef-b6bc-1ae855ced88f" (UID: "2aba01e8-085c-4bef-b6bc-1ae855ced88f"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.171787 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "2aba01e8-085c-4bef-b6bc-1ae855ced88f" (UID: "2aba01e8-085c-4bef-b6bc-1ae855ced88f"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.210892 4776 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.210931 4776 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.210941 4776 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.210952 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24dh6\" (UniqueName: \"kubernetes.io/projected/2aba01e8-085c-4bef-b6bc-1ae855ced88f-kube-api-access-24dh6\") on node \"crc\" DevicePath \"\"" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.210963 4776 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.210972 4776 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.210981 4776 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.210990 4776 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.210999 4776 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aba01e8-085c-4bef-b6bc-1ae855ced88f-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.590849 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" event={"ID":"2aba01e8-085c-4bef-b6bc-1ae855ced88f","Type":"ContainerDied","Data":"9706d8a8ad261d49f0ff4cb351ba83a70c78b9b1f162b5e524669577ff3882aa"} Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.590901 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9706d8a8ad261d49f0ff4cb351ba83a70c78b9b1f162b5e524669577ff3882aa" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.590914 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-bh42p" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.778888 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-vdhqb"] Nov 25 11:45:45 crc kubenswrapper[4776]: E1125 11:45:45.779690 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aba01e8-085c-4bef-b6bc-1ae855ced88f" containerName="nova-cell1-openstack-openstack-cell1" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.779707 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aba01e8-085c-4bef-b6bc-1ae855ced88f" containerName="nova-cell1-openstack-openstack-cell1" Nov 25 11:45:45 crc kubenswrapper[4776]: E1125 11:45:45.780081 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dd6c817-f320-44c4-b3d9-46551476961a" containerName="collect-profiles" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.780094 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dd6c817-f320-44c4-b3d9-46551476961a" containerName="collect-profiles" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.780405 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aba01e8-085c-4bef-b6bc-1ae855ced88f" containerName="nova-cell1-openstack-openstack-cell1" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.780436 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dd6c817-f320-44c4-b3d9-46551476961a" containerName="collect-profiles" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.781588 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.786583 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.786588 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.787028 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dknh7" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.787878 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.788644 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.813367 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-vdhqb"] Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.824153 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-vdhqb\" (UID: \"13669d03-2d4a-4bed-8386-cfcae6272ffb\") " pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.824589 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-inventory\") pod \"telemetry-openstack-openstack-cell1-vdhqb\" (UID: \"13669d03-2d4a-4bed-8386-cfcae6272ffb\") " pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.824716 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-vdhqb\" (UID: \"13669d03-2d4a-4bed-8386-cfcae6272ffb\") " pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.824835 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkjr4\" (UniqueName: \"kubernetes.io/projected/13669d03-2d4a-4bed-8386-cfcae6272ffb-kube-api-access-tkjr4\") pod \"telemetry-openstack-openstack-cell1-vdhqb\" (UID: \"13669d03-2d4a-4bed-8386-cfcae6272ffb\") " pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.824969 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-vdhqb\" (UID: \"13669d03-2d4a-4bed-8386-cfcae6272ffb\") " pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.825390 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-ssh-key\") pod \"telemetry-openstack-openstack-cell1-vdhqb\" (UID: \"13669d03-2d4a-4bed-8386-cfcae6272ffb\") " pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.825532 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-vdhqb\" (UID: \"13669d03-2d4a-4bed-8386-cfcae6272ffb\") " pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" Nov 25 11:45:45 crc kubenswrapper[4776]: E1125 11:45:45.896213 4776 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2aba01e8_085c_4bef_b6bc_1ae855ced88f.slice/crio-9706d8a8ad261d49f0ff4cb351ba83a70c78b9b1f162b5e524669577ff3882aa\": RecentStats: unable to find data in memory cache]" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.927614 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-ssh-key\") pod \"telemetry-openstack-openstack-cell1-vdhqb\" (UID: \"13669d03-2d4a-4bed-8386-cfcae6272ffb\") " pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.927904 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-vdhqb\" (UID: \"13669d03-2d4a-4bed-8386-cfcae6272ffb\") " pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.928045 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-vdhqb\" (UID: \"13669d03-2d4a-4bed-8386-cfcae6272ffb\") " pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.928303 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-inventory\") pod \"telemetry-openstack-openstack-cell1-vdhqb\" (UID: \"13669d03-2d4a-4bed-8386-cfcae6272ffb\") " pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.928423 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-vdhqb\" (UID: \"13669d03-2d4a-4bed-8386-cfcae6272ffb\") " pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.928518 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkjr4\" (UniqueName: \"kubernetes.io/projected/13669d03-2d4a-4bed-8386-cfcae6272ffb-kube-api-access-tkjr4\") pod \"telemetry-openstack-openstack-cell1-vdhqb\" (UID: \"13669d03-2d4a-4bed-8386-cfcae6272ffb\") " pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.928606 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-vdhqb\" (UID: \"13669d03-2d4a-4bed-8386-cfcae6272ffb\") " pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.933235 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-vdhqb\" (UID: \"13669d03-2d4a-4bed-8386-cfcae6272ffb\") " pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.933586 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-vdhqb\" (UID: \"13669d03-2d4a-4bed-8386-cfcae6272ffb\") " pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.933721 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-ssh-key\") pod \"telemetry-openstack-openstack-cell1-vdhqb\" (UID: \"13669d03-2d4a-4bed-8386-cfcae6272ffb\") " pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.934126 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-inventory\") pod \"telemetry-openstack-openstack-cell1-vdhqb\" (UID: \"13669d03-2d4a-4bed-8386-cfcae6272ffb\") " pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.934861 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-vdhqb\" (UID: \"13669d03-2d4a-4bed-8386-cfcae6272ffb\") " pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.938792 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-vdhqb\" (UID: \"13669d03-2d4a-4bed-8386-cfcae6272ffb\") " pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" Nov 25 11:45:45 crc kubenswrapper[4776]: I1125 11:45:45.945893 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkjr4\" (UniqueName: \"kubernetes.io/projected/13669d03-2d4a-4bed-8386-cfcae6272ffb-kube-api-access-tkjr4\") pod \"telemetry-openstack-openstack-cell1-vdhqb\" (UID: \"13669d03-2d4a-4bed-8386-cfcae6272ffb\") " pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" Nov 25 11:45:46 crc kubenswrapper[4776]: I1125 11:45:46.115709 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" Nov 25 11:45:46 crc kubenswrapper[4776]: I1125 11:45:46.672545 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-vdhqb"] Nov 25 11:45:46 crc kubenswrapper[4776]: W1125 11:45:46.674332 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13669d03_2d4a_4bed_8386_cfcae6272ffb.slice/crio-e8175a2702b6c724f8a699d024dc8e08935732e4f97e51d566858127880fe57b WatchSource:0}: Error finding container e8175a2702b6c724f8a699d024dc8e08935732e4f97e51d566858127880fe57b: Status 404 returned error can't find the container with id e8175a2702b6c724f8a699d024dc8e08935732e4f97e51d566858127880fe57b Nov 25 11:45:46 crc kubenswrapper[4776]: I1125 11:45:46.677841 4776 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 11:45:47 crc kubenswrapper[4776]: I1125 11:45:47.618101 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" event={"ID":"13669d03-2d4a-4bed-8386-cfcae6272ffb","Type":"ContainerStarted","Data":"e8175a2702b6c724f8a699d024dc8e08935732e4f97e51d566858127880fe57b"} Nov 25 11:45:48 crc kubenswrapper[4776]: I1125 11:45:48.632517 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" event={"ID":"13669d03-2d4a-4bed-8386-cfcae6272ffb","Type":"ContainerStarted","Data":"f8d5508c15fd69d2d87e6f9fddb92f43b03d03faae8b6c9de46793f010fea17a"} Nov 25 11:45:48 crc kubenswrapper[4776]: I1125 11:45:48.668918 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" podStartSLOduration=2.927660558 podStartE2EDuration="3.668896046s" podCreationTimestamp="2025-11-25 11:45:45 +0000 UTC" firstStartedPulling="2025-11-25 11:45:46.677517051 +0000 UTC m=+8491.718576604" lastFinishedPulling="2025-11-25 11:45:47.418752529 +0000 UTC m=+8492.459812092" observedRunningTime="2025-11-25 11:45:48.665357528 +0000 UTC m=+8493.706417121" watchObservedRunningTime="2025-11-25 11:45:48.668896046 +0000 UTC m=+8493.709955599" Nov 25 11:47:47 crc kubenswrapper[4776]: I1125 11:47:47.818031 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:47:47 crc kubenswrapper[4776]: I1125 11:47:47.818641 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:48:17 crc kubenswrapper[4776]: I1125 11:48:17.818249 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:48:17 crc kubenswrapper[4776]: I1125 11:48:17.818821 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:48:47 crc kubenswrapper[4776]: I1125 11:48:47.818510 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:48:47 crc kubenswrapper[4776]: I1125 11:48:47.819199 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:48:47 crc kubenswrapper[4776]: I1125 11:48:47.819287 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 11:48:47 crc kubenswrapper[4776]: I1125 11:48:47.821485 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b3060280eff383f881b2e41e47023c8d8874f894afc80e5917b1ff4623cd5d74"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 11:48:47 crc kubenswrapper[4776]: I1125 11:48:47.821627 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://b3060280eff383f881b2e41e47023c8d8874f894afc80e5917b1ff4623cd5d74" gracePeriod=600 Nov 25 11:48:48 crc kubenswrapper[4776]: I1125 11:48:48.862385 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="b3060280eff383f881b2e41e47023c8d8874f894afc80e5917b1ff4623cd5d74" exitCode=0 Nov 25 11:48:48 crc kubenswrapper[4776]: I1125 11:48:48.862445 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"b3060280eff383f881b2e41e47023c8d8874f894afc80e5917b1ff4623cd5d74"} Nov 25 11:48:48 crc kubenswrapper[4776]: I1125 11:48:48.862790 4776 scope.go:117] "RemoveContainer" containerID="3c1cfed665db753ed2420406a803b607ef5b6397a584a9bd8a0fe236412dd42b" Nov 25 11:48:49 crc kubenswrapper[4776]: I1125 11:48:49.874705 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965"} Nov 25 11:48:49 crc kubenswrapper[4776]: I1125 11:48:49.915231 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-c2d7k"] Nov 25 11:48:49 crc kubenswrapper[4776]: I1125 11:48:49.917734 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c2d7k" Nov 25 11:48:49 crc kubenswrapper[4776]: I1125 11:48:49.940058 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c2d7k"] Nov 25 11:48:49 crc kubenswrapper[4776]: I1125 11:48:49.995710 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d46339c-f862-49cf-b4e9-fde20e8688de-catalog-content\") pod \"redhat-operators-c2d7k\" (UID: \"5d46339c-f862-49cf-b4e9-fde20e8688de\") " pod="openshift-marketplace/redhat-operators-c2d7k" Nov 25 11:48:49 crc kubenswrapper[4776]: I1125 11:48:49.995830 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pp4nb\" (UniqueName: \"kubernetes.io/projected/5d46339c-f862-49cf-b4e9-fde20e8688de-kube-api-access-pp4nb\") pod \"redhat-operators-c2d7k\" (UID: \"5d46339c-f862-49cf-b4e9-fde20e8688de\") " pod="openshift-marketplace/redhat-operators-c2d7k" Nov 25 11:48:49 crc kubenswrapper[4776]: I1125 11:48:49.995879 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d46339c-f862-49cf-b4e9-fde20e8688de-utilities\") pod \"redhat-operators-c2d7k\" (UID: \"5d46339c-f862-49cf-b4e9-fde20e8688de\") " pod="openshift-marketplace/redhat-operators-c2d7k" Nov 25 11:48:50 crc kubenswrapper[4776]: I1125 11:48:50.098023 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d46339c-f862-49cf-b4e9-fde20e8688de-catalog-content\") pod \"redhat-operators-c2d7k\" (UID: \"5d46339c-f862-49cf-b4e9-fde20e8688de\") " pod="openshift-marketplace/redhat-operators-c2d7k" Nov 25 11:48:50 crc kubenswrapper[4776]: I1125 11:48:50.098631 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d46339c-f862-49cf-b4e9-fde20e8688de-catalog-content\") pod \"redhat-operators-c2d7k\" (UID: \"5d46339c-f862-49cf-b4e9-fde20e8688de\") " pod="openshift-marketplace/redhat-operators-c2d7k" Nov 25 11:48:50 crc kubenswrapper[4776]: I1125 11:48:50.098812 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pp4nb\" (UniqueName: \"kubernetes.io/projected/5d46339c-f862-49cf-b4e9-fde20e8688de-kube-api-access-pp4nb\") pod \"redhat-operators-c2d7k\" (UID: \"5d46339c-f862-49cf-b4e9-fde20e8688de\") " pod="openshift-marketplace/redhat-operators-c2d7k" Nov 25 11:48:50 crc kubenswrapper[4776]: I1125 11:48:50.098942 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d46339c-f862-49cf-b4e9-fde20e8688de-utilities\") pod \"redhat-operators-c2d7k\" (UID: \"5d46339c-f862-49cf-b4e9-fde20e8688de\") " pod="openshift-marketplace/redhat-operators-c2d7k" Nov 25 11:48:50 crc kubenswrapper[4776]: I1125 11:48:50.099392 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d46339c-f862-49cf-b4e9-fde20e8688de-utilities\") pod \"redhat-operators-c2d7k\" (UID: \"5d46339c-f862-49cf-b4e9-fde20e8688de\") " pod="openshift-marketplace/redhat-operators-c2d7k" Nov 25 11:48:50 crc kubenswrapper[4776]: I1125 11:48:50.121676 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pp4nb\" (UniqueName: \"kubernetes.io/projected/5d46339c-f862-49cf-b4e9-fde20e8688de-kube-api-access-pp4nb\") pod \"redhat-operators-c2d7k\" (UID: \"5d46339c-f862-49cf-b4e9-fde20e8688de\") " pod="openshift-marketplace/redhat-operators-c2d7k" Nov 25 11:48:50 crc kubenswrapper[4776]: I1125 11:48:50.256098 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c2d7k" Nov 25 11:48:50 crc kubenswrapper[4776]: I1125 11:48:50.737445 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c2d7k"] Nov 25 11:48:50 crc kubenswrapper[4776]: I1125 11:48:50.888910 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c2d7k" event={"ID":"5d46339c-f862-49cf-b4e9-fde20e8688de","Type":"ContainerStarted","Data":"eeb9cc0112b70cc9eebe65821a93d9f6423723c808f4e601e1f335774ffbd0c2"} Nov 25 11:48:51 crc kubenswrapper[4776]: I1125 11:48:51.899410 4776 generic.go:334] "Generic (PLEG): container finished" podID="5d46339c-f862-49cf-b4e9-fde20e8688de" containerID="df078ac879078387127e8a572684e5c759a13e9c6f2c32f86eda539f296a3ee5" exitCode=0 Nov 25 11:48:51 crc kubenswrapper[4776]: I1125 11:48:51.899615 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c2d7k" event={"ID":"5d46339c-f862-49cf-b4e9-fde20e8688de","Type":"ContainerDied","Data":"df078ac879078387127e8a572684e5c759a13e9c6f2c32f86eda539f296a3ee5"} Nov 25 11:49:04 crc kubenswrapper[4776]: I1125 11:49:04.031864 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c2d7k" event={"ID":"5d46339c-f862-49cf-b4e9-fde20e8688de","Type":"ContainerStarted","Data":"9b5247cfb1debb63cb9a23c998c064414f4dd7ed92818af4406b27566aa9cb0f"} Nov 25 11:49:05 crc kubenswrapper[4776]: I1125 11:49:05.044527 4776 generic.go:334] "Generic (PLEG): container finished" podID="5d46339c-f862-49cf-b4e9-fde20e8688de" containerID="9b5247cfb1debb63cb9a23c998c064414f4dd7ed92818af4406b27566aa9cb0f" exitCode=0 Nov 25 11:49:05 crc kubenswrapper[4776]: I1125 11:49:05.044585 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c2d7k" event={"ID":"5d46339c-f862-49cf-b4e9-fde20e8688de","Type":"ContainerDied","Data":"9b5247cfb1debb63cb9a23c998c064414f4dd7ed92818af4406b27566aa9cb0f"} Nov 25 11:49:06 crc kubenswrapper[4776]: I1125 11:49:06.056305 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c2d7k" event={"ID":"5d46339c-f862-49cf-b4e9-fde20e8688de","Type":"ContainerStarted","Data":"1d6b88f37c903e8f794d3621671c036f410252d9f3457bc5d5bda8130d3d5985"} Nov 25 11:49:06 crc kubenswrapper[4776]: I1125 11:49:06.085467 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-c2d7k" podStartSLOduration=3.501223811 podStartE2EDuration="17.085448355s" podCreationTimestamp="2025-11-25 11:48:49 +0000 UTC" firstStartedPulling="2025-11-25 11:48:51.90168795 +0000 UTC m=+8676.942747503" lastFinishedPulling="2025-11-25 11:49:05.485912494 +0000 UTC m=+8690.526972047" observedRunningTime="2025-11-25 11:49:06.075168738 +0000 UTC m=+8691.116228291" watchObservedRunningTime="2025-11-25 11:49:06.085448355 +0000 UTC m=+8691.126507908" Nov 25 11:49:10 crc kubenswrapper[4776]: I1125 11:49:10.257258 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-c2d7k" Nov 25 11:49:10 crc kubenswrapper[4776]: I1125 11:49:10.257826 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-c2d7k" Nov 25 11:49:11 crc kubenswrapper[4776]: I1125 11:49:11.309950 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-c2d7k" podUID="5d46339c-f862-49cf-b4e9-fde20e8688de" containerName="registry-server" probeResult="failure" output=< Nov 25 11:49:11 crc kubenswrapper[4776]: timeout: failed to connect service ":50051" within 1s Nov 25 11:49:11 crc kubenswrapper[4776]: > Nov 25 11:49:17 crc kubenswrapper[4776]: I1125 11:49:17.001017 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7ckr9"] Nov 25 11:49:17 crc kubenswrapper[4776]: I1125 11:49:17.009507 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7ckr9" Nov 25 11:49:17 crc kubenswrapper[4776]: I1125 11:49:17.015318 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7ckr9"] Nov 25 11:49:17 crc kubenswrapper[4776]: I1125 11:49:17.168791 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxbt6\" (UniqueName: \"kubernetes.io/projected/67720d0b-3037-48a3-81d4-1d01c2753961-kube-api-access-mxbt6\") pod \"community-operators-7ckr9\" (UID: \"67720d0b-3037-48a3-81d4-1d01c2753961\") " pod="openshift-marketplace/community-operators-7ckr9" Nov 25 11:49:17 crc kubenswrapper[4776]: I1125 11:49:17.168847 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67720d0b-3037-48a3-81d4-1d01c2753961-catalog-content\") pod \"community-operators-7ckr9\" (UID: \"67720d0b-3037-48a3-81d4-1d01c2753961\") " pod="openshift-marketplace/community-operators-7ckr9" Nov 25 11:49:17 crc kubenswrapper[4776]: I1125 11:49:17.168898 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67720d0b-3037-48a3-81d4-1d01c2753961-utilities\") pod \"community-operators-7ckr9\" (UID: \"67720d0b-3037-48a3-81d4-1d01c2753961\") " pod="openshift-marketplace/community-operators-7ckr9" Nov 25 11:49:17 crc kubenswrapper[4776]: I1125 11:49:17.271410 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxbt6\" (UniqueName: \"kubernetes.io/projected/67720d0b-3037-48a3-81d4-1d01c2753961-kube-api-access-mxbt6\") pod \"community-operators-7ckr9\" (UID: \"67720d0b-3037-48a3-81d4-1d01c2753961\") " pod="openshift-marketplace/community-operators-7ckr9" Nov 25 11:49:17 crc kubenswrapper[4776]: I1125 11:49:17.271483 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67720d0b-3037-48a3-81d4-1d01c2753961-catalog-content\") pod \"community-operators-7ckr9\" (UID: \"67720d0b-3037-48a3-81d4-1d01c2753961\") " pod="openshift-marketplace/community-operators-7ckr9" Nov 25 11:49:17 crc kubenswrapper[4776]: I1125 11:49:17.271576 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67720d0b-3037-48a3-81d4-1d01c2753961-utilities\") pod \"community-operators-7ckr9\" (UID: \"67720d0b-3037-48a3-81d4-1d01c2753961\") " pod="openshift-marketplace/community-operators-7ckr9" Nov 25 11:49:17 crc kubenswrapper[4776]: I1125 11:49:17.272011 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67720d0b-3037-48a3-81d4-1d01c2753961-catalog-content\") pod \"community-operators-7ckr9\" (UID: \"67720d0b-3037-48a3-81d4-1d01c2753961\") " pod="openshift-marketplace/community-operators-7ckr9" Nov 25 11:49:17 crc kubenswrapper[4776]: I1125 11:49:17.272149 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67720d0b-3037-48a3-81d4-1d01c2753961-utilities\") pod \"community-operators-7ckr9\" (UID: \"67720d0b-3037-48a3-81d4-1d01c2753961\") " pod="openshift-marketplace/community-operators-7ckr9" Nov 25 11:49:17 crc kubenswrapper[4776]: I1125 11:49:17.293672 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxbt6\" (UniqueName: \"kubernetes.io/projected/67720d0b-3037-48a3-81d4-1d01c2753961-kube-api-access-mxbt6\") pod \"community-operators-7ckr9\" (UID: \"67720d0b-3037-48a3-81d4-1d01c2753961\") " pod="openshift-marketplace/community-operators-7ckr9" Nov 25 11:49:17 crc kubenswrapper[4776]: I1125 11:49:17.339013 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7ckr9" Nov 25 11:49:17 crc kubenswrapper[4776]: I1125 11:49:17.902559 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7ckr9"] Nov 25 11:49:18 crc kubenswrapper[4776]: I1125 11:49:18.189400 4776 generic.go:334] "Generic (PLEG): container finished" podID="67720d0b-3037-48a3-81d4-1d01c2753961" containerID="ef23b68bd82a3ceeb078d2a71fee964d6ccefa1662ea457c1dcaeda1f14e5a1c" exitCode=0 Nov 25 11:49:18 crc kubenswrapper[4776]: I1125 11:49:18.189484 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7ckr9" event={"ID":"67720d0b-3037-48a3-81d4-1d01c2753961","Type":"ContainerDied","Data":"ef23b68bd82a3ceeb078d2a71fee964d6ccefa1662ea457c1dcaeda1f14e5a1c"} Nov 25 11:49:18 crc kubenswrapper[4776]: I1125 11:49:18.189705 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7ckr9" event={"ID":"67720d0b-3037-48a3-81d4-1d01c2753961","Type":"ContainerStarted","Data":"b08af7851593a3094fb2fff9bd93a8a9759bfd2fbd9e17b739f58278895feaa6"} Nov 25 11:49:19 crc kubenswrapper[4776]: I1125 11:49:19.204973 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7ckr9" event={"ID":"67720d0b-3037-48a3-81d4-1d01c2753961","Type":"ContainerStarted","Data":"a1603964a8e670ff3ea3ca9cc39f5f66aed36977d9e04b417b09490c326246ab"} Nov 25 11:49:20 crc kubenswrapper[4776]: I1125 11:49:20.317486 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-c2d7k" Nov 25 11:49:20 crc kubenswrapper[4776]: I1125 11:49:20.369330 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-c2d7k" Nov 25 11:49:21 crc kubenswrapper[4776]: I1125 11:49:21.242801 4776 generic.go:334] "Generic (PLEG): container finished" podID="67720d0b-3037-48a3-81d4-1d01c2753961" containerID="a1603964a8e670ff3ea3ca9cc39f5f66aed36977d9e04b417b09490c326246ab" exitCode=0 Nov 25 11:49:21 crc kubenswrapper[4776]: I1125 11:49:21.242875 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7ckr9" event={"ID":"67720d0b-3037-48a3-81d4-1d01c2753961","Type":"ContainerDied","Data":"a1603964a8e670ff3ea3ca9cc39f5f66aed36977d9e04b417b09490c326246ab"} Nov 25 11:49:21 crc kubenswrapper[4776]: I1125 11:49:21.794599 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c2d7k"] Nov 25 11:49:22 crc kubenswrapper[4776]: I1125 11:49:22.167383 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-54gpf"] Nov 25 11:49:22 crc kubenswrapper[4776]: I1125 11:49:22.168033 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-54gpf" podUID="7066eebe-a597-40d1-b9f1-f6cde8df2666" containerName="registry-server" containerID="cri-o://bffa44dbf4705824967a378be71a548630361ec0d949a1793858a16bda1292c8" gracePeriod=2 Nov 25 11:49:22 crc kubenswrapper[4776]: I1125 11:49:22.260013 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7ckr9" event={"ID":"67720d0b-3037-48a3-81d4-1d01c2753961","Type":"ContainerStarted","Data":"92d05a781978fe787630fa4a349572de94fabe571b00b157c7f015b9bd342dff"} Nov 25 11:49:22 crc kubenswrapper[4776]: I1125 11:49:22.294375 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7ckr9" podStartSLOduration=2.550818647 podStartE2EDuration="6.294354203s" podCreationTimestamp="2025-11-25 11:49:16 +0000 UTC" firstStartedPulling="2025-11-25 11:49:18.192486216 +0000 UTC m=+8703.233545769" lastFinishedPulling="2025-11-25 11:49:21.936021772 +0000 UTC m=+8706.977081325" observedRunningTime="2025-11-25 11:49:22.279803878 +0000 UTC m=+8707.320863431" watchObservedRunningTime="2025-11-25 11:49:22.294354203 +0000 UTC m=+8707.335413756" Nov 25 11:49:22 crc kubenswrapper[4776]: I1125 11:49:22.937330 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-54gpf" Nov 25 11:49:23 crc kubenswrapper[4776]: I1125 11:49:23.055966 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxt6g\" (UniqueName: \"kubernetes.io/projected/7066eebe-a597-40d1-b9f1-f6cde8df2666-kube-api-access-lxt6g\") pod \"7066eebe-a597-40d1-b9f1-f6cde8df2666\" (UID: \"7066eebe-a597-40d1-b9f1-f6cde8df2666\") " Nov 25 11:49:23 crc kubenswrapper[4776]: I1125 11:49:23.056343 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7066eebe-a597-40d1-b9f1-f6cde8df2666-catalog-content\") pod \"7066eebe-a597-40d1-b9f1-f6cde8df2666\" (UID: \"7066eebe-a597-40d1-b9f1-f6cde8df2666\") " Nov 25 11:49:23 crc kubenswrapper[4776]: I1125 11:49:23.056438 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7066eebe-a597-40d1-b9f1-f6cde8df2666-utilities\") pod \"7066eebe-a597-40d1-b9f1-f6cde8df2666\" (UID: \"7066eebe-a597-40d1-b9f1-f6cde8df2666\") " Nov 25 11:49:23 crc kubenswrapper[4776]: I1125 11:49:23.058430 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7066eebe-a597-40d1-b9f1-f6cde8df2666-utilities" (OuterVolumeSpecName: "utilities") pod "7066eebe-a597-40d1-b9f1-f6cde8df2666" (UID: "7066eebe-a597-40d1-b9f1-f6cde8df2666"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:49:23 crc kubenswrapper[4776]: I1125 11:49:23.059613 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7066eebe-a597-40d1-b9f1-f6cde8df2666-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:49:23 crc kubenswrapper[4776]: I1125 11:49:23.088619 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7066eebe-a597-40d1-b9f1-f6cde8df2666-kube-api-access-lxt6g" (OuterVolumeSpecName: "kube-api-access-lxt6g") pod "7066eebe-a597-40d1-b9f1-f6cde8df2666" (UID: "7066eebe-a597-40d1-b9f1-f6cde8df2666"). InnerVolumeSpecName "kube-api-access-lxt6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:49:23 crc kubenswrapper[4776]: I1125 11:49:23.154122 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7066eebe-a597-40d1-b9f1-f6cde8df2666-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7066eebe-a597-40d1-b9f1-f6cde8df2666" (UID: "7066eebe-a597-40d1-b9f1-f6cde8df2666"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:49:23 crc kubenswrapper[4776]: I1125 11:49:23.161837 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7066eebe-a597-40d1-b9f1-f6cde8df2666-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:49:23 crc kubenswrapper[4776]: I1125 11:49:23.161872 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxt6g\" (UniqueName: \"kubernetes.io/projected/7066eebe-a597-40d1-b9f1-f6cde8df2666-kube-api-access-lxt6g\") on node \"crc\" DevicePath \"\"" Nov 25 11:49:23 crc kubenswrapper[4776]: I1125 11:49:23.278019 4776 generic.go:334] "Generic (PLEG): container finished" podID="7066eebe-a597-40d1-b9f1-f6cde8df2666" containerID="bffa44dbf4705824967a378be71a548630361ec0d949a1793858a16bda1292c8" exitCode=0 Nov 25 11:49:23 crc kubenswrapper[4776]: I1125 11:49:23.278118 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-54gpf" event={"ID":"7066eebe-a597-40d1-b9f1-f6cde8df2666","Type":"ContainerDied","Data":"bffa44dbf4705824967a378be71a548630361ec0d949a1793858a16bda1292c8"} Nov 25 11:49:23 crc kubenswrapper[4776]: I1125 11:49:23.278462 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-54gpf" event={"ID":"7066eebe-a597-40d1-b9f1-f6cde8df2666","Type":"ContainerDied","Data":"e92483d2c7f3f642b9a4a96360cf677f08ee178e32e95ef42619c4836c943020"} Nov 25 11:49:23 crc kubenswrapper[4776]: I1125 11:49:23.278494 4776 scope.go:117] "RemoveContainer" containerID="bffa44dbf4705824967a378be71a548630361ec0d949a1793858a16bda1292c8" Nov 25 11:49:23 crc kubenswrapper[4776]: I1125 11:49:23.278133 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-54gpf" Nov 25 11:49:23 crc kubenswrapper[4776]: I1125 11:49:23.320664 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-54gpf"] Nov 25 11:49:23 crc kubenswrapper[4776]: I1125 11:49:23.322315 4776 scope.go:117] "RemoveContainer" containerID="a62c20d9b3b7300c478a5443dd50d616686b5715fcec5645c01d6e2c237f34f1" Nov 25 11:49:23 crc kubenswrapper[4776]: I1125 11:49:23.339419 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-54gpf"] Nov 25 11:49:23 crc kubenswrapper[4776]: I1125 11:49:23.676516 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7066eebe-a597-40d1-b9f1-f6cde8df2666" path="/var/lib/kubelet/pods/7066eebe-a597-40d1-b9f1-f6cde8df2666/volumes" Nov 25 11:49:23 crc kubenswrapper[4776]: I1125 11:49:23.684822 4776 scope.go:117] "RemoveContainer" containerID="5687b5e1d8cdedc8c37fe871b0e6a016cb89ff745077d03b02410dfd40ff560b" Nov 25 11:49:23 crc kubenswrapper[4776]: I1125 11:49:23.739874 4776 scope.go:117] "RemoveContainer" containerID="bffa44dbf4705824967a378be71a548630361ec0d949a1793858a16bda1292c8" Nov 25 11:49:23 crc kubenswrapper[4776]: E1125 11:49:23.740398 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bffa44dbf4705824967a378be71a548630361ec0d949a1793858a16bda1292c8\": container with ID starting with bffa44dbf4705824967a378be71a548630361ec0d949a1793858a16bda1292c8 not found: ID does not exist" containerID="bffa44dbf4705824967a378be71a548630361ec0d949a1793858a16bda1292c8" Nov 25 11:49:23 crc kubenswrapper[4776]: I1125 11:49:23.740440 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bffa44dbf4705824967a378be71a548630361ec0d949a1793858a16bda1292c8"} err="failed to get container status \"bffa44dbf4705824967a378be71a548630361ec0d949a1793858a16bda1292c8\": rpc error: code = NotFound desc = could not find container \"bffa44dbf4705824967a378be71a548630361ec0d949a1793858a16bda1292c8\": container with ID starting with bffa44dbf4705824967a378be71a548630361ec0d949a1793858a16bda1292c8 not found: ID does not exist" Nov 25 11:49:23 crc kubenswrapper[4776]: I1125 11:49:23.740487 4776 scope.go:117] "RemoveContainer" containerID="a62c20d9b3b7300c478a5443dd50d616686b5715fcec5645c01d6e2c237f34f1" Nov 25 11:49:23 crc kubenswrapper[4776]: E1125 11:49:23.740801 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a62c20d9b3b7300c478a5443dd50d616686b5715fcec5645c01d6e2c237f34f1\": container with ID starting with a62c20d9b3b7300c478a5443dd50d616686b5715fcec5645c01d6e2c237f34f1 not found: ID does not exist" containerID="a62c20d9b3b7300c478a5443dd50d616686b5715fcec5645c01d6e2c237f34f1" Nov 25 11:49:23 crc kubenswrapper[4776]: I1125 11:49:23.740832 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a62c20d9b3b7300c478a5443dd50d616686b5715fcec5645c01d6e2c237f34f1"} err="failed to get container status \"a62c20d9b3b7300c478a5443dd50d616686b5715fcec5645c01d6e2c237f34f1\": rpc error: code = NotFound desc = could not find container \"a62c20d9b3b7300c478a5443dd50d616686b5715fcec5645c01d6e2c237f34f1\": container with ID starting with a62c20d9b3b7300c478a5443dd50d616686b5715fcec5645c01d6e2c237f34f1 not found: ID does not exist" Nov 25 11:49:23 crc kubenswrapper[4776]: I1125 11:49:23.740851 4776 scope.go:117] "RemoveContainer" containerID="5687b5e1d8cdedc8c37fe871b0e6a016cb89ff745077d03b02410dfd40ff560b" Nov 25 11:49:23 crc kubenswrapper[4776]: E1125 11:49:23.741230 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5687b5e1d8cdedc8c37fe871b0e6a016cb89ff745077d03b02410dfd40ff560b\": container with ID starting with 5687b5e1d8cdedc8c37fe871b0e6a016cb89ff745077d03b02410dfd40ff560b not found: ID does not exist" containerID="5687b5e1d8cdedc8c37fe871b0e6a016cb89ff745077d03b02410dfd40ff560b" Nov 25 11:49:23 crc kubenswrapper[4776]: I1125 11:49:23.741278 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5687b5e1d8cdedc8c37fe871b0e6a016cb89ff745077d03b02410dfd40ff560b"} err="failed to get container status \"5687b5e1d8cdedc8c37fe871b0e6a016cb89ff745077d03b02410dfd40ff560b\": rpc error: code = NotFound desc = could not find container \"5687b5e1d8cdedc8c37fe871b0e6a016cb89ff745077d03b02410dfd40ff560b\": container with ID starting with 5687b5e1d8cdedc8c37fe871b0e6a016cb89ff745077d03b02410dfd40ff560b not found: ID does not exist" Nov 25 11:49:27 crc kubenswrapper[4776]: I1125 11:49:27.339221 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7ckr9" Nov 25 11:49:27 crc kubenswrapper[4776]: I1125 11:49:27.340911 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7ckr9" Nov 25 11:49:28 crc kubenswrapper[4776]: I1125 11:49:28.399586 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-7ckr9" podUID="67720d0b-3037-48a3-81d4-1d01c2753961" containerName="registry-server" probeResult="failure" output=< Nov 25 11:49:28 crc kubenswrapper[4776]: timeout: failed to connect service ":50051" within 1s Nov 25 11:49:28 crc kubenswrapper[4776]: > Nov 25 11:49:38 crc kubenswrapper[4776]: I1125 11:49:38.908564 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-7ckr9" podUID="67720d0b-3037-48a3-81d4-1d01c2753961" containerName="registry-server" probeResult="failure" output=< Nov 25 11:49:38 crc kubenswrapper[4776]: timeout: failed to connect service ":50051" within 1s Nov 25 11:49:38 crc kubenswrapper[4776]: > Nov 25 11:49:47 crc kubenswrapper[4776]: I1125 11:49:47.392490 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7ckr9" Nov 25 11:49:47 crc kubenswrapper[4776]: I1125 11:49:47.449021 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7ckr9" Nov 25 11:49:48 crc kubenswrapper[4776]: I1125 11:49:48.195906 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7ckr9"] Nov 25 11:49:48 crc kubenswrapper[4776]: I1125 11:49:48.521957 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7ckr9" podUID="67720d0b-3037-48a3-81d4-1d01c2753961" containerName="registry-server" containerID="cri-o://92d05a781978fe787630fa4a349572de94fabe571b00b157c7f015b9bd342dff" gracePeriod=2 Nov 25 11:49:49 crc kubenswrapper[4776]: I1125 11:49:49.533115 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7ckr9" Nov 25 11:49:49 crc kubenswrapper[4776]: I1125 11:49:49.533309 4776 generic.go:334] "Generic (PLEG): container finished" podID="67720d0b-3037-48a3-81d4-1d01c2753961" containerID="92d05a781978fe787630fa4a349572de94fabe571b00b157c7f015b9bd342dff" exitCode=0 Nov 25 11:49:49 crc kubenswrapper[4776]: I1125 11:49:49.533335 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7ckr9" event={"ID":"67720d0b-3037-48a3-81d4-1d01c2753961","Type":"ContainerDied","Data":"92d05a781978fe787630fa4a349572de94fabe571b00b157c7f015b9bd342dff"} Nov 25 11:49:49 crc kubenswrapper[4776]: I1125 11:49:49.534344 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7ckr9" event={"ID":"67720d0b-3037-48a3-81d4-1d01c2753961","Type":"ContainerDied","Data":"b08af7851593a3094fb2fff9bd93a8a9759bfd2fbd9e17b739f58278895feaa6"} Nov 25 11:49:49 crc kubenswrapper[4776]: I1125 11:49:49.534380 4776 scope.go:117] "RemoveContainer" containerID="92d05a781978fe787630fa4a349572de94fabe571b00b157c7f015b9bd342dff" Nov 25 11:49:49 crc kubenswrapper[4776]: I1125 11:49:49.570349 4776 scope.go:117] "RemoveContainer" containerID="a1603964a8e670ff3ea3ca9cc39f5f66aed36977d9e04b417b09490c326246ab" Nov 25 11:49:49 crc kubenswrapper[4776]: I1125 11:49:49.603187 4776 scope.go:117] "RemoveContainer" containerID="ef23b68bd82a3ceeb078d2a71fee964d6ccefa1662ea457c1dcaeda1f14e5a1c" Nov 25 11:49:49 crc kubenswrapper[4776]: I1125 11:49:49.641205 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxbt6\" (UniqueName: \"kubernetes.io/projected/67720d0b-3037-48a3-81d4-1d01c2753961-kube-api-access-mxbt6\") pod \"67720d0b-3037-48a3-81d4-1d01c2753961\" (UID: \"67720d0b-3037-48a3-81d4-1d01c2753961\") " Nov 25 11:49:49 crc kubenswrapper[4776]: I1125 11:49:49.641259 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67720d0b-3037-48a3-81d4-1d01c2753961-catalog-content\") pod \"67720d0b-3037-48a3-81d4-1d01c2753961\" (UID: \"67720d0b-3037-48a3-81d4-1d01c2753961\") " Nov 25 11:49:49 crc kubenswrapper[4776]: I1125 11:49:49.641543 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67720d0b-3037-48a3-81d4-1d01c2753961-utilities\") pod \"67720d0b-3037-48a3-81d4-1d01c2753961\" (UID: \"67720d0b-3037-48a3-81d4-1d01c2753961\") " Nov 25 11:49:49 crc kubenswrapper[4776]: I1125 11:49:49.642302 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67720d0b-3037-48a3-81d4-1d01c2753961-utilities" (OuterVolumeSpecName: "utilities") pod "67720d0b-3037-48a3-81d4-1d01c2753961" (UID: "67720d0b-3037-48a3-81d4-1d01c2753961"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:49:49 crc kubenswrapper[4776]: I1125 11:49:49.643670 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67720d0b-3037-48a3-81d4-1d01c2753961-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:49:49 crc kubenswrapper[4776]: I1125 11:49:49.652412 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67720d0b-3037-48a3-81d4-1d01c2753961-kube-api-access-mxbt6" (OuterVolumeSpecName: "kube-api-access-mxbt6") pod "67720d0b-3037-48a3-81d4-1d01c2753961" (UID: "67720d0b-3037-48a3-81d4-1d01c2753961"). InnerVolumeSpecName "kube-api-access-mxbt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:49:49 crc kubenswrapper[4776]: I1125 11:49:49.699653 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67720d0b-3037-48a3-81d4-1d01c2753961-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "67720d0b-3037-48a3-81d4-1d01c2753961" (UID: "67720d0b-3037-48a3-81d4-1d01c2753961"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:49:49 crc kubenswrapper[4776]: I1125 11:49:49.725554 4776 scope.go:117] "RemoveContainer" containerID="92d05a781978fe787630fa4a349572de94fabe571b00b157c7f015b9bd342dff" Nov 25 11:49:49 crc kubenswrapper[4776]: E1125 11:49:49.726136 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92d05a781978fe787630fa4a349572de94fabe571b00b157c7f015b9bd342dff\": container with ID starting with 92d05a781978fe787630fa4a349572de94fabe571b00b157c7f015b9bd342dff not found: ID does not exist" containerID="92d05a781978fe787630fa4a349572de94fabe571b00b157c7f015b9bd342dff" Nov 25 11:49:49 crc kubenswrapper[4776]: I1125 11:49:49.726175 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92d05a781978fe787630fa4a349572de94fabe571b00b157c7f015b9bd342dff"} err="failed to get container status \"92d05a781978fe787630fa4a349572de94fabe571b00b157c7f015b9bd342dff\": rpc error: code = NotFound desc = could not find container \"92d05a781978fe787630fa4a349572de94fabe571b00b157c7f015b9bd342dff\": container with ID starting with 92d05a781978fe787630fa4a349572de94fabe571b00b157c7f015b9bd342dff not found: ID does not exist" Nov 25 11:49:49 crc kubenswrapper[4776]: I1125 11:49:49.726217 4776 scope.go:117] "RemoveContainer" containerID="a1603964a8e670ff3ea3ca9cc39f5f66aed36977d9e04b417b09490c326246ab" Nov 25 11:49:49 crc kubenswrapper[4776]: E1125 11:49:49.729800 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1603964a8e670ff3ea3ca9cc39f5f66aed36977d9e04b417b09490c326246ab\": container with ID starting with a1603964a8e670ff3ea3ca9cc39f5f66aed36977d9e04b417b09490c326246ab not found: ID does not exist" containerID="a1603964a8e670ff3ea3ca9cc39f5f66aed36977d9e04b417b09490c326246ab" Nov 25 11:49:49 crc kubenswrapper[4776]: I1125 11:49:49.729832 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1603964a8e670ff3ea3ca9cc39f5f66aed36977d9e04b417b09490c326246ab"} err="failed to get container status \"a1603964a8e670ff3ea3ca9cc39f5f66aed36977d9e04b417b09490c326246ab\": rpc error: code = NotFound desc = could not find container \"a1603964a8e670ff3ea3ca9cc39f5f66aed36977d9e04b417b09490c326246ab\": container with ID starting with a1603964a8e670ff3ea3ca9cc39f5f66aed36977d9e04b417b09490c326246ab not found: ID does not exist" Nov 25 11:49:49 crc kubenswrapper[4776]: I1125 11:49:49.729847 4776 scope.go:117] "RemoveContainer" containerID="ef23b68bd82a3ceeb078d2a71fee964d6ccefa1662ea457c1dcaeda1f14e5a1c" Nov 25 11:49:49 crc kubenswrapper[4776]: E1125 11:49:49.730301 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef23b68bd82a3ceeb078d2a71fee964d6ccefa1662ea457c1dcaeda1f14e5a1c\": container with ID starting with ef23b68bd82a3ceeb078d2a71fee964d6ccefa1662ea457c1dcaeda1f14e5a1c not found: ID does not exist" containerID="ef23b68bd82a3ceeb078d2a71fee964d6ccefa1662ea457c1dcaeda1f14e5a1c" Nov 25 11:49:49 crc kubenswrapper[4776]: I1125 11:49:49.730353 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef23b68bd82a3ceeb078d2a71fee964d6ccefa1662ea457c1dcaeda1f14e5a1c"} err="failed to get container status \"ef23b68bd82a3ceeb078d2a71fee964d6ccefa1662ea457c1dcaeda1f14e5a1c\": rpc error: code = NotFound desc = could not find container \"ef23b68bd82a3ceeb078d2a71fee964d6ccefa1662ea457c1dcaeda1f14e5a1c\": container with ID starting with ef23b68bd82a3ceeb078d2a71fee964d6ccefa1662ea457c1dcaeda1f14e5a1c not found: ID does not exist" Nov 25 11:49:49 crc kubenswrapper[4776]: I1125 11:49:49.746936 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxbt6\" (UniqueName: \"kubernetes.io/projected/67720d0b-3037-48a3-81d4-1d01c2753961-kube-api-access-mxbt6\") on node \"crc\" DevicePath \"\"" Nov 25 11:49:49 crc kubenswrapper[4776]: I1125 11:49:49.747254 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67720d0b-3037-48a3-81d4-1d01c2753961-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:49:50 crc kubenswrapper[4776]: I1125 11:49:50.551718 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7ckr9" Nov 25 11:49:50 crc kubenswrapper[4776]: I1125 11:49:50.589959 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7ckr9"] Nov 25 11:49:50 crc kubenswrapper[4776]: I1125 11:49:50.602977 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7ckr9"] Nov 25 11:49:51 crc kubenswrapper[4776]: I1125 11:49:51.683781 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67720d0b-3037-48a3-81d4-1d01c2753961" path="/var/lib/kubelet/pods/67720d0b-3037-48a3-81d4-1d01c2753961/volumes" Nov 25 11:50:05 crc kubenswrapper[4776]: I1125 11:50:05.695028 4776 generic.go:334] "Generic (PLEG): container finished" podID="13669d03-2d4a-4bed-8386-cfcae6272ffb" containerID="f8d5508c15fd69d2d87e6f9fddb92f43b03d03faae8b6c9de46793f010fea17a" exitCode=0 Nov 25 11:50:05 crc kubenswrapper[4776]: I1125 11:50:05.695125 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" event={"ID":"13669d03-2d4a-4bed-8386-cfcae6272ffb","Type":"ContainerDied","Data":"f8d5508c15fd69d2d87e6f9fddb92f43b03d03faae8b6c9de46793f010fea17a"} Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.180064 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.346735 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-ceilometer-compute-config-data-0\") pod \"13669d03-2d4a-4bed-8386-cfcae6272ffb\" (UID: \"13669d03-2d4a-4bed-8386-cfcae6272ffb\") " Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.346846 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-inventory\") pod \"13669d03-2d4a-4bed-8386-cfcae6272ffb\" (UID: \"13669d03-2d4a-4bed-8386-cfcae6272ffb\") " Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.346942 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-ceilometer-compute-config-data-2\") pod \"13669d03-2d4a-4bed-8386-cfcae6272ffb\" (UID: \"13669d03-2d4a-4bed-8386-cfcae6272ffb\") " Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.346976 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-telemetry-combined-ca-bundle\") pod \"13669d03-2d4a-4bed-8386-cfcae6272ffb\" (UID: \"13669d03-2d4a-4bed-8386-cfcae6272ffb\") " Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.347036 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-ssh-key\") pod \"13669d03-2d4a-4bed-8386-cfcae6272ffb\" (UID: \"13669d03-2d4a-4bed-8386-cfcae6272ffb\") " Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.347082 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-ceilometer-compute-config-data-1\") pod \"13669d03-2d4a-4bed-8386-cfcae6272ffb\" (UID: \"13669d03-2d4a-4bed-8386-cfcae6272ffb\") " Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.347171 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkjr4\" (UniqueName: \"kubernetes.io/projected/13669d03-2d4a-4bed-8386-cfcae6272ffb-kube-api-access-tkjr4\") pod \"13669d03-2d4a-4bed-8386-cfcae6272ffb\" (UID: \"13669d03-2d4a-4bed-8386-cfcae6272ffb\") " Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.354484 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "13669d03-2d4a-4bed-8386-cfcae6272ffb" (UID: "13669d03-2d4a-4bed-8386-cfcae6272ffb"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.354549 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13669d03-2d4a-4bed-8386-cfcae6272ffb-kube-api-access-tkjr4" (OuterVolumeSpecName: "kube-api-access-tkjr4") pod "13669d03-2d4a-4bed-8386-cfcae6272ffb" (UID: "13669d03-2d4a-4bed-8386-cfcae6272ffb"). InnerVolumeSpecName "kube-api-access-tkjr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.383971 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "13669d03-2d4a-4bed-8386-cfcae6272ffb" (UID: "13669d03-2d4a-4bed-8386-cfcae6272ffb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.385229 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "13669d03-2d4a-4bed-8386-cfcae6272ffb" (UID: "13669d03-2d4a-4bed-8386-cfcae6272ffb"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.386467 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "13669d03-2d4a-4bed-8386-cfcae6272ffb" (UID: "13669d03-2d4a-4bed-8386-cfcae6272ffb"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.390344 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "13669d03-2d4a-4bed-8386-cfcae6272ffb" (UID: "13669d03-2d4a-4bed-8386-cfcae6272ffb"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.396207 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-inventory" (OuterVolumeSpecName: "inventory") pod "13669d03-2d4a-4bed-8386-cfcae6272ffb" (UID: "13669d03-2d4a-4bed-8386-cfcae6272ffb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.451817 4776 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.451881 4776 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.451900 4776 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.451916 4776 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.451935 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkjr4\" (UniqueName: \"kubernetes.io/projected/13669d03-2d4a-4bed-8386-cfcae6272ffb-kube-api-access-tkjr4\") on node \"crc\" DevicePath \"\"" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.451952 4776 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.451971 4776 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/13669d03-2d4a-4bed-8386-cfcae6272ffb-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.721604 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" event={"ID":"13669d03-2d4a-4bed-8386-cfcae6272ffb","Type":"ContainerDied","Data":"e8175a2702b6c724f8a699d024dc8e08935732e4f97e51d566858127880fe57b"} Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.721654 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8175a2702b6c724f8a699d024dc8e08935732e4f97e51d566858127880fe57b" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.721820 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-vdhqb" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.813874 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-x7ctw"] Nov 25 11:50:07 crc kubenswrapper[4776]: E1125 11:50:07.814518 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7066eebe-a597-40d1-b9f1-f6cde8df2666" containerName="extract-utilities" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.814540 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="7066eebe-a597-40d1-b9f1-f6cde8df2666" containerName="extract-utilities" Nov 25 11:50:07 crc kubenswrapper[4776]: E1125 11:50:07.814568 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13669d03-2d4a-4bed-8386-cfcae6272ffb" containerName="telemetry-openstack-openstack-cell1" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.814576 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="13669d03-2d4a-4bed-8386-cfcae6272ffb" containerName="telemetry-openstack-openstack-cell1" Nov 25 11:50:07 crc kubenswrapper[4776]: E1125 11:50:07.814589 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7066eebe-a597-40d1-b9f1-f6cde8df2666" containerName="registry-server" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.814598 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="7066eebe-a597-40d1-b9f1-f6cde8df2666" containerName="registry-server" Nov 25 11:50:07 crc kubenswrapper[4776]: E1125 11:50:07.814630 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7066eebe-a597-40d1-b9f1-f6cde8df2666" containerName="extract-content" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.814648 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="7066eebe-a597-40d1-b9f1-f6cde8df2666" containerName="extract-content" Nov 25 11:50:07 crc kubenswrapper[4776]: E1125 11:50:07.814660 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67720d0b-3037-48a3-81d4-1d01c2753961" containerName="extract-utilities" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.814668 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="67720d0b-3037-48a3-81d4-1d01c2753961" containerName="extract-utilities" Nov 25 11:50:07 crc kubenswrapper[4776]: E1125 11:50:07.814683 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67720d0b-3037-48a3-81d4-1d01c2753961" containerName="registry-server" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.814690 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="67720d0b-3037-48a3-81d4-1d01c2753961" containerName="registry-server" Nov 25 11:50:07 crc kubenswrapper[4776]: E1125 11:50:07.814709 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67720d0b-3037-48a3-81d4-1d01c2753961" containerName="extract-content" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.814716 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="67720d0b-3037-48a3-81d4-1d01c2753961" containerName="extract-content" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.814992 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="67720d0b-3037-48a3-81d4-1d01c2753961" containerName="registry-server" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.815008 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="7066eebe-a597-40d1-b9f1-f6cde8df2666" containerName="registry-server" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.815046 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="13669d03-2d4a-4bed-8386-cfcae6272ffb" containerName="telemetry-openstack-openstack-cell1" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.816301 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-x7ctw" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.819895 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dknh7" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.820120 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.820267 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.820431 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-sriov-agent-neutron-config" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.821622 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.825695 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-x7ctw"] Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.962814 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6r62x\" (UniqueName: \"kubernetes.io/projected/c622a8cc-2543-4ba5-b53f-32d680331106-kube-api-access-6r62x\") pod \"neutron-sriov-openstack-openstack-cell1-x7ctw\" (UID: \"c622a8cc-2543-4ba5-b53f-32d680331106\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x7ctw" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.963369 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c622a8cc-2543-4ba5-b53f-32d680331106-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-x7ctw\" (UID: \"c622a8cc-2543-4ba5-b53f-32d680331106\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x7ctw" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.964965 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c622a8cc-2543-4ba5-b53f-32d680331106-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-x7ctw\" (UID: \"c622a8cc-2543-4ba5-b53f-32d680331106\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x7ctw" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.965147 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c622a8cc-2543-4ba5-b53f-32d680331106-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-x7ctw\" (UID: \"c622a8cc-2543-4ba5-b53f-32d680331106\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x7ctw" Nov 25 11:50:07 crc kubenswrapper[4776]: I1125 11:50:07.965252 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c622a8cc-2543-4ba5-b53f-32d680331106-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-x7ctw\" (UID: \"c622a8cc-2543-4ba5-b53f-32d680331106\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x7ctw" Nov 25 11:50:08 crc kubenswrapper[4776]: I1125 11:50:08.067433 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c622a8cc-2543-4ba5-b53f-32d680331106-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-x7ctw\" (UID: \"c622a8cc-2543-4ba5-b53f-32d680331106\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x7ctw" Nov 25 11:50:08 crc kubenswrapper[4776]: I1125 11:50:08.067588 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6r62x\" (UniqueName: \"kubernetes.io/projected/c622a8cc-2543-4ba5-b53f-32d680331106-kube-api-access-6r62x\") pod \"neutron-sriov-openstack-openstack-cell1-x7ctw\" (UID: \"c622a8cc-2543-4ba5-b53f-32d680331106\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x7ctw" Nov 25 11:50:08 crc kubenswrapper[4776]: I1125 11:50:08.067617 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c622a8cc-2543-4ba5-b53f-32d680331106-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-x7ctw\" (UID: \"c622a8cc-2543-4ba5-b53f-32d680331106\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x7ctw" Nov 25 11:50:08 crc kubenswrapper[4776]: I1125 11:50:08.067646 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c622a8cc-2543-4ba5-b53f-32d680331106-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-x7ctw\" (UID: \"c622a8cc-2543-4ba5-b53f-32d680331106\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x7ctw" Nov 25 11:50:08 crc kubenswrapper[4776]: I1125 11:50:08.067781 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c622a8cc-2543-4ba5-b53f-32d680331106-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-x7ctw\" (UID: \"c622a8cc-2543-4ba5-b53f-32d680331106\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x7ctw" Nov 25 11:50:08 crc kubenswrapper[4776]: I1125 11:50:08.072019 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c622a8cc-2543-4ba5-b53f-32d680331106-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-x7ctw\" (UID: \"c622a8cc-2543-4ba5-b53f-32d680331106\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x7ctw" Nov 25 11:50:08 crc kubenswrapper[4776]: I1125 11:50:08.073843 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c622a8cc-2543-4ba5-b53f-32d680331106-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-x7ctw\" (UID: \"c622a8cc-2543-4ba5-b53f-32d680331106\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x7ctw" Nov 25 11:50:08 crc kubenswrapper[4776]: I1125 11:50:08.077872 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c622a8cc-2543-4ba5-b53f-32d680331106-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-x7ctw\" (UID: \"c622a8cc-2543-4ba5-b53f-32d680331106\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x7ctw" Nov 25 11:50:08 crc kubenswrapper[4776]: I1125 11:50:08.088585 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c622a8cc-2543-4ba5-b53f-32d680331106-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-x7ctw\" (UID: \"c622a8cc-2543-4ba5-b53f-32d680331106\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x7ctw" Nov 25 11:50:08 crc kubenswrapper[4776]: I1125 11:50:08.092823 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6r62x\" (UniqueName: \"kubernetes.io/projected/c622a8cc-2543-4ba5-b53f-32d680331106-kube-api-access-6r62x\") pod \"neutron-sriov-openstack-openstack-cell1-x7ctw\" (UID: \"c622a8cc-2543-4ba5-b53f-32d680331106\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-x7ctw" Nov 25 11:50:08 crc kubenswrapper[4776]: I1125 11:50:08.148567 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-x7ctw" Nov 25 11:50:08 crc kubenswrapper[4776]: I1125 11:50:08.726049 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-x7ctw"] Nov 25 11:50:08 crc kubenswrapper[4776]: W1125 11:50:08.735367 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc622a8cc_2543_4ba5_b53f_32d680331106.slice/crio-7970c5baa0242def0a7a5ee6f8b2d0e5477a409bedaba052e8176898c4de7d33 WatchSource:0}: Error finding container 7970c5baa0242def0a7a5ee6f8b2d0e5477a409bedaba052e8176898c4de7d33: Status 404 returned error can't find the container with id 7970c5baa0242def0a7a5ee6f8b2d0e5477a409bedaba052e8176898c4de7d33 Nov 25 11:50:09 crc kubenswrapper[4776]: I1125 11:50:09.756508 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-x7ctw" event={"ID":"c622a8cc-2543-4ba5-b53f-32d680331106","Type":"ContainerStarted","Data":"99c42b87324b208d20242c15110d042604354f0ba4aa57ae0867225ffc06e420"} Nov 25 11:50:09 crc kubenswrapper[4776]: I1125 11:50:09.756973 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-x7ctw" event={"ID":"c622a8cc-2543-4ba5-b53f-32d680331106","Type":"ContainerStarted","Data":"7970c5baa0242def0a7a5ee6f8b2d0e5477a409bedaba052e8176898c4de7d33"} Nov 25 11:50:09 crc kubenswrapper[4776]: I1125 11:50:09.780136 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-sriov-openstack-openstack-cell1-x7ctw" podStartSLOduration=2.32499501 podStartE2EDuration="2.780116429s" podCreationTimestamp="2025-11-25 11:50:07 +0000 UTC" firstStartedPulling="2025-11-25 11:50:08.738306339 +0000 UTC m=+8753.779365892" lastFinishedPulling="2025-11-25 11:50:09.193427758 +0000 UTC m=+8754.234487311" observedRunningTime="2025-11-25 11:50:09.77696987 +0000 UTC m=+8754.818029423" watchObservedRunningTime="2025-11-25 11:50:09.780116429 +0000 UTC m=+8754.821175982" Nov 25 11:50:10 crc kubenswrapper[4776]: I1125 11:50:10.086860 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-c2x7m"] Nov 25 11:50:10 crc kubenswrapper[4776]: I1125 11:50:10.090865 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c2x7m" Nov 25 11:50:10 crc kubenswrapper[4776]: I1125 11:50:10.097668 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c2x7m"] Nov 25 11:50:10 crc kubenswrapper[4776]: I1125 11:50:10.220413 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f46b22e-2c01-40b1-9e6e-4fb013683a65-utilities\") pod \"certified-operators-c2x7m\" (UID: \"2f46b22e-2c01-40b1-9e6e-4fb013683a65\") " pod="openshift-marketplace/certified-operators-c2x7m" Nov 25 11:50:10 crc kubenswrapper[4776]: I1125 11:50:10.220467 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qr7j\" (UniqueName: \"kubernetes.io/projected/2f46b22e-2c01-40b1-9e6e-4fb013683a65-kube-api-access-4qr7j\") pod \"certified-operators-c2x7m\" (UID: \"2f46b22e-2c01-40b1-9e6e-4fb013683a65\") " pod="openshift-marketplace/certified-operators-c2x7m" Nov 25 11:50:10 crc kubenswrapper[4776]: I1125 11:50:10.220590 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f46b22e-2c01-40b1-9e6e-4fb013683a65-catalog-content\") pod \"certified-operators-c2x7m\" (UID: \"2f46b22e-2c01-40b1-9e6e-4fb013683a65\") " pod="openshift-marketplace/certified-operators-c2x7m" Nov 25 11:50:10 crc kubenswrapper[4776]: I1125 11:50:10.322455 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f46b22e-2c01-40b1-9e6e-4fb013683a65-utilities\") pod \"certified-operators-c2x7m\" (UID: \"2f46b22e-2c01-40b1-9e6e-4fb013683a65\") " pod="openshift-marketplace/certified-operators-c2x7m" Nov 25 11:50:10 crc kubenswrapper[4776]: I1125 11:50:10.322824 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qr7j\" (UniqueName: \"kubernetes.io/projected/2f46b22e-2c01-40b1-9e6e-4fb013683a65-kube-api-access-4qr7j\") pod \"certified-operators-c2x7m\" (UID: \"2f46b22e-2c01-40b1-9e6e-4fb013683a65\") " pod="openshift-marketplace/certified-operators-c2x7m" Nov 25 11:50:10 crc kubenswrapper[4776]: I1125 11:50:10.322955 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f46b22e-2c01-40b1-9e6e-4fb013683a65-catalog-content\") pod \"certified-operators-c2x7m\" (UID: \"2f46b22e-2c01-40b1-9e6e-4fb013683a65\") " pod="openshift-marketplace/certified-operators-c2x7m" Nov 25 11:50:10 crc kubenswrapper[4776]: I1125 11:50:10.323035 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f46b22e-2c01-40b1-9e6e-4fb013683a65-utilities\") pod \"certified-operators-c2x7m\" (UID: \"2f46b22e-2c01-40b1-9e6e-4fb013683a65\") " pod="openshift-marketplace/certified-operators-c2x7m" Nov 25 11:50:10 crc kubenswrapper[4776]: I1125 11:50:10.323410 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f46b22e-2c01-40b1-9e6e-4fb013683a65-catalog-content\") pod \"certified-operators-c2x7m\" (UID: \"2f46b22e-2c01-40b1-9e6e-4fb013683a65\") " pod="openshift-marketplace/certified-operators-c2x7m" Nov 25 11:50:10 crc kubenswrapper[4776]: I1125 11:50:10.358870 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qr7j\" (UniqueName: \"kubernetes.io/projected/2f46b22e-2c01-40b1-9e6e-4fb013683a65-kube-api-access-4qr7j\") pod \"certified-operators-c2x7m\" (UID: \"2f46b22e-2c01-40b1-9e6e-4fb013683a65\") " pod="openshift-marketplace/certified-operators-c2x7m" Nov 25 11:50:10 crc kubenswrapper[4776]: I1125 11:50:10.422719 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c2x7m" Nov 25 11:50:10 crc kubenswrapper[4776]: I1125 11:50:10.967937 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c2x7m"] Nov 25 11:50:11 crc kubenswrapper[4776]: I1125 11:50:11.776970 4776 generic.go:334] "Generic (PLEG): container finished" podID="2f46b22e-2c01-40b1-9e6e-4fb013683a65" containerID="7ce177561ee68c609abc363f3723fba9f17d9b17512d6a352b4f740e74be718b" exitCode=0 Nov 25 11:50:11 crc kubenswrapper[4776]: I1125 11:50:11.777102 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c2x7m" event={"ID":"2f46b22e-2c01-40b1-9e6e-4fb013683a65","Type":"ContainerDied","Data":"7ce177561ee68c609abc363f3723fba9f17d9b17512d6a352b4f740e74be718b"} Nov 25 11:50:11 crc kubenswrapper[4776]: I1125 11:50:11.777347 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c2x7m" event={"ID":"2f46b22e-2c01-40b1-9e6e-4fb013683a65","Type":"ContainerStarted","Data":"6a012c8bf718b6c6058ead53fccceb780fffeaa0c775913363202ac0d5e38ae9"} Nov 25 11:50:13 crc kubenswrapper[4776]: I1125 11:50:13.800921 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c2x7m" event={"ID":"2f46b22e-2c01-40b1-9e6e-4fb013683a65","Type":"ContainerStarted","Data":"ea8d86e746bc491d7af4fe5920266f59856bc42497380e275a8581f2ed44d8d4"} Nov 25 11:50:14 crc kubenswrapper[4776]: I1125 11:50:14.815234 4776 generic.go:334] "Generic (PLEG): container finished" podID="2f46b22e-2c01-40b1-9e6e-4fb013683a65" containerID="ea8d86e746bc491d7af4fe5920266f59856bc42497380e275a8581f2ed44d8d4" exitCode=0 Nov 25 11:50:14 crc kubenswrapper[4776]: I1125 11:50:14.815439 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c2x7m" event={"ID":"2f46b22e-2c01-40b1-9e6e-4fb013683a65","Type":"ContainerDied","Data":"ea8d86e746bc491d7af4fe5920266f59856bc42497380e275a8581f2ed44d8d4"} Nov 25 11:50:15 crc kubenswrapper[4776]: I1125 11:50:15.830180 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c2x7m" event={"ID":"2f46b22e-2c01-40b1-9e6e-4fb013683a65","Type":"ContainerStarted","Data":"2584f68840e6778543c1c77a789353c40689b036086a54b96de7f3efa85ca422"} Nov 25 11:50:15 crc kubenswrapper[4776]: I1125 11:50:15.858782 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-c2x7m" podStartSLOduration=2.240351597 podStartE2EDuration="5.858763914s" podCreationTimestamp="2025-11-25 11:50:10 +0000 UTC" firstStartedPulling="2025-11-25 11:50:11.778678943 +0000 UTC m=+8756.819738496" lastFinishedPulling="2025-11-25 11:50:15.39709126 +0000 UTC m=+8760.438150813" observedRunningTime="2025-11-25 11:50:15.852471326 +0000 UTC m=+8760.893530889" watchObservedRunningTime="2025-11-25 11:50:15.858763914 +0000 UTC m=+8760.899823457" Nov 25 11:50:20 crc kubenswrapper[4776]: I1125 11:50:20.423787 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-c2x7m" Nov 25 11:50:20 crc kubenswrapper[4776]: I1125 11:50:20.424105 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-c2x7m" Nov 25 11:50:20 crc kubenswrapper[4776]: I1125 11:50:20.498515 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-c2x7m" Nov 25 11:50:20 crc kubenswrapper[4776]: I1125 11:50:20.926454 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-c2x7m" Nov 25 11:50:20 crc kubenswrapper[4776]: I1125 11:50:20.980047 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c2x7m"] Nov 25 11:50:22 crc kubenswrapper[4776]: I1125 11:50:22.894368 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-c2x7m" podUID="2f46b22e-2c01-40b1-9e6e-4fb013683a65" containerName="registry-server" containerID="cri-o://2584f68840e6778543c1c77a789353c40689b036086a54b96de7f3efa85ca422" gracePeriod=2 Nov 25 11:50:23 crc kubenswrapper[4776]: I1125 11:50:23.364304 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c2x7m" Nov 25 11:50:23 crc kubenswrapper[4776]: I1125 11:50:23.527794 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f46b22e-2c01-40b1-9e6e-4fb013683a65-catalog-content\") pod \"2f46b22e-2c01-40b1-9e6e-4fb013683a65\" (UID: \"2f46b22e-2c01-40b1-9e6e-4fb013683a65\") " Nov 25 11:50:23 crc kubenswrapper[4776]: I1125 11:50:23.528129 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qr7j\" (UniqueName: \"kubernetes.io/projected/2f46b22e-2c01-40b1-9e6e-4fb013683a65-kube-api-access-4qr7j\") pod \"2f46b22e-2c01-40b1-9e6e-4fb013683a65\" (UID: \"2f46b22e-2c01-40b1-9e6e-4fb013683a65\") " Nov 25 11:50:23 crc kubenswrapper[4776]: I1125 11:50:23.529622 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f46b22e-2c01-40b1-9e6e-4fb013683a65-utilities\") pod \"2f46b22e-2c01-40b1-9e6e-4fb013683a65\" (UID: \"2f46b22e-2c01-40b1-9e6e-4fb013683a65\") " Nov 25 11:50:23 crc kubenswrapper[4776]: I1125 11:50:23.530431 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f46b22e-2c01-40b1-9e6e-4fb013683a65-utilities" (OuterVolumeSpecName: "utilities") pod "2f46b22e-2c01-40b1-9e6e-4fb013683a65" (UID: "2f46b22e-2c01-40b1-9e6e-4fb013683a65"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:50:23 crc kubenswrapper[4776]: I1125 11:50:23.531038 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f46b22e-2c01-40b1-9e6e-4fb013683a65-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:50:23 crc kubenswrapper[4776]: I1125 11:50:23.536566 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f46b22e-2c01-40b1-9e6e-4fb013683a65-kube-api-access-4qr7j" (OuterVolumeSpecName: "kube-api-access-4qr7j") pod "2f46b22e-2c01-40b1-9e6e-4fb013683a65" (UID: "2f46b22e-2c01-40b1-9e6e-4fb013683a65"). InnerVolumeSpecName "kube-api-access-4qr7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:50:23 crc kubenswrapper[4776]: I1125 11:50:23.633034 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qr7j\" (UniqueName: \"kubernetes.io/projected/2f46b22e-2c01-40b1-9e6e-4fb013683a65-kube-api-access-4qr7j\") on node \"crc\" DevicePath \"\"" Nov 25 11:50:23 crc kubenswrapper[4776]: I1125 11:50:23.907360 4776 generic.go:334] "Generic (PLEG): container finished" podID="2f46b22e-2c01-40b1-9e6e-4fb013683a65" containerID="2584f68840e6778543c1c77a789353c40689b036086a54b96de7f3efa85ca422" exitCode=0 Nov 25 11:50:23 crc kubenswrapper[4776]: I1125 11:50:23.907424 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c2x7m" Nov 25 11:50:23 crc kubenswrapper[4776]: I1125 11:50:23.907438 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c2x7m" event={"ID":"2f46b22e-2c01-40b1-9e6e-4fb013683a65","Type":"ContainerDied","Data":"2584f68840e6778543c1c77a789353c40689b036086a54b96de7f3efa85ca422"} Nov 25 11:50:23 crc kubenswrapper[4776]: I1125 11:50:23.907806 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c2x7m" event={"ID":"2f46b22e-2c01-40b1-9e6e-4fb013683a65","Type":"ContainerDied","Data":"6a012c8bf718b6c6058ead53fccceb780fffeaa0c775913363202ac0d5e38ae9"} Nov 25 11:50:23 crc kubenswrapper[4776]: I1125 11:50:23.907832 4776 scope.go:117] "RemoveContainer" containerID="2584f68840e6778543c1c77a789353c40689b036086a54b96de7f3efa85ca422" Nov 25 11:50:23 crc kubenswrapper[4776]: I1125 11:50:23.932160 4776 scope.go:117] "RemoveContainer" containerID="ea8d86e746bc491d7af4fe5920266f59856bc42497380e275a8581f2ed44d8d4" Nov 25 11:50:23 crc kubenswrapper[4776]: I1125 11:50:23.967880 4776 scope.go:117] "RemoveContainer" containerID="7ce177561ee68c609abc363f3723fba9f17d9b17512d6a352b4f740e74be718b" Nov 25 11:50:24 crc kubenswrapper[4776]: I1125 11:50:24.007585 4776 scope.go:117] "RemoveContainer" containerID="2584f68840e6778543c1c77a789353c40689b036086a54b96de7f3efa85ca422" Nov 25 11:50:24 crc kubenswrapper[4776]: E1125 11:50:24.008176 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2584f68840e6778543c1c77a789353c40689b036086a54b96de7f3efa85ca422\": container with ID starting with 2584f68840e6778543c1c77a789353c40689b036086a54b96de7f3efa85ca422 not found: ID does not exist" containerID="2584f68840e6778543c1c77a789353c40689b036086a54b96de7f3efa85ca422" Nov 25 11:50:24 crc kubenswrapper[4776]: I1125 11:50:24.008233 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2584f68840e6778543c1c77a789353c40689b036086a54b96de7f3efa85ca422"} err="failed to get container status \"2584f68840e6778543c1c77a789353c40689b036086a54b96de7f3efa85ca422\": rpc error: code = NotFound desc = could not find container \"2584f68840e6778543c1c77a789353c40689b036086a54b96de7f3efa85ca422\": container with ID starting with 2584f68840e6778543c1c77a789353c40689b036086a54b96de7f3efa85ca422 not found: ID does not exist" Nov 25 11:50:24 crc kubenswrapper[4776]: I1125 11:50:24.008266 4776 scope.go:117] "RemoveContainer" containerID="ea8d86e746bc491d7af4fe5920266f59856bc42497380e275a8581f2ed44d8d4" Nov 25 11:50:24 crc kubenswrapper[4776]: E1125 11:50:24.008771 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea8d86e746bc491d7af4fe5920266f59856bc42497380e275a8581f2ed44d8d4\": container with ID starting with ea8d86e746bc491d7af4fe5920266f59856bc42497380e275a8581f2ed44d8d4 not found: ID does not exist" containerID="ea8d86e746bc491d7af4fe5920266f59856bc42497380e275a8581f2ed44d8d4" Nov 25 11:50:24 crc kubenswrapper[4776]: I1125 11:50:24.008814 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea8d86e746bc491d7af4fe5920266f59856bc42497380e275a8581f2ed44d8d4"} err="failed to get container status \"ea8d86e746bc491d7af4fe5920266f59856bc42497380e275a8581f2ed44d8d4\": rpc error: code = NotFound desc = could not find container \"ea8d86e746bc491d7af4fe5920266f59856bc42497380e275a8581f2ed44d8d4\": container with ID starting with ea8d86e746bc491d7af4fe5920266f59856bc42497380e275a8581f2ed44d8d4 not found: ID does not exist" Nov 25 11:50:24 crc kubenswrapper[4776]: I1125 11:50:24.008862 4776 scope.go:117] "RemoveContainer" containerID="7ce177561ee68c609abc363f3723fba9f17d9b17512d6a352b4f740e74be718b" Nov 25 11:50:24 crc kubenswrapper[4776]: E1125 11:50:24.009128 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ce177561ee68c609abc363f3723fba9f17d9b17512d6a352b4f740e74be718b\": container with ID starting with 7ce177561ee68c609abc363f3723fba9f17d9b17512d6a352b4f740e74be718b not found: ID does not exist" containerID="7ce177561ee68c609abc363f3723fba9f17d9b17512d6a352b4f740e74be718b" Nov 25 11:50:24 crc kubenswrapper[4776]: I1125 11:50:24.009159 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ce177561ee68c609abc363f3723fba9f17d9b17512d6a352b4f740e74be718b"} err="failed to get container status \"7ce177561ee68c609abc363f3723fba9f17d9b17512d6a352b4f740e74be718b\": rpc error: code = NotFound desc = could not find container \"7ce177561ee68c609abc363f3723fba9f17d9b17512d6a352b4f740e74be718b\": container with ID starting with 7ce177561ee68c609abc363f3723fba9f17d9b17512d6a352b4f740e74be718b not found: ID does not exist" Nov 25 11:50:24 crc kubenswrapper[4776]: I1125 11:50:24.092638 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f46b22e-2c01-40b1-9e6e-4fb013683a65-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2f46b22e-2c01-40b1-9e6e-4fb013683a65" (UID: "2f46b22e-2c01-40b1-9e6e-4fb013683a65"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:50:24 crc kubenswrapper[4776]: I1125 11:50:24.146779 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f46b22e-2c01-40b1-9e6e-4fb013683a65-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:50:24 crc kubenswrapper[4776]: I1125 11:50:24.245421 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c2x7m"] Nov 25 11:50:24 crc kubenswrapper[4776]: I1125 11:50:24.255895 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-c2x7m"] Nov 25 11:50:25 crc kubenswrapper[4776]: I1125 11:50:25.695961 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f46b22e-2c01-40b1-9e6e-4fb013683a65" path="/var/lib/kubelet/pods/2f46b22e-2c01-40b1-9e6e-4fb013683a65/volumes" Nov 25 11:51:17 crc kubenswrapper[4776]: I1125 11:51:17.818660 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:51:17 crc kubenswrapper[4776]: I1125 11:51:17.819175 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:51:47 crc kubenswrapper[4776]: I1125 11:51:47.818135 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:51:47 crc kubenswrapper[4776]: I1125 11:51:47.818701 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:51:53 crc kubenswrapper[4776]: I1125 11:51:53.616913 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fvcw5"] Nov 25 11:51:53 crc kubenswrapper[4776]: E1125 11:51:53.617866 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f46b22e-2c01-40b1-9e6e-4fb013683a65" containerName="extract-content" Nov 25 11:51:53 crc kubenswrapper[4776]: I1125 11:51:53.617941 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f46b22e-2c01-40b1-9e6e-4fb013683a65" containerName="extract-content" Nov 25 11:51:53 crc kubenswrapper[4776]: E1125 11:51:53.617970 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f46b22e-2c01-40b1-9e6e-4fb013683a65" containerName="extract-utilities" Nov 25 11:51:53 crc kubenswrapper[4776]: I1125 11:51:53.617979 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f46b22e-2c01-40b1-9e6e-4fb013683a65" containerName="extract-utilities" Nov 25 11:51:53 crc kubenswrapper[4776]: E1125 11:51:53.617998 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f46b22e-2c01-40b1-9e6e-4fb013683a65" containerName="registry-server" Nov 25 11:51:53 crc kubenswrapper[4776]: I1125 11:51:53.618005 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f46b22e-2c01-40b1-9e6e-4fb013683a65" containerName="registry-server" Nov 25 11:51:53 crc kubenswrapper[4776]: I1125 11:51:53.618239 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f46b22e-2c01-40b1-9e6e-4fb013683a65" containerName="registry-server" Nov 25 11:51:53 crc kubenswrapper[4776]: I1125 11:51:53.619767 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fvcw5" Nov 25 11:51:53 crc kubenswrapper[4776]: I1125 11:51:53.631666 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fvcw5"] Nov 25 11:51:53 crc kubenswrapper[4776]: I1125 11:51:53.761166 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58414be8-5a6a-4577-8ecd-568b1a196dc5-catalog-content\") pod \"redhat-marketplace-fvcw5\" (UID: \"58414be8-5a6a-4577-8ecd-568b1a196dc5\") " pod="openshift-marketplace/redhat-marketplace-fvcw5" Nov 25 11:51:53 crc kubenswrapper[4776]: I1125 11:51:53.761509 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58414be8-5a6a-4577-8ecd-568b1a196dc5-utilities\") pod \"redhat-marketplace-fvcw5\" (UID: \"58414be8-5a6a-4577-8ecd-568b1a196dc5\") " pod="openshift-marketplace/redhat-marketplace-fvcw5" Nov 25 11:51:53 crc kubenswrapper[4776]: I1125 11:51:53.761550 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27f4h\" (UniqueName: \"kubernetes.io/projected/58414be8-5a6a-4577-8ecd-568b1a196dc5-kube-api-access-27f4h\") pod \"redhat-marketplace-fvcw5\" (UID: \"58414be8-5a6a-4577-8ecd-568b1a196dc5\") " pod="openshift-marketplace/redhat-marketplace-fvcw5" Nov 25 11:51:53 crc kubenswrapper[4776]: I1125 11:51:53.864033 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58414be8-5a6a-4577-8ecd-568b1a196dc5-utilities\") pod \"redhat-marketplace-fvcw5\" (UID: \"58414be8-5a6a-4577-8ecd-568b1a196dc5\") " pod="openshift-marketplace/redhat-marketplace-fvcw5" Nov 25 11:51:53 crc kubenswrapper[4776]: I1125 11:51:53.864119 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27f4h\" (UniqueName: \"kubernetes.io/projected/58414be8-5a6a-4577-8ecd-568b1a196dc5-kube-api-access-27f4h\") pod \"redhat-marketplace-fvcw5\" (UID: \"58414be8-5a6a-4577-8ecd-568b1a196dc5\") " pod="openshift-marketplace/redhat-marketplace-fvcw5" Nov 25 11:51:53 crc kubenswrapper[4776]: I1125 11:51:53.864265 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58414be8-5a6a-4577-8ecd-568b1a196dc5-catalog-content\") pod \"redhat-marketplace-fvcw5\" (UID: \"58414be8-5a6a-4577-8ecd-568b1a196dc5\") " pod="openshift-marketplace/redhat-marketplace-fvcw5" Nov 25 11:51:53 crc kubenswrapper[4776]: I1125 11:51:53.864577 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58414be8-5a6a-4577-8ecd-568b1a196dc5-utilities\") pod \"redhat-marketplace-fvcw5\" (UID: \"58414be8-5a6a-4577-8ecd-568b1a196dc5\") " pod="openshift-marketplace/redhat-marketplace-fvcw5" Nov 25 11:51:53 crc kubenswrapper[4776]: I1125 11:51:53.864593 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58414be8-5a6a-4577-8ecd-568b1a196dc5-catalog-content\") pod \"redhat-marketplace-fvcw5\" (UID: \"58414be8-5a6a-4577-8ecd-568b1a196dc5\") " pod="openshift-marketplace/redhat-marketplace-fvcw5" Nov 25 11:51:53 crc kubenswrapper[4776]: I1125 11:51:53.882727 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27f4h\" (UniqueName: \"kubernetes.io/projected/58414be8-5a6a-4577-8ecd-568b1a196dc5-kube-api-access-27f4h\") pod \"redhat-marketplace-fvcw5\" (UID: \"58414be8-5a6a-4577-8ecd-568b1a196dc5\") " pod="openshift-marketplace/redhat-marketplace-fvcw5" Nov 25 11:51:53 crc kubenswrapper[4776]: I1125 11:51:53.940954 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fvcw5" Nov 25 11:51:54 crc kubenswrapper[4776]: I1125 11:51:54.417904 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fvcw5"] Nov 25 11:51:54 crc kubenswrapper[4776]: I1125 11:51:54.869195 4776 generic.go:334] "Generic (PLEG): container finished" podID="58414be8-5a6a-4577-8ecd-568b1a196dc5" containerID="92a7492d5e9e8e5ffba5da680a1aad365dbab4d5e584301dd98bf42077793b3c" exitCode=0 Nov 25 11:51:54 crc kubenswrapper[4776]: I1125 11:51:54.869238 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fvcw5" event={"ID":"58414be8-5a6a-4577-8ecd-568b1a196dc5","Type":"ContainerDied","Data":"92a7492d5e9e8e5ffba5da680a1aad365dbab4d5e584301dd98bf42077793b3c"} Nov 25 11:51:54 crc kubenswrapper[4776]: I1125 11:51:54.869282 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fvcw5" event={"ID":"58414be8-5a6a-4577-8ecd-568b1a196dc5","Type":"ContainerStarted","Data":"ede88636d068f37ca20f88cafa553b75643a4e154b1419c840090c72c25b03c2"} Nov 25 11:51:54 crc kubenswrapper[4776]: I1125 11:51:54.872048 4776 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 11:51:56 crc kubenswrapper[4776]: I1125 11:51:56.893109 4776 generic.go:334] "Generic (PLEG): container finished" podID="58414be8-5a6a-4577-8ecd-568b1a196dc5" containerID="c9e81ac1d40aeb9506988c19ddfe750c56693e75cf8a5a24c3b086eefec59a08" exitCode=0 Nov 25 11:51:56 crc kubenswrapper[4776]: I1125 11:51:56.893220 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fvcw5" event={"ID":"58414be8-5a6a-4577-8ecd-568b1a196dc5","Type":"ContainerDied","Data":"c9e81ac1d40aeb9506988c19ddfe750c56693e75cf8a5a24c3b086eefec59a08"} Nov 25 11:51:57 crc kubenswrapper[4776]: I1125 11:51:57.905824 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fvcw5" event={"ID":"58414be8-5a6a-4577-8ecd-568b1a196dc5","Type":"ContainerStarted","Data":"47c4550c4bb5c16184a938b3059bfa034d6ec5ba315907a0700cfd5d89cebb76"} Nov 25 11:51:57 crc kubenswrapper[4776]: I1125 11:51:57.927720 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fvcw5" podStartSLOduration=2.43081451 podStartE2EDuration="4.927698808s" podCreationTimestamp="2025-11-25 11:51:53 +0000 UTC" firstStartedPulling="2025-11-25 11:51:54.871564279 +0000 UTC m=+8859.912623822" lastFinishedPulling="2025-11-25 11:51:57.368448577 +0000 UTC m=+8862.409508120" observedRunningTime="2025-11-25 11:51:57.922766314 +0000 UTC m=+8862.963825867" watchObservedRunningTime="2025-11-25 11:51:57.927698808 +0000 UTC m=+8862.968758361" Nov 25 11:52:03 crc kubenswrapper[4776]: I1125 11:52:03.942322 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fvcw5" Nov 25 11:52:03 crc kubenswrapper[4776]: I1125 11:52:03.942871 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fvcw5" Nov 25 11:52:04 crc kubenswrapper[4776]: I1125 11:52:04.003039 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fvcw5" Nov 25 11:52:04 crc kubenswrapper[4776]: I1125 11:52:04.069030 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fvcw5" Nov 25 11:52:04 crc kubenswrapper[4776]: I1125 11:52:04.243411 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fvcw5"] Nov 25 11:52:06 crc kubenswrapper[4776]: I1125 11:52:06.024860 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fvcw5" podUID="58414be8-5a6a-4577-8ecd-568b1a196dc5" containerName="registry-server" containerID="cri-o://47c4550c4bb5c16184a938b3059bfa034d6ec5ba315907a0700cfd5d89cebb76" gracePeriod=2 Nov 25 11:52:07 crc kubenswrapper[4776]: I1125 11:52:07.034870 4776 generic.go:334] "Generic (PLEG): container finished" podID="58414be8-5a6a-4577-8ecd-568b1a196dc5" containerID="47c4550c4bb5c16184a938b3059bfa034d6ec5ba315907a0700cfd5d89cebb76" exitCode=0 Nov 25 11:52:07 crc kubenswrapper[4776]: I1125 11:52:07.034930 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fvcw5" event={"ID":"58414be8-5a6a-4577-8ecd-568b1a196dc5","Type":"ContainerDied","Data":"47c4550c4bb5c16184a938b3059bfa034d6ec5ba315907a0700cfd5d89cebb76"} Nov 25 11:52:09 crc kubenswrapper[4776]: I1125 11:52:09.102035 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fvcw5" Nov 25 11:52:09 crc kubenswrapper[4776]: I1125 11:52:09.128873 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fvcw5" event={"ID":"58414be8-5a6a-4577-8ecd-568b1a196dc5","Type":"ContainerDied","Data":"ede88636d068f37ca20f88cafa553b75643a4e154b1419c840090c72c25b03c2"} Nov 25 11:52:09 crc kubenswrapper[4776]: I1125 11:52:09.128954 4776 scope.go:117] "RemoveContainer" containerID="47c4550c4bb5c16184a938b3059bfa034d6ec5ba315907a0700cfd5d89cebb76" Nov 25 11:52:09 crc kubenswrapper[4776]: I1125 11:52:09.128982 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fvcw5" Nov 25 11:52:09 crc kubenswrapper[4776]: I1125 11:52:09.131185 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58414be8-5a6a-4577-8ecd-568b1a196dc5-catalog-content\") pod \"58414be8-5a6a-4577-8ecd-568b1a196dc5\" (UID: \"58414be8-5a6a-4577-8ecd-568b1a196dc5\") " Nov 25 11:52:09 crc kubenswrapper[4776]: I1125 11:52:09.131302 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27f4h\" (UniqueName: \"kubernetes.io/projected/58414be8-5a6a-4577-8ecd-568b1a196dc5-kube-api-access-27f4h\") pod \"58414be8-5a6a-4577-8ecd-568b1a196dc5\" (UID: \"58414be8-5a6a-4577-8ecd-568b1a196dc5\") " Nov 25 11:52:09 crc kubenswrapper[4776]: I1125 11:52:09.131601 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58414be8-5a6a-4577-8ecd-568b1a196dc5-utilities\") pod \"58414be8-5a6a-4577-8ecd-568b1a196dc5\" (UID: \"58414be8-5a6a-4577-8ecd-568b1a196dc5\") " Nov 25 11:52:09 crc kubenswrapper[4776]: I1125 11:52:09.133092 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58414be8-5a6a-4577-8ecd-568b1a196dc5-utilities" (OuterVolumeSpecName: "utilities") pod "58414be8-5a6a-4577-8ecd-568b1a196dc5" (UID: "58414be8-5a6a-4577-8ecd-568b1a196dc5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:52:09 crc kubenswrapper[4776]: I1125 11:52:09.137486 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58414be8-5a6a-4577-8ecd-568b1a196dc5-kube-api-access-27f4h" (OuterVolumeSpecName: "kube-api-access-27f4h") pod "58414be8-5a6a-4577-8ecd-568b1a196dc5" (UID: "58414be8-5a6a-4577-8ecd-568b1a196dc5"). InnerVolumeSpecName "kube-api-access-27f4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:52:09 crc kubenswrapper[4776]: I1125 11:52:09.158591 4776 scope.go:117] "RemoveContainer" containerID="c9e81ac1d40aeb9506988c19ddfe750c56693e75cf8a5a24c3b086eefec59a08" Nov 25 11:52:09 crc kubenswrapper[4776]: I1125 11:52:09.220033 4776 scope.go:117] "RemoveContainer" containerID="92a7492d5e9e8e5ffba5da680a1aad365dbab4d5e584301dd98bf42077793b3c" Nov 25 11:52:09 crc kubenswrapper[4776]: I1125 11:52:09.234729 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58414be8-5a6a-4577-8ecd-568b1a196dc5-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 11:52:09 crc kubenswrapper[4776]: I1125 11:52:09.234766 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27f4h\" (UniqueName: \"kubernetes.io/projected/58414be8-5a6a-4577-8ecd-568b1a196dc5-kube-api-access-27f4h\") on node \"crc\" DevicePath \"\"" Nov 25 11:52:09 crc kubenswrapper[4776]: I1125 11:52:09.675208 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58414be8-5a6a-4577-8ecd-568b1a196dc5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "58414be8-5a6a-4577-8ecd-568b1a196dc5" (UID: "58414be8-5a6a-4577-8ecd-568b1a196dc5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:52:09 crc kubenswrapper[4776]: I1125 11:52:09.748672 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58414be8-5a6a-4577-8ecd-568b1a196dc5-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 11:52:09 crc kubenswrapper[4776]: I1125 11:52:09.773390 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fvcw5"] Nov 25 11:52:09 crc kubenswrapper[4776]: I1125 11:52:09.785593 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fvcw5"] Nov 25 11:52:11 crc kubenswrapper[4776]: I1125 11:52:11.678442 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58414be8-5a6a-4577-8ecd-568b1a196dc5" path="/var/lib/kubelet/pods/58414be8-5a6a-4577-8ecd-568b1a196dc5/volumes" Nov 25 11:52:17 crc kubenswrapper[4776]: I1125 11:52:17.818289 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:52:17 crc kubenswrapper[4776]: I1125 11:52:17.819895 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:52:17 crc kubenswrapper[4776]: I1125 11:52:17.819972 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 11:52:17 crc kubenswrapper[4776]: I1125 11:52:17.821409 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 11:52:17 crc kubenswrapper[4776]: I1125 11:52:17.821498 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965" gracePeriod=600 Nov 25 11:52:19 crc kubenswrapper[4776]: E1125 11:52:19.104614 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:52:19 crc kubenswrapper[4776]: I1125 11:52:19.264996 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965" exitCode=0 Nov 25 11:52:19 crc kubenswrapper[4776]: I1125 11:52:19.265044 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965"} Nov 25 11:52:19 crc kubenswrapper[4776]: I1125 11:52:19.265404 4776 scope.go:117] "RemoveContainer" containerID="b3060280eff383f881b2e41e47023c8d8874f894afc80e5917b1ff4623cd5d74" Nov 25 11:52:19 crc kubenswrapper[4776]: I1125 11:52:19.266038 4776 scope.go:117] "RemoveContainer" containerID="63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965" Nov 25 11:52:19 crc kubenswrapper[4776]: E1125 11:52:19.266616 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:52:33 crc kubenswrapper[4776]: I1125 11:52:33.663042 4776 scope.go:117] "RemoveContainer" containerID="63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965" Nov 25 11:52:33 crc kubenswrapper[4776]: E1125 11:52:33.663740 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:52:44 crc kubenswrapper[4776]: I1125 11:52:44.663157 4776 scope.go:117] "RemoveContainer" containerID="63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965" Nov 25 11:52:44 crc kubenswrapper[4776]: E1125 11:52:44.664451 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:52:57 crc kubenswrapper[4776]: I1125 11:52:57.662874 4776 scope.go:117] "RemoveContainer" containerID="63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965" Nov 25 11:52:57 crc kubenswrapper[4776]: E1125 11:52:57.663644 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:53:10 crc kubenswrapper[4776]: I1125 11:53:10.662601 4776 scope.go:117] "RemoveContainer" containerID="63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965" Nov 25 11:53:10 crc kubenswrapper[4776]: E1125 11:53:10.663602 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:53:25 crc kubenswrapper[4776]: I1125 11:53:25.669560 4776 scope.go:117] "RemoveContainer" containerID="63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965" Nov 25 11:53:25 crc kubenswrapper[4776]: E1125 11:53:25.670182 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:53:37 crc kubenswrapper[4776]: I1125 11:53:37.663362 4776 scope.go:117] "RemoveContainer" containerID="63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965" Nov 25 11:53:37 crc kubenswrapper[4776]: E1125 11:53:37.664719 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:53:51 crc kubenswrapper[4776]: I1125 11:53:51.663282 4776 scope.go:117] "RemoveContainer" containerID="63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965" Nov 25 11:53:51 crc kubenswrapper[4776]: E1125 11:53:51.664279 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:54:02 crc kubenswrapper[4776]: I1125 11:54:02.662566 4776 scope.go:117] "RemoveContainer" containerID="63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965" Nov 25 11:54:02 crc kubenswrapper[4776]: E1125 11:54:02.663440 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:54:13 crc kubenswrapper[4776]: I1125 11:54:13.663935 4776 scope.go:117] "RemoveContainer" containerID="63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965" Nov 25 11:54:13 crc kubenswrapper[4776]: E1125 11:54:13.665264 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:54:19 crc kubenswrapper[4776]: I1125 11:54:19.507662 4776 generic.go:334] "Generic (PLEG): container finished" podID="c622a8cc-2543-4ba5-b53f-32d680331106" containerID="99c42b87324b208d20242c15110d042604354f0ba4aa57ae0867225ffc06e420" exitCode=0 Nov 25 11:54:19 crc kubenswrapper[4776]: I1125 11:54:19.507781 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-x7ctw" event={"ID":"c622a8cc-2543-4ba5-b53f-32d680331106","Type":"ContainerDied","Data":"99c42b87324b208d20242c15110d042604354f0ba4aa57ae0867225ffc06e420"} Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.087565 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-x7ctw" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.264243 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c622a8cc-2543-4ba5-b53f-32d680331106-neutron-sriov-combined-ca-bundle\") pod \"c622a8cc-2543-4ba5-b53f-32d680331106\" (UID: \"c622a8cc-2543-4ba5-b53f-32d680331106\") " Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.264518 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c622a8cc-2543-4ba5-b53f-32d680331106-ssh-key\") pod \"c622a8cc-2543-4ba5-b53f-32d680331106\" (UID: \"c622a8cc-2543-4ba5-b53f-32d680331106\") " Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.264695 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c622a8cc-2543-4ba5-b53f-32d680331106-inventory\") pod \"c622a8cc-2543-4ba5-b53f-32d680331106\" (UID: \"c622a8cc-2543-4ba5-b53f-32d680331106\") " Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.264756 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c622a8cc-2543-4ba5-b53f-32d680331106-neutron-sriov-agent-neutron-config-0\") pod \"c622a8cc-2543-4ba5-b53f-32d680331106\" (UID: \"c622a8cc-2543-4ba5-b53f-32d680331106\") " Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.264893 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6r62x\" (UniqueName: \"kubernetes.io/projected/c622a8cc-2543-4ba5-b53f-32d680331106-kube-api-access-6r62x\") pod \"c622a8cc-2543-4ba5-b53f-32d680331106\" (UID: \"c622a8cc-2543-4ba5-b53f-32d680331106\") " Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.271498 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c622a8cc-2543-4ba5-b53f-32d680331106-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "c622a8cc-2543-4ba5-b53f-32d680331106" (UID: "c622a8cc-2543-4ba5-b53f-32d680331106"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.271640 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c622a8cc-2543-4ba5-b53f-32d680331106-kube-api-access-6r62x" (OuterVolumeSpecName: "kube-api-access-6r62x") pod "c622a8cc-2543-4ba5-b53f-32d680331106" (UID: "c622a8cc-2543-4ba5-b53f-32d680331106"). InnerVolumeSpecName "kube-api-access-6r62x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.295703 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c622a8cc-2543-4ba5-b53f-32d680331106-inventory" (OuterVolumeSpecName: "inventory") pod "c622a8cc-2543-4ba5-b53f-32d680331106" (UID: "c622a8cc-2543-4ba5-b53f-32d680331106"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.298168 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c622a8cc-2543-4ba5-b53f-32d680331106-neutron-sriov-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-sriov-agent-neutron-config-0") pod "c622a8cc-2543-4ba5-b53f-32d680331106" (UID: "c622a8cc-2543-4ba5-b53f-32d680331106"). InnerVolumeSpecName "neutron-sriov-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.305831 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c622a8cc-2543-4ba5-b53f-32d680331106-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c622a8cc-2543-4ba5-b53f-32d680331106" (UID: "c622a8cc-2543-4ba5-b53f-32d680331106"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.368352 4776 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c622a8cc-2543-4ba5-b53f-32d680331106-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.368407 4776 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c622a8cc-2543-4ba5-b53f-32d680331106-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.368425 4776 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c622a8cc-2543-4ba5-b53f-32d680331106-neutron-sriov-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.368442 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6r62x\" (UniqueName: \"kubernetes.io/projected/c622a8cc-2543-4ba5-b53f-32d680331106-kube-api-access-6r62x\") on node \"crc\" DevicePath \"\"" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.368455 4776 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c622a8cc-2543-4ba5-b53f-32d680331106-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.534951 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-x7ctw" event={"ID":"c622a8cc-2543-4ba5-b53f-32d680331106","Type":"ContainerDied","Data":"7970c5baa0242def0a7a5ee6f8b2d0e5477a409bedaba052e8176898c4de7d33"} Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.535020 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7970c5baa0242def0a7a5ee6f8b2d0e5477a409bedaba052e8176898c4de7d33" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.535147 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-x7ctw" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.695833 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-k4kxp"] Nov 25 11:54:21 crc kubenswrapper[4776]: E1125 11:54:21.696916 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58414be8-5a6a-4577-8ecd-568b1a196dc5" containerName="registry-server" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.696949 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="58414be8-5a6a-4577-8ecd-568b1a196dc5" containerName="registry-server" Nov 25 11:54:21 crc kubenswrapper[4776]: E1125 11:54:21.696980 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58414be8-5a6a-4577-8ecd-568b1a196dc5" containerName="extract-content" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.696989 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="58414be8-5a6a-4577-8ecd-568b1a196dc5" containerName="extract-content" Nov 25 11:54:21 crc kubenswrapper[4776]: E1125 11:54:21.697007 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58414be8-5a6a-4577-8ecd-568b1a196dc5" containerName="extract-utilities" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.697015 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="58414be8-5a6a-4577-8ecd-568b1a196dc5" containerName="extract-utilities" Nov 25 11:54:21 crc kubenswrapper[4776]: E1125 11:54:21.697044 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c622a8cc-2543-4ba5-b53f-32d680331106" containerName="neutron-sriov-openstack-openstack-cell1" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.697052 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="c622a8cc-2543-4ba5-b53f-32d680331106" containerName="neutron-sriov-openstack-openstack-cell1" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.697340 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="58414be8-5a6a-4577-8ecd-568b1a196dc5" containerName="registry-server" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.697379 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="c622a8cc-2543-4ba5-b53f-32d680331106" containerName="neutron-sriov-openstack-openstack-cell1" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.698409 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-k4kxp" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.701385 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-dhcp-agent-neutron-config" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.704823 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-k4kxp"] Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.705870 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dknh7" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.706144 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.707645 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.707842 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.782588 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a0bc66d-b624-4636-baec-30c555632825-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-k4kxp\" (UID: \"1a0bc66d-b624-4636-baec-30c555632825\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-k4kxp" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.782811 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a0bc66d-b624-4636-baec-30c555632825-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-k4kxp\" (UID: \"1a0bc66d-b624-4636-baec-30c555632825\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-k4kxp" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.782882 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sl8d\" (UniqueName: \"kubernetes.io/projected/1a0bc66d-b624-4636-baec-30c555632825-kube-api-access-6sl8d\") pod \"neutron-dhcp-openstack-openstack-cell1-k4kxp\" (UID: \"1a0bc66d-b624-4636-baec-30c555632825\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-k4kxp" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.782937 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a0bc66d-b624-4636-baec-30c555632825-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-k4kxp\" (UID: \"1a0bc66d-b624-4636-baec-30c555632825\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-k4kxp" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.783053 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1a0bc66d-b624-4636-baec-30c555632825-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-k4kxp\" (UID: \"1a0bc66d-b624-4636-baec-30c555632825\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-k4kxp" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.883872 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a0bc66d-b624-4636-baec-30c555632825-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-k4kxp\" (UID: \"1a0bc66d-b624-4636-baec-30c555632825\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-k4kxp" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.884475 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sl8d\" (UniqueName: \"kubernetes.io/projected/1a0bc66d-b624-4636-baec-30c555632825-kube-api-access-6sl8d\") pod \"neutron-dhcp-openstack-openstack-cell1-k4kxp\" (UID: \"1a0bc66d-b624-4636-baec-30c555632825\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-k4kxp" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.884515 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a0bc66d-b624-4636-baec-30c555632825-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-k4kxp\" (UID: \"1a0bc66d-b624-4636-baec-30c555632825\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-k4kxp" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.884572 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1a0bc66d-b624-4636-baec-30c555632825-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-k4kxp\" (UID: \"1a0bc66d-b624-4636-baec-30c555632825\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-k4kxp" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.884674 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a0bc66d-b624-4636-baec-30c555632825-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-k4kxp\" (UID: \"1a0bc66d-b624-4636-baec-30c555632825\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-k4kxp" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.888824 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a0bc66d-b624-4636-baec-30c555632825-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-k4kxp\" (UID: \"1a0bc66d-b624-4636-baec-30c555632825\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-k4kxp" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.890030 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1a0bc66d-b624-4636-baec-30c555632825-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-k4kxp\" (UID: \"1a0bc66d-b624-4636-baec-30c555632825\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-k4kxp" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.890817 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a0bc66d-b624-4636-baec-30c555632825-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-k4kxp\" (UID: \"1a0bc66d-b624-4636-baec-30c555632825\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-k4kxp" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.902772 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a0bc66d-b624-4636-baec-30c555632825-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-k4kxp\" (UID: \"1a0bc66d-b624-4636-baec-30c555632825\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-k4kxp" Nov 25 11:54:21 crc kubenswrapper[4776]: I1125 11:54:21.911034 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sl8d\" (UniqueName: \"kubernetes.io/projected/1a0bc66d-b624-4636-baec-30c555632825-kube-api-access-6sl8d\") pod \"neutron-dhcp-openstack-openstack-cell1-k4kxp\" (UID: \"1a0bc66d-b624-4636-baec-30c555632825\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-k4kxp" Nov 25 11:54:22 crc kubenswrapper[4776]: I1125 11:54:22.024647 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-k4kxp" Nov 25 11:54:22 crc kubenswrapper[4776]: I1125 11:54:22.641819 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-k4kxp"] Nov 25 11:54:23 crc kubenswrapper[4776]: I1125 11:54:23.560719 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-k4kxp" event={"ID":"1a0bc66d-b624-4636-baec-30c555632825","Type":"ContainerStarted","Data":"e0febf58d6ed3d7903fcad7edde98f6a1a81c618fe200c1cd3e92b3eac52a8f3"} Nov 25 11:54:23 crc kubenswrapper[4776]: I1125 11:54:23.561059 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-k4kxp" event={"ID":"1a0bc66d-b624-4636-baec-30c555632825","Type":"ContainerStarted","Data":"a1493730ddf291cf1672039a91218c0d3ad5b030086b5bb7aa304650dfa6da96"} Nov 25 11:54:23 crc kubenswrapper[4776]: I1125 11:54:23.579059 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dhcp-openstack-openstack-cell1-k4kxp" podStartSLOduration=2.064155112 podStartE2EDuration="2.579036221s" podCreationTimestamp="2025-11-25 11:54:21 +0000 UTC" firstStartedPulling="2025-11-25 11:54:22.66553362 +0000 UTC m=+9007.706593173" lastFinishedPulling="2025-11-25 11:54:23.180414729 +0000 UTC m=+9008.221474282" observedRunningTime="2025-11-25 11:54:23.575652216 +0000 UTC m=+9008.616711769" watchObservedRunningTime="2025-11-25 11:54:23.579036221 +0000 UTC m=+9008.620095774" Nov 25 11:54:24 crc kubenswrapper[4776]: I1125 11:54:24.663150 4776 scope.go:117] "RemoveContainer" containerID="63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965" Nov 25 11:54:24 crc kubenswrapper[4776]: E1125 11:54:24.663568 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:54:37 crc kubenswrapper[4776]: I1125 11:54:37.663215 4776 scope.go:117] "RemoveContainer" containerID="63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965" Nov 25 11:54:37 crc kubenswrapper[4776]: E1125 11:54:37.663915 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:54:51 crc kubenswrapper[4776]: I1125 11:54:51.663350 4776 scope.go:117] "RemoveContainer" containerID="63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965" Nov 25 11:54:51 crc kubenswrapper[4776]: E1125 11:54:51.666159 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:55:06 crc kubenswrapper[4776]: I1125 11:55:06.664211 4776 scope.go:117] "RemoveContainer" containerID="63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965" Nov 25 11:55:06 crc kubenswrapper[4776]: E1125 11:55:06.666525 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:55:21 crc kubenswrapper[4776]: I1125 11:55:21.662639 4776 scope.go:117] "RemoveContainer" containerID="63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965" Nov 25 11:55:21 crc kubenswrapper[4776]: E1125 11:55:21.663373 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:55:32 crc kubenswrapper[4776]: I1125 11:55:32.665376 4776 scope.go:117] "RemoveContainer" containerID="63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965" Nov 25 11:55:32 crc kubenswrapper[4776]: E1125 11:55:32.666865 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:55:47 crc kubenswrapper[4776]: I1125 11:55:47.663241 4776 scope.go:117] "RemoveContainer" containerID="63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965" Nov 25 11:55:47 crc kubenswrapper[4776]: E1125 11:55:47.664099 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:55:58 crc kubenswrapper[4776]: I1125 11:55:58.662859 4776 scope.go:117] "RemoveContainer" containerID="63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965" Nov 25 11:55:58 crc kubenswrapper[4776]: E1125 11:55:58.663607 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:56:13 crc kubenswrapper[4776]: I1125 11:56:13.662822 4776 scope.go:117] "RemoveContainer" containerID="63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965" Nov 25 11:56:13 crc kubenswrapper[4776]: E1125 11:56:13.663797 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:56:25 crc kubenswrapper[4776]: I1125 11:56:25.664387 4776 scope.go:117] "RemoveContainer" containerID="63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965" Nov 25 11:56:25 crc kubenswrapper[4776]: E1125 11:56:25.665535 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:56:37 crc kubenswrapper[4776]: I1125 11:56:37.663042 4776 scope.go:117] "RemoveContainer" containerID="63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965" Nov 25 11:56:37 crc kubenswrapper[4776]: E1125 11:56:37.664418 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:56:51 crc kubenswrapper[4776]: I1125 11:56:51.663286 4776 scope.go:117] "RemoveContainer" containerID="63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965" Nov 25 11:56:51 crc kubenswrapper[4776]: E1125 11:56:51.664636 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:57:02 crc kubenswrapper[4776]: I1125 11:57:02.663742 4776 scope.go:117] "RemoveContainer" containerID="63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965" Nov 25 11:57:02 crc kubenswrapper[4776]: E1125 11:57:02.666149 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:57:15 crc kubenswrapper[4776]: I1125 11:57:15.665288 4776 scope.go:117] "RemoveContainer" containerID="63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965" Nov 25 11:57:15 crc kubenswrapper[4776]: E1125 11:57:15.666354 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 11:57:30 crc kubenswrapper[4776]: I1125 11:57:30.662908 4776 scope.go:117] "RemoveContainer" containerID="63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965" Nov 25 11:57:31 crc kubenswrapper[4776]: I1125 11:57:31.816231 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"8e060fbd41a6ceeaf0a55e5585d31b0083930b3310803a6e50fc5823301c1b89"} Nov 25 11:57:41 crc kubenswrapper[4776]: I1125 11:57:41.921279 4776 generic.go:334] "Generic (PLEG): container finished" podID="1a0bc66d-b624-4636-baec-30c555632825" containerID="e0febf58d6ed3d7903fcad7edde98f6a1a81c618fe200c1cd3e92b3eac52a8f3" exitCode=0 Nov 25 11:57:41 crc kubenswrapper[4776]: I1125 11:57:41.921353 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-k4kxp" event={"ID":"1a0bc66d-b624-4636-baec-30c555632825","Type":"ContainerDied","Data":"e0febf58d6ed3d7903fcad7edde98f6a1a81c618fe200c1cd3e92b3eac52a8f3"} Nov 25 11:57:43 crc kubenswrapper[4776]: I1125 11:57:43.488014 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-k4kxp" Nov 25 11:57:43 crc kubenswrapper[4776]: I1125 11:57:43.538490 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a0bc66d-b624-4636-baec-30c555632825-ssh-key\") pod \"1a0bc66d-b624-4636-baec-30c555632825\" (UID: \"1a0bc66d-b624-4636-baec-30c555632825\") " Nov 25 11:57:43 crc kubenswrapper[4776]: I1125 11:57:43.538689 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a0bc66d-b624-4636-baec-30c555632825-neutron-dhcp-combined-ca-bundle\") pod \"1a0bc66d-b624-4636-baec-30c555632825\" (UID: \"1a0bc66d-b624-4636-baec-30c555632825\") " Nov 25 11:57:43 crc kubenswrapper[4776]: I1125 11:57:43.538821 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1a0bc66d-b624-4636-baec-30c555632825-neutron-dhcp-agent-neutron-config-0\") pod \"1a0bc66d-b624-4636-baec-30c555632825\" (UID: \"1a0bc66d-b624-4636-baec-30c555632825\") " Nov 25 11:57:43 crc kubenswrapper[4776]: I1125 11:57:43.538873 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sl8d\" (UniqueName: \"kubernetes.io/projected/1a0bc66d-b624-4636-baec-30c555632825-kube-api-access-6sl8d\") pod \"1a0bc66d-b624-4636-baec-30c555632825\" (UID: \"1a0bc66d-b624-4636-baec-30c555632825\") " Nov 25 11:57:43 crc kubenswrapper[4776]: I1125 11:57:43.538918 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a0bc66d-b624-4636-baec-30c555632825-inventory\") pod \"1a0bc66d-b624-4636-baec-30c555632825\" (UID: \"1a0bc66d-b624-4636-baec-30c555632825\") " Nov 25 11:57:43 crc kubenswrapper[4776]: I1125 11:57:43.551746 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a0bc66d-b624-4636-baec-30c555632825-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "1a0bc66d-b624-4636-baec-30c555632825" (UID: "1a0bc66d-b624-4636-baec-30c555632825"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:57:43 crc kubenswrapper[4776]: I1125 11:57:43.558973 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a0bc66d-b624-4636-baec-30c555632825-kube-api-access-6sl8d" (OuterVolumeSpecName: "kube-api-access-6sl8d") pod "1a0bc66d-b624-4636-baec-30c555632825" (UID: "1a0bc66d-b624-4636-baec-30c555632825"). InnerVolumeSpecName "kube-api-access-6sl8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:57:43 crc kubenswrapper[4776]: I1125 11:57:43.602737 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a0bc66d-b624-4636-baec-30c555632825-inventory" (OuterVolumeSpecName: "inventory") pod "1a0bc66d-b624-4636-baec-30c555632825" (UID: "1a0bc66d-b624-4636-baec-30c555632825"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:57:43 crc kubenswrapper[4776]: I1125 11:57:43.606314 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a0bc66d-b624-4636-baec-30c555632825-neutron-dhcp-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-dhcp-agent-neutron-config-0") pod "1a0bc66d-b624-4636-baec-30c555632825" (UID: "1a0bc66d-b624-4636-baec-30c555632825"). InnerVolumeSpecName "neutron-dhcp-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:57:43 crc kubenswrapper[4776]: I1125 11:57:43.626623 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a0bc66d-b624-4636-baec-30c555632825-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1a0bc66d-b624-4636-baec-30c555632825" (UID: "1a0bc66d-b624-4636-baec-30c555632825"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:57:43 crc kubenswrapper[4776]: I1125 11:57:43.641696 4776 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a0bc66d-b624-4636-baec-30c555632825-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:57:43 crc kubenswrapper[4776]: I1125 11:57:43.641739 4776 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1a0bc66d-b624-4636-baec-30c555632825-neutron-dhcp-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 11:57:43 crc kubenswrapper[4776]: I1125 11:57:43.641754 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sl8d\" (UniqueName: \"kubernetes.io/projected/1a0bc66d-b624-4636-baec-30c555632825-kube-api-access-6sl8d\") on node \"crc\" DevicePath \"\"" Nov 25 11:57:43 crc kubenswrapper[4776]: I1125 11:57:43.641767 4776 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a0bc66d-b624-4636-baec-30c555632825-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 11:57:43 crc kubenswrapper[4776]: I1125 11:57:43.641778 4776 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a0bc66d-b624-4636-baec-30c555632825-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 11:57:43 crc kubenswrapper[4776]: I1125 11:57:43.947332 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-k4kxp" event={"ID":"1a0bc66d-b624-4636-baec-30c555632825","Type":"ContainerDied","Data":"a1493730ddf291cf1672039a91218c0d3ad5b030086b5bb7aa304650dfa6da96"} Nov 25 11:57:43 crc kubenswrapper[4776]: I1125 11:57:43.947386 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1493730ddf291cf1672039a91218c0d3ad5b030086b5bb7aa304650dfa6da96" Nov 25 11:57:43 crc kubenswrapper[4776]: I1125 11:57:43.947445 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-k4kxp" Nov 25 11:58:06 crc kubenswrapper[4776]: I1125 11:58:06.952774 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 11:58:06 crc kubenswrapper[4776]: I1125 11:58:06.953346 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="9c39e31e-7255-4c8e-90d9-5e0cc5d3910a" containerName="nova-cell0-conductor-conductor" containerID="cri-o://15c598d4474dfc8c30f350b528152717780cc311dc924d946511163c8d02d979" gracePeriod=30 Nov 25 11:58:06 crc kubenswrapper[4776]: I1125 11:58:06.990691 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 11:58:06 crc kubenswrapper[4776]: I1125 11:58:06.990937 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="ec8114a1-e74a-498c-a180-5d20733326f5" containerName="nova-cell1-conductor-conductor" containerID="cri-o://45a9c50d84c62aa7c775647665d279cdfcddd2f980985e31b8d45ecaf948d334" gracePeriod=30 Nov 25 11:58:07 crc kubenswrapper[4776]: E1125 11:58:07.751753 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="45a9c50d84c62aa7c775647665d279cdfcddd2f980985e31b8d45ecaf948d334" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 25 11:58:07 crc kubenswrapper[4776]: E1125 11:58:07.757275 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="45a9c50d84c62aa7c775647665d279cdfcddd2f980985e31b8d45ecaf948d334" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 25 11:58:07 crc kubenswrapper[4776]: E1125 11:58:07.760881 4776 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="45a9c50d84c62aa7c775647665d279cdfcddd2f980985e31b8d45ecaf948d334" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Nov 25 11:58:07 crc kubenswrapper[4776]: E1125 11:58:07.760966 4776 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="ec8114a1-e74a-498c-a180-5d20733326f5" containerName="nova-cell1-conductor-conductor" Nov 25 11:58:07 crc kubenswrapper[4776]: I1125 11:58:07.828127 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 11:58:07 crc kubenswrapper[4776]: I1125 11:58:07.828395 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d854edbe-4354-4149-9d58-b503e44bd1c8" containerName="nova-api-log" containerID="cri-o://d6e46cd49b2c18d00f7c5566b97c6a00bed98218a03cc5997bbd1e486a0d50d1" gracePeriod=30 Nov 25 11:58:07 crc kubenswrapper[4776]: I1125 11:58:07.828905 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d854edbe-4354-4149-9d58-b503e44bd1c8" containerName="nova-api-api" containerID="cri-o://e2071c42555d92e072b179f72bc660dfec1e153919f6300461d43023f7c18e70" gracePeriod=30 Nov 25 11:58:07 crc kubenswrapper[4776]: I1125 11:58:07.851912 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 11:58:07 crc kubenswrapper[4776]: I1125 11:58:07.852359 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14" containerName="nova-metadata-log" containerID="cri-o://d691f3e8528edb91b33ba0d8b46b7550b59a4896095da1369db835bf1ea9af57" gracePeriod=30 Nov 25 11:58:07 crc kubenswrapper[4776]: I1125 11:58:07.853208 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14" containerName="nova-metadata-metadata" containerID="cri-o://235b27dba37c09496e9794f40807ee09d92277a36d96b376f20e24a9121f25d0" gracePeriod=30 Nov 25 11:58:07 crc kubenswrapper[4776]: I1125 11:58:07.869444 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 11:58:07 crc kubenswrapper[4776]: I1125 11:58:07.869749 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3" containerName="nova-scheduler-scheduler" containerID="cri-o://457893b901e04fac4aa2a523915fbb0b5c6a60a040f98930992bfbfa4f443310" gracePeriod=30 Nov 25 11:58:08 crc kubenswrapper[4776]: I1125 11:58:08.244867 4776 generic.go:334] "Generic (PLEG): container finished" podID="99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14" containerID="d691f3e8528edb91b33ba0d8b46b7550b59a4896095da1369db835bf1ea9af57" exitCode=143 Nov 25 11:58:08 crc kubenswrapper[4776]: I1125 11:58:08.244976 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14","Type":"ContainerDied","Data":"d691f3e8528edb91b33ba0d8b46b7550b59a4896095da1369db835bf1ea9af57"} Nov 25 11:58:08 crc kubenswrapper[4776]: I1125 11:58:08.246739 4776 generic.go:334] "Generic (PLEG): container finished" podID="d854edbe-4354-4149-9d58-b503e44bd1c8" containerID="d6e46cd49b2c18d00f7c5566b97c6a00bed98218a03cc5997bbd1e486a0d50d1" exitCode=143 Nov 25 11:58:08 crc kubenswrapper[4776]: I1125 11:58:08.246800 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d854edbe-4354-4149-9d58-b503e44bd1c8","Type":"ContainerDied","Data":"d6e46cd49b2c18d00f7c5566b97c6a00bed98218a03cc5997bbd1e486a0d50d1"} Nov 25 11:58:08 crc kubenswrapper[4776]: I1125 11:58:08.694883 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 25 11:58:08 crc kubenswrapper[4776]: I1125 11:58:08.768968 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c39e31e-7255-4c8e-90d9-5e0cc5d3910a-config-data\") pod \"9c39e31e-7255-4c8e-90d9-5e0cc5d3910a\" (UID: \"9c39e31e-7255-4c8e-90d9-5e0cc5d3910a\") " Nov 25 11:58:08 crc kubenswrapper[4776]: I1125 11:58:08.769458 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bh96c\" (UniqueName: \"kubernetes.io/projected/9c39e31e-7255-4c8e-90d9-5e0cc5d3910a-kube-api-access-bh96c\") pod \"9c39e31e-7255-4c8e-90d9-5e0cc5d3910a\" (UID: \"9c39e31e-7255-4c8e-90d9-5e0cc5d3910a\") " Nov 25 11:58:08 crc kubenswrapper[4776]: I1125 11:58:08.769532 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c39e31e-7255-4c8e-90d9-5e0cc5d3910a-combined-ca-bundle\") pod \"9c39e31e-7255-4c8e-90d9-5e0cc5d3910a\" (UID: \"9c39e31e-7255-4c8e-90d9-5e0cc5d3910a\") " Nov 25 11:58:08 crc kubenswrapper[4776]: I1125 11:58:08.780041 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c39e31e-7255-4c8e-90d9-5e0cc5d3910a-kube-api-access-bh96c" (OuterVolumeSpecName: "kube-api-access-bh96c") pod "9c39e31e-7255-4c8e-90d9-5e0cc5d3910a" (UID: "9c39e31e-7255-4c8e-90d9-5e0cc5d3910a"). InnerVolumeSpecName "kube-api-access-bh96c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:58:08 crc kubenswrapper[4776]: I1125 11:58:08.813186 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c39e31e-7255-4c8e-90d9-5e0cc5d3910a-config-data" (OuterVolumeSpecName: "config-data") pod "9c39e31e-7255-4c8e-90d9-5e0cc5d3910a" (UID: "9c39e31e-7255-4c8e-90d9-5e0cc5d3910a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:58:08 crc kubenswrapper[4776]: I1125 11:58:08.813290 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c39e31e-7255-4c8e-90d9-5e0cc5d3910a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c39e31e-7255-4c8e-90d9-5e0cc5d3910a" (UID: "9c39e31e-7255-4c8e-90d9-5e0cc5d3910a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:58:08 crc kubenswrapper[4776]: I1125 11:58:08.872417 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bh96c\" (UniqueName: \"kubernetes.io/projected/9c39e31e-7255-4c8e-90d9-5e0cc5d3910a-kube-api-access-bh96c\") on node \"crc\" DevicePath \"\"" Nov 25 11:58:08 crc kubenswrapper[4776]: I1125 11:58:08.872462 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c39e31e-7255-4c8e-90d9-5e0cc5d3910a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:58:08 crc kubenswrapper[4776]: I1125 11:58:08.872476 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c39e31e-7255-4c8e-90d9-5e0cc5d3910a-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:58:09 crc kubenswrapper[4776]: I1125 11:58:09.261122 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 25 11:58:09 crc kubenswrapper[4776]: I1125 11:58:09.261045 4776 generic.go:334] "Generic (PLEG): container finished" podID="9c39e31e-7255-4c8e-90d9-5e0cc5d3910a" containerID="15c598d4474dfc8c30f350b528152717780cc311dc924d946511163c8d02d979" exitCode=0 Nov 25 11:58:09 crc kubenswrapper[4776]: I1125 11:58:09.261164 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"9c39e31e-7255-4c8e-90d9-5e0cc5d3910a","Type":"ContainerDied","Data":"15c598d4474dfc8c30f350b528152717780cc311dc924d946511163c8d02d979"} Nov 25 11:58:09 crc kubenswrapper[4776]: I1125 11:58:09.261220 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"9c39e31e-7255-4c8e-90d9-5e0cc5d3910a","Type":"ContainerDied","Data":"77e62d592424186ef952b8e5f3b09300e547a04fc1000b176e7e9dec9ddcfecc"} Nov 25 11:58:09 crc kubenswrapper[4776]: I1125 11:58:09.261245 4776 scope.go:117] "RemoveContainer" containerID="15c598d4474dfc8c30f350b528152717780cc311dc924d946511163c8d02d979" Nov 25 11:58:09 crc kubenswrapper[4776]: I1125 11:58:09.296390 4776 scope.go:117] "RemoveContainer" containerID="15c598d4474dfc8c30f350b528152717780cc311dc924d946511163c8d02d979" Nov 25 11:58:09 crc kubenswrapper[4776]: E1125 11:58:09.297403 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15c598d4474dfc8c30f350b528152717780cc311dc924d946511163c8d02d979\": container with ID starting with 15c598d4474dfc8c30f350b528152717780cc311dc924d946511163c8d02d979 not found: ID does not exist" containerID="15c598d4474dfc8c30f350b528152717780cc311dc924d946511163c8d02d979" Nov 25 11:58:09 crc kubenswrapper[4776]: I1125 11:58:09.297464 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15c598d4474dfc8c30f350b528152717780cc311dc924d946511163c8d02d979"} err="failed to get container status \"15c598d4474dfc8c30f350b528152717780cc311dc924d946511163c8d02d979\": rpc error: code = NotFound desc = could not find container \"15c598d4474dfc8c30f350b528152717780cc311dc924d946511163c8d02d979\": container with ID starting with 15c598d4474dfc8c30f350b528152717780cc311dc924d946511163c8d02d979 not found: ID does not exist" Nov 25 11:58:09 crc kubenswrapper[4776]: I1125 11:58:09.299378 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 11:58:09 crc kubenswrapper[4776]: I1125 11:58:09.316003 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 11:58:09 crc kubenswrapper[4776]: I1125 11:58:09.329035 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 11:58:09 crc kubenswrapper[4776]: E1125 11:58:09.329618 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a0bc66d-b624-4636-baec-30c555632825" containerName="neutron-dhcp-openstack-openstack-cell1" Nov 25 11:58:09 crc kubenswrapper[4776]: I1125 11:58:09.329638 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a0bc66d-b624-4636-baec-30c555632825" containerName="neutron-dhcp-openstack-openstack-cell1" Nov 25 11:58:09 crc kubenswrapper[4776]: E1125 11:58:09.329704 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c39e31e-7255-4c8e-90d9-5e0cc5d3910a" containerName="nova-cell0-conductor-conductor" Nov 25 11:58:09 crc kubenswrapper[4776]: I1125 11:58:09.329712 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c39e31e-7255-4c8e-90d9-5e0cc5d3910a" containerName="nova-cell0-conductor-conductor" Nov 25 11:58:09 crc kubenswrapper[4776]: I1125 11:58:09.329916 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c39e31e-7255-4c8e-90d9-5e0cc5d3910a" containerName="nova-cell0-conductor-conductor" Nov 25 11:58:09 crc kubenswrapper[4776]: I1125 11:58:09.329935 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a0bc66d-b624-4636-baec-30c555632825" containerName="neutron-dhcp-openstack-openstack-cell1" Nov 25 11:58:09 crc kubenswrapper[4776]: I1125 11:58:09.330807 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 25 11:58:09 crc kubenswrapper[4776]: I1125 11:58:09.333200 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 25 11:58:09 crc kubenswrapper[4776]: I1125 11:58:09.342674 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 11:58:09 crc kubenswrapper[4776]: I1125 11:58:09.382969 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0945270e-80b0-44ca-9c2b-642abacd57bf-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"0945270e-80b0-44ca-9c2b-642abacd57bf\") " pod="openstack/nova-cell0-conductor-0" Nov 25 11:58:09 crc kubenswrapper[4776]: I1125 11:58:09.383021 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0945270e-80b0-44ca-9c2b-642abacd57bf-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"0945270e-80b0-44ca-9c2b-642abacd57bf\") " pod="openstack/nova-cell0-conductor-0" Nov 25 11:58:09 crc kubenswrapper[4776]: I1125 11:58:09.383090 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22p8v\" (UniqueName: \"kubernetes.io/projected/0945270e-80b0-44ca-9c2b-642abacd57bf-kube-api-access-22p8v\") pod \"nova-cell0-conductor-0\" (UID: \"0945270e-80b0-44ca-9c2b-642abacd57bf\") " pod="openstack/nova-cell0-conductor-0" Nov 25 11:58:09 crc kubenswrapper[4776]: I1125 11:58:09.486574 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0945270e-80b0-44ca-9c2b-642abacd57bf-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"0945270e-80b0-44ca-9c2b-642abacd57bf\") " pod="openstack/nova-cell0-conductor-0" Nov 25 11:58:09 crc kubenswrapper[4776]: I1125 11:58:09.487381 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0945270e-80b0-44ca-9c2b-642abacd57bf-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"0945270e-80b0-44ca-9c2b-642abacd57bf\") " pod="openstack/nova-cell0-conductor-0" Nov 25 11:58:09 crc kubenswrapper[4776]: I1125 11:58:09.487562 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22p8v\" (UniqueName: \"kubernetes.io/projected/0945270e-80b0-44ca-9c2b-642abacd57bf-kube-api-access-22p8v\") pod \"nova-cell0-conductor-0\" (UID: \"0945270e-80b0-44ca-9c2b-642abacd57bf\") " pod="openstack/nova-cell0-conductor-0" Nov 25 11:58:09 crc kubenswrapper[4776]: I1125 11:58:09.492042 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0945270e-80b0-44ca-9c2b-642abacd57bf-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"0945270e-80b0-44ca-9c2b-642abacd57bf\") " pod="openstack/nova-cell0-conductor-0" Nov 25 11:58:09 crc kubenswrapper[4776]: I1125 11:58:09.492131 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0945270e-80b0-44ca-9c2b-642abacd57bf-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"0945270e-80b0-44ca-9c2b-642abacd57bf\") " pod="openstack/nova-cell0-conductor-0" Nov 25 11:58:09 crc kubenswrapper[4776]: I1125 11:58:09.507330 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22p8v\" (UniqueName: \"kubernetes.io/projected/0945270e-80b0-44ca-9c2b-642abacd57bf-kube-api-access-22p8v\") pod \"nova-cell0-conductor-0\" (UID: \"0945270e-80b0-44ca-9c2b-642abacd57bf\") " pod="openstack/nova-cell0-conductor-0" Nov 25 11:58:09 crc kubenswrapper[4776]: I1125 11:58:09.661613 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 25 11:58:09 crc kubenswrapper[4776]: I1125 11:58:09.679758 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c39e31e-7255-4c8e-90d9-5e0cc5d3910a" path="/var/lib/kubelet/pods/9c39e31e-7255-4c8e-90d9-5e0cc5d3910a/volumes" Nov 25 11:58:10 crc kubenswrapper[4776]: I1125 11:58:10.172886 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 11:58:10 crc kubenswrapper[4776]: W1125 11:58:10.186322 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0945270e_80b0_44ca_9c2b_642abacd57bf.slice/crio-ffacc637a01389a667898c214e4edebe8524bd2413ab0e8e756429d91f21deeb WatchSource:0}: Error finding container ffacc637a01389a667898c214e4edebe8524bd2413ab0e8e756429d91f21deeb: Status 404 returned error can't find the container with id ffacc637a01389a667898c214e4edebe8524bd2413ab0e8e756429d91f21deeb Nov 25 11:58:10 crc kubenswrapper[4776]: I1125 11:58:10.274323 4776 generic.go:334] "Generic (PLEG): container finished" podID="f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3" containerID="457893b901e04fac4aa2a523915fbb0b5c6a60a040f98930992bfbfa4f443310" exitCode=0 Nov 25 11:58:10 crc kubenswrapper[4776]: I1125 11:58:10.274629 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3","Type":"ContainerDied","Data":"457893b901e04fac4aa2a523915fbb0b5c6a60a040f98930992bfbfa4f443310"} Nov 25 11:58:10 crc kubenswrapper[4776]: I1125 11:58:10.274664 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3","Type":"ContainerDied","Data":"0a0f498048728567e4f4279aa87099f59624abc65b7d75b4c49de048c13e9b2d"} Nov 25 11:58:10 crc kubenswrapper[4776]: I1125 11:58:10.274677 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a0f498048728567e4f4279aa87099f59624abc65b7d75b4c49de048c13e9b2d" Nov 25 11:58:10 crc kubenswrapper[4776]: I1125 11:58:10.281610 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"0945270e-80b0-44ca-9c2b-642abacd57bf","Type":"ContainerStarted","Data":"ffacc637a01389a667898c214e4edebe8524bd2413ab0e8e756429d91f21deeb"} Nov 25 11:58:10 crc kubenswrapper[4776]: I1125 11:58:10.377082 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 11:58:10 crc kubenswrapper[4776]: I1125 11:58:10.409771 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3-combined-ca-bundle\") pod \"f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3\" (UID: \"f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3\") " Nov 25 11:58:10 crc kubenswrapper[4776]: I1125 11:58:10.409905 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jctt\" (UniqueName: \"kubernetes.io/projected/f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3-kube-api-access-8jctt\") pod \"f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3\" (UID: \"f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3\") " Nov 25 11:58:10 crc kubenswrapper[4776]: I1125 11:58:10.410001 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3-config-data\") pod \"f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3\" (UID: \"f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3\") " Nov 25 11:58:10 crc kubenswrapper[4776]: I1125 11:58:10.417288 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3-kube-api-access-8jctt" (OuterVolumeSpecName: "kube-api-access-8jctt") pod "f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3" (UID: "f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3"). InnerVolumeSpecName "kube-api-access-8jctt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:58:10 crc kubenswrapper[4776]: I1125 11:58:10.460922 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3" (UID: "f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:58:10 crc kubenswrapper[4776]: I1125 11:58:10.479767 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3-config-data" (OuterVolumeSpecName: "config-data") pod "f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3" (UID: "f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:58:10 crc kubenswrapper[4776]: I1125 11:58:10.512713 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:58:10 crc kubenswrapper[4776]: I1125 11:58:10.512759 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jctt\" (UniqueName: \"kubernetes.io/projected/f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3-kube-api-access-8jctt\") on node \"crc\" DevicePath \"\"" Nov 25 11:58:10 crc kubenswrapper[4776]: I1125 11:58:10.512771 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.198007 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.99:8775/\": dial tcp 10.217.1.99:8775: connect: connection refused" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.198260 4776 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.99:8775/\": dial tcp 10.217.1.99:8775: connect: connection refused" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.315417 4776 generic.go:334] "Generic (PLEG): container finished" podID="ec8114a1-e74a-498c-a180-5d20733326f5" containerID="45a9c50d84c62aa7c775647665d279cdfcddd2f980985e31b8d45ecaf948d334" exitCode=0 Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.315538 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ec8114a1-e74a-498c-a180-5d20733326f5","Type":"ContainerDied","Data":"45a9c50d84c62aa7c775647665d279cdfcddd2f980985e31b8d45ecaf948d334"} Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.315580 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ec8114a1-e74a-498c-a180-5d20733326f5","Type":"ContainerDied","Data":"6bc8806f176397f277181f32f6b7f9283d79e4f837faf4e59ebf91978e984ff7"} Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.315594 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6bc8806f176397f277181f32f6b7f9283d79e4f837faf4e59ebf91978e984ff7" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.321661 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14","Type":"ContainerDied","Data":"235b27dba37c09496e9794f40807ee09d92277a36d96b376f20e24a9121f25d0"} Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.321611 4776 generic.go:334] "Generic (PLEG): container finished" podID="99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14" containerID="235b27dba37c09496e9794f40807ee09d92277a36d96b376f20e24a9121f25d0" exitCode=0 Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.324844 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.326041 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"0945270e-80b0-44ca-9c2b-642abacd57bf","Type":"ContainerStarted","Data":"d553b01b8b41e26fce6f5da920e5d2eaf96ded31ad76df1e1118559119c21dc8"} Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.326196 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.357053 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.357016026 podStartE2EDuration="2.357016026s" podCreationTimestamp="2025-11-25 11:58:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:58:11.343983939 +0000 UTC m=+9236.385043492" watchObservedRunningTime="2025-11-25 11:58:11.357016026 +0000 UTC m=+9236.398075579" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.474954 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.498471 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.500151 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.516521 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.537225 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 11:58:11 crc kubenswrapper[4776]: E1125 11:58:11.537839 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14" containerName="nova-metadata-log" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.537875 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14" containerName="nova-metadata-log" Nov 25 11:58:11 crc kubenswrapper[4776]: E1125 11:58:11.537894 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14" containerName="nova-metadata-metadata" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.537907 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14" containerName="nova-metadata-metadata" Nov 25 11:58:11 crc kubenswrapper[4776]: E1125 11:58:11.537947 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3" containerName="nova-scheduler-scheduler" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.537959 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3" containerName="nova-scheduler-scheduler" Nov 25 11:58:11 crc kubenswrapper[4776]: E1125 11:58:11.538022 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec8114a1-e74a-498c-a180-5d20733326f5" containerName="nova-cell1-conductor-conductor" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.538033 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec8114a1-e74a-498c-a180-5d20733326f5" containerName="nova-cell1-conductor-conductor" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.538345 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec8114a1-e74a-498c-a180-5d20733326f5" containerName="nova-cell1-conductor-conductor" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.538374 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14" containerName="nova-metadata-log" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.538390 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3" containerName="nova-scheduler-scheduler" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.538410 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14" containerName="nova-metadata-metadata" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.539389 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.540760 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec8114a1-e74a-498c-a180-5d20733326f5-combined-ca-bundle\") pod \"ec8114a1-e74a-498c-a180-5d20733326f5\" (UID: \"ec8114a1-e74a-498c-a180-5d20733326f5\") " Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.540830 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plwlp\" (UniqueName: \"kubernetes.io/projected/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-kube-api-access-plwlp\") pod \"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14\" (UID: \"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14\") " Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.540919 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-config-data\") pod \"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14\" (UID: \"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14\") " Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.540975 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-logs\") pod \"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14\" (UID: \"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14\") " Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.541044 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec8114a1-e74a-498c-a180-5d20733326f5-config-data\") pod \"ec8114a1-e74a-498c-a180-5d20733326f5\" (UID: \"ec8114a1-e74a-498c-a180-5d20733326f5\") " Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.541115 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-nova-metadata-tls-certs\") pod \"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14\" (UID: \"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14\") " Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.541309 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fgbm\" (UniqueName: \"kubernetes.io/projected/ec8114a1-e74a-498c-a180-5d20733326f5-kube-api-access-9fgbm\") pod \"ec8114a1-e74a-498c-a180-5d20733326f5\" (UID: \"ec8114a1-e74a-498c-a180-5d20733326f5\") " Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.541358 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-combined-ca-bundle\") pod \"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14\" (UID: \"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14\") " Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.541796 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-logs" (OuterVolumeSpecName: "logs") pod "99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14" (UID: "99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.542087 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-logs\") on node \"crc\" DevicePath \"\"" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.545500 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.555357 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-kube-api-access-plwlp" (OuterVolumeSpecName: "kube-api-access-plwlp") pod "99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14" (UID: "99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14"). InnerVolumeSpecName "kube-api-access-plwlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.579171 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.582453 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec8114a1-e74a-498c-a180-5d20733326f5-kube-api-access-9fgbm" (OuterVolumeSpecName: "kube-api-access-9fgbm") pod "ec8114a1-e74a-498c-a180-5d20733326f5" (UID: "ec8114a1-e74a-498c-a180-5d20733326f5"). InnerVolumeSpecName "kube-api-access-9fgbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.625880 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14" (UID: "99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.626918 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-config-data" (OuterVolumeSpecName: "config-data") pod "99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14" (UID: "99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.645287 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec8114a1-e74a-498c-a180-5d20733326f5-config-data" (OuterVolumeSpecName: "config-data") pod "ec8114a1-e74a-498c-a180-5d20733326f5" (UID: "ec8114a1-e74a-498c-a180-5d20733326f5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.647788 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qshl\" (UniqueName: \"kubernetes.io/projected/47179497-a4ae-4629-8a62-f1875880b8df-kube-api-access-9qshl\") pod \"nova-scheduler-0\" (UID: \"47179497-a4ae-4629-8a62-f1875880b8df\") " pod="openstack/nova-scheduler-0" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.647935 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47179497-a4ae-4629-8a62-f1875880b8df-config-data\") pod \"nova-scheduler-0\" (UID: \"47179497-a4ae-4629-8a62-f1875880b8df\") " pod="openstack/nova-scheduler-0" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.647998 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47179497-a4ae-4629-8a62-f1875880b8df-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"47179497-a4ae-4629-8a62-f1875880b8df\") " pod="openstack/nova-scheduler-0" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.648148 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plwlp\" (UniqueName: \"kubernetes.io/projected/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-kube-api-access-plwlp\") on node \"crc\" DevicePath \"\"" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.648164 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.648172 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec8114a1-e74a-498c-a180-5d20733326f5-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.648182 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fgbm\" (UniqueName: \"kubernetes.io/projected/ec8114a1-e74a-498c-a180-5d20733326f5-kube-api-access-9fgbm\") on node \"crc\" DevicePath \"\"" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.648191 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.648606 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14" (UID: "99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.649161 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec8114a1-e74a-498c-a180-5d20733326f5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ec8114a1-e74a-498c-a180-5d20733326f5" (UID: "ec8114a1-e74a-498c-a180-5d20733326f5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.674328 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3" path="/var/lib/kubelet/pods/f6277e15-c6b7-4f7d-af0a-4ef6d362c2a3/volumes" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.750433 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qshl\" (UniqueName: \"kubernetes.io/projected/47179497-a4ae-4629-8a62-f1875880b8df-kube-api-access-9qshl\") pod \"nova-scheduler-0\" (UID: \"47179497-a4ae-4629-8a62-f1875880b8df\") " pod="openstack/nova-scheduler-0" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.750625 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47179497-a4ae-4629-8a62-f1875880b8df-config-data\") pod \"nova-scheduler-0\" (UID: \"47179497-a4ae-4629-8a62-f1875880b8df\") " pod="openstack/nova-scheduler-0" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.750680 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47179497-a4ae-4629-8a62-f1875880b8df-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"47179497-a4ae-4629-8a62-f1875880b8df\") " pod="openstack/nova-scheduler-0" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.750800 4776 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.750820 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec8114a1-e74a-498c-a180-5d20733326f5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.754321 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47179497-a4ae-4629-8a62-f1875880b8df-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"47179497-a4ae-4629-8a62-f1875880b8df\") " pod="openstack/nova-scheduler-0" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.754462 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47179497-a4ae-4629-8a62-f1875880b8df-config-data\") pod \"nova-scheduler-0\" (UID: \"47179497-a4ae-4629-8a62-f1875880b8df\") " pod="openstack/nova-scheduler-0" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.768243 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qshl\" (UniqueName: \"kubernetes.io/projected/47179497-a4ae-4629-8a62-f1875880b8df-kube-api-access-9qshl\") pod \"nova-scheduler-0\" (UID: \"47179497-a4ae-4629-8a62-f1875880b8df\") " pod="openstack/nova-scheduler-0" Nov 25 11:58:11 crc kubenswrapper[4776]: I1125 11:58:11.893049 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.337751 4776 generic.go:334] "Generic (PLEG): container finished" podID="d854edbe-4354-4149-9d58-b503e44bd1c8" containerID="e2071c42555d92e072b179f72bc660dfec1e153919f6300461d43023f7c18e70" exitCode=0 Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.337802 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d854edbe-4354-4149-9d58-b503e44bd1c8","Type":"ContainerDied","Data":"e2071c42555d92e072b179f72bc660dfec1e153919f6300461d43023f7c18e70"} Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.337856 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d854edbe-4354-4149-9d58-b503e44bd1c8","Type":"ContainerDied","Data":"d0b260d7fdb4c737a5daa056f7515b10e6b2d43212d0556af09a948ed987e75b"} Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.337868 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0b260d7fdb4c737a5daa056f7515b10e6b2d43212d0556af09a948ed987e75b" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.339814 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14","Type":"ContainerDied","Data":"621286f1ba37497c9f2b5734b368462b55e27bdd382dc8a4ed2c4cf1dadf295c"} Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.339843 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.339859 4776 scope.go:117] "RemoveContainer" containerID="235b27dba37c09496e9794f40807ee09d92277a36d96b376f20e24a9121f25d0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.339985 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.367484 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.387407 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.401010 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.403866 4776 scope.go:117] "RemoveContainer" containerID="d691f3e8528edb91b33ba0d8b46b7550b59a4896095da1369db835bf1ea9af57" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.414254 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.426678 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.440143 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 25 11:58:12 crc kubenswrapper[4776]: E1125 11:58:12.440829 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d854edbe-4354-4149-9d58-b503e44bd1c8" containerName="nova-api-api" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.440854 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="d854edbe-4354-4149-9d58-b503e44bd1c8" containerName="nova-api-api" Nov 25 11:58:12 crc kubenswrapper[4776]: E1125 11:58:12.440872 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d854edbe-4354-4149-9d58-b503e44bd1c8" containerName="nova-api-log" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.440880 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="d854edbe-4354-4149-9d58-b503e44bd1c8" containerName="nova-api-log" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.441248 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="d854edbe-4354-4149-9d58-b503e44bd1c8" containerName="nova-api-log" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.441275 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="d854edbe-4354-4149-9d58-b503e44bd1c8" containerName="nova-api-api" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.442780 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.452025 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.452569 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.452642 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.468021 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d854edbe-4354-4149-9d58-b503e44bd1c8-combined-ca-bundle\") pod \"d854edbe-4354-4149-9d58-b503e44bd1c8\" (UID: \"d854edbe-4354-4149-9d58-b503e44bd1c8\") " Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.468395 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d854edbe-4354-4149-9d58-b503e44bd1c8-config-data\") pod \"d854edbe-4354-4149-9d58-b503e44bd1c8\" (UID: \"d854edbe-4354-4149-9d58-b503e44bd1c8\") " Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.468580 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d854edbe-4354-4149-9d58-b503e44bd1c8-internal-tls-certs\") pod \"d854edbe-4354-4149-9d58-b503e44bd1c8\" (UID: \"d854edbe-4354-4149-9d58-b503e44bd1c8\") " Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.468712 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmzcc\" (UniqueName: \"kubernetes.io/projected/d854edbe-4354-4149-9d58-b503e44bd1c8-kube-api-access-jmzcc\") pod \"d854edbe-4354-4149-9d58-b503e44bd1c8\" (UID: \"d854edbe-4354-4149-9d58-b503e44bd1c8\") " Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.468838 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d854edbe-4354-4149-9d58-b503e44bd1c8-public-tls-certs\") pod \"d854edbe-4354-4149-9d58-b503e44bd1c8\" (UID: \"d854edbe-4354-4149-9d58-b503e44bd1c8\") " Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.469182 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d854edbe-4354-4149-9d58-b503e44bd1c8-logs\") pod \"d854edbe-4354-4149-9d58-b503e44bd1c8\" (UID: \"d854edbe-4354-4149-9d58-b503e44bd1c8\") " Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.469675 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw6t4\" (UniqueName: \"kubernetes.io/projected/39189bd1-98eb-48a8-b087-ac932aa0dace-kube-api-access-kw6t4\") pod \"nova-metadata-0\" (UID: \"39189bd1-98eb-48a8-b087-ac932aa0dace\") " pod="openstack/nova-metadata-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.469877 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/39189bd1-98eb-48a8-b087-ac932aa0dace-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"39189bd1-98eb-48a8-b087-ac932aa0dace\") " pod="openstack/nova-metadata-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.470038 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39189bd1-98eb-48a8-b087-ac932aa0dace-config-data\") pod \"nova-metadata-0\" (UID: \"39189bd1-98eb-48a8-b087-ac932aa0dace\") " pod="openstack/nova-metadata-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.470384 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39189bd1-98eb-48a8-b087-ac932aa0dace-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"39189bd1-98eb-48a8-b087-ac932aa0dace\") " pod="openstack/nova-metadata-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.470632 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39189bd1-98eb-48a8-b087-ac932aa0dace-logs\") pod \"nova-metadata-0\" (UID: \"39189bd1-98eb-48a8-b087-ac932aa0dace\") " pod="openstack/nova-metadata-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.469735 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d854edbe-4354-4149-9d58-b503e44bd1c8-logs" (OuterVolumeSpecName: "logs") pod "d854edbe-4354-4149-9d58-b503e44bd1c8" (UID: "d854edbe-4354-4149-9d58-b503e44bd1c8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.478281 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.480373 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.487228 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.509049 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.515140 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d854edbe-4354-4149-9d58-b503e44bd1c8-kube-api-access-jmzcc" (OuterVolumeSpecName: "kube-api-access-jmzcc") pod "d854edbe-4354-4149-9d58-b503e44bd1c8" (UID: "d854edbe-4354-4149-9d58-b503e44bd1c8"). InnerVolumeSpecName "kube-api-access-jmzcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.528667 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d854edbe-4354-4149-9d58-b503e44bd1c8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d854edbe-4354-4149-9d58-b503e44bd1c8" (UID: "d854edbe-4354-4149-9d58-b503e44bd1c8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.534803 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 11:58:12 crc kubenswrapper[4776]: W1125 11:58:12.537758 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47179497_a4ae_4629_8a62_f1875880b8df.slice/crio-99fa330ac3f1ff03e0c9685f35946302419f8ae6377c7d3b9ffed9e881c56dac WatchSource:0}: Error finding container 99fa330ac3f1ff03e0c9685f35946302419f8ae6377c7d3b9ffed9e881c56dac: Status 404 returned error can't find the container with id 99fa330ac3f1ff03e0c9685f35946302419f8ae6377c7d3b9ffed9e881c56dac Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.573109 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw6t4\" (UniqueName: \"kubernetes.io/projected/39189bd1-98eb-48a8-b087-ac932aa0dace-kube-api-access-kw6t4\") pod \"nova-metadata-0\" (UID: \"39189bd1-98eb-48a8-b087-ac932aa0dace\") " pod="openstack/nova-metadata-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.573174 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/39189bd1-98eb-48a8-b087-ac932aa0dace-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"39189bd1-98eb-48a8-b087-ac932aa0dace\") " pod="openstack/nova-metadata-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.573217 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-874sj\" (UniqueName: \"kubernetes.io/projected/8a2e68e3-1379-4a66-9b36-bd74718b49ab-kube-api-access-874sj\") pod \"nova-cell1-conductor-0\" (UID: \"8a2e68e3-1379-4a66-9b36-bd74718b49ab\") " pod="openstack/nova-cell1-conductor-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.573261 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39189bd1-98eb-48a8-b087-ac932aa0dace-config-data\") pod \"nova-metadata-0\" (UID: \"39189bd1-98eb-48a8-b087-ac932aa0dace\") " pod="openstack/nova-metadata-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.573322 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a2e68e3-1379-4a66-9b36-bd74718b49ab-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"8a2e68e3-1379-4a66-9b36-bd74718b49ab\") " pod="openstack/nova-cell1-conductor-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.573384 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39189bd1-98eb-48a8-b087-ac932aa0dace-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"39189bd1-98eb-48a8-b087-ac932aa0dace\") " pod="openstack/nova-metadata-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.573460 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a2e68e3-1379-4a66-9b36-bd74718b49ab-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"8a2e68e3-1379-4a66-9b36-bd74718b49ab\") " pod="openstack/nova-cell1-conductor-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.573551 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39189bd1-98eb-48a8-b087-ac932aa0dace-logs\") pod \"nova-metadata-0\" (UID: \"39189bd1-98eb-48a8-b087-ac932aa0dace\") " pod="openstack/nova-metadata-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.573637 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d854edbe-4354-4149-9d58-b503e44bd1c8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.573660 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmzcc\" (UniqueName: \"kubernetes.io/projected/d854edbe-4354-4149-9d58-b503e44bd1c8-kube-api-access-jmzcc\") on node \"crc\" DevicePath \"\"" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.573673 4776 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d854edbe-4354-4149-9d58-b503e44bd1c8-logs\") on node \"crc\" DevicePath \"\"" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.574122 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/39189bd1-98eb-48a8-b087-ac932aa0dace-logs\") pod \"nova-metadata-0\" (UID: \"39189bd1-98eb-48a8-b087-ac932aa0dace\") " pod="openstack/nova-metadata-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.579748 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39189bd1-98eb-48a8-b087-ac932aa0dace-config-data\") pod \"nova-metadata-0\" (UID: \"39189bd1-98eb-48a8-b087-ac932aa0dace\") " pod="openstack/nova-metadata-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.581206 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/39189bd1-98eb-48a8-b087-ac932aa0dace-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"39189bd1-98eb-48a8-b087-ac932aa0dace\") " pod="openstack/nova-metadata-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.584707 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d854edbe-4354-4149-9d58-b503e44bd1c8-config-data" (OuterVolumeSpecName: "config-data") pod "d854edbe-4354-4149-9d58-b503e44bd1c8" (UID: "d854edbe-4354-4149-9d58-b503e44bd1c8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.594752 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39189bd1-98eb-48a8-b087-ac932aa0dace-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"39189bd1-98eb-48a8-b087-ac932aa0dace\") " pod="openstack/nova-metadata-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.604692 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d854edbe-4354-4149-9d58-b503e44bd1c8-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d854edbe-4354-4149-9d58-b503e44bd1c8" (UID: "d854edbe-4354-4149-9d58-b503e44bd1c8"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.606651 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d854edbe-4354-4149-9d58-b503e44bd1c8-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d854edbe-4354-4149-9d58-b503e44bd1c8" (UID: "d854edbe-4354-4149-9d58-b503e44bd1c8"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.608330 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw6t4\" (UniqueName: \"kubernetes.io/projected/39189bd1-98eb-48a8-b087-ac932aa0dace-kube-api-access-kw6t4\") pod \"nova-metadata-0\" (UID: \"39189bd1-98eb-48a8-b087-ac932aa0dace\") " pod="openstack/nova-metadata-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.675297 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a2e68e3-1379-4a66-9b36-bd74718b49ab-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"8a2e68e3-1379-4a66-9b36-bd74718b49ab\") " pod="openstack/nova-cell1-conductor-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.675980 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-874sj\" (UniqueName: \"kubernetes.io/projected/8a2e68e3-1379-4a66-9b36-bd74718b49ab-kube-api-access-874sj\") pod \"nova-cell1-conductor-0\" (UID: \"8a2e68e3-1379-4a66-9b36-bd74718b49ab\") " pod="openstack/nova-cell1-conductor-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.676097 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a2e68e3-1379-4a66-9b36-bd74718b49ab-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"8a2e68e3-1379-4a66-9b36-bd74718b49ab\") " pod="openstack/nova-cell1-conductor-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.676175 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d854edbe-4354-4149-9d58-b503e44bd1c8-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.676188 4776 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d854edbe-4354-4149-9d58-b503e44bd1c8-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.676201 4776 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d854edbe-4354-4149-9d58-b503e44bd1c8-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.679789 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a2e68e3-1379-4a66-9b36-bd74718b49ab-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"8a2e68e3-1379-4a66-9b36-bd74718b49ab\") " pod="openstack/nova-cell1-conductor-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.680122 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a2e68e3-1379-4a66-9b36-bd74718b49ab-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"8a2e68e3-1379-4a66-9b36-bd74718b49ab\") " pod="openstack/nova-cell1-conductor-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.691450 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-874sj\" (UniqueName: \"kubernetes.io/projected/8a2e68e3-1379-4a66-9b36-bd74718b49ab-kube-api-access-874sj\") pod \"nova-cell1-conductor-0\" (UID: \"8a2e68e3-1379-4a66-9b36-bd74718b49ab\") " pod="openstack/nova-cell1-conductor-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.780634 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 11:58:12 crc kubenswrapper[4776]: I1125 11:58:12.973411 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.163711 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.351167 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"39189bd1-98eb-48a8-b087-ac932aa0dace","Type":"ContainerStarted","Data":"7c684f7c0a6d99644b7d1699ce6fedf239b8adaab6fd8579c80ab852e1058622"} Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.352619 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"47179497-a4ae-4629-8a62-f1875880b8df","Type":"ContainerStarted","Data":"469b9b7c1f064870ab4288a2bc6c17fc120acd0a208a73ed00cf0740f1d46f46"} Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.352681 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"47179497-a4ae-4629-8a62-f1875880b8df","Type":"ContainerStarted","Data":"99fa330ac3f1ff03e0c9685f35946302419f8ae6377c7d3b9ffed9e881c56dac"} Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.353474 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.380309 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.380287892 podStartE2EDuration="2.380287892s" podCreationTimestamp="2025-11-25 11:58:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:58:13.366209818 +0000 UTC m=+9238.407269371" watchObservedRunningTime="2025-11-25 11:58:13.380287892 +0000 UTC m=+9238.421347445" Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.416040 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.432974 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.446486 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.448937 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.457619 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.459589 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.461247 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.469291 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.554965 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.609171 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ae65eee-ea73-403b-a154-e6ed7865d548-logs\") pod \"nova-api-0\" (UID: \"4ae65eee-ea73-403b-a154-e6ed7865d548\") " pod="openstack/nova-api-0" Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.609225 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgxw8\" (UniqueName: \"kubernetes.io/projected/4ae65eee-ea73-403b-a154-e6ed7865d548-kube-api-access-zgxw8\") pod \"nova-api-0\" (UID: \"4ae65eee-ea73-403b-a154-e6ed7865d548\") " pod="openstack/nova-api-0" Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.609367 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ae65eee-ea73-403b-a154-e6ed7865d548-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4ae65eee-ea73-403b-a154-e6ed7865d548\") " pod="openstack/nova-api-0" Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.609488 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ae65eee-ea73-403b-a154-e6ed7865d548-config-data\") pod \"nova-api-0\" (UID: \"4ae65eee-ea73-403b-a154-e6ed7865d548\") " pod="openstack/nova-api-0" Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.609601 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ae65eee-ea73-403b-a154-e6ed7865d548-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4ae65eee-ea73-403b-a154-e6ed7865d548\") " pod="openstack/nova-api-0" Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.609803 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ae65eee-ea73-403b-a154-e6ed7865d548-public-tls-certs\") pod \"nova-api-0\" (UID: \"4ae65eee-ea73-403b-a154-e6ed7865d548\") " pod="openstack/nova-api-0" Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.676360 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14" path="/var/lib/kubelet/pods/99c9e61f-5ae7-41bf-800a-2fcb3d9f2e14/volumes" Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.677608 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d854edbe-4354-4149-9d58-b503e44bd1c8" path="/var/lib/kubelet/pods/d854edbe-4354-4149-9d58-b503e44bd1c8/volumes" Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.680884 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec8114a1-e74a-498c-a180-5d20733326f5" path="/var/lib/kubelet/pods/ec8114a1-e74a-498c-a180-5d20733326f5/volumes" Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.712939 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ae65eee-ea73-403b-a154-e6ed7865d548-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4ae65eee-ea73-403b-a154-e6ed7865d548\") " pod="openstack/nova-api-0" Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.713104 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ae65eee-ea73-403b-a154-e6ed7865d548-public-tls-certs\") pod \"nova-api-0\" (UID: \"4ae65eee-ea73-403b-a154-e6ed7865d548\") " pod="openstack/nova-api-0" Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.713191 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ae65eee-ea73-403b-a154-e6ed7865d548-logs\") pod \"nova-api-0\" (UID: \"4ae65eee-ea73-403b-a154-e6ed7865d548\") " pod="openstack/nova-api-0" Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.713217 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgxw8\" (UniqueName: \"kubernetes.io/projected/4ae65eee-ea73-403b-a154-e6ed7865d548-kube-api-access-zgxw8\") pod \"nova-api-0\" (UID: \"4ae65eee-ea73-403b-a154-e6ed7865d548\") " pod="openstack/nova-api-0" Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.713292 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ae65eee-ea73-403b-a154-e6ed7865d548-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4ae65eee-ea73-403b-a154-e6ed7865d548\") " pod="openstack/nova-api-0" Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.713363 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ae65eee-ea73-403b-a154-e6ed7865d548-config-data\") pod \"nova-api-0\" (UID: \"4ae65eee-ea73-403b-a154-e6ed7865d548\") " pod="openstack/nova-api-0" Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.713759 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ae65eee-ea73-403b-a154-e6ed7865d548-logs\") pod \"nova-api-0\" (UID: \"4ae65eee-ea73-403b-a154-e6ed7865d548\") " pod="openstack/nova-api-0" Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.717162 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ae65eee-ea73-403b-a154-e6ed7865d548-public-tls-certs\") pod \"nova-api-0\" (UID: \"4ae65eee-ea73-403b-a154-e6ed7865d548\") " pod="openstack/nova-api-0" Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.717576 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ae65eee-ea73-403b-a154-e6ed7865d548-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4ae65eee-ea73-403b-a154-e6ed7865d548\") " pod="openstack/nova-api-0" Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.717760 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ae65eee-ea73-403b-a154-e6ed7865d548-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4ae65eee-ea73-403b-a154-e6ed7865d548\") " pod="openstack/nova-api-0" Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.718483 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ae65eee-ea73-403b-a154-e6ed7865d548-config-data\") pod \"nova-api-0\" (UID: \"4ae65eee-ea73-403b-a154-e6ed7865d548\") " pod="openstack/nova-api-0" Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.730481 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgxw8\" (UniqueName: \"kubernetes.io/projected/4ae65eee-ea73-403b-a154-e6ed7865d548-kube-api-access-zgxw8\") pod \"nova-api-0\" (UID: \"4ae65eee-ea73-403b-a154-e6ed7865d548\") " pod="openstack/nova-api-0" Nov 25 11:58:13 crc kubenswrapper[4776]: I1125 11:58:13.787266 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 11:58:14 crc kubenswrapper[4776]: I1125 11:58:14.351640 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 11:58:14 crc kubenswrapper[4776]: I1125 11:58:14.382998 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"8a2e68e3-1379-4a66-9b36-bd74718b49ab","Type":"ContainerStarted","Data":"bc39b25a560317e94fa05e9d9ae9298df60c615e17f045f950fb1c7442aa4e0f"} Nov 25 11:58:14 crc kubenswrapper[4776]: I1125 11:58:14.383052 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"8a2e68e3-1379-4a66-9b36-bd74718b49ab","Type":"ContainerStarted","Data":"cda131a3fe8758f3032b76149b11658c8ec1a8dad4a537365da807f385311307"} Nov 25 11:58:14 crc kubenswrapper[4776]: I1125 11:58:14.384586 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 25 11:58:14 crc kubenswrapper[4776]: I1125 11:58:14.392245 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"39189bd1-98eb-48a8-b087-ac932aa0dace","Type":"ContainerStarted","Data":"10d02555d8d3d52005dc149a96ee5ff32e5c427f58f39debb4fd87a65ee89a09"} Nov 25 11:58:14 crc kubenswrapper[4776]: I1125 11:58:14.392290 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"39189bd1-98eb-48a8-b087-ac932aa0dace","Type":"ContainerStarted","Data":"7668b8ce9762a6b3487517cb6581e9005d6bddd9a8dbd25355423f34d1578e15"} Nov 25 11:58:14 crc kubenswrapper[4776]: I1125 11:58:14.410891 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.410870259 podStartE2EDuration="2.410870259s" podCreationTimestamp="2025-11-25 11:58:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:58:14.410445268 +0000 UTC m=+9239.451504821" watchObservedRunningTime="2025-11-25 11:58:14.410870259 +0000 UTC m=+9239.451929812" Nov 25 11:58:14 crc kubenswrapper[4776]: I1125 11:58:14.454743 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.454722269 podStartE2EDuration="2.454722269s" podCreationTimestamp="2025-11-25 11:58:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:58:14.445520318 +0000 UTC m=+9239.486579871" watchObservedRunningTime="2025-11-25 11:58:14.454722269 +0000 UTC m=+9239.495781822" Nov 25 11:58:15 crc kubenswrapper[4776]: I1125 11:58:15.405479 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4ae65eee-ea73-403b-a154-e6ed7865d548","Type":"ContainerStarted","Data":"1af3598a458b0ace5f9a6d109ba5c1a55a5bbf00f8500dc34f459080a3a1b8e5"} Nov 25 11:58:15 crc kubenswrapper[4776]: I1125 11:58:15.405764 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4ae65eee-ea73-403b-a154-e6ed7865d548","Type":"ContainerStarted","Data":"9116bd992ade58ba46afc8af192dafc35b0fe7ea2898e4d316f40979384accb7"} Nov 25 11:58:15 crc kubenswrapper[4776]: I1125 11:58:15.405775 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4ae65eee-ea73-403b-a154-e6ed7865d548","Type":"ContainerStarted","Data":"7ea14fa4087105645d47bc6a12c538b442617c753eeccb9761341b03d63b49a3"} Nov 25 11:58:15 crc kubenswrapper[4776]: I1125 11:58:15.435226 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.43520001 podStartE2EDuration="2.43520001s" podCreationTimestamp="2025-11-25 11:58:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 11:58:15.429975339 +0000 UTC m=+9240.471034902" watchObservedRunningTime="2025-11-25 11:58:15.43520001 +0000 UTC m=+9240.476259563" Nov 25 11:58:16 crc kubenswrapper[4776]: I1125 11:58:16.894231 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 25 11:58:17 crc kubenswrapper[4776]: I1125 11:58:17.781873 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 11:58:17 crc kubenswrapper[4776]: I1125 11:58:17.782468 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 11:58:20 crc kubenswrapper[4776]: I1125 11:58:20.082594 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 25 11:58:21 crc kubenswrapper[4776]: I1125 11:58:21.894263 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 25 11:58:21 crc kubenswrapper[4776]: I1125 11:58:21.930433 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 25 11:58:22 crc kubenswrapper[4776]: I1125 11:58:22.509096 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 25 11:58:22 crc kubenswrapper[4776]: I1125 11:58:22.780759 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 11:58:22 crc kubenswrapper[4776]: I1125 11:58:22.781894 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 11:58:23 crc kubenswrapper[4776]: I1125 11:58:23.007823 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 25 11:58:23 crc kubenswrapper[4776]: I1125 11:58:23.787633 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 11:58:23 crc kubenswrapper[4776]: I1125 11:58:23.787716 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 11:58:23 crc kubenswrapper[4776]: I1125 11:58:23.794280 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="39189bd1-98eb-48a8-b087-ac932aa0dace" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 11:58:23 crc kubenswrapper[4776]: I1125 11:58:23.794731 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="39189bd1-98eb-48a8-b087-ac932aa0dace" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 11:58:24 crc kubenswrapper[4776]: I1125 11:58:24.803370 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4ae65eee-ea73-403b-a154-e6ed7865d548" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.206:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 11:58:24 crc kubenswrapper[4776]: I1125 11:58:24.803378 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4ae65eee-ea73-403b-a154-e6ed7865d548" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.206:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 11:58:32 crc kubenswrapper[4776]: I1125 11:58:32.787827 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 25 11:58:32 crc kubenswrapper[4776]: I1125 11:58:32.789110 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 25 11:58:32 crc kubenswrapper[4776]: I1125 11:58:32.793590 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 25 11:58:32 crc kubenswrapper[4776]: I1125 11:58:32.795255 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 25 11:58:33 crc kubenswrapper[4776]: I1125 11:58:33.795619 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 11:58:33 crc kubenswrapper[4776]: I1125 11:58:33.795803 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 11:58:33 crc kubenswrapper[4776]: I1125 11:58:33.796700 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 11:58:33 crc kubenswrapper[4776]: I1125 11:58:33.797169 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 11:58:33 crc kubenswrapper[4776]: I1125 11:58:33.805412 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 11:58:33 crc kubenswrapper[4776]: I1125 11:58:33.806456 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.025570 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk"] Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.027929 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.032950 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.033062 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-dknh7" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.033253 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.033258 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.033357 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.033589 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.034499 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.050659 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk"] Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.085010 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/1db10c36-1c34-47f5-a872-a82d15d88299-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.085110 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.085159 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.085185 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.085209 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssr88\" (UniqueName: \"kubernetes.io/projected/1db10c36-1c34-47f5-a872-a82d15d88299-kube-api-access-ssr88\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.085228 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.085270 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.085330 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.085381 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.187550 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssr88\" (UniqueName: \"kubernetes.io/projected/1db10c36-1c34-47f5-a872-a82d15d88299-kube-api-access-ssr88\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.187602 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.187652 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.187719 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.187779 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.187832 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/1db10c36-1c34-47f5-a872-a82d15d88299-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.187879 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.187921 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.187938 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.190526 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/1db10c36-1c34-47f5-a872-a82d15d88299-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.194538 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.195681 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.196086 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.196104 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.197001 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.197743 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.203903 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.209504 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssr88\" (UniqueName: \"kubernetes.io/projected/1db10c36-1c34-47f5-a872-a82d15d88299-kube-api-access-ssr88\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:35 crc kubenswrapper[4776]: I1125 11:58:35.393148 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 11:58:36 crc kubenswrapper[4776]: I1125 11:58:36.042983 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk"] Nov 25 11:58:36 crc kubenswrapper[4776]: W1125 11:58:36.047721 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1db10c36_1c34_47f5_a872_a82d15d88299.slice/crio-e6a5ccf05e0c7c8ad99d27f5bf5d99cab8c3979387f222c407bda5e73ec12d65 WatchSource:0}: Error finding container e6a5ccf05e0c7c8ad99d27f5bf5d99cab8c3979387f222c407bda5e73ec12d65: Status 404 returned error can't find the container with id e6a5ccf05e0c7c8ad99d27f5bf5d99cab8c3979387f222c407bda5e73ec12d65 Nov 25 11:58:36 crc kubenswrapper[4776]: I1125 11:58:36.051537 4776 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 11:58:36 crc kubenswrapper[4776]: I1125 11:58:36.631388 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" event={"ID":"1db10c36-1c34-47f5-a872-a82d15d88299","Type":"ContainerStarted","Data":"e6a5ccf05e0c7c8ad99d27f5bf5d99cab8c3979387f222c407bda5e73ec12d65"} Nov 25 11:58:37 crc kubenswrapper[4776]: I1125 11:58:37.649819 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" event={"ID":"1db10c36-1c34-47f5-a872-a82d15d88299","Type":"ContainerStarted","Data":"348c62a0a95daa8f9cf57bb1fd0e254d304868ccfdf61b5064ce0a023c6be675"} Nov 25 11:58:37 crc kubenswrapper[4776]: I1125 11:58:37.677960 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" podStartSLOduration=3.22734153 podStartE2EDuration="3.677940276s" podCreationTimestamp="2025-11-25 11:58:34 +0000 UTC" firstStartedPulling="2025-11-25 11:58:36.051336793 +0000 UTC m=+9261.092396346" lastFinishedPulling="2025-11-25 11:58:36.501935529 +0000 UTC m=+9261.542995092" observedRunningTime="2025-11-25 11:58:37.676193852 +0000 UTC m=+9262.717253415" watchObservedRunningTime="2025-11-25 11:58:37.677940276 +0000 UTC m=+9262.718999829" Nov 25 11:58:42 crc kubenswrapper[4776]: I1125 11:58:42.354924 4776 scope.go:117] "RemoveContainer" containerID="457893b901e04fac4aa2a523915fbb0b5c6a60a040f98930992bfbfa4f443310" Nov 25 11:58:42 crc kubenswrapper[4776]: I1125 11:58:42.385737 4776 scope.go:117] "RemoveContainer" containerID="e2071c42555d92e072b179f72bc660dfec1e153919f6300461d43023f7c18e70" Nov 25 11:58:42 crc kubenswrapper[4776]: I1125 11:58:42.414922 4776 scope.go:117] "RemoveContainer" containerID="45a9c50d84c62aa7c775647665d279cdfcddd2f980985e31b8d45ecaf948d334" Nov 25 11:58:42 crc kubenswrapper[4776]: I1125 11:58:42.449789 4776 scope.go:117] "RemoveContainer" containerID="d6e46cd49b2c18d00f7c5566b97c6a00bed98218a03cc5997bbd1e486a0d50d1" Nov 25 11:59:44 crc kubenswrapper[4776]: I1125 11:59:44.140120 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-n7fjw"] Nov 25 11:59:44 crc kubenswrapper[4776]: I1125 11:59:44.143981 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n7fjw" Nov 25 11:59:44 crc kubenswrapper[4776]: I1125 11:59:44.154794 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n7fjw"] Nov 25 11:59:44 crc kubenswrapper[4776]: I1125 11:59:44.223654 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wg5t9\" (UniqueName: \"kubernetes.io/projected/393868fc-3b2b-4aad-8621-530e5ba64268-kube-api-access-wg5t9\") pod \"redhat-operators-n7fjw\" (UID: \"393868fc-3b2b-4aad-8621-530e5ba64268\") " pod="openshift-marketplace/redhat-operators-n7fjw" Nov 25 11:59:44 crc kubenswrapper[4776]: I1125 11:59:44.223807 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/393868fc-3b2b-4aad-8621-530e5ba64268-utilities\") pod \"redhat-operators-n7fjw\" (UID: \"393868fc-3b2b-4aad-8621-530e5ba64268\") " pod="openshift-marketplace/redhat-operators-n7fjw" Nov 25 11:59:44 crc kubenswrapper[4776]: I1125 11:59:44.223834 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/393868fc-3b2b-4aad-8621-530e5ba64268-catalog-content\") pod \"redhat-operators-n7fjw\" (UID: \"393868fc-3b2b-4aad-8621-530e5ba64268\") " pod="openshift-marketplace/redhat-operators-n7fjw" Nov 25 11:59:44 crc kubenswrapper[4776]: I1125 11:59:44.325873 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/393868fc-3b2b-4aad-8621-530e5ba64268-utilities\") pod \"redhat-operators-n7fjw\" (UID: \"393868fc-3b2b-4aad-8621-530e5ba64268\") " pod="openshift-marketplace/redhat-operators-n7fjw" Nov 25 11:59:44 crc kubenswrapper[4776]: I1125 11:59:44.326289 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/393868fc-3b2b-4aad-8621-530e5ba64268-catalog-content\") pod \"redhat-operators-n7fjw\" (UID: \"393868fc-3b2b-4aad-8621-530e5ba64268\") " pod="openshift-marketplace/redhat-operators-n7fjw" Nov 25 11:59:44 crc kubenswrapper[4776]: I1125 11:59:44.326452 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/393868fc-3b2b-4aad-8621-530e5ba64268-utilities\") pod \"redhat-operators-n7fjw\" (UID: \"393868fc-3b2b-4aad-8621-530e5ba64268\") " pod="openshift-marketplace/redhat-operators-n7fjw" Nov 25 11:59:44 crc kubenswrapper[4776]: I1125 11:59:44.326480 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg5t9\" (UniqueName: \"kubernetes.io/projected/393868fc-3b2b-4aad-8621-530e5ba64268-kube-api-access-wg5t9\") pod \"redhat-operators-n7fjw\" (UID: \"393868fc-3b2b-4aad-8621-530e5ba64268\") " pod="openshift-marketplace/redhat-operators-n7fjw" Nov 25 11:59:44 crc kubenswrapper[4776]: I1125 11:59:44.327087 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/393868fc-3b2b-4aad-8621-530e5ba64268-catalog-content\") pod \"redhat-operators-n7fjw\" (UID: \"393868fc-3b2b-4aad-8621-530e5ba64268\") " pod="openshift-marketplace/redhat-operators-n7fjw" Nov 25 11:59:44 crc kubenswrapper[4776]: I1125 11:59:44.351868 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg5t9\" (UniqueName: \"kubernetes.io/projected/393868fc-3b2b-4aad-8621-530e5ba64268-kube-api-access-wg5t9\") pod \"redhat-operators-n7fjw\" (UID: \"393868fc-3b2b-4aad-8621-530e5ba64268\") " pod="openshift-marketplace/redhat-operators-n7fjw" Nov 25 11:59:44 crc kubenswrapper[4776]: I1125 11:59:44.468039 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n7fjw" Nov 25 11:59:44 crc kubenswrapper[4776]: I1125 11:59:44.981151 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n7fjw"] Nov 25 11:59:45 crc kubenswrapper[4776]: I1125 11:59:45.383247 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n7fjw" event={"ID":"393868fc-3b2b-4aad-8621-530e5ba64268","Type":"ContainerStarted","Data":"fa0ba5753ce17ddfafa4d13ffcbfdee679bee87b4520d89475e12ff429bde536"} Nov 25 11:59:45 crc kubenswrapper[4776]: I1125 11:59:45.383714 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n7fjw" event={"ID":"393868fc-3b2b-4aad-8621-530e5ba64268","Type":"ContainerStarted","Data":"4f8054b2bbb3383029d316d773488fbc9811b9bdd70cc5924200bf035712bf80"} Nov 25 11:59:46 crc kubenswrapper[4776]: I1125 11:59:46.398315 4776 generic.go:334] "Generic (PLEG): container finished" podID="393868fc-3b2b-4aad-8621-530e5ba64268" containerID="fa0ba5753ce17ddfafa4d13ffcbfdee679bee87b4520d89475e12ff429bde536" exitCode=0 Nov 25 11:59:46 crc kubenswrapper[4776]: I1125 11:59:46.398705 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n7fjw" event={"ID":"393868fc-3b2b-4aad-8621-530e5ba64268","Type":"ContainerDied","Data":"fa0ba5753ce17ddfafa4d13ffcbfdee679bee87b4520d89475e12ff429bde536"} Nov 25 11:59:47 crc kubenswrapper[4776]: I1125 11:59:47.819381 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 11:59:47 crc kubenswrapper[4776]: I1125 11:59:47.819854 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 11:59:48 crc kubenswrapper[4776]: I1125 11:59:48.427490 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n7fjw" event={"ID":"393868fc-3b2b-4aad-8621-530e5ba64268","Type":"ContainerStarted","Data":"3ee570da3e051ced2570f4231bb916714f037f71cce01519e7cf649d623644b4"} Nov 25 11:59:49 crc kubenswrapper[4776]: I1125 11:59:49.722576 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f8dn9"] Nov 25 11:59:49 crc kubenswrapper[4776]: I1125 11:59:49.725578 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f8dn9" Nov 25 11:59:49 crc kubenswrapper[4776]: I1125 11:59:49.746987 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f8dn9"] Nov 25 11:59:49 crc kubenswrapper[4776]: I1125 11:59:49.886278 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jwn9\" (UniqueName: \"kubernetes.io/projected/769ca824-3fc6-47bc-8d13-8a8d0dfb8449-kube-api-access-9jwn9\") pod \"community-operators-f8dn9\" (UID: \"769ca824-3fc6-47bc-8d13-8a8d0dfb8449\") " pod="openshift-marketplace/community-operators-f8dn9" Nov 25 11:59:49 crc kubenswrapper[4776]: I1125 11:59:49.886540 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/769ca824-3fc6-47bc-8d13-8a8d0dfb8449-utilities\") pod \"community-operators-f8dn9\" (UID: \"769ca824-3fc6-47bc-8d13-8a8d0dfb8449\") " pod="openshift-marketplace/community-operators-f8dn9" Nov 25 11:59:49 crc kubenswrapper[4776]: I1125 11:59:49.886764 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/769ca824-3fc6-47bc-8d13-8a8d0dfb8449-catalog-content\") pod \"community-operators-f8dn9\" (UID: \"769ca824-3fc6-47bc-8d13-8a8d0dfb8449\") " pod="openshift-marketplace/community-operators-f8dn9" Nov 25 11:59:49 crc kubenswrapper[4776]: I1125 11:59:49.988687 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/769ca824-3fc6-47bc-8d13-8a8d0dfb8449-utilities\") pod \"community-operators-f8dn9\" (UID: \"769ca824-3fc6-47bc-8d13-8a8d0dfb8449\") " pod="openshift-marketplace/community-operators-f8dn9" Nov 25 11:59:49 crc kubenswrapper[4776]: I1125 11:59:49.988781 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/769ca824-3fc6-47bc-8d13-8a8d0dfb8449-catalog-content\") pod \"community-operators-f8dn9\" (UID: \"769ca824-3fc6-47bc-8d13-8a8d0dfb8449\") " pod="openshift-marketplace/community-operators-f8dn9" Nov 25 11:59:49 crc kubenswrapper[4776]: I1125 11:59:49.988962 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jwn9\" (UniqueName: \"kubernetes.io/projected/769ca824-3fc6-47bc-8d13-8a8d0dfb8449-kube-api-access-9jwn9\") pod \"community-operators-f8dn9\" (UID: \"769ca824-3fc6-47bc-8d13-8a8d0dfb8449\") " pod="openshift-marketplace/community-operators-f8dn9" Nov 25 11:59:49 crc kubenswrapper[4776]: I1125 11:59:49.989596 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/769ca824-3fc6-47bc-8d13-8a8d0dfb8449-utilities\") pod \"community-operators-f8dn9\" (UID: \"769ca824-3fc6-47bc-8d13-8a8d0dfb8449\") " pod="openshift-marketplace/community-operators-f8dn9" Nov 25 11:59:49 crc kubenswrapper[4776]: I1125 11:59:49.989646 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/769ca824-3fc6-47bc-8d13-8a8d0dfb8449-catalog-content\") pod \"community-operators-f8dn9\" (UID: \"769ca824-3fc6-47bc-8d13-8a8d0dfb8449\") " pod="openshift-marketplace/community-operators-f8dn9" Nov 25 11:59:50 crc kubenswrapper[4776]: I1125 11:59:50.023274 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jwn9\" (UniqueName: \"kubernetes.io/projected/769ca824-3fc6-47bc-8d13-8a8d0dfb8449-kube-api-access-9jwn9\") pod \"community-operators-f8dn9\" (UID: \"769ca824-3fc6-47bc-8d13-8a8d0dfb8449\") " pod="openshift-marketplace/community-operators-f8dn9" Nov 25 11:59:50 crc kubenswrapper[4776]: I1125 11:59:50.052853 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f8dn9" Nov 25 11:59:50 crc kubenswrapper[4776]: I1125 11:59:50.770671 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f8dn9"] Nov 25 11:59:50 crc kubenswrapper[4776]: W1125 11:59:50.776740 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod769ca824_3fc6_47bc_8d13_8a8d0dfb8449.slice/crio-15e14170722ebc1747325bc6329a2ddc347c2e3ccaec38aa3ec6651ecf2ba5fa WatchSource:0}: Error finding container 15e14170722ebc1747325bc6329a2ddc347c2e3ccaec38aa3ec6651ecf2ba5fa: Status 404 returned error can't find the container with id 15e14170722ebc1747325bc6329a2ddc347c2e3ccaec38aa3ec6651ecf2ba5fa Nov 25 11:59:51 crc kubenswrapper[4776]: I1125 11:59:51.461523 4776 generic.go:334] "Generic (PLEG): container finished" podID="769ca824-3fc6-47bc-8d13-8a8d0dfb8449" containerID="1de9d50a05b9e641da13f7c4a37a3782e69837425dc3031b3ec35e6b7c092127" exitCode=0 Nov 25 11:59:51 crc kubenswrapper[4776]: I1125 11:59:51.461576 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8dn9" event={"ID":"769ca824-3fc6-47bc-8d13-8a8d0dfb8449","Type":"ContainerDied","Data":"1de9d50a05b9e641da13f7c4a37a3782e69837425dc3031b3ec35e6b7c092127"} Nov 25 11:59:51 crc kubenswrapper[4776]: I1125 11:59:51.461858 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8dn9" event={"ID":"769ca824-3fc6-47bc-8d13-8a8d0dfb8449","Type":"ContainerStarted","Data":"15e14170722ebc1747325bc6329a2ddc347c2e3ccaec38aa3ec6651ecf2ba5fa"} Nov 25 11:59:53 crc kubenswrapper[4776]: I1125 11:59:53.481604 4776 generic.go:334] "Generic (PLEG): container finished" podID="393868fc-3b2b-4aad-8621-530e5ba64268" containerID="3ee570da3e051ced2570f4231bb916714f037f71cce01519e7cf649d623644b4" exitCode=0 Nov 25 11:59:53 crc kubenswrapper[4776]: I1125 11:59:53.481716 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n7fjw" event={"ID":"393868fc-3b2b-4aad-8621-530e5ba64268","Type":"ContainerDied","Data":"3ee570da3e051ced2570f4231bb916714f037f71cce01519e7cf649d623644b4"} Nov 25 11:59:53 crc kubenswrapper[4776]: I1125 11:59:53.487477 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8dn9" event={"ID":"769ca824-3fc6-47bc-8d13-8a8d0dfb8449","Type":"ContainerStarted","Data":"fce685d4154bee9c802d787f03fd0271a651ae4caf82c122db83c3485a569724"} Nov 25 11:59:54 crc kubenswrapper[4776]: I1125 11:59:54.500096 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n7fjw" event={"ID":"393868fc-3b2b-4aad-8621-530e5ba64268","Type":"ContainerStarted","Data":"d45573d6e1f2ed34b1a535d0b040271fbc6280f71131740685efe02f38a49cb8"} Nov 25 11:59:54 crc kubenswrapper[4776]: I1125 11:59:54.529288 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-n7fjw" podStartSLOduration=2.9688753930000003 podStartE2EDuration="10.529264838s" podCreationTimestamp="2025-11-25 11:59:44 +0000 UTC" firstStartedPulling="2025-11-25 11:59:46.402118033 +0000 UTC m=+9331.443177586" lastFinishedPulling="2025-11-25 11:59:53.962507478 +0000 UTC m=+9339.003567031" observedRunningTime="2025-11-25 11:59:54.521910724 +0000 UTC m=+9339.562970287" watchObservedRunningTime="2025-11-25 11:59:54.529264838 +0000 UTC m=+9339.570324391" Nov 25 11:59:56 crc kubenswrapper[4776]: I1125 11:59:56.521984 4776 generic.go:334] "Generic (PLEG): container finished" podID="769ca824-3fc6-47bc-8d13-8a8d0dfb8449" containerID="fce685d4154bee9c802d787f03fd0271a651ae4caf82c122db83c3485a569724" exitCode=0 Nov 25 11:59:56 crc kubenswrapper[4776]: I1125 11:59:56.522063 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8dn9" event={"ID":"769ca824-3fc6-47bc-8d13-8a8d0dfb8449","Type":"ContainerDied","Data":"fce685d4154bee9c802d787f03fd0271a651ae4caf82c122db83c3485a569724"} Nov 25 11:59:57 crc kubenswrapper[4776]: I1125 11:59:57.536356 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8dn9" event={"ID":"769ca824-3fc6-47bc-8d13-8a8d0dfb8449","Type":"ContainerStarted","Data":"bd03f5a84d03b505e2aafdcb33c64394a93a13e698ed6235b411d260427ca376"} Nov 25 12:00:00 crc kubenswrapper[4776]: I1125 12:00:00.055060 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f8dn9" Nov 25 12:00:00 crc kubenswrapper[4776]: I1125 12:00:00.055163 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f8dn9" Nov 25 12:00:00 crc kubenswrapper[4776]: I1125 12:00:00.157983 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f8dn9" podStartSLOduration=5.484899114 podStartE2EDuration="11.157958406s" podCreationTimestamp="2025-11-25 11:59:49 +0000 UTC" firstStartedPulling="2025-11-25 11:59:51.463587678 +0000 UTC m=+9336.504647231" lastFinishedPulling="2025-11-25 11:59:57.13664697 +0000 UTC m=+9342.177706523" observedRunningTime="2025-11-25 11:59:58.570457834 +0000 UTC m=+9343.611517407" watchObservedRunningTime="2025-11-25 12:00:00.157958406 +0000 UTC m=+9345.199017959" Nov 25 12:00:00 crc kubenswrapper[4776]: I1125 12:00:00.158238 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401200-cq5h9"] Nov 25 12:00:00 crc kubenswrapper[4776]: I1125 12:00:00.159913 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-cq5h9" Nov 25 12:00:00 crc kubenswrapper[4776]: I1125 12:00:00.161718 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 12:00:00 crc kubenswrapper[4776]: I1125 12:00:00.162359 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 12:00:00 crc kubenswrapper[4776]: I1125 12:00:00.171243 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401200-cq5h9"] Nov 25 12:00:00 crc kubenswrapper[4776]: I1125 12:00:00.254359 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bz8b\" (UniqueName: \"kubernetes.io/projected/1ebb4b80-22bd-440c-b70e-e533e9dc6b23-kube-api-access-9bz8b\") pod \"collect-profiles-29401200-cq5h9\" (UID: \"1ebb4b80-22bd-440c-b70e-e533e9dc6b23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-cq5h9" Nov 25 12:00:00 crc kubenswrapper[4776]: I1125 12:00:00.254446 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1ebb4b80-22bd-440c-b70e-e533e9dc6b23-config-volume\") pod \"collect-profiles-29401200-cq5h9\" (UID: \"1ebb4b80-22bd-440c-b70e-e533e9dc6b23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-cq5h9" Nov 25 12:00:00 crc kubenswrapper[4776]: I1125 12:00:00.254510 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1ebb4b80-22bd-440c-b70e-e533e9dc6b23-secret-volume\") pod \"collect-profiles-29401200-cq5h9\" (UID: \"1ebb4b80-22bd-440c-b70e-e533e9dc6b23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-cq5h9" Nov 25 12:00:00 crc kubenswrapper[4776]: I1125 12:00:00.357131 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bz8b\" (UniqueName: \"kubernetes.io/projected/1ebb4b80-22bd-440c-b70e-e533e9dc6b23-kube-api-access-9bz8b\") pod \"collect-profiles-29401200-cq5h9\" (UID: \"1ebb4b80-22bd-440c-b70e-e533e9dc6b23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-cq5h9" Nov 25 12:00:00 crc kubenswrapper[4776]: I1125 12:00:00.357249 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1ebb4b80-22bd-440c-b70e-e533e9dc6b23-config-volume\") pod \"collect-profiles-29401200-cq5h9\" (UID: \"1ebb4b80-22bd-440c-b70e-e533e9dc6b23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-cq5h9" Nov 25 12:00:00 crc kubenswrapper[4776]: I1125 12:00:00.357338 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1ebb4b80-22bd-440c-b70e-e533e9dc6b23-secret-volume\") pod \"collect-profiles-29401200-cq5h9\" (UID: \"1ebb4b80-22bd-440c-b70e-e533e9dc6b23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-cq5h9" Nov 25 12:00:00 crc kubenswrapper[4776]: I1125 12:00:00.358346 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1ebb4b80-22bd-440c-b70e-e533e9dc6b23-config-volume\") pod \"collect-profiles-29401200-cq5h9\" (UID: \"1ebb4b80-22bd-440c-b70e-e533e9dc6b23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-cq5h9" Nov 25 12:00:00 crc kubenswrapper[4776]: I1125 12:00:00.368172 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1ebb4b80-22bd-440c-b70e-e533e9dc6b23-secret-volume\") pod \"collect-profiles-29401200-cq5h9\" (UID: \"1ebb4b80-22bd-440c-b70e-e533e9dc6b23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-cq5h9" Nov 25 12:00:00 crc kubenswrapper[4776]: I1125 12:00:00.375854 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bz8b\" (UniqueName: \"kubernetes.io/projected/1ebb4b80-22bd-440c-b70e-e533e9dc6b23-kube-api-access-9bz8b\") pod \"collect-profiles-29401200-cq5h9\" (UID: \"1ebb4b80-22bd-440c-b70e-e533e9dc6b23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-cq5h9" Nov 25 12:00:00 crc kubenswrapper[4776]: I1125 12:00:00.528391 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-cq5h9" Nov 25 12:00:01 crc kubenswrapper[4776]: I1125 12:00:01.056695 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401200-cq5h9"] Nov 25 12:00:01 crc kubenswrapper[4776]: W1125 12:00:01.059472 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ebb4b80_22bd_440c_b70e_e533e9dc6b23.slice/crio-e57173553ae06aa72dca3d735a9503d45aedd73266c5f29f3f005d57fb55b5b1 WatchSource:0}: Error finding container e57173553ae06aa72dca3d735a9503d45aedd73266c5f29f3f005d57fb55b5b1: Status 404 returned error can't find the container with id e57173553ae06aa72dca3d735a9503d45aedd73266c5f29f3f005d57fb55b5b1 Nov 25 12:00:01 crc kubenswrapper[4776]: I1125 12:00:01.138170 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-f8dn9" podUID="769ca824-3fc6-47bc-8d13-8a8d0dfb8449" containerName="registry-server" probeResult="failure" output=< Nov 25 12:00:01 crc kubenswrapper[4776]: timeout: failed to connect service ":50051" within 1s Nov 25 12:00:01 crc kubenswrapper[4776]: > Nov 25 12:00:01 crc kubenswrapper[4776]: I1125 12:00:01.596269 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-cq5h9" event={"ID":"1ebb4b80-22bd-440c-b70e-e533e9dc6b23","Type":"ContainerStarted","Data":"e57173553ae06aa72dca3d735a9503d45aedd73266c5f29f3f005d57fb55b5b1"} Nov 25 12:00:02 crc kubenswrapper[4776]: I1125 12:00:02.608571 4776 generic.go:334] "Generic (PLEG): container finished" podID="1ebb4b80-22bd-440c-b70e-e533e9dc6b23" containerID="abf15b877079b487ef67413198fef457e8c6ab0f6f955728a5483a2cd0a999ea" exitCode=0 Nov 25 12:00:02 crc kubenswrapper[4776]: I1125 12:00:02.608624 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-cq5h9" event={"ID":"1ebb4b80-22bd-440c-b70e-e533e9dc6b23","Type":"ContainerDied","Data":"abf15b877079b487ef67413198fef457e8c6ab0f6f955728a5483a2cd0a999ea"} Nov 25 12:00:04 crc kubenswrapper[4776]: I1125 12:00:04.014328 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-cq5h9" Nov 25 12:00:04 crc kubenswrapper[4776]: I1125 12:00:04.165492 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1ebb4b80-22bd-440c-b70e-e533e9dc6b23-secret-volume\") pod \"1ebb4b80-22bd-440c-b70e-e533e9dc6b23\" (UID: \"1ebb4b80-22bd-440c-b70e-e533e9dc6b23\") " Nov 25 12:00:04 crc kubenswrapper[4776]: I1125 12:00:04.165775 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bz8b\" (UniqueName: \"kubernetes.io/projected/1ebb4b80-22bd-440c-b70e-e533e9dc6b23-kube-api-access-9bz8b\") pod \"1ebb4b80-22bd-440c-b70e-e533e9dc6b23\" (UID: \"1ebb4b80-22bd-440c-b70e-e533e9dc6b23\") " Nov 25 12:00:04 crc kubenswrapper[4776]: I1125 12:00:04.165921 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1ebb4b80-22bd-440c-b70e-e533e9dc6b23-config-volume\") pod \"1ebb4b80-22bd-440c-b70e-e533e9dc6b23\" (UID: \"1ebb4b80-22bd-440c-b70e-e533e9dc6b23\") " Nov 25 12:00:04 crc kubenswrapper[4776]: I1125 12:00:04.166863 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ebb4b80-22bd-440c-b70e-e533e9dc6b23-config-volume" (OuterVolumeSpecName: "config-volume") pod "1ebb4b80-22bd-440c-b70e-e533e9dc6b23" (UID: "1ebb4b80-22bd-440c-b70e-e533e9dc6b23"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:00:04 crc kubenswrapper[4776]: I1125 12:00:04.172122 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ebb4b80-22bd-440c-b70e-e533e9dc6b23-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1ebb4b80-22bd-440c-b70e-e533e9dc6b23" (UID: "1ebb4b80-22bd-440c-b70e-e533e9dc6b23"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:00:04 crc kubenswrapper[4776]: I1125 12:00:04.183291 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ebb4b80-22bd-440c-b70e-e533e9dc6b23-kube-api-access-9bz8b" (OuterVolumeSpecName: "kube-api-access-9bz8b") pod "1ebb4b80-22bd-440c-b70e-e533e9dc6b23" (UID: "1ebb4b80-22bd-440c-b70e-e533e9dc6b23"). InnerVolumeSpecName "kube-api-access-9bz8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:00:04 crc kubenswrapper[4776]: I1125 12:00:04.269222 4776 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1ebb4b80-22bd-440c-b70e-e533e9dc6b23-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 12:00:04 crc kubenswrapper[4776]: I1125 12:00:04.269264 4776 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1ebb4b80-22bd-440c-b70e-e533e9dc6b23-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 12:00:04 crc kubenswrapper[4776]: I1125 12:00:04.269274 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bz8b\" (UniqueName: \"kubernetes.io/projected/1ebb4b80-22bd-440c-b70e-e533e9dc6b23-kube-api-access-9bz8b\") on node \"crc\" DevicePath \"\"" Nov 25 12:00:04 crc kubenswrapper[4776]: I1125 12:00:04.468166 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-n7fjw" Nov 25 12:00:04 crc kubenswrapper[4776]: I1125 12:00:04.468232 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-n7fjw" Nov 25 12:00:04 crc kubenswrapper[4776]: I1125 12:00:04.641969 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-cq5h9" event={"ID":"1ebb4b80-22bd-440c-b70e-e533e9dc6b23","Type":"ContainerDied","Data":"e57173553ae06aa72dca3d735a9503d45aedd73266c5f29f3f005d57fb55b5b1"} Nov 25 12:00:04 crc kubenswrapper[4776]: I1125 12:00:04.642673 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e57173553ae06aa72dca3d735a9503d45aedd73266c5f29f3f005d57fb55b5b1" Nov 25 12:00:04 crc kubenswrapper[4776]: I1125 12:00:04.642041 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401200-cq5h9" Nov 25 12:00:05 crc kubenswrapper[4776]: I1125 12:00:05.104831 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401155-c8xlr"] Nov 25 12:00:05 crc kubenswrapper[4776]: I1125 12:00:05.115743 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401155-c8xlr"] Nov 25 12:00:05 crc kubenswrapper[4776]: I1125 12:00:05.522147 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-n7fjw" podUID="393868fc-3b2b-4aad-8621-530e5ba64268" containerName="registry-server" probeResult="failure" output=< Nov 25 12:00:05 crc kubenswrapper[4776]: timeout: failed to connect service ":50051" within 1s Nov 25 12:00:05 crc kubenswrapper[4776]: > Nov 25 12:00:05 crc kubenswrapper[4776]: I1125 12:00:05.688573 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd8e5bf7-c842-4966-94d7-f33b21d20be3" path="/var/lib/kubelet/pods/fd8e5bf7-c842-4966-94d7-f33b21d20be3/volumes" Nov 25 12:00:10 crc kubenswrapper[4776]: I1125 12:00:10.106732 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f8dn9" Nov 25 12:00:10 crc kubenswrapper[4776]: I1125 12:00:10.155383 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f8dn9" Nov 25 12:00:10 crc kubenswrapper[4776]: I1125 12:00:10.350228 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f8dn9"] Nov 25 12:00:11 crc kubenswrapper[4776]: I1125 12:00:11.723248 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-f8dn9" podUID="769ca824-3fc6-47bc-8d13-8a8d0dfb8449" containerName="registry-server" containerID="cri-o://bd03f5a84d03b505e2aafdcb33c64394a93a13e698ed6235b411d260427ca376" gracePeriod=2 Nov 25 12:00:12 crc kubenswrapper[4776]: I1125 12:00:12.734832 4776 generic.go:334] "Generic (PLEG): container finished" podID="769ca824-3fc6-47bc-8d13-8a8d0dfb8449" containerID="bd03f5a84d03b505e2aafdcb33c64394a93a13e698ed6235b411d260427ca376" exitCode=0 Nov 25 12:00:12 crc kubenswrapper[4776]: I1125 12:00:12.734952 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8dn9" event={"ID":"769ca824-3fc6-47bc-8d13-8a8d0dfb8449","Type":"ContainerDied","Data":"bd03f5a84d03b505e2aafdcb33c64394a93a13e698ed6235b411d260427ca376"} Nov 25 12:00:12 crc kubenswrapper[4776]: I1125 12:00:12.921041 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f8dn9" Nov 25 12:00:13 crc kubenswrapper[4776]: I1125 12:00:13.082400 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/769ca824-3fc6-47bc-8d13-8a8d0dfb8449-catalog-content\") pod \"769ca824-3fc6-47bc-8d13-8a8d0dfb8449\" (UID: \"769ca824-3fc6-47bc-8d13-8a8d0dfb8449\") " Nov 25 12:00:13 crc kubenswrapper[4776]: I1125 12:00:13.083018 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/769ca824-3fc6-47bc-8d13-8a8d0dfb8449-utilities\") pod \"769ca824-3fc6-47bc-8d13-8a8d0dfb8449\" (UID: \"769ca824-3fc6-47bc-8d13-8a8d0dfb8449\") " Nov 25 12:00:13 crc kubenswrapper[4776]: I1125 12:00:13.083159 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jwn9\" (UniqueName: \"kubernetes.io/projected/769ca824-3fc6-47bc-8d13-8a8d0dfb8449-kube-api-access-9jwn9\") pod \"769ca824-3fc6-47bc-8d13-8a8d0dfb8449\" (UID: \"769ca824-3fc6-47bc-8d13-8a8d0dfb8449\") " Nov 25 12:00:13 crc kubenswrapper[4776]: I1125 12:00:13.083876 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/769ca824-3fc6-47bc-8d13-8a8d0dfb8449-utilities" (OuterVolumeSpecName: "utilities") pod "769ca824-3fc6-47bc-8d13-8a8d0dfb8449" (UID: "769ca824-3fc6-47bc-8d13-8a8d0dfb8449"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:00:13 crc kubenswrapper[4776]: I1125 12:00:13.088311 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/769ca824-3fc6-47bc-8d13-8a8d0dfb8449-kube-api-access-9jwn9" (OuterVolumeSpecName: "kube-api-access-9jwn9") pod "769ca824-3fc6-47bc-8d13-8a8d0dfb8449" (UID: "769ca824-3fc6-47bc-8d13-8a8d0dfb8449"). InnerVolumeSpecName "kube-api-access-9jwn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:00:13 crc kubenswrapper[4776]: I1125 12:00:13.136399 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/769ca824-3fc6-47bc-8d13-8a8d0dfb8449-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "769ca824-3fc6-47bc-8d13-8a8d0dfb8449" (UID: "769ca824-3fc6-47bc-8d13-8a8d0dfb8449"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:00:13 crc kubenswrapper[4776]: I1125 12:00:13.185914 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/769ca824-3fc6-47bc-8d13-8a8d0dfb8449-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:00:13 crc kubenswrapper[4776]: I1125 12:00:13.185945 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/769ca824-3fc6-47bc-8d13-8a8d0dfb8449-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:00:13 crc kubenswrapper[4776]: I1125 12:00:13.185954 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jwn9\" (UniqueName: \"kubernetes.io/projected/769ca824-3fc6-47bc-8d13-8a8d0dfb8449-kube-api-access-9jwn9\") on node \"crc\" DevicePath \"\"" Nov 25 12:00:13 crc kubenswrapper[4776]: I1125 12:00:13.751144 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f8dn9" event={"ID":"769ca824-3fc6-47bc-8d13-8a8d0dfb8449","Type":"ContainerDied","Data":"15e14170722ebc1747325bc6329a2ddc347c2e3ccaec38aa3ec6651ecf2ba5fa"} Nov 25 12:00:13 crc kubenswrapper[4776]: I1125 12:00:13.751211 4776 scope.go:117] "RemoveContainer" containerID="bd03f5a84d03b505e2aafdcb33c64394a93a13e698ed6235b411d260427ca376" Nov 25 12:00:13 crc kubenswrapper[4776]: I1125 12:00:13.751245 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f8dn9" Nov 25 12:00:13 crc kubenswrapper[4776]: I1125 12:00:13.786770 4776 scope.go:117] "RemoveContainer" containerID="fce685d4154bee9c802d787f03fd0271a651ae4caf82c122db83c3485a569724" Nov 25 12:00:13 crc kubenswrapper[4776]: I1125 12:00:13.797095 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f8dn9"] Nov 25 12:00:13 crc kubenswrapper[4776]: I1125 12:00:13.811140 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-f8dn9"] Nov 25 12:00:13 crc kubenswrapper[4776]: I1125 12:00:13.814968 4776 scope.go:117] "RemoveContainer" containerID="1de9d50a05b9e641da13f7c4a37a3782e69837425dc3031b3ec35e6b7c092127" Nov 25 12:00:15 crc kubenswrapper[4776]: I1125 12:00:15.515878 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-n7fjw" podUID="393868fc-3b2b-4aad-8621-530e5ba64268" containerName="registry-server" probeResult="failure" output=< Nov 25 12:00:15 crc kubenswrapper[4776]: timeout: failed to connect service ":50051" within 1s Nov 25 12:00:15 crc kubenswrapper[4776]: > Nov 25 12:00:15 crc kubenswrapper[4776]: I1125 12:00:15.674296 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="769ca824-3fc6-47bc-8d13-8a8d0dfb8449" path="/var/lib/kubelet/pods/769ca824-3fc6-47bc-8d13-8a8d0dfb8449/volumes" Nov 25 12:00:17 crc kubenswrapper[4776]: I1125 12:00:17.818454 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:00:17 crc kubenswrapper[4776]: I1125 12:00:17.818722 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:00:24 crc kubenswrapper[4776]: I1125 12:00:24.532568 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-n7fjw" Nov 25 12:00:24 crc kubenswrapper[4776]: I1125 12:00:24.588459 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-n7fjw" Nov 25 12:00:24 crc kubenswrapper[4776]: I1125 12:00:24.777628 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n7fjw"] Nov 25 12:00:25 crc kubenswrapper[4776]: I1125 12:00:25.882465 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-n7fjw" podUID="393868fc-3b2b-4aad-8621-530e5ba64268" containerName="registry-server" containerID="cri-o://d45573d6e1f2ed34b1a535d0b040271fbc6280f71131740685efe02f38a49cb8" gracePeriod=2 Nov 25 12:00:26 crc kubenswrapper[4776]: I1125 12:00:26.897440 4776 generic.go:334] "Generic (PLEG): container finished" podID="393868fc-3b2b-4aad-8621-530e5ba64268" containerID="d45573d6e1f2ed34b1a535d0b040271fbc6280f71131740685efe02f38a49cb8" exitCode=0 Nov 25 12:00:26 crc kubenswrapper[4776]: I1125 12:00:26.897526 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n7fjw" event={"ID":"393868fc-3b2b-4aad-8621-530e5ba64268","Type":"ContainerDied","Data":"d45573d6e1f2ed34b1a535d0b040271fbc6280f71131740685efe02f38a49cb8"} Nov 25 12:00:26 crc kubenswrapper[4776]: I1125 12:00:26.897778 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n7fjw" event={"ID":"393868fc-3b2b-4aad-8621-530e5ba64268","Type":"ContainerDied","Data":"4f8054b2bbb3383029d316d773488fbc9811b9bdd70cc5924200bf035712bf80"} Nov 25 12:00:26 crc kubenswrapper[4776]: I1125 12:00:26.897795 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f8054b2bbb3383029d316d773488fbc9811b9bdd70cc5924200bf035712bf80" Nov 25 12:00:26 crc kubenswrapper[4776]: I1125 12:00:26.921882 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n7fjw" Nov 25 12:00:27 crc kubenswrapper[4776]: I1125 12:00:27.019133 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/393868fc-3b2b-4aad-8621-530e5ba64268-catalog-content\") pod \"393868fc-3b2b-4aad-8621-530e5ba64268\" (UID: \"393868fc-3b2b-4aad-8621-530e5ba64268\") " Nov 25 12:00:27 crc kubenswrapper[4776]: I1125 12:00:27.019203 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/393868fc-3b2b-4aad-8621-530e5ba64268-utilities\") pod \"393868fc-3b2b-4aad-8621-530e5ba64268\" (UID: \"393868fc-3b2b-4aad-8621-530e5ba64268\") " Nov 25 12:00:27 crc kubenswrapper[4776]: I1125 12:00:27.019557 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wg5t9\" (UniqueName: \"kubernetes.io/projected/393868fc-3b2b-4aad-8621-530e5ba64268-kube-api-access-wg5t9\") pod \"393868fc-3b2b-4aad-8621-530e5ba64268\" (UID: \"393868fc-3b2b-4aad-8621-530e5ba64268\") " Nov 25 12:00:27 crc kubenswrapper[4776]: I1125 12:00:27.020245 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/393868fc-3b2b-4aad-8621-530e5ba64268-utilities" (OuterVolumeSpecName: "utilities") pod "393868fc-3b2b-4aad-8621-530e5ba64268" (UID: "393868fc-3b2b-4aad-8621-530e5ba64268"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:00:27 crc kubenswrapper[4776]: I1125 12:00:27.027081 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/393868fc-3b2b-4aad-8621-530e5ba64268-kube-api-access-wg5t9" (OuterVolumeSpecName: "kube-api-access-wg5t9") pod "393868fc-3b2b-4aad-8621-530e5ba64268" (UID: "393868fc-3b2b-4aad-8621-530e5ba64268"). InnerVolumeSpecName "kube-api-access-wg5t9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:00:27 crc kubenswrapper[4776]: I1125 12:00:27.117089 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/393868fc-3b2b-4aad-8621-530e5ba64268-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "393868fc-3b2b-4aad-8621-530e5ba64268" (UID: "393868fc-3b2b-4aad-8621-530e5ba64268"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:00:27 crc kubenswrapper[4776]: I1125 12:00:27.123377 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/393868fc-3b2b-4aad-8621-530e5ba64268-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:00:27 crc kubenswrapper[4776]: I1125 12:00:27.123431 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/393868fc-3b2b-4aad-8621-530e5ba64268-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:00:27 crc kubenswrapper[4776]: I1125 12:00:27.123443 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wg5t9\" (UniqueName: \"kubernetes.io/projected/393868fc-3b2b-4aad-8621-530e5ba64268-kube-api-access-wg5t9\") on node \"crc\" DevicePath \"\"" Nov 25 12:00:27 crc kubenswrapper[4776]: I1125 12:00:27.913259 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n7fjw" Nov 25 12:00:27 crc kubenswrapper[4776]: I1125 12:00:27.961300 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n7fjw"] Nov 25 12:00:27 crc kubenswrapper[4776]: I1125 12:00:27.975629 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-n7fjw"] Nov 25 12:00:29 crc kubenswrapper[4776]: I1125 12:00:29.676391 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="393868fc-3b2b-4aad-8621-530e5ba64268" path="/var/lib/kubelet/pods/393868fc-3b2b-4aad-8621-530e5ba64268/volumes" Nov 25 12:00:42 crc kubenswrapper[4776]: I1125 12:00:42.637922 4776 scope.go:117] "RemoveContainer" containerID="6d8f1800b7959498a8813e39a75ec0bdcb10bc2cab2d105342f6a5f186cf4107" Nov 25 12:00:47 crc kubenswrapper[4776]: I1125 12:00:47.818777 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:00:47 crc kubenswrapper[4776]: I1125 12:00:47.819765 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:00:47 crc kubenswrapper[4776]: I1125 12:00:47.819851 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 12:00:47 crc kubenswrapper[4776]: I1125 12:00:47.821336 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8e060fbd41a6ceeaf0a55e5585d31b0083930b3310803a6e50fc5823301c1b89"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 12:00:47 crc kubenswrapper[4776]: I1125 12:00:47.821441 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://8e060fbd41a6ceeaf0a55e5585d31b0083930b3310803a6e50fc5823301c1b89" gracePeriod=600 Nov 25 12:00:48 crc kubenswrapper[4776]: I1125 12:00:48.088296 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dz69w"] Nov 25 12:00:48 crc kubenswrapper[4776]: E1125 12:00:48.089460 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="393868fc-3b2b-4aad-8621-530e5ba64268" containerName="extract-utilities" Nov 25 12:00:48 crc kubenswrapper[4776]: I1125 12:00:48.089487 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="393868fc-3b2b-4aad-8621-530e5ba64268" containerName="extract-utilities" Nov 25 12:00:48 crc kubenswrapper[4776]: E1125 12:00:48.089539 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="393868fc-3b2b-4aad-8621-530e5ba64268" containerName="registry-server" Nov 25 12:00:48 crc kubenswrapper[4776]: I1125 12:00:48.089549 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="393868fc-3b2b-4aad-8621-530e5ba64268" containerName="registry-server" Nov 25 12:00:48 crc kubenswrapper[4776]: E1125 12:00:48.089570 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ebb4b80-22bd-440c-b70e-e533e9dc6b23" containerName="collect-profiles" Nov 25 12:00:48 crc kubenswrapper[4776]: I1125 12:00:48.089578 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ebb4b80-22bd-440c-b70e-e533e9dc6b23" containerName="collect-profiles" Nov 25 12:00:48 crc kubenswrapper[4776]: E1125 12:00:48.089604 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="769ca824-3fc6-47bc-8d13-8a8d0dfb8449" containerName="extract-content" Nov 25 12:00:48 crc kubenswrapper[4776]: I1125 12:00:48.089612 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="769ca824-3fc6-47bc-8d13-8a8d0dfb8449" containerName="extract-content" Nov 25 12:00:48 crc kubenswrapper[4776]: E1125 12:00:48.089630 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="769ca824-3fc6-47bc-8d13-8a8d0dfb8449" containerName="extract-utilities" Nov 25 12:00:48 crc kubenswrapper[4776]: I1125 12:00:48.089638 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="769ca824-3fc6-47bc-8d13-8a8d0dfb8449" containerName="extract-utilities" Nov 25 12:00:48 crc kubenswrapper[4776]: E1125 12:00:48.089659 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="769ca824-3fc6-47bc-8d13-8a8d0dfb8449" containerName="registry-server" Nov 25 12:00:48 crc kubenswrapper[4776]: I1125 12:00:48.089666 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="769ca824-3fc6-47bc-8d13-8a8d0dfb8449" containerName="registry-server" Nov 25 12:00:48 crc kubenswrapper[4776]: E1125 12:00:48.089690 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="393868fc-3b2b-4aad-8621-530e5ba64268" containerName="extract-content" Nov 25 12:00:48 crc kubenswrapper[4776]: I1125 12:00:48.089700 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="393868fc-3b2b-4aad-8621-530e5ba64268" containerName="extract-content" Nov 25 12:00:48 crc kubenswrapper[4776]: I1125 12:00:48.091388 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="769ca824-3fc6-47bc-8d13-8a8d0dfb8449" containerName="registry-server" Nov 25 12:00:48 crc kubenswrapper[4776]: I1125 12:00:48.091439 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="393868fc-3b2b-4aad-8621-530e5ba64268" containerName="registry-server" Nov 25 12:00:48 crc kubenswrapper[4776]: I1125 12:00:48.091465 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ebb4b80-22bd-440c-b70e-e533e9dc6b23" containerName="collect-profiles" Nov 25 12:00:48 crc kubenswrapper[4776]: I1125 12:00:48.093882 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dz69w" Nov 25 12:00:48 crc kubenswrapper[4776]: I1125 12:00:48.098823 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dz69w"] Nov 25 12:00:48 crc kubenswrapper[4776]: I1125 12:00:48.145144 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="8e060fbd41a6ceeaf0a55e5585d31b0083930b3310803a6e50fc5823301c1b89" exitCode=0 Nov 25 12:00:48 crc kubenswrapper[4776]: I1125 12:00:48.145209 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"8e060fbd41a6ceeaf0a55e5585d31b0083930b3310803a6e50fc5823301c1b89"} Nov 25 12:00:48 crc kubenswrapper[4776]: I1125 12:00:48.145245 4776 scope.go:117] "RemoveContainer" containerID="63c4f1439259b1e6eebffc31f746dd85e12cc4c8255311ded06696c418c17965" Nov 25 12:00:48 crc kubenswrapper[4776]: I1125 12:00:48.170891 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f2878a8-b88d-4a1f-9e46-3b31e155f1ea-catalog-content\") pod \"certified-operators-dz69w\" (UID: \"1f2878a8-b88d-4a1f-9e46-3b31e155f1ea\") " pod="openshift-marketplace/certified-operators-dz69w" Nov 25 12:00:48 crc kubenswrapper[4776]: I1125 12:00:48.171052 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f2878a8-b88d-4a1f-9e46-3b31e155f1ea-utilities\") pod \"certified-operators-dz69w\" (UID: \"1f2878a8-b88d-4a1f-9e46-3b31e155f1ea\") " pod="openshift-marketplace/certified-operators-dz69w" Nov 25 12:00:48 crc kubenswrapper[4776]: I1125 12:00:48.171344 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slptp\" (UniqueName: \"kubernetes.io/projected/1f2878a8-b88d-4a1f-9e46-3b31e155f1ea-kube-api-access-slptp\") pod \"certified-operators-dz69w\" (UID: \"1f2878a8-b88d-4a1f-9e46-3b31e155f1ea\") " pod="openshift-marketplace/certified-operators-dz69w" Nov 25 12:00:48 crc kubenswrapper[4776]: I1125 12:00:48.273401 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slptp\" (UniqueName: \"kubernetes.io/projected/1f2878a8-b88d-4a1f-9e46-3b31e155f1ea-kube-api-access-slptp\") pod \"certified-operators-dz69w\" (UID: \"1f2878a8-b88d-4a1f-9e46-3b31e155f1ea\") " pod="openshift-marketplace/certified-operators-dz69w" Nov 25 12:00:48 crc kubenswrapper[4776]: I1125 12:00:48.273576 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f2878a8-b88d-4a1f-9e46-3b31e155f1ea-catalog-content\") pod \"certified-operators-dz69w\" (UID: \"1f2878a8-b88d-4a1f-9e46-3b31e155f1ea\") " pod="openshift-marketplace/certified-operators-dz69w" Nov 25 12:00:48 crc kubenswrapper[4776]: I1125 12:00:48.273649 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f2878a8-b88d-4a1f-9e46-3b31e155f1ea-utilities\") pod \"certified-operators-dz69w\" (UID: \"1f2878a8-b88d-4a1f-9e46-3b31e155f1ea\") " pod="openshift-marketplace/certified-operators-dz69w" Nov 25 12:00:48 crc kubenswrapper[4776]: I1125 12:00:48.274059 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f2878a8-b88d-4a1f-9e46-3b31e155f1ea-catalog-content\") pod \"certified-operators-dz69w\" (UID: \"1f2878a8-b88d-4a1f-9e46-3b31e155f1ea\") " pod="openshift-marketplace/certified-operators-dz69w" Nov 25 12:00:48 crc kubenswrapper[4776]: I1125 12:00:48.274149 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f2878a8-b88d-4a1f-9e46-3b31e155f1ea-utilities\") pod \"certified-operators-dz69w\" (UID: \"1f2878a8-b88d-4a1f-9e46-3b31e155f1ea\") " pod="openshift-marketplace/certified-operators-dz69w" Nov 25 12:00:48 crc kubenswrapper[4776]: I1125 12:00:48.657767 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slptp\" (UniqueName: \"kubernetes.io/projected/1f2878a8-b88d-4a1f-9e46-3b31e155f1ea-kube-api-access-slptp\") pod \"certified-operators-dz69w\" (UID: \"1f2878a8-b88d-4a1f-9e46-3b31e155f1ea\") " pod="openshift-marketplace/certified-operators-dz69w" Nov 25 12:00:48 crc kubenswrapper[4776]: I1125 12:00:48.724672 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dz69w" Nov 25 12:00:49 crc kubenswrapper[4776]: I1125 12:00:49.157992 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d"} Nov 25 12:00:49 crc kubenswrapper[4776]: I1125 12:00:49.248178 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dz69w"] Nov 25 12:00:50 crc kubenswrapper[4776]: I1125 12:00:50.172561 4776 generic.go:334] "Generic (PLEG): container finished" podID="1f2878a8-b88d-4a1f-9e46-3b31e155f1ea" containerID="2a94fc7cc414382e9100fb6fbee4de58ddda5243112676d64b855d752e80920d" exitCode=0 Nov 25 12:00:50 crc kubenswrapper[4776]: I1125 12:00:50.172849 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dz69w" event={"ID":"1f2878a8-b88d-4a1f-9e46-3b31e155f1ea","Type":"ContainerDied","Data":"2a94fc7cc414382e9100fb6fbee4de58ddda5243112676d64b855d752e80920d"} Nov 25 12:00:50 crc kubenswrapper[4776]: I1125 12:00:50.173510 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dz69w" event={"ID":"1f2878a8-b88d-4a1f-9e46-3b31e155f1ea","Type":"ContainerStarted","Data":"ac7f43a5b506e902d321603478cbf10167c886f904ff2358d2bdf32790acb368"} Nov 25 12:00:53 crc kubenswrapper[4776]: I1125 12:00:53.206667 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dz69w" event={"ID":"1f2878a8-b88d-4a1f-9e46-3b31e155f1ea","Type":"ContainerStarted","Data":"0481e30b215d7782e3337c8d235b578d47cb9aaf0203ac7bea341b4df50b5cec"} Nov 25 12:00:54 crc kubenswrapper[4776]: I1125 12:00:54.218282 4776 generic.go:334] "Generic (PLEG): container finished" podID="1f2878a8-b88d-4a1f-9e46-3b31e155f1ea" containerID="0481e30b215d7782e3337c8d235b578d47cb9aaf0203ac7bea341b4df50b5cec" exitCode=0 Nov 25 12:00:54 crc kubenswrapper[4776]: I1125 12:00:54.218396 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dz69w" event={"ID":"1f2878a8-b88d-4a1f-9e46-3b31e155f1ea","Type":"ContainerDied","Data":"0481e30b215d7782e3337c8d235b578d47cb9aaf0203ac7bea341b4df50b5cec"} Nov 25 12:00:56 crc kubenswrapper[4776]: I1125 12:00:56.267480 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dz69w" event={"ID":"1f2878a8-b88d-4a1f-9e46-3b31e155f1ea","Type":"ContainerStarted","Data":"38f095361ebf50809d8e0b387bcf639e27a08e1db3b4e10f042c24370449d87a"} Nov 25 12:00:56 crc kubenswrapper[4776]: I1125 12:00:56.301313 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dz69w" podStartSLOduration=3.5207372120000002 podStartE2EDuration="8.301293739s" podCreationTimestamp="2025-11-25 12:00:48 +0000 UTC" firstStartedPulling="2025-11-25 12:00:50.175860768 +0000 UTC m=+9395.216920331" lastFinishedPulling="2025-11-25 12:00:54.956417305 +0000 UTC m=+9399.997476858" observedRunningTime="2025-11-25 12:00:56.288621741 +0000 UTC m=+9401.329681304" watchObservedRunningTime="2025-11-25 12:00:56.301293739 +0000 UTC m=+9401.342353292" Nov 25 12:00:58 crc kubenswrapper[4776]: I1125 12:00:58.725273 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dz69w" Nov 25 12:00:58 crc kubenswrapper[4776]: I1125 12:00:58.725846 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dz69w" Nov 25 12:00:59 crc kubenswrapper[4776]: I1125 12:00:59.022058 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dz69w" Nov 25 12:01:00 crc kubenswrapper[4776]: I1125 12:01:00.156234 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29401201-8vgn6"] Nov 25 12:01:00 crc kubenswrapper[4776]: I1125 12:01:00.158312 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29401201-8vgn6" Nov 25 12:01:00 crc kubenswrapper[4776]: I1125 12:01:00.178172 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29401201-8vgn6"] Nov 25 12:01:00 crc kubenswrapper[4776]: I1125 12:01:00.283633 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zk9wf\" (UniqueName: \"kubernetes.io/projected/841c9ff3-7f79-424b-9da0-4bca60584e2a-kube-api-access-zk9wf\") pod \"keystone-cron-29401201-8vgn6\" (UID: \"841c9ff3-7f79-424b-9da0-4bca60584e2a\") " pod="openstack/keystone-cron-29401201-8vgn6" Nov 25 12:01:00 crc kubenswrapper[4776]: I1125 12:01:00.283701 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/841c9ff3-7f79-424b-9da0-4bca60584e2a-fernet-keys\") pod \"keystone-cron-29401201-8vgn6\" (UID: \"841c9ff3-7f79-424b-9da0-4bca60584e2a\") " pod="openstack/keystone-cron-29401201-8vgn6" Nov 25 12:01:00 crc kubenswrapper[4776]: I1125 12:01:00.283792 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/841c9ff3-7f79-424b-9da0-4bca60584e2a-combined-ca-bundle\") pod \"keystone-cron-29401201-8vgn6\" (UID: \"841c9ff3-7f79-424b-9da0-4bca60584e2a\") " pod="openstack/keystone-cron-29401201-8vgn6" Nov 25 12:01:00 crc kubenswrapper[4776]: I1125 12:01:00.283921 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/841c9ff3-7f79-424b-9da0-4bca60584e2a-config-data\") pod \"keystone-cron-29401201-8vgn6\" (UID: \"841c9ff3-7f79-424b-9da0-4bca60584e2a\") " pod="openstack/keystone-cron-29401201-8vgn6" Nov 25 12:01:00 crc kubenswrapper[4776]: I1125 12:01:00.386256 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/841c9ff3-7f79-424b-9da0-4bca60584e2a-combined-ca-bundle\") pod \"keystone-cron-29401201-8vgn6\" (UID: \"841c9ff3-7f79-424b-9da0-4bca60584e2a\") " pod="openstack/keystone-cron-29401201-8vgn6" Nov 25 12:01:00 crc kubenswrapper[4776]: I1125 12:01:00.386403 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/841c9ff3-7f79-424b-9da0-4bca60584e2a-config-data\") pod \"keystone-cron-29401201-8vgn6\" (UID: \"841c9ff3-7f79-424b-9da0-4bca60584e2a\") " pod="openstack/keystone-cron-29401201-8vgn6" Nov 25 12:01:00 crc kubenswrapper[4776]: I1125 12:01:00.386485 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zk9wf\" (UniqueName: \"kubernetes.io/projected/841c9ff3-7f79-424b-9da0-4bca60584e2a-kube-api-access-zk9wf\") pod \"keystone-cron-29401201-8vgn6\" (UID: \"841c9ff3-7f79-424b-9da0-4bca60584e2a\") " pod="openstack/keystone-cron-29401201-8vgn6" Nov 25 12:01:00 crc kubenswrapper[4776]: I1125 12:01:00.386513 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/841c9ff3-7f79-424b-9da0-4bca60584e2a-fernet-keys\") pod \"keystone-cron-29401201-8vgn6\" (UID: \"841c9ff3-7f79-424b-9da0-4bca60584e2a\") " pod="openstack/keystone-cron-29401201-8vgn6" Nov 25 12:01:00 crc kubenswrapper[4776]: I1125 12:01:00.392561 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/841c9ff3-7f79-424b-9da0-4bca60584e2a-fernet-keys\") pod \"keystone-cron-29401201-8vgn6\" (UID: \"841c9ff3-7f79-424b-9da0-4bca60584e2a\") " pod="openstack/keystone-cron-29401201-8vgn6" Nov 25 12:01:00 crc kubenswrapper[4776]: I1125 12:01:00.392689 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/841c9ff3-7f79-424b-9da0-4bca60584e2a-combined-ca-bundle\") pod \"keystone-cron-29401201-8vgn6\" (UID: \"841c9ff3-7f79-424b-9da0-4bca60584e2a\") " pod="openstack/keystone-cron-29401201-8vgn6" Nov 25 12:01:00 crc kubenswrapper[4776]: I1125 12:01:00.395882 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/841c9ff3-7f79-424b-9da0-4bca60584e2a-config-data\") pod \"keystone-cron-29401201-8vgn6\" (UID: \"841c9ff3-7f79-424b-9da0-4bca60584e2a\") " pod="openstack/keystone-cron-29401201-8vgn6" Nov 25 12:01:00 crc kubenswrapper[4776]: I1125 12:01:00.401980 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zk9wf\" (UniqueName: \"kubernetes.io/projected/841c9ff3-7f79-424b-9da0-4bca60584e2a-kube-api-access-zk9wf\") pod \"keystone-cron-29401201-8vgn6\" (UID: \"841c9ff3-7f79-424b-9da0-4bca60584e2a\") " pod="openstack/keystone-cron-29401201-8vgn6" Nov 25 12:01:00 crc kubenswrapper[4776]: I1125 12:01:00.488209 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29401201-8vgn6" Nov 25 12:01:00 crc kubenswrapper[4776]: I1125 12:01:00.986743 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29401201-8vgn6"] Nov 25 12:01:02 crc kubenswrapper[4776]: I1125 12:01:02.331151 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29401201-8vgn6" event={"ID":"841c9ff3-7f79-424b-9da0-4bca60584e2a","Type":"ContainerStarted","Data":"9b02203e4d5878fb0710febd68cedf9cecb35a379785f9fd1ca73f989fb9e434"} Nov 25 12:01:02 crc kubenswrapper[4776]: I1125 12:01:02.331532 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29401201-8vgn6" event={"ID":"841c9ff3-7f79-424b-9da0-4bca60584e2a","Type":"ContainerStarted","Data":"f169501ad03a0b247ce087656bec7a8e22aed4a118a3683cc76f5f6f19850e77"} Nov 25 12:01:02 crc kubenswrapper[4776]: I1125 12:01:02.355052 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29401201-8vgn6" podStartSLOduration=2.35503089 podStartE2EDuration="2.35503089s" podCreationTimestamp="2025-11-25 12:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:01:02.347975443 +0000 UTC m=+9407.389034996" watchObservedRunningTime="2025-11-25 12:01:02.35503089 +0000 UTC m=+9407.396090443" Nov 25 12:01:06 crc kubenswrapper[4776]: I1125 12:01:06.381037 4776 generic.go:334] "Generic (PLEG): container finished" podID="841c9ff3-7f79-424b-9da0-4bca60584e2a" containerID="9b02203e4d5878fb0710febd68cedf9cecb35a379785f9fd1ca73f989fb9e434" exitCode=0 Nov 25 12:01:06 crc kubenswrapper[4776]: I1125 12:01:06.381111 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29401201-8vgn6" event={"ID":"841c9ff3-7f79-424b-9da0-4bca60584e2a","Type":"ContainerDied","Data":"9b02203e4d5878fb0710febd68cedf9cecb35a379785f9fd1ca73f989fb9e434"} Nov 25 12:01:07 crc kubenswrapper[4776]: I1125 12:01:07.810678 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29401201-8vgn6" Nov 25 12:01:07 crc kubenswrapper[4776]: I1125 12:01:07.980222 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/841c9ff3-7f79-424b-9da0-4bca60584e2a-fernet-keys\") pod \"841c9ff3-7f79-424b-9da0-4bca60584e2a\" (UID: \"841c9ff3-7f79-424b-9da0-4bca60584e2a\") " Nov 25 12:01:07 crc kubenswrapper[4776]: I1125 12:01:07.980311 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/841c9ff3-7f79-424b-9da0-4bca60584e2a-config-data\") pod \"841c9ff3-7f79-424b-9da0-4bca60584e2a\" (UID: \"841c9ff3-7f79-424b-9da0-4bca60584e2a\") " Nov 25 12:01:07 crc kubenswrapper[4776]: I1125 12:01:07.980472 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/841c9ff3-7f79-424b-9da0-4bca60584e2a-combined-ca-bundle\") pod \"841c9ff3-7f79-424b-9da0-4bca60584e2a\" (UID: \"841c9ff3-7f79-424b-9da0-4bca60584e2a\") " Nov 25 12:01:07 crc kubenswrapper[4776]: I1125 12:01:07.980530 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zk9wf\" (UniqueName: \"kubernetes.io/projected/841c9ff3-7f79-424b-9da0-4bca60584e2a-kube-api-access-zk9wf\") pod \"841c9ff3-7f79-424b-9da0-4bca60584e2a\" (UID: \"841c9ff3-7f79-424b-9da0-4bca60584e2a\") " Nov 25 12:01:07 crc kubenswrapper[4776]: I1125 12:01:07.986561 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/841c9ff3-7f79-424b-9da0-4bca60584e2a-kube-api-access-zk9wf" (OuterVolumeSpecName: "kube-api-access-zk9wf") pod "841c9ff3-7f79-424b-9da0-4bca60584e2a" (UID: "841c9ff3-7f79-424b-9da0-4bca60584e2a"). InnerVolumeSpecName "kube-api-access-zk9wf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:01:07 crc kubenswrapper[4776]: I1125 12:01:07.995972 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/841c9ff3-7f79-424b-9da0-4bca60584e2a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "841c9ff3-7f79-424b-9da0-4bca60584e2a" (UID: "841c9ff3-7f79-424b-9da0-4bca60584e2a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:01:08 crc kubenswrapper[4776]: I1125 12:01:08.021277 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/841c9ff3-7f79-424b-9da0-4bca60584e2a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "841c9ff3-7f79-424b-9da0-4bca60584e2a" (UID: "841c9ff3-7f79-424b-9da0-4bca60584e2a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:01:08 crc kubenswrapper[4776]: I1125 12:01:08.041933 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/841c9ff3-7f79-424b-9da0-4bca60584e2a-config-data" (OuterVolumeSpecName: "config-data") pod "841c9ff3-7f79-424b-9da0-4bca60584e2a" (UID: "841c9ff3-7f79-424b-9da0-4bca60584e2a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:01:08 crc kubenswrapper[4776]: I1125 12:01:08.084662 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zk9wf\" (UniqueName: \"kubernetes.io/projected/841c9ff3-7f79-424b-9da0-4bca60584e2a-kube-api-access-zk9wf\") on node \"crc\" DevicePath \"\"" Nov 25 12:01:08 crc kubenswrapper[4776]: I1125 12:01:08.084696 4776 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/841c9ff3-7f79-424b-9da0-4bca60584e2a-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 12:01:08 crc kubenswrapper[4776]: I1125 12:01:08.084708 4776 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/841c9ff3-7f79-424b-9da0-4bca60584e2a-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 12:01:08 crc kubenswrapper[4776]: I1125 12:01:08.084715 4776 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/841c9ff3-7f79-424b-9da0-4bca60584e2a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 12:01:08 crc kubenswrapper[4776]: I1125 12:01:08.403405 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29401201-8vgn6" event={"ID":"841c9ff3-7f79-424b-9da0-4bca60584e2a","Type":"ContainerDied","Data":"f169501ad03a0b247ce087656bec7a8e22aed4a118a3683cc76f5f6f19850e77"} Nov 25 12:01:08 crc kubenswrapper[4776]: I1125 12:01:08.403447 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f169501ad03a0b247ce087656bec7a8e22aed4a118a3683cc76f5f6f19850e77" Nov 25 12:01:08 crc kubenswrapper[4776]: I1125 12:01:08.403497 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29401201-8vgn6" Nov 25 12:01:08 crc kubenswrapper[4776]: I1125 12:01:08.769445 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dz69w" Nov 25 12:01:08 crc kubenswrapper[4776]: I1125 12:01:08.825649 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dz69w"] Nov 25 12:01:09 crc kubenswrapper[4776]: I1125 12:01:09.412887 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dz69w" podUID="1f2878a8-b88d-4a1f-9e46-3b31e155f1ea" containerName="registry-server" containerID="cri-o://38f095361ebf50809d8e0b387bcf639e27a08e1db3b4e10f042c24370449d87a" gracePeriod=2 Nov 25 12:01:10 crc kubenswrapper[4776]: I1125 12:01:10.426005 4776 generic.go:334] "Generic (PLEG): container finished" podID="1f2878a8-b88d-4a1f-9e46-3b31e155f1ea" containerID="38f095361ebf50809d8e0b387bcf639e27a08e1db3b4e10f042c24370449d87a" exitCode=0 Nov 25 12:01:10 crc kubenswrapper[4776]: I1125 12:01:10.426081 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dz69w" event={"ID":"1f2878a8-b88d-4a1f-9e46-3b31e155f1ea","Type":"ContainerDied","Data":"38f095361ebf50809d8e0b387bcf639e27a08e1db3b4e10f042c24370449d87a"} Nov 25 12:01:10 crc kubenswrapper[4776]: I1125 12:01:10.702267 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dz69w" Nov 25 12:01:10 crc kubenswrapper[4776]: I1125 12:01:10.848378 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f2878a8-b88d-4a1f-9e46-3b31e155f1ea-catalog-content\") pod \"1f2878a8-b88d-4a1f-9e46-3b31e155f1ea\" (UID: \"1f2878a8-b88d-4a1f-9e46-3b31e155f1ea\") " Nov 25 12:01:10 crc kubenswrapper[4776]: I1125 12:01:10.848475 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slptp\" (UniqueName: \"kubernetes.io/projected/1f2878a8-b88d-4a1f-9e46-3b31e155f1ea-kube-api-access-slptp\") pod \"1f2878a8-b88d-4a1f-9e46-3b31e155f1ea\" (UID: \"1f2878a8-b88d-4a1f-9e46-3b31e155f1ea\") " Nov 25 12:01:10 crc kubenswrapper[4776]: I1125 12:01:10.848701 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f2878a8-b88d-4a1f-9e46-3b31e155f1ea-utilities\") pod \"1f2878a8-b88d-4a1f-9e46-3b31e155f1ea\" (UID: \"1f2878a8-b88d-4a1f-9e46-3b31e155f1ea\") " Nov 25 12:01:10 crc kubenswrapper[4776]: I1125 12:01:10.849524 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f2878a8-b88d-4a1f-9e46-3b31e155f1ea-utilities" (OuterVolumeSpecName: "utilities") pod "1f2878a8-b88d-4a1f-9e46-3b31e155f1ea" (UID: "1f2878a8-b88d-4a1f-9e46-3b31e155f1ea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:01:10 crc kubenswrapper[4776]: I1125 12:01:10.855903 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f2878a8-b88d-4a1f-9e46-3b31e155f1ea-kube-api-access-slptp" (OuterVolumeSpecName: "kube-api-access-slptp") pod "1f2878a8-b88d-4a1f-9e46-3b31e155f1ea" (UID: "1f2878a8-b88d-4a1f-9e46-3b31e155f1ea"). InnerVolumeSpecName "kube-api-access-slptp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:01:10 crc kubenswrapper[4776]: I1125 12:01:10.895937 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f2878a8-b88d-4a1f-9e46-3b31e155f1ea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1f2878a8-b88d-4a1f-9e46-3b31e155f1ea" (UID: "1f2878a8-b88d-4a1f-9e46-3b31e155f1ea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:01:10 crc kubenswrapper[4776]: I1125 12:01:10.952292 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slptp\" (UniqueName: \"kubernetes.io/projected/1f2878a8-b88d-4a1f-9e46-3b31e155f1ea-kube-api-access-slptp\") on node \"crc\" DevicePath \"\"" Nov 25 12:01:10 crc kubenswrapper[4776]: I1125 12:01:10.952340 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f2878a8-b88d-4a1f-9e46-3b31e155f1ea-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:01:10 crc kubenswrapper[4776]: I1125 12:01:10.952350 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f2878a8-b88d-4a1f-9e46-3b31e155f1ea-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:01:11 crc kubenswrapper[4776]: I1125 12:01:11.446204 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dz69w" event={"ID":"1f2878a8-b88d-4a1f-9e46-3b31e155f1ea","Type":"ContainerDied","Data":"ac7f43a5b506e902d321603478cbf10167c886f904ff2358d2bdf32790acb368"} Nov 25 12:01:11 crc kubenswrapper[4776]: I1125 12:01:11.446277 4776 scope.go:117] "RemoveContainer" containerID="38f095361ebf50809d8e0b387bcf639e27a08e1db3b4e10f042c24370449d87a" Nov 25 12:01:11 crc kubenswrapper[4776]: I1125 12:01:11.446319 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dz69w" Nov 25 12:01:11 crc kubenswrapper[4776]: I1125 12:01:11.471415 4776 scope.go:117] "RemoveContainer" containerID="0481e30b215d7782e3337c8d235b578d47cb9aaf0203ac7bea341b4df50b5cec" Nov 25 12:01:11 crc kubenswrapper[4776]: I1125 12:01:11.489430 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dz69w"] Nov 25 12:01:11 crc kubenswrapper[4776]: I1125 12:01:11.500876 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dz69w"] Nov 25 12:01:11 crc kubenswrapper[4776]: I1125 12:01:11.526695 4776 scope.go:117] "RemoveContainer" containerID="2a94fc7cc414382e9100fb6fbee4de58ddda5243112676d64b855d752e80920d" Nov 25 12:01:11 crc kubenswrapper[4776]: I1125 12:01:11.673257 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f2878a8-b88d-4a1f-9e46-3b31e155f1ea" path="/var/lib/kubelet/pods/1f2878a8-b88d-4a1f-9e46-3b31e155f1ea/volumes" Nov 25 12:01:54 crc kubenswrapper[4776]: I1125 12:01:54.768367 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qz2nk"] Nov 25 12:01:54 crc kubenswrapper[4776]: E1125 12:01:54.769498 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f2878a8-b88d-4a1f-9e46-3b31e155f1ea" containerName="extract-utilities" Nov 25 12:01:54 crc kubenswrapper[4776]: I1125 12:01:54.769517 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f2878a8-b88d-4a1f-9e46-3b31e155f1ea" containerName="extract-utilities" Nov 25 12:01:54 crc kubenswrapper[4776]: E1125 12:01:54.769552 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f2878a8-b88d-4a1f-9e46-3b31e155f1ea" containerName="extract-content" Nov 25 12:01:54 crc kubenswrapper[4776]: I1125 12:01:54.769561 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f2878a8-b88d-4a1f-9e46-3b31e155f1ea" containerName="extract-content" Nov 25 12:01:54 crc kubenswrapper[4776]: E1125 12:01:54.769604 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f2878a8-b88d-4a1f-9e46-3b31e155f1ea" containerName="registry-server" Nov 25 12:01:54 crc kubenswrapper[4776]: I1125 12:01:54.769613 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f2878a8-b88d-4a1f-9e46-3b31e155f1ea" containerName="registry-server" Nov 25 12:01:54 crc kubenswrapper[4776]: E1125 12:01:54.769637 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="841c9ff3-7f79-424b-9da0-4bca60584e2a" containerName="keystone-cron" Nov 25 12:01:54 crc kubenswrapper[4776]: I1125 12:01:54.769645 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="841c9ff3-7f79-424b-9da0-4bca60584e2a" containerName="keystone-cron" Nov 25 12:01:54 crc kubenswrapper[4776]: I1125 12:01:54.769920 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f2878a8-b88d-4a1f-9e46-3b31e155f1ea" containerName="registry-server" Nov 25 12:01:54 crc kubenswrapper[4776]: I1125 12:01:54.769933 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="841c9ff3-7f79-424b-9da0-4bca60584e2a" containerName="keystone-cron" Nov 25 12:01:54 crc kubenswrapper[4776]: I1125 12:01:54.771938 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qz2nk" Nov 25 12:01:54 crc kubenswrapper[4776]: I1125 12:01:54.784814 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qz2nk"] Nov 25 12:01:54 crc kubenswrapper[4776]: I1125 12:01:54.879851 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59b1c546-56a1-4255-b939-02e7b5e989ff-utilities\") pod \"redhat-marketplace-qz2nk\" (UID: \"59b1c546-56a1-4255-b939-02e7b5e989ff\") " pod="openshift-marketplace/redhat-marketplace-qz2nk" Nov 25 12:01:54 crc kubenswrapper[4776]: I1125 12:01:54.879968 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59b1c546-56a1-4255-b939-02e7b5e989ff-catalog-content\") pod \"redhat-marketplace-qz2nk\" (UID: \"59b1c546-56a1-4255-b939-02e7b5e989ff\") " pod="openshift-marketplace/redhat-marketplace-qz2nk" Nov 25 12:01:54 crc kubenswrapper[4776]: I1125 12:01:54.880028 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrvsz\" (UniqueName: \"kubernetes.io/projected/59b1c546-56a1-4255-b939-02e7b5e989ff-kube-api-access-lrvsz\") pod \"redhat-marketplace-qz2nk\" (UID: \"59b1c546-56a1-4255-b939-02e7b5e989ff\") " pod="openshift-marketplace/redhat-marketplace-qz2nk" Nov 25 12:01:54 crc kubenswrapper[4776]: I1125 12:01:54.983909 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59b1c546-56a1-4255-b939-02e7b5e989ff-utilities\") pod \"redhat-marketplace-qz2nk\" (UID: \"59b1c546-56a1-4255-b939-02e7b5e989ff\") " pod="openshift-marketplace/redhat-marketplace-qz2nk" Nov 25 12:01:54 crc kubenswrapper[4776]: I1125 12:01:54.984450 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59b1c546-56a1-4255-b939-02e7b5e989ff-catalog-content\") pod \"redhat-marketplace-qz2nk\" (UID: \"59b1c546-56a1-4255-b939-02e7b5e989ff\") " pod="openshift-marketplace/redhat-marketplace-qz2nk" Nov 25 12:01:54 crc kubenswrapper[4776]: I1125 12:01:54.984562 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59b1c546-56a1-4255-b939-02e7b5e989ff-utilities\") pod \"redhat-marketplace-qz2nk\" (UID: \"59b1c546-56a1-4255-b939-02e7b5e989ff\") " pod="openshift-marketplace/redhat-marketplace-qz2nk" Nov 25 12:01:54 crc kubenswrapper[4776]: I1125 12:01:54.984605 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrvsz\" (UniqueName: \"kubernetes.io/projected/59b1c546-56a1-4255-b939-02e7b5e989ff-kube-api-access-lrvsz\") pod \"redhat-marketplace-qz2nk\" (UID: \"59b1c546-56a1-4255-b939-02e7b5e989ff\") " pod="openshift-marketplace/redhat-marketplace-qz2nk" Nov 25 12:01:54 crc kubenswrapper[4776]: I1125 12:01:54.985006 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59b1c546-56a1-4255-b939-02e7b5e989ff-catalog-content\") pod \"redhat-marketplace-qz2nk\" (UID: \"59b1c546-56a1-4255-b939-02e7b5e989ff\") " pod="openshift-marketplace/redhat-marketplace-qz2nk" Nov 25 12:01:55 crc kubenswrapper[4776]: I1125 12:01:55.019296 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrvsz\" (UniqueName: \"kubernetes.io/projected/59b1c546-56a1-4255-b939-02e7b5e989ff-kube-api-access-lrvsz\") pod \"redhat-marketplace-qz2nk\" (UID: \"59b1c546-56a1-4255-b939-02e7b5e989ff\") " pod="openshift-marketplace/redhat-marketplace-qz2nk" Nov 25 12:01:55 crc kubenswrapper[4776]: I1125 12:01:55.098921 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qz2nk" Nov 25 12:01:55 crc kubenswrapper[4776]: I1125 12:01:55.675646 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qz2nk"] Nov 25 12:01:55 crc kubenswrapper[4776]: I1125 12:01:55.948458 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qz2nk" event={"ID":"59b1c546-56a1-4255-b939-02e7b5e989ff","Type":"ContainerStarted","Data":"4338f75139dd658c073efb6a1fd956142cf298bdb8b410f97c0a6ff11d2a277d"} Nov 25 12:01:56 crc kubenswrapper[4776]: I1125 12:01:56.965290 4776 generic.go:334] "Generic (PLEG): container finished" podID="59b1c546-56a1-4255-b939-02e7b5e989ff" containerID="0b3d9e824e9a54957e4cfee3a9d843cc8c9dfa97f8ca13ca7dfee4895340f9e2" exitCode=0 Nov 25 12:01:56 crc kubenswrapper[4776]: I1125 12:01:56.966096 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qz2nk" event={"ID":"59b1c546-56a1-4255-b939-02e7b5e989ff","Type":"ContainerDied","Data":"0b3d9e824e9a54957e4cfee3a9d843cc8c9dfa97f8ca13ca7dfee4895340f9e2"} Nov 25 12:02:00 crc kubenswrapper[4776]: I1125 12:02:00.011856 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qz2nk" event={"ID":"59b1c546-56a1-4255-b939-02e7b5e989ff","Type":"ContainerStarted","Data":"3dc8ba28c17f4d84ac2228bd7f24cc0f677bd19c21c80be413213c401a8266ec"} Nov 25 12:02:02 crc kubenswrapper[4776]: I1125 12:02:02.044419 4776 generic.go:334] "Generic (PLEG): container finished" podID="59b1c546-56a1-4255-b939-02e7b5e989ff" containerID="3dc8ba28c17f4d84ac2228bd7f24cc0f677bd19c21c80be413213c401a8266ec" exitCode=0 Nov 25 12:02:02 crc kubenswrapper[4776]: I1125 12:02:02.044559 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qz2nk" event={"ID":"59b1c546-56a1-4255-b939-02e7b5e989ff","Type":"ContainerDied","Data":"3dc8ba28c17f4d84ac2228bd7f24cc0f677bd19c21c80be413213c401a8266ec"} Nov 25 12:02:07 crc kubenswrapper[4776]: I1125 12:02:07.105240 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qz2nk" event={"ID":"59b1c546-56a1-4255-b939-02e7b5e989ff","Type":"ContainerStarted","Data":"6df9bb8593d400b0504b4469bd1c07a7b0311fc1505be659a7b3f10e9b044b2b"} Nov 25 12:02:08 crc kubenswrapper[4776]: I1125 12:02:08.141519 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qz2nk" podStartSLOduration=6.378363914 podStartE2EDuration="14.141496536s" podCreationTimestamp="2025-11-25 12:01:54 +0000 UTC" firstStartedPulling="2025-11-25 12:01:56.968260782 +0000 UTC m=+9462.009320335" lastFinishedPulling="2025-11-25 12:02:04.731393394 +0000 UTC m=+9469.772452957" observedRunningTime="2025-11-25 12:02:08.139043204 +0000 UTC m=+9473.180102757" watchObservedRunningTime="2025-11-25 12:02:08.141496536 +0000 UTC m=+9473.182556089" Nov 25 12:02:15 crc kubenswrapper[4776]: I1125 12:02:15.099256 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qz2nk" Nov 25 12:02:15 crc kubenswrapper[4776]: I1125 12:02:15.099933 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qz2nk" Nov 25 12:02:15 crc kubenswrapper[4776]: I1125 12:02:15.612236 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qz2nk" Nov 25 12:02:15 crc kubenswrapper[4776]: I1125 12:02:15.685383 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qz2nk" Nov 25 12:02:15 crc kubenswrapper[4776]: I1125 12:02:15.859231 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qz2nk"] Nov 25 12:02:17 crc kubenswrapper[4776]: I1125 12:02:17.218499 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qz2nk" podUID="59b1c546-56a1-4255-b939-02e7b5e989ff" containerName="registry-server" containerID="cri-o://6df9bb8593d400b0504b4469bd1c07a7b0311fc1505be659a7b3f10e9b044b2b" gracePeriod=2 Nov 25 12:02:18 crc kubenswrapper[4776]: I1125 12:02:18.245053 4776 generic.go:334] "Generic (PLEG): container finished" podID="59b1c546-56a1-4255-b939-02e7b5e989ff" containerID="6df9bb8593d400b0504b4469bd1c07a7b0311fc1505be659a7b3f10e9b044b2b" exitCode=0 Nov 25 12:02:18 crc kubenswrapper[4776]: I1125 12:02:18.245247 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qz2nk" event={"ID":"59b1c546-56a1-4255-b939-02e7b5e989ff","Type":"ContainerDied","Data":"6df9bb8593d400b0504b4469bd1c07a7b0311fc1505be659a7b3f10e9b044b2b"} Nov 25 12:02:18 crc kubenswrapper[4776]: I1125 12:02:18.631503 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qz2nk" Nov 25 12:02:18 crc kubenswrapper[4776]: I1125 12:02:18.728770 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrvsz\" (UniqueName: \"kubernetes.io/projected/59b1c546-56a1-4255-b939-02e7b5e989ff-kube-api-access-lrvsz\") pod \"59b1c546-56a1-4255-b939-02e7b5e989ff\" (UID: \"59b1c546-56a1-4255-b939-02e7b5e989ff\") " Nov 25 12:02:18 crc kubenswrapper[4776]: I1125 12:02:18.728936 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59b1c546-56a1-4255-b939-02e7b5e989ff-utilities\") pod \"59b1c546-56a1-4255-b939-02e7b5e989ff\" (UID: \"59b1c546-56a1-4255-b939-02e7b5e989ff\") " Nov 25 12:02:18 crc kubenswrapper[4776]: I1125 12:02:18.729131 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59b1c546-56a1-4255-b939-02e7b5e989ff-catalog-content\") pod \"59b1c546-56a1-4255-b939-02e7b5e989ff\" (UID: \"59b1c546-56a1-4255-b939-02e7b5e989ff\") " Nov 25 12:02:18 crc kubenswrapper[4776]: I1125 12:02:18.730352 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59b1c546-56a1-4255-b939-02e7b5e989ff-utilities" (OuterVolumeSpecName: "utilities") pod "59b1c546-56a1-4255-b939-02e7b5e989ff" (UID: "59b1c546-56a1-4255-b939-02e7b5e989ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:02:18 crc kubenswrapper[4776]: I1125 12:02:18.738038 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59b1c546-56a1-4255-b939-02e7b5e989ff-kube-api-access-lrvsz" (OuterVolumeSpecName: "kube-api-access-lrvsz") pod "59b1c546-56a1-4255-b939-02e7b5e989ff" (UID: "59b1c546-56a1-4255-b939-02e7b5e989ff"). InnerVolumeSpecName "kube-api-access-lrvsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:02:18 crc kubenswrapper[4776]: I1125 12:02:18.746857 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59b1c546-56a1-4255-b939-02e7b5e989ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "59b1c546-56a1-4255-b939-02e7b5e989ff" (UID: "59b1c546-56a1-4255-b939-02e7b5e989ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:02:18 crc kubenswrapper[4776]: I1125 12:02:18.832800 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59b1c546-56a1-4255-b939-02e7b5e989ff-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:02:18 crc kubenswrapper[4776]: I1125 12:02:18.832840 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrvsz\" (UniqueName: \"kubernetes.io/projected/59b1c546-56a1-4255-b939-02e7b5e989ff-kube-api-access-lrvsz\") on node \"crc\" DevicePath \"\"" Nov 25 12:02:18 crc kubenswrapper[4776]: I1125 12:02:18.832853 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59b1c546-56a1-4255-b939-02e7b5e989ff-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:02:19 crc kubenswrapper[4776]: I1125 12:02:19.262025 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qz2nk" event={"ID":"59b1c546-56a1-4255-b939-02e7b5e989ff","Type":"ContainerDied","Data":"4338f75139dd658c073efb6a1fd956142cf298bdb8b410f97c0a6ff11d2a277d"} Nov 25 12:02:19 crc kubenswrapper[4776]: I1125 12:02:19.262139 4776 scope.go:117] "RemoveContainer" containerID="6df9bb8593d400b0504b4469bd1c07a7b0311fc1505be659a7b3f10e9b044b2b" Nov 25 12:02:19 crc kubenswrapper[4776]: I1125 12:02:19.262175 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qz2nk" Nov 25 12:02:19 crc kubenswrapper[4776]: I1125 12:02:19.293215 4776 scope.go:117] "RemoveContainer" containerID="3dc8ba28c17f4d84ac2228bd7f24cc0f677bd19c21c80be413213c401a8266ec" Nov 25 12:02:19 crc kubenswrapper[4776]: I1125 12:02:19.332370 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qz2nk"] Nov 25 12:02:19 crc kubenswrapper[4776]: I1125 12:02:19.341845 4776 scope.go:117] "RemoveContainer" containerID="0b3d9e824e9a54957e4cfee3a9d843cc8c9dfa97f8ca13ca7dfee4895340f9e2" Nov 25 12:02:19 crc kubenswrapper[4776]: I1125 12:02:19.351464 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qz2nk"] Nov 25 12:02:19 crc kubenswrapper[4776]: I1125 12:02:19.678806 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59b1c546-56a1-4255-b939-02e7b5e989ff" path="/var/lib/kubelet/pods/59b1c546-56a1-4255-b939-02e7b5e989ff/volumes" Nov 25 12:03:17 crc kubenswrapper[4776]: I1125 12:03:17.817911 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:03:17 crc kubenswrapper[4776]: I1125 12:03:17.818463 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:03:47 crc kubenswrapper[4776]: I1125 12:03:47.820056 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:03:47 crc kubenswrapper[4776]: I1125 12:03:47.820628 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:03:51 crc kubenswrapper[4776]: I1125 12:03:51.219313 4776 generic.go:334] "Generic (PLEG): container finished" podID="1db10c36-1c34-47f5-a872-a82d15d88299" containerID="348c62a0a95daa8f9cf57bb1fd0e254d304868ccfdf61b5064ce0a023c6be675" exitCode=0 Nov 25 12:03:51 crc kubenswrapper[4776]: I1125 12:03:51.219434 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" event={"ID":"1db10c36-1c34-47f5-a872-a82d15d88299","Type":"ContainerDied","Data":"348c62a0a95daa8f9cf57bb1fd0e254d304868ccfdf61b5064ce0a023c6be675"} Nov 25 12:03:52 crc kubenswrapper[4776]: I1125 12:03:52.709912 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 12:03:52 crc kubenswrapper[4776]: I1125 12:03:52.818232 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-inventory\") pod \"1db10c36-1c34-47f5-a872-a82d15d88299\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " Nov 25 12:03:52 crc kubenswrapper[4776]: I1125 12:03:52.818282 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ssr88\" (UniqueName: \"kubernetes.io/projected/1db10c36-1c34-47f5-a872-a82d15d88299-kube-api-access-ssr88\") pod \"1db10c36-1c34-47f5-a872-a82d15d88299\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " Nov 25 12:03:52 crc kubenswrapper[4776]: I1125 12:03:52.818365 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-cell1-compute-config-1\") pod \"1db10c36-1c34-47f5-a872-a82d15d88299\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " Nov 25 12:03:52 crc kubenswrapper[4776]: I1125 12:03:52.818418 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-cell1-compute-config-0\") pod \"1db10c36-1c34-47f5-a872-a82d15d88299\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " Nov 25 12:03:52 crc kubenswrapper[4776]: I1125 12:03:52.818439 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/1db10c36-1c34-47f5-a872-a82d15d88299-nova-cells-global-config-0\") pod \"1db10c36-1c34-47f5-a872-a82d15d88299\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " Nov 25 12:03:52 crc kubenswrapper[4776]: I1125 12:03:52.818537 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-migration-ssh-key-0\") pod \"1db10c36-1c34-47f5-a872-a82d15d88299\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " Nov 25 12:03:52 crc kubenswrapper[4776]: I1125 12:03:52.818622 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-cell1-combined-ca-bundle\") pod \"1db10c36-1c34-47f5-a872-a82d15d88299\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " Nov 25 12:03:52 crc kubenswrapper[4776]: I1125 12:03:52.818641 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-migration-ssh-key-1\") pod \"1db10c36-1c34-47f5-a872-a82d15d88299\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " Nov 25 12:03:52 crc kubenswrapper[4776]: I1125 12:03:52.818734 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-ssh-key\") pod \"1db10c36-1c34-47f5-a872-a82d15d88299\" (UID: \"1db10c36-1c34-47f5-a872-a82d15d88299\") " Nov 25 12:03:52 crc kubenswrapper[4776]: I1125 12:03:52.827388 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "1db10c36-1c34-47f5-a872-a82d15d88299" (UID: "1db10c36-1c34-47f5-a872-a82d15d88299"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:03:52 crc kubenswrapper[4776]: I1125 12:03:52.829428 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1db10c36-1c34-47f5-a872-a82d15d88299-kube-api-access-ssr88" (OuterVolumeSpecName: "kube-api-access-ssr88") pod "1db10c36-1c34-47f5-a872-a82d15d88299" (UID: "1db10c36-1c34-47f5-a872-a82d15d88299"). InnerVolumeSpecName "kube-api-access-ssr88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:03:52 crc kubenswrapper[4776]: I1125 12:03:52.856811 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "1db10c36-1c34-47f5-a872-a82d15d88299" (UID: "1db10c36-1c34-47f5-a872-a82d15d88299"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:03:52 crc kubenswrapper[4776]: I1125 12:03:52.857486 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-inventory" (OuterVolumeSpecName: "inventory") pod "1db10c36-1c34-47f5-a872-a82d15d88299" (UID: "1db10c36-1c34-47f5-a872-a82d15d88299"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:03:52 crc kubenswrapper[4776]: I1125 12:03:52.866558 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "1db10c36-1c34-47f5-a872-a82d15d88299" (UID: "1db10c36-1c34-47f5-a872-a82d15d88299"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:03:52 crc kubenswrapper[4776]: I1125 12:03:52.871305 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "1db10c36-1c34-47f5-a872-a82d15d88299" (UID: "1db10c36-1c34-47f5-a872-a82d15d88299"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:03:52 crc kubenswrapper[4776]: I1125 12:03:52.874115 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1db10c36-1c34-47f5-a872-a82d15d88299" (UID: "1db10c36-1c34-47f5-a872-a82d15d88299"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:03:52 crc kubenswrapper[4776]: I1125 12:03:52.887360 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1db10c36-1c34-47f5-a872-a82d15d88299-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "1db10c36-1c34-47f5-a872-a82d15d88299" (UID: "1db10c36-1c34-47f5-a872-a82d15d88299"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:03:52 crc kubenswrapper[4776]: I1125 12:03:52.897259 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "1db10c36-1c34-47f5-a872-a82d15d88299" (UID: "1db10c36-1c34-47f5-a872-a82d15d88299"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:03:52 crc kubenswrapper[4776]: I1125 12:03:52.921198 4776 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 12:03:52 crc kubenswrapper[4776]: I1125 12:03:52.921234 4776 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 25 12:03:52 crc kubenswrapper[4776]: I1125 12:03:52.921244 4776 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 12:03:52 crc kubenswrapper[4776]: I1125 12:03:52.921254 4776 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 12:03:52 crc kubenswrapper[4776]: I1125 12:03:52.921263 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ssr88\" (UniqueName: \"kubernetes.io/projected/1db10c36-1c34-47f5-a872-a82d15d88299-kube-api-access-ssr88\") on node \"crc\" DevicePath \"\"" Nov 25 12:03:52 crc kubenswrapper[4776]: I1125 12:03:52.921274 4776 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 25 12:03:52 crc kubenswrapper[4776]: I1125 12:03:52.921284 4776 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 12:03:52 crc kubenswrapper[4776]: I1125 12:03:52.921293 4776 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/1db10c36-1c34-47f5-a872-a82d15d88299-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 12:03:52 crc kubenswrapper[4776]: I1125 12:03:52.921302 4776 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1db10c36-1c34-47f5-a872-a82d15d88299-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 25 12:03:53 crc kubenswrapper[4776]: I1125 12:03:53.288339 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" event={"ID":"1db10c36-1c34-47f5-a872-a82d15d88299","Type":"ContainerDied","Data":"e6a5ccf05e0c7c8ad99d27f5bf5d99cab8c3979387f222c407bda5e73ec12d65"} Nov 25 12:03:53 crc kubenswrapper[4776]: I1125 12:03:53.288406 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6a5ccf05e0c7c8ad99d27f5bf5d99cab8c3979387f222c407bda5e73ec12d65" Nov 25 12:03:53 crc kubenswrapper[4776]: I1125 12:03:53.288512 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk" Nov 25 12:04:17 crc kubenswrapper[4776]: I1125 12:04:17.821616 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:04:17 crc kubenswrapper[4776]: I1125 12:04:17.822540 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:04:17 crc kubenswrapper[4776]: I1125 12:04:17.822665 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 12:04:17 crc kubenswrapper[4776]: I1125 12:04:17.823994 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 12:04:17 crc kubenswrapper[4776]: I1125 12:04:17.824097 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d" gracePeriod=600 Nov 25 12:04:17 crc kubenswrapper[4776]: E1125 12:04:17.976051 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:04:18 crc kubenswrapper[4776]: I1125 12:04:18.632052 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d" exitCode=0 Nov 25 12:04:18 crc kubenswrapper[4776]: I1125 12:04:18.632133 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d"} Nov 25 12:04:18 crc kubenswrapper[4776]: I1125 12:04:18.632209 4776 scope.go:117] "RemoveContainer" containerID="8e060fbd41a6ceeaf0a55e5585d31b0083930b3310803a6e50fc5823301c1b89" Nov 25 12:04:18 crc kubenswrapper[4776]: I1125 12:04:18.632832 4776 scope.go:117] "RemoveContainer" containerID="b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d" Nov 25 12:04:18 crc kubenswrapper[4776]: E1125 12:04:18.633290 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:04:29 crc kubenswrapper[4776]: I1125 12:04:29.662411 4776 scope.go:117] "RemoveContainer" containerID="b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d" Nov 25 12:04:29 crc kubenswrapper[4776]: E1125 12:04:29.665171 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:04:43 crc kubenswrapper[4776]: I1125 12:04:43.663562 4776 scope.go:117] "RemoveContainer" containerID="b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d" Nov 25 12:04:43 crc kubenswrapper[4776]: E1125 12:04:43.665222 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:04:55 crc kubenswrapper[4776]: I1125 12:04:55.672662 4776 scope.go:117] "RemoveContainer" containerID="b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d" Nov 25 12:04:55 crc kubenswrapper[4776]: E1125 12:04:55.673557 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:05:06 crc kubenswrapper[4776]: I1125 12:05:06.662668 4776 scope.go:117] "RemoveContainer" containerID="b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d" Nov 25 12:05:06 crc kubenswrapper[4776]: E1125 12:05:06.665934 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:05:17 crc kubenswrapper[4776]: I1125 12:05:17.670340 4776 scope.go:117] "RemoveContainer" containerID="b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d" Nov 25 12:05:17 crc kubenswrapper[4776]: E1125 12:05:17.673029 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:05:30 crc kubenswrapper[4776]: I1125 12:05:30.663384 4776 scope.go:117] "RemoveContainer" containerID="b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d" Nov 25 12:05:30 crc kubenswrapper[4776]: E1125 12:05:30.664577 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:05:43 crc kubenswrapper[4776]: I1125 12:05:43.663522 4776 scope.go:117] "RemoveContainer" containerID="b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d" Nov 25 12:05:43 crc kubenswrapper[4776]: E1125 12:05:43.664961 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:05:51 crc kubenswrapper[4776]: I1125 12:05:51.267714 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Nov 25 12:05:51 crc kubenswrapper[4776]: I1125 12:05:51.268407 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-copy-data" podUID="d4a1804d-1d0c-4976-840a-02335c611592" containerName="adoption" containerID="cri-o://398d97eaeb03fef58b8fdf1ee14c3011b9e92d960abf3bfbc7e3c61f4f792c94" gracePeriod=30 Nov 25 12:05:54 crc kubenswrapper[4776]: I1125 12:05:54.663300 4776 scope.go:117] "RemoveContainer" containerID="b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d" Nov 25 12:05:54 crc kubenswrapper[4776]: E1125 12:05:54.664272 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:06:05 crc kubenswrapper[4776]: I1125 12:06:05.669994 4776 scope.go:117] "RemoveContainer" containerID="b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d" Nov 25 12:06:05 crc kubenswrapper[4776]: E1125 12:06:05.671261 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:06:17 crc kubenswrapper[4776]: I1125 12:06:17.663591 4776 scope.go:117] "RemoveContainer" containerID="b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d" Nov 25 12:06:17 crc kubenswrapper[4776]: E1125 12:06:17.667814 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:06:21 crc kubenswrapper[4776]: I1125 12:06:21.416700 4776 generic.go:334] "Generic (PLEG): container finished" podID="d4a1804d-1d0c-4976-840a-02335c611592" containerID="398d97eaeb03fef58b8fdf1ee14c3011b9e92d960abf3bfbc7e3c61f4f792c94" exitCode=137 Nov 25 12:06:21 crc kubenswrapper[4776]: I1125 12:06:21.416863 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"d4a1804d-1d0c-4976-840a-02335c611592","Type":"ContainerDied","Data":"398d97eaeb03fef58b8fdf1ee14c3011b9e92d960abf3bfbc7e3c61f4f792c94"} Nov 25 12:06:21 crc kubenswrapper[4776]: I1125 12:06:21.763128 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Nov 25 12:06:21 crc kubenswrapper[4776]: I1125 12:06:21.870805 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mariadb-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-39e1751a-fb9e-4783-809a-8a50b438fb75\") pod \"d4a1804d-1d0c-4976-840a-02335c611592\" (UID: \"d4a1804d-1d0c-4976-840a-02335c611592\") " Nov 25 12:06:21 crc kubenswrapper[4776]: I1125 12:06:21.871013 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f89bv\" (UniqueName: \"kubernetes.io/projected/d4a1804d-1d0c-4976-840a-02335c611592-kube-api-access-f89bv\") pod \"d4a1804d-1d0c-4976-840a-02335c611592\" (UID: \"d4a1804d-1d0c-4976-840a-02335c611592\") " Nov 25 12:06:21 crc kubenswrapper[4776]: I1125 12:06:21.877784 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4a1804d-1d0c-4976-840a-02335c611592-kube-api-access-f89bv" (OuterVolumeSpecName: "kube-api-access-f89bv") pod "d4a1804d-1d0c-4976-840a-02335c611592" (UID: "d4a1804d-1d0c-4976-840a-02335c611592"). InnerVolumeSpecName "kube-api-access-f89bv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:06:21 crc kubenswrapper[4776]: I1125 12:06:21.905109 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-39e1751a-fb9e-4783-809a-8a50b438fb75" (OuterVolumeSpecName: "mariadb-data") pod "d4a1804d-1d0c-4976-840a-02335c611592" (UID: "d4a1804d-1d0c-4976-840a-02335c611592"). InnerVolumeSpecName "pvc-39e1751a-fb9e-4783-809a-8a50b438fb75". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 12:06:21 crc kubenswrapper[4776]: I1125 12:06:21.975121 4776 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-39e1751a-fb9e-4783-809a-8a50b438fb75\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-39e1751a-fb9e-4783-809a-8a50b438fb75\") on node \"crc\" " Nov 25 12:06:21 crc kubenswrapper[4776]: I1125 12:06:21.975171 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f89bv\" (UniqueName: \"kubernetes.io/projected/d4a1804d-1d0c-4976-840a-02335c611592-kube-api-access-f89bv\") on node \"crc\" DevicePath \"\"" Nov 25 12:06:22 crc kubenswrapper[4776]: I1125 12:06:22.016730 4776 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 25 12:06:22 crc kubenswrapper[4776]: I1125 12:06:22.057464 4776 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-39e1751a-fb9e-4783-809a-8a50b438fb75" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-39e1751a-fb9e-4783-809a-8a50b438fb75") on node "crc" Nov 25 12:06:22 crc kubenswrapper[4776]: I1125 12:06:22.077480 4776 reconciler_common.go:293] "Volume detached for volume \"pvc-39e1751a-fb9e-4783-809a-8a50b438fb75\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-39e1751a-fb9e-4783-809a-8a50b438fb75\") on node \"crc\" DevicePath \"\"" Nov 25 12:06:22 crc kubenswrapper[4776]: I1125 12:06:22.428495 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"d4a1804d-1d0c-4976-840a-02335c611592","Type":"ContainerDied","Data":"2304778b2312639657687cd5bcf5b78dc0faba9ba827ba060427e9a53c57d948"} Nov 25 12:06:22 crc kubenswrapper[4776]: I1125 12:06:22.429862 4776 scope.go:117] "RemoveContainer" containerID="398d97eaeb03fef58b8fdf1ee14c3011b9e92d960abf3bfbc7e3c61f4f792c94" Nov 25 12:06:22 crc kubenswrapper[4776]: I1125 12:06:22.428544 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Nov 25 12:06:22 crc kubenswrapper[4776]: I1125 12:06:22.474183 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Nov 25 12:06:22 crc kubenswrapper[4776]: I1125 12:06:22.484793 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-copy-data"] Nov 25 12:06:23 crc kubenswrapper[4776]: I1125 12:06:23.106511 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Nov 25 12:06:23 crc kubenswrapper[4776]: I1125 12:06:23.107406 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-copy-data" podUID="bb36b96d-a860-41e0-9cea-59479eb5ce28" containerName="adoption" containerID="cri-o://ca02947c0f670b534f946ae0b95ced2d33b13e1cabb85697d6bd977eebd187d3" gracePeriod=30 Nov 25 12:06:23 crc kubenswrapper[4776]: I1125 12:06:23.682970 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4a1804d-1d0c-4976-840a-02335c611592" path="/var/lib/kubelet/pods/d4a1804d-1d0c-4976-840a-02335c611592/volumes" Nov 25 12:06:28 crc kubenswrapper[4776]: I1125 12:06:28.663258 4776 scope.go:117] "RemoveContainer" containerID="b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d" Nov 25 12:06:28 crc kubenswrapper[4776]: E1125 12:06:28.664663 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:06:42 crc kubenswrapper[4776]: I1125 12:06:42.662779 4776 scope.go:117] "RemoveContainer" containerID="b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d" Nov 25 12:06:42 crc kubenswrapper[4776]: E1125 12:06:42.663763 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:06:42 crc kubenswrapper[4776]: I1125 12:06:42.907267 4776 scope.go:117] "RemoveContainer" containerID="3ee570da3e051ced2570f4231bb916714f037f71cce01519e7cf649d623644b4" Nov 25 12:06:42 crc kubenswrapper[4776]: I1125 12:06:42.946486 4776 scope.go:117] "RemoveContainer" containerID="d45573d6e1f2ed34b1a535d0b040271fbc6280f71131740685efe02f38a49cb8" Nov 25 12:06:43 crc kubenswrapper[4776]: I1125 12:06:43.028042 4776 scope.go:117] "RemoveContainer" containerID="fa0ba5753ce17ddfafa4d13ffcbfdee679bee87b4520d89475e12ff429bde536" Nov 25 12:06:53 crc kubenswrapper[4776]: I1125 12:06:53.812926 4776 generic.go:334] "Generic (PLEG): container finished" podID="bb36b96d-a860-41e0-9cea-59479eb5ce28" containerID="ca02947c0f670b534f946ae0b95ced2d33b13e1cabb85697d6bd977eebd187d3" exitCode=137 Nov 25 12:06:53 crc kubenswrapper[4776]: I1125 12:06:53.813621 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"bb36b96d-a860-41e0-9cea-59479eb5ce28","Type":"ContainerDied","Data":"ca02947c0f670b534f946ae0b95ced2d33b13e1cabb85697d6bd977eebd187d3"} Nov 25 12:06:53 crc kubenswrapper[4776]: I1125 12:06:53.954689 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Nov 25 12:06:54 crc kubenswrapper[4776]: I1125 12:06:54.027786 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84s25\" (UniqueName: \"kubernetes.io/projected/bb36b96d-a860-41e0-9cea-59479eb5ce28-kube-api-access-84s25\") pod \"bb36b96d-a860-41e0-9cea-59479eb5ce28\" (UID: \"bb36b96d-a860-41e0-9cea-59479eb5ce28\") " Nov 25 12:06:54 crc kubenswrapper[4776]: I1125 12:06:54.032543 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-69f0de1e-12d2-4c9e-a4b1-29b517947901\") pod \"bb36b96d-a860-41e0-9cea-59479eb5ce28\" (UID: \"bb36b96d-a860-41e0-9cea-59479eb5ce28\") " Nov 25 12:06:54 crc kubenswrapper[4776]: I1125 12:06:54.032611 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/bb36b96d-a860-41e0-9cea-59479eb5ce28-ovn-data-cert\") pod \"bb36b96d-a860-41e0-9cea-59479eb5ce28\" (UID: \"bb36b96d-a860-41e0-9cea-59479eb5ce28\") " Nov 25 12:06:54 crc kubenswrapper[4776]: I1125 12:06:54.042275 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb36b96d-a860-41e0-9cea-59479eb5ce28-ovn-data-cert" (OuterVolumeSpecName: "ovn-data-cert") pod "bb36b96d-a860-41e0-9cea-59479eb5ce28" (UID: "bb36b96d-a860-41e0-9cea-59479eb5ce28"). InnerVolumeSpecName "ovn-data-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:06:54 crc kubenswrapper[4776]: I1125 12:06:54.042493 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb36b96d-a860-41e0-9cea-59479eb5ce28-kube-api-access-84s25" (OuterVolumeSpecName: "kube-api-access-84s25") pod "bb36b96d-a860-41e0-9cea-59479eb5ce28" (UID: "bb36b96d-a860-41e0-9cea-59479eb5ce28"). InnerVolumeSpecName "kube-api-access-84s25". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:06:54 crc kubenswrapper[4776]: I1125 12:06:54.068671 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-69f0de1e-12d2-4c9e-a4b1-29b517947901" (OuterVolumeSpecName: "ovn-data") pod "bb36b96d-a860-41e0-9cea-59479eb5ce28" (UID: "bb36b96d-a860-41e0-9cea-59479eb5ce28"). InnerVolumeSpecName "pvc-69f0de1e-12d2-4c9e-a4b1-29b517947901". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 12:06:54 crc kubenswrapper[4776]: I1125 12:06:54.137817 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84s25\" (UniqueName: \"kubernetes.io/projected/bb36b96d-a860-41e0-9cea-59479eb5ce28-kube-api-access-84s25\") on node \"crc\" DevicePath \"\"" Nov 25 12:06:54 crc kubenswrapper[4776]: I1125 12:06:54.137925 4776 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-69f0de1e-12d2-4c9e-a4b1-29b517947901\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-69f0de1e-12d2-4c9e-a4b1-29b517947901\") on node \"crc\" " Nov 25 12:06:54 crc kubenswrapper[4776]: I1125 12:06:54.137947 4776 reconciler_common.go:293] "Volume detached for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/bb36b96d-a860-41e0-9cea-59479eb5ce28-ovn-data-cert\") on node \"crc\" DevicePath \"\"" Nov 25 12:06:54 crc kubenswrapper[4776]: I1125 12:06:54.213884 4776 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Nov 25 12:06:54 crc kubenswrapper[4776]: I1125 12:06:54.214116 4776 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-69f0de1e-12d2-4c9e-a4b1-29b517947901" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-69f0de1e-12d2-4c9e-a4b1-29b517947901") on node "crc" Nov 25 12:06:54 crc kubenswrapper[4776]: I1125 12:06:54.241117 4776 reconciler_common.go:293] "Volume detached for volume \"pvc-69f0de1e-12d2-4c9e-a4b1-29b517947901\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-69f0de1e-12d2-4c9e-a4b1-29b517947901\") on node \"crc\" DevicePath \"\"" Nov 25 12:06:54 crc kubenswrapper[4776]: I1125 12:06:54.663324 4776 scope.go:117] "RemoveContainer" containerID="b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d" Nov 25 12:06:54 crc kubenswrapper[4776]: E1125 12:06:54.665298 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:06:54 crc kubenswrapper[4776]: I1125 12:06:54.830922 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"bb36b96d-a860-41e0-9cea-59479eb5ce28","Type":"ContainerDied","Data":"a814c0d7928e235b44086052c1869b9f5022c1e710d9aa180203f31305c76eb2"} Nov 25 12:06:54 crc kubenswrapper[4776]: I1125 12:06:54.830991 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Nov 25 12:06:54 crc kubenswrapper[4776]: I1125 12:06:54.831003 4776 scope.go:117] "RemoveContainer" containerID="ca02947c0f670b534f946ae0b95ced2d33b13e1cabb85697d6bd977eebd187d3" Nov 25 12:06:54 crc kubenswrapper[4776]: I1125 12:06:54.874671 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Nov 25 12:06:54 crc kubenswrapper[4776]: I1125 12:06:54.886697 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-copy-data"] Nov 25 12:06:55 crc kubenswrapper[4776]: I1125 12:06:55.737702 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb36b96d-a860-41e0-9cea-59479eb5ce28" path="/var/lib/kubelet/pods/bb36b96d-a860-41e0-9cea-59479eb5ce28/volumes" Nov 25 12:07:08 crc kubenswrapper[4776]: I1125 12:07:08.663014 4776 scope.go:117] "RemoveContainer" containerID="b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d" Nov 25 12:07:08 crc kubenswrapper[4776]: E1125 12:07:08.663993 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:07:19 crc kubenswrapper[4776]: I1125 12:07:19.664120 4776 scope.go:117] "RemoveContainer" containerID="b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d" Nov 25 12:07:19 crc kubenswrapper[4776]: E1125 12:07:19.665494 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:07:33 crc kubenswrapper[4776]: I1125 12:07:33.662906 4776 scope.go:117] "RemoveContainer" containerID="b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d" Nov 25 12:07:33 crc kubenswrapper[4776]: E1125 12:07:33.663962 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:07:48 crc kubenswrapper[4776]: I1125 12:07:48.663195 4776 scope.go:117] "RemoveContainer" containerID="b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d" Nov 25 12:07:48 crc kubenswrapper[4776]: E1125 12:07:48.664103 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:08:02 crc kubenswrapper[4776]: I1125 12:08:02.664957 4776 scope.go:117] "RemoveContainer" containerID="b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d" Nov 25 12:08:02 crc kubenswrapper[4776]: E1125 12:08:02.665718 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:08:04 crc kubenswrapper[4776]: I1125 12:08:04.489172 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7b7jq/must-gather-wmmkf"] Nov 25 12:08:04 crc kubenswrapper[4776]: E1125 12:08:04.490201 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1db10c36-1c34-47f5-a872-a82d15d88299" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Nov 25 12:08:04 crc kubenswrapper[4776]: I1125 12:08:04.490217 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="1db10c36-1c34-47f5-a872-a82d15d88299" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Nov 25 12:08:04 crc kubenswrapper[4776]: E1125 12:08:04.490227 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59b1c546-56a1-4255-b939-02e7b5e989ff" containerName="extract-utilities" Nov 25 12:08:04 crc kubenswrapper[4776]: I1125 12:08:04.490235 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="59b1c546-56a1-4255-b939-02e7b5e989ff" containerName="extract-utilities" Nov 25 12:08:04 crc kubenswrapper[4776]: E1125 12:08:04.490252 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb36b96d-a860-41e0-9cea-59479eb5ce28" containerName="adoption" Nov 25 12:08:04 crc kubenswrapper[4776]: I1125 12:08:04.490258 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb36b96d-a860-41e0-9cea-59479eb5ce28" containerName="adoption" Nov 25 12:08:04 crc kubenswrapper[4776]: E1125 12:08:04.490265 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4a1804d-1d0c-4976-840a-02335c611592" containerName="adoption" Nov 25 12:08:04 crc kubenswrapper[4776]: I1125 12:08:04.490272 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4a1804d-1d0c-4976-840a-02335c611592" containerName="adoption" Nov 25 12:08:04 crc kubenswrapper[4776]: E1125 12:08:04.490284 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59b1c546-56a1-4255-b939-02e7b5e989ff" containerName="registry-server" Nov 25 12:08:04 crc kubenswrapper[4776]: I1125 12:08:04.490289 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="59b1c546-56a1-4255-b939-02e7b5e989ff" containerName="registry-server" Nov 25 12:08:04 crc kubenswrapper[4776]: E1125 12:08:04.490317 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59b1c546-56a1-4255-b939-02e7b5e989ff" containerName="extract-content" Nov 25 12:08:04 crc kubenswrapper[4776]: I1125 12:08:04.490322 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="59b1c546-56a1-4255-b939-02e7b5e989ff" containerName="extract-content" Nov 25 12:08:04 crc kubenswrapper[4776]: I1125 12:08:04.490520 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="1db10c36-1c34-47f5-a872-a82d15d88299" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Nov 25 12:08:04 crc kubenswrapper[4776]: I1125 12:08:04.490536 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4a1804d-1d0c-4976-840a-02335c611592" containerName="adoption" Nov 25 12:08:04 crc kubenswrapper[4776]: I1125 12:08:04.490550 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb36b96d-a860-41e0-9cea-59479eb5ce28" containerName="adoption" Nov 25 12:08:04 crc kubenswrapper[4776]: I1125 12:08:04.490560 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="59b1c546-56a1-4255-b939-02e7b5e989ff" containerName="registry-server" Nov 25 12:08:04 crc kubenswrapper[4776]: I1125 12:08:04.493742 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b7jq/must-gather-wmmkf" Nov 25 12:08:04 crc kubenswrapper[4776]: I1125 12:08:04.503277 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-7b7jq"/"openshift-service-ca.crt" Nov 25 12:08:04 crc kubenswrapper[4776]: I1125 12:08:04.503529 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-7b7jq"/"default-dockercfg-kwcsg" Nov 25 12:08:04 crc kubenswrapper[4776]: I1125 12:08:04.503688 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-7b7jq"/"kube-root-ca.crt" Nov 25 12:08:04 crc kubenswrapper[4776]: I1125 12:08:04.535769 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-7b7jq/must-gather-wmmkf"] Nov 25 12:08:04 crc kubenswrapper[4776]: I1125 12:08:04.587520 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl455\" (UniqueName: \"kubernetes.io/projected/7256708a-6fd0-414d-902c-105d6efc08e8-kube-api-access-tl455\") pod \"must-gather-wmmkf\" (UID: \"7256708a-6fd0-414d-902c-105d6efc08e8\") " pod="openshift-must-gather-7b7jq/must-gather-wmmkf" Nov 25 12:08:04 crc kubenswrapper[4776]: I1125 12:08:04.587909 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7256708a-6fd0-414d-902c-105d6efc08e8-must-gather-output\") pod \"must-gather-wmmkf\" (UID: \"7256708a-6fd0-414d-902c-105d6efc08e8\") " pod="openshift-must-gather-7b7jq/must-gather-wmmkf" Nov 25 12:08:04 crc kubenswrapper[4776]: I1125 12:08:04.689750 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl455\" (UniqueName: \"kubernetes.io/projected/7256708a-6fd0-414d-902c-105d6efc08e8-kube-api-access-tl455\") pod \"must-gather-wmmkf\" (UID: \"7256708a-6fd0-414d-902c-105d6efc08e8\") " pod="openshift-must-gather-7b7jq/must-gather-wmmkf" Nov 25 12:08:04 crc kubenswrapper[4776]: I1125 12:08:04.689943 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7256708a-6fd0-414d-902c-105d6efc08e8-must-gather-output\") pod \"must-gather-wmmkf\" (UID: \"7256708a-6fd0-414d-902c-105d6efc08e8\") " pod="openshift-must-gather-7b7jq/must-gather-wmmkf" Nov 25 12:08:04 crc kubenswrapper[4776]: I1125 12:08:04.690462 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7256708a-6fd0-414d-902c-105d6efc08e8-must-gather-output\") pod \"must-gather-wmmkf\" (UID: \"7256708a-6fd0-414d-902c-105d6efc08e8\") " pod="openshift-must-gather-7b7jq/must-gather-wmmkf" Nov 25 12:08:04 crc kubenswrapper[4776]: I1125 12:08:04.723850 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl455\" (UniqueName: \"kubernetes.io/projected/7256708a-6fd0-414d-902c-105d6efc08e8-kube-api-access-tl455\") pod \"must-gather-wmmkf\" (UID: \"7256708a-6fd0-414d-902c-105d6efc08e8\") " pod="openshift-must-gather-7b7jq/must-gather-wmmkf" Nov 25 12:08:04 crc kubenswrapper[4776]: I1125 12:08:04.829462 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b7jq/must-gather-wmmkf" Nov 25 12:08:05 crc kubenswrapper[4776]: I1125 12:08:05.396234 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-7b7jq/must-gather-wmmkf"] Nov 25 12:08:05 crc kubenswrapper[4776]: I1125 12:08:05.400534 4776 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 12:08:05 crc kubenswrapper[4776]: I1125 12:08:05.682912 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7b7jq/must-gather-wmmkf" event={"ID":"7256708a-6fd0-414d-902c-105d6efc08e8","Type":"ContainerStarted","Data":"2a2d015ee0c8cc7e57254b65cd464880db25600e2062e287ad6f70b06bbc3d92"} Nov 25 12:08:13 crc kubenswrapper[4776]: I1125 12:08:13.773480 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7b7jq/must-gather-wmmkf" event={"ID":"7256708a-6fd0-414d-902c-105d6efc08e8","Type":"ContainerStarted","Data":"a6559546a505d58e7870ea9ef3612997d38c54df3b2a8d5b62720da5a4dd70a6"} Nov 25 12:08:13 crc kubenswrapper[4776]: I1125 12:08:13.774088 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7b7jq/must-gather-wmmkf" event={"ID":"7256708a-6fd0-414d-902c-105d6efc08e8","Type":"ContainerStarted","Data":"5a1cce6bea9b448d7b7c9493aa43fb0c00de36186e83fd5eb96b09b4ba3a735d"} Nov 25 12:08:13 crc kubenswrapper[4776]: I1125 12:08:13.804567 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-7b7jq/must-gather-wmmkf" podStartSLOduration=3.595923757 podStartE2EDuration="9.804537985s" podCreationTimestamp="2025-11-25 12:08:04 +0000 UTC" firstStartedPulling="2025-11-25 12:08:05.400279256 +0000 UTC m=+9830.441338809" lastFinishedPulling="2025-11-25 12:08:11.608893474 +0000 UTC m=+9836.649953037" observedRunningTime="2025-11-25 12:08:13.788923153 +0000 UTC m=+9838.829982706" watchObservedRunningTime="2025-11-25 12:08:13.804537985 +0000 UTC m=+9838.845597558" Nov 25 12:08:14 crc kubenswrapper[4776]: I1125 12:08:14.662281 4776 scope.go:117] "RemoveContainer" containerID="b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d" Nov 25 12:08:14 crc kubenswrapper[4776]: E1125 12:08:14.662836 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:08:17 crc kubenswrapper[4776]: I1125 12:08:17.162905 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7b7jq/crc-debug-5xj98"] Nov 25 12:08:17 crc kubenswrapper[4776]: I1125 12:08:17.166587 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b7jq/crc-debug-5xj98" Nov 25 12:08:17 crc kubenswrapper[4776]: I1125 12:08:17.248754 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs8cw\" (UniqueName: \"kubernetes.io/projected/5da06a58-941d-4405-8d81-3894f8366813-kube-api-access-xs8cw\") pod \"crc-debug-5xj98\" (UID: \"5da06a58-941d-4405-8d81-3894f8366813\") " pod="openshift-must-gather-7b7jq/crc-debug-5xj98" Nov 25 12:08:17 crc kubenswrapper[4776]: I1125 12:08:17.248937 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5da06a58-941d-4405-8d81-3894f8366813-host\") pod \"crc-debug-5xj98\" (UID: \"5da06a58-941d-4405-8d81-3894f8366813\") " pod="openshift-must-gather-7b7jq/crc-debug-5xj98" Nov 25 12:08:17 crc kubenswrapper[4776]: I1125 12:08:17.351807 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5da06a58-941d-4405-8d81-3894f8366813-host\") pod \"crc-debug-5xj98\" (UID: \"5da06a58-941d-4405-8d81-3894f8366813\") " pod="openshift-must-gather-7b7jq/crc-debug-5xj98" Nov 25 12:08:17 crc kubenswrapper[4776]: I1125 12:08:17.351967 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5da06a58-941d-4405-8d81-3894f8366813-host\") pod \"crc-debug-5xj98\" (UID: \"5da06a58-941d-4405-8d81-3894f8366813\") " pod="openshift-must-gather-7b7jq/crc-debug-5xj98" Nov 25 12:08:17 crc kubenswrapper[4776]: I1125 12:08:17.351995 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs8cw\" (UniqueName: \"kubernetes.io/projected/5da06a58-941d-4405-8d81-3894f8366813-kube-api-access-xs8cw\") pod \"crc-debug-5xj98\" (UID: \"5da06a58-941d-4405-8d81-3894f8366813\") " pod="openshift-must-gather-7b7jq/crc-debug-5xj98" Nov 25 12:08:17 crc kubenswrapper[4776]: I1125 12:08:17.376744 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs8cw\" (UniqueName: \"kubernetes.io/projected/5da06a58-941d-4405-8d81-3894f8366813-kube-api-access-xs8cw\") pod \"crc-debug-5xj98\" (UID: \"5da06a58-941d-4405-8d81-3894f8366813\") " pod="openshift-must-gather-7b7jq/crc-debug-5xj98" Nov 25 12:08:17 crc kubenswrapper[4776]: I1125 12:08:17.490037 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b7jq/crc-debug-5xj98" Nov 25 12:08:17 crc kubenswrapper[4776]: W1125 12:08:17.535698 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5da06a58_941d_4405_8d81_3894f8366813.slice/crio-43d840cd863bf46cefe5f01f466f023afa86bb2df93a7f26ccc7641ab86b80c1 WatchSource:0}: Error finding container 43d840cd863bf46cefe5f01f466f023afa86bb2df93a7f26ccc7641ab86b80c1: Status 404 returned error can't find the container with id 43d840cd863bf46cefe5f01f466f023afa86bb2df93a7f26ccc7641ab86b80c1 Nov 25 12:08:17 crc kubenswrapper[4776]: I1125 12:08:17.824295 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7b7jq/crc-debug-5xj98" event={"ID":"5da06a58-941d-4405-8d81-3894f8366813","Type":"ContainerStarted","Data":"43d840cd863bf46cefe5f01f466f023afa86bb2df93a7f26ccc7641ab86b80c1"} Nov 25 12:08:26 crc kubenswrapper[4776]: I1125 12:08:26.662315 4776 scope.go:117] "RemoveContainer" containerID="b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d" Nov 25 12:08:26 crc kubenswrapper[4776]: E1125 12:08:26.663287 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:08:33 crc kubenswrapper[4776]: I1125 12:08:33.131318 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7b7jq/crc-debug-5xj98" event={"ID":"5da06a58-941d-4405-8d81-3894f8366813","Type":"ContainerStarted","Data":"85802b706fdad3ac3557bac1ed60645880cb1f8c7ea11d97d02f3461a00ecd51"} Nov 25 12:08:33 crc kubenswrapper[4776]: I1125 12:08:33.166703 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-7b7jq/crc-debug-5xj98" podStartSLOduration=1.278004997 podStartE2EDuration="16.166683444s" podCreationTimestamp="2025-11-25 12:08:17 +0000 UTC" firstStartedPulling="2025-11-25 12:08:17.542523184 +0000 UTC m=+9842.583582747" lastFinishedPulling="2025-11-25 12:08:32.431201641 +0000 UTC m=+9857.472261194" observedRunningTime="2025-11-25 12:08:33.156233651 +0000 UTC m=+9858.197293204" watchObservedRunningTime="2025-11-25 12:08:33.166683444 +0000 UTC m=+9858.207742997" Nov 25 12:08:41 crc kubenswrapper[4776]: I1125 12:08:41.662349 4776 scope.go:117] "RemoveContainer" containerID="b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d" Nov 25 12:08:41 crc kubenswrapper[4776]: E1125 12:08:41.664454 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:08:53 crc kubenswrapper[4776]: I1125 12:08:53.662375 4776 scope.go:117] "RemoveContainer" containerID="b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d" Nov 25 12:08:53 crc kubenswrapper[4776]: E1125 12:08:53.663336 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:09:07 crc kubenswrapper[4776]: I1125 12:09:07.663271 4776 scope.go:117] "RemoveContainer" containerID="b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d" Nov 25 12:09:07 crc kubenswrapper[4776]: E1125 12:09:07.664513 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:09:22 crc kubenswrapper[4776]: I1125 12:09:22.663261 4776 scope.go:117] "RemoveContainer" containerID="b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d" Nov 25 12:09:23 crc kubenswrapper[4776]: I1125 12:09:23.787577 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"5b13db60744f199580371de408731797095f5765187d672f3989cfe03fd42a7f"} Nov 25 12:09:25 crc kubenswrapper[4776]: I1125 12:09:25.811547 4776 generic.go:334] "Generic (PLEG): container finished" podID="5da06a58-941d-4405-8d81-3894f8366813" containerID="85802b706fdad3ac3557bac1ed60645880cb1f8c7ea11d97d02f3461a00ecd51" exitCode=0 Nov 25 12:09:25 crc kubenswrapper[4776]: I1125 12:09:25.812178 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7b7jq/crc-debug-5xj98" event={"ID":"5da06a58-941d-4405-8d81-3894f8366813","Type":"ContainerDied","Data":"85802b706fdad3ac3557bac1ed60645880cb1f8c7ea11d97d02f3461a00ecd51"} Nov 25 12:09:26 crc kubenswrapper[4776]: I1125 12:09:26.960008 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b7jq/crc-debug-5xj98" Nov 25 12:09:27 crc kubenswrapper[4776]: I1125 12:09:27.003598 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7b7jq/crc-debug-5xj98"] Nov 25 12:09:27 crc kubenswrapper[4776]: I1125 12:09:27.014410 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7b7jq/crc-debug-5xj98"] Nov 25 12:09:27 crc kubenswrapper[4776]: I1125 12:09:27.033582 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5da06a58-941d-4405-8d81-3894f8366813-host\") pod \"5da06a58-941d-4405-8d81-3894f8366813\" (UID: \"5da06a58-941d-4405-8d81-3894f8366813\") " Nov 25 12:09:27 crc kubenswrapper[4776]: I1125 12:09:27.033714 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5da06a58-941d-4405-8d81-3894f8366813-host" (OuterVolumeSpecName: "host") pod "5da06a58-941d-4405-8d81-3894f8366813" (UID: "5da06a58-941d-4405-8d81-3894f8366813"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:09:27 crc kubenswrapper[4776]: I1125 12:09:27.033969 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xs8cw\" (UniqueName: \"kubernetes.io/projected/5da06a58-941d-4405-8d81-3894f8366813-kube-api-access-xs8cw\") pod \"5da06a58-941d-4405-8d81-3894f8366813\" (UID: \"5da06a58-941d-4405-8d81-3894f8366813\") " Nov 25 12:09:27 crc kubenswrapper[4776]: I1125 12:09:27.034764 4776 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5da06a58-941d-4405-8d81-3894f8366813-host\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:27 crc kubenswrapper[4776]: I1125 12:09:27.042525 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5da06a58-941d-4405-8d81-3894f8366813-kube-api-access-xs8cw" (OuterVolumeSpecName: "kube-api-access-xs8cw") pod "5da06a58-941d-4405-8d81-3894f8366813" (UID: "5da06a58-941d-4405-8d81-3894f8366813"). InnerVolumeSpecName "kube-api-access-xs8cw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:27 crc kubenswrapper[4776]: I1125 12:09:27.136649 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xs8cw\" (UniqueName: \"kubernetes.io/projected/5da06a58-941d-4405-8d81-3894f8366813-kube-api-access-xs8cw\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:27 crc kubenswrapper[4776]: I1125 12:09:27.677256 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5da06a58-941d-4405-8d81-3894f8366813" path="/var/lib/kubelet/pods/5da06a58-941d-4405-8d81-3894f8366813/volumes" Nov 25 12:09:27 crc kubenswrapper[4776]: I1125 12:09:27.831903 4776 scope.go:117] "RemoveContainer" containerID="85802b706fdad3ac3557bac1ed60645880cb1f8c7ea11d97d02f3461a00ecd51" Nov 25 12:09:27 crc kubenswrapper[4776]: I1125 12:09:27.831959 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b7jq/crc-debug-5xj98" Nov 25 12:09:28 crc kubenswrapper[4776]: I1125 12:09:28.176971 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7b7jq/crc-debug-fj8x6"] Nov 25 12:09:28 crc kubenswrapper[4776]: E1125 12:09:28.178686 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5da06a58-941d-4405-8d81-3894f8366813" containerName="container-00" Nov 25 12:09:28 crc kubenswrapper[4776]: I1125 12:09:28.178817 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="5da06a58-941d-4405-8d81-3894f8366813" containerName="container-00" Nov 25 12:09:28 crc kubenswrapper[4776]: I1125 12:09:28.179230 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="5da06a58-941d-4405-8d81-3894f8366813" containerName="container-00" Nov 25 12:09:28 crc kubenswrapper[4776]: I1125 12:09:28.180496 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b7jq/crc-debug-fj8x6" Nov 25 12:09:28 crc kubenswrapper[4776]: I1125 12:09:28.367695 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6psmw\" (UniqueName: \"kubernetes.io/projected/4599ac89-1fca-4d1c-a6d4-704552d448fa-kube-api-access-6psmw\") pod \"crc-debug-fj8x6\" (UID: \"4599ac89-1fca-4d1c-a6d4-704552d448fa\") " pod="openshift-must-gather-7b7jq/crc-debug-fj8x6" Nov 25 12:09:28 crc kubenswrapper[4776]: I1125 12:09:28.368265 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4599ac89-1fca-4d1c-a6d4-704552d448fa-host\") pod \"crc-debug-fj8x6\" (UID: \"4599ac89-1fca-4d1c-a6d4-704552d448fa\") " pod="openshift-must-gather-7b7jq/crc-debug-fj8x6" Nov 25 12:09:28 crc kubenswrapper[4776]: I1125 12:09:28.470664 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6psmw\" (UniqueName: \"kubernetes.io/projected/4599ac89-1fca-4d1c-a6d4-704552d448fa-kube-api-access-6psmw\") pod \"crc-debug-fj8x6\" (UID: \"4599ac89-1fca-4d1c-a6d4-704552d448fa\") " pod="openshift-must-gather-7b7jq/crc-debug-fj8x6" Nov 25 12:09:28 crc kubenswrapper[4776]: I1125 12:09:28.471270 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4599ac89-1fca-4d1c-a6d4-704552d448fa-host\") pod \"crc-debug-fj8x6\" (UID: \"4599ac89-1fca-4d1c-a6d4-704552d448fa\") " pod="openshift-must-gather-7b7jq/crc-debug-fj8x6" Nov 25 12:09:28 crc kubenswrapper[4776]: I1125 12:09:28.471414 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4599ac89-1fca-4d1c-a6d4-704552d448fa-host\") pod \"crc-debug-fj8x6\" (UID: \"4599ac89-1fca-4d1c-a6d4-704552d448fa\") " pod="openshift-must-gather-7b7jq/crc-debug-fj8x6" Nov 25 12:09:28 crc kubenswrapper[4776]: I1125 12:09:28.491993 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6psmw\" (UniqueName: \"kubernetes.io/projected/4599ac89-1fca-4d1c-a6d4-704552d448fa-kube-api-access-6psmw\") pod \"crc-debug-fj8x6\" (UID: \"4599ac89-1fca-4d1c-a6d4-704552d448fa\") " pod="openshift-must-gather-7b7jq/crc-debug-fj8x6" Nov 25 12:09:28 crc kubenswrapper[4776]: I1125 12:09:28.499109 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b7jq/crc-debug-fj8x6" Nov 25 12:09:28 crc kubenswrapper[4776]: I1125 12:09:28.852455 4776 generic.go:334] "Generic (PLEG): container finished" podID="4599ac89-1fca-4d1c-a6d4-704552d448fa" containerID="44c56021b9793ef5cfa8517002bc26b67b04bc391f842ac0ec555488a20a5691" exitCode=0 Nov 25 12:09:28 crc kubenswrapper[4776]: I1125 12:09:28.852774 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7b7jq/crc-debug-fj8x6" event={"ID":"4599ac89-1fca-4d1c-a6d4-704552d448fa","Type":"ContainerDied","Data":"44c56021b9793ef5cfa8517002bc26b67b04bc391f842ac0ec555488a20a5691"} Nov 25 12:09:28 crc kubenswrapper[4776]: I1125 12:09:28.852809 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7b7jq/crc-debug-fj8x6" event={"ID":"4599ac89-1fca-4d1c-a6d4-704552d448fa","Type":"ContainerStarted","Data":"28eb18fe0ef37d0fa3319456ad705d8bfd6c84422f7f969b028016934288b29c"} Nov 25 12:09:29 crc kubenswrapper[4776]: I1125 12:09:29.321698 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7b7jq/crc-debug-fj8x6"] Nov 25 12:09:29 crc kubenswrapper[4776]: I1125 12:09:29.329692 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7b7jq/crc-debug-fj8x6"] Nov 25 12:09:29 crc kubenswrapper[4776]: I1125 12:09:29.978957 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b7jq/crc-debug-fj8x6" Nov 25 12:09:30 crc kubenswrapper[4776]: I1125 12:09:30.110963 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4599ac89-1fca-4d1c-a6d4-704552d448fa-host\") pod \"4599ac89-1fca-4d1c-a6d4-704552d448fa\" (UID: \"4599ac89-1fca-4d1c-a6d4-704552d448fa\") " Nov 25 12:09:30 crc kubenswrapper[4776]: I1125 12:09:30.111276 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6psmw\" (UniqueName: \"kubernetes.io/projected/4599ac89-1fca-4d1c-a6d4-704552d448fa-kube-api-access-6psmw\") pod \"4599ac89-1fca-4d1c-a6d4-704552d448fa\" (UID: \"4599ac89-1fca-4d1c-a6d4-704552d448fa\") " Nov 25 12:09:30 crc kubenswrapper[4776]: I1125 12:09:30.111625 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4599ac89-1fca-4d1c-a6d4-704552d448fa-host" (OuterVolumeSpecName: "host") pod "4599ac89-1fca-4d1c-a6d4-704552d448fa" (UID: "4599ac89-1fca-4d1c-a6d4-704552d448fa"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:09:30 crc kubenswrapper[4776]: I1125 12:09:30.112677 4776 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4599ac89-1fca-4d1c-a6d4-704552d448fa-host\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:30 crc kubenswrapper[4776]: I1125 12:09:30.122972 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4599ac89-1fca-4d1c-a6d4-704552d448fa-kube-api-access-6psmw" (OuterVolumeSpecName: "kube-api-access-6psmw") pod "4599ac89-1fca-4d1c-a6d4-704552d448fa" (UID: "4599ac89-1fca-4d1c-a6d4-704552d448fa"). InnerVolumeSpecName "kube-api-access-6psmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:30 crc kubenswrapper[4776]: I1125 12:09:30.215107 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6psmw\" (UniqueName: \"kubernetes.io/projected/4599ac89-1fca-4d1c-a6d4-704552d448fa-kube-api-access-6psmw\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:30 crc kubenswrapper[4776]: I1125 12:09:30.537208 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7b7jq/crc-debug-hlhvv"] Nov 25 12:09:30 crc kubenswrapper[4776]: E1125 12:09:30.537824 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4599ac89-1fca-4d1c-a6d4-704552d448fa" containerName="container-00" Nov 25 12:09:30 crc kubenswrapper[4776]: I1125 12:09:30.537842 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="4599ac89-1fca-4d1c-a6d4-704552d448fa" containerName="container-00" Nov 25 12:09:30 crc kubenswrapper[4776]: I1125 12:09:30.538115 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="4599ac89-1fca-4d1c-a6d4-704552d448fa" containerName="container-00" Nov 25 12:09:30 crc kubenswrapper[4776]: I1125 12:09:30.539150 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b7jq/crc-debug-hlhvv" Nov 25 12:09:30 crc kubenswrapper[4776]: I1125 12:09:30.624749 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6dc82b8c-19f0-4f97-9f47-f9ea444442bd-host\") pod \"crc-debug-hlhvv\" (UID: \"6dc82b8c-19f0-4f97-9f47-f9ea444442bd\") " pod="openshift-must-gather-7b7jq/crc-debug-hlhvv" Nov 25 12:09:30 crc kubenswrapper[4776]: I1125 12:09:30.624823 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcrrt\" (UniqueName: \"kubernetes.io/projected/6dc82b8c-19f0-4f97-9f47-f9ea444442bd-kube-api-access-hcrrt\") pod \"crc-debug-hlhvv\" (UID: \"6dc82b8c-19f0-4f97-9f47-f9ea444442bd\") " pod="openshift-must-gather-7b7jq/crc-debug-hlhvv" Nov 25 12:09:30 crc kubenswrapper[4776]: I1125 12:09:30.728126 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6dc82b8c-19f0-4f97-9f47-f9ea444442bd-host\") pod \"crc-debug-hlhvv\" (UID: \"6dc82b8c-19f0-4f97-9f47-f9ea444442bd\") " pod="openshift-must-gather-7b7jq/crc-debug-hlhvv" Nov 25 12:09:30 crc kubenswrapper[4776]: I1125 12:09:30.728204 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcrrt\" (UniqueName: \"kubernetes.io/projected/6dc82b8c-19f0-4f97-9f47-f9ea444442bd-kube-api-access-hcrrt\") pod \"crc-debug-hlhvv\" (UID: \"6dc82b8c-19f0-4f97-9f47-f9ea444442bd\") " pod="openshift-must-gather-7b7jq/crc-debug-hlhvv" Nov 25 12:09:30 crc kubenswrapper[4776]: I1125 12:09:30.728286 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6dc82b8c-19f0-4f97-9f47-f9ea444442bd-host\") pod \"crc-debug-hlhvv\" (UID: \"6dc82b8c-19f0-4f97-9f47-f9ea444442bd\") " pod="openshift-must-gather-7b7jq/crc-debug-hlhvv" Nov 25 12:09:30 crc kubenswrapper[4776]: I1125 12:09:30.750060 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcrrt\" (UniqueName: \"kubernetes.io/projected/6dc82b8c-19f0-4f97-9f47-f9ea444442bd-kube-api-access-hcrrt\") pod \"crc-debug-hlhvv\" (UID: \"6dc82b8c-19f0-4f97-9f47-f9ea444442bd\") " pod="openshift-must-gather-7b7jq/crc-debug-hlhvv" Nov 25 12:09:30 crc kubenswrapper[4776]: I1125 12:09:30.857728 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b7jq/crc-debug-hlhvv" Nov 25 12:09:30 crc kubenswrapper[4776]: I1125 12:09:30.875275 4776 scope.go:117] "RemoveContainer" containerID="44c56021b9793ef5cfa8517002bc26b67b04bc391f842ac0ec555488a20a5691" Nov 25 12:09:30 crc kubenswrapper[4776]: I1125 12:09:30.875406 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b7jq/crc-debug-fj8x6" Nov 25 12:09:30 crc kubenswrapper[4776]: W1125 12:09:30.893146 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6dc82b8c_19f0_4f97_9f47_f9ea444442bd.slice/crio-0f29a25fece5bdeccd8d8192c6481c5044480c635dc43406861fc6ca02610cc0 WatchSource:0}: Error finding container 0f29a25fece5bdeccd8d8192c6481c5044480c635dc43406861fc6ca02610cc0: Status 404 returned error can't find the container with id 0f29a25fece5bdeccd8d8192c6481c5044480c635dc43406861fc6ca02610cc0 Nov 25 12:09:31 crc kubenswrapper[4776]: I1125 12:09:31.690637 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4599ac89-1fca-4d1c-a6d4-704552d448fa" path="/var/lib/kubelet/pods/4599ac89-1fca-4d1c-a6d4-704552d448fa/volumes" Nov 25 12:09:31 crc kubenswrapper[4776]: I1125 12:09:31.889115 4776 generic.go:334] "Generic (PLEG): container finished" podID="6dc82b8c-19f0-4f97-9f47-f9ea444442bd" containerID="7ee1ebd8ac2f35cad1b0e8fe66216d5239f1d6e9b68dbec50cd76b3003dce442" exitCode=0 Nov 25 12:09:31 crc kubenswrapper[4776]: I1125 12:09:31.889212 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7b7jq/crc-debug-hlhvv" event={"ID":"6dc82b8c-19f0-4f97-9f47-f9ea444442bd","Type":"ContainerDied","Data":"7ee1ebd8ac2f35cad1b0e8fe66216d5239f1d6e9b68dbec50cd76b3003dce442"} Nov 25 12:09:31 crc kubenswrapper[4776]: I1125 12:09:31.889376 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7b7jq/crc-debug-hlhvv" event={"ID":"6dc82b8c-19f0-4f97-9f47-f9ea444442bd","Type":"ContainerStarted","Data":"0f29a25fece5bdeccd8d8192c6481c5044480c635dc43406861fc6ca02610cc0"} Nov 25 12:09:31 crc kubenswrapper[4776]: I1125 12:09:31.940704 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7b7jq/crc-debug-hlhvv"] Nov 25 12:09:31 crc kubenswrapper[4776]: I1125 12:09:31.953719 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7b7jq/crc-debug-hlhvv"] Nov 25 12:09:33 crc kubenswrapper[4776]: I1125 12:09:33.042804 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b7jq/crc-debug-hlhvv" Nov 25 12:09:33 crc kubenswrapper[4776]: I1125 12:09:33.186879 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6dc82b8c-19f0-4f97-9f47-f9ea444442bd-host\") pod \"6dc82b8c-19f0-4f97-9f47-f9ea444442bd\" (UID: \"6dc82b8c-19f0-4f97-9f47-f9ea444442bd\") " Nov 25 12:09:33 crc kubenswrapper[4776]: I1125 12:09:33.186960 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcrrt\" (UniqueName: \"kubernetes.io/projected/6dc82b8c-19f0-4f97-9f47-f9ea444442bd-kube-api-access-hcrrt\") pod \"6dc82b8c-19f0-4f97-9f47-f9ea444442bd\" (UID: \"6dc82b8c-19f0-4f97-9f47-f9ea444442bd\") " Nov 25 12:09:33 crc kubenswrapper[4776]: I1125 12:09:33.187336 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6dc82b8c-19f0-4f97-9f47-f9ea444442bd-host" (OuterVolumeSpecName: "host") pod "6dc82b8c-19f0-4f97-9f47-f9ea444442bd" (UID: "6dc82b8c-19f0-4f97-9f47-f9ea444442bd"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 12:09:33 crc kubenswrapper[4776]: I1125 12:09:33.187977 4776 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6dc82b8c-19f0-4f97-9f47-f9ea444442bd-host\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:33 crc kubenswrapper[4776]: I1125 12:09:33.192336 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dc82b8c-19f0-4f97-9f47-f9ea444442bd-kube-api-access-hcrrt" (OuterVolumeSpecName: "kube-api-access-hcrrt") pod "6dc82b8c-19f0-4f97-9f47-f9ea444442bd" (UID: "6dc82b8c-19f0-4f97-9f47-f9ea444442bd"). InnerVolumeSpecName "kube-api-access-hcrrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:09:33 crc kubenswrapper[4776]: I1125 12:09:33.289413 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcrrt\" (UniqueName: \"kubernetes.io/projected/6dc82b8c-19f0-4f97-9f47-f9ea444442bd-kube-api-access-hcrrt\") on node \"crc\" DevicePath \"\"" Nov 25 12:09:33 crc kubenswrapper[4776]: I1125 12:09:33.674233 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6dc82b8c-19f0-4f97-9f47-f9ea444442bd" path="/var/lib/kubelet/pods/6dc82b8c-19f0-4f97-9f47-f9ea444442bd/volumes" Nov 25 12:09:33 crc kubenswrapper[4776]: I1125 12:09:33.913499 4776 scope.go:117] "RemoveContainer" containerID="7ee1ebd8ac2f35cad1b0e8fe66216d5239f1d6e9b68dbec50cd76b3003dce442" Nov 25 12:09:33 crc kubenswrapper[4776]: I1125 12:09:33.913545 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b7jq/crc-debug-hlhvv" Nov 25 12:10:03 crc kubenswrapper[4776]: I1125 12:10:03.554918 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2b8wt"] Nov 25 12:10:03 crc kubenswrapper[4776]: E1125 12:10:03.556011 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dc82b8c-19f0-4f97-9f47-f9ea444442bd" containerName="container-00" Nov 25 12:10:03 crc kubenswrapper[4776]: I1125 12:10:03.556024 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dc82b8c-19f0-4f97-9f47-f9ea444442bd" containerName="container-00" Nov 25 12:10:03 crc kubenswrapper[4776]: I1125 12:10:03.556265 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dc82b8c-19f0-4f97-9f47-f9ea444442bd" containerName="container-00" Nov 25 12:10:03 crc kubenswrapper[4776]: I1125 12:10:03.558147 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2b8wt" Nov 25 12:10:03 crc kubenswrapper[4776]: I1125 12:10:03.573348 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2b8wt"] Nov 25 12:10:03 crc kubenswrapper[4776]: I1125 12:10:03.692925 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgt55\" (UniqueName: \"kubernetes.io/projected/2ff941eb-20b9-484d-9ba2-f28f83f78a64-kube-api-access-cgt55\") pod \"redhat-operators-2b8wt\" (UID: \"2ff941eb-20b9-484d-9ba2-f28f83f78a64\") " pod="openshift-marketplace/redhat-operators-2b8wt" Nov 25 12:10:03 crc kubenswrapper[4776]: I1125 12:10:03.694057 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ff941eb-20b9-484d-9ba2-f28f83f78a64-catalog-content\") pod \"redhat-operators-2b8wt\" (UID: \"2ff941eb-20b9-484d-9ba2-f28f83f78a64\") " pod="openshift-marketplace/redhat-operators-2b8wt" Nov 25 12:10:03 crc kubenswrapper[4776]: I1125 12:10:03.694344 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ff941eb-20b9-484d-9ba2-f28f83f78a64-utilities\") pod \"redhat-operators-2b8wt\" (UID: \"2ff941eb-20b9-484d-9ba2-f28f83f78a64\") " pod="openshift-marketplace/redhat-operators-2b8wt" Nov 25 12:10:03 crc kubenswrapper[4776]: I1125 12:10:03.800742 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ff941eb-20b9-484d-9ba2-f28f83f78a64-catalog-content\") pod \"redhat-operators-2b8wt\" (UID: \"2ff941eb-20b9-484d-9ba2-f28f83f78a64\") " pod="openshift-marketplace/redhat-operators-2b8wt" Nov 25 12:10:03 crc kubenswrapper[4776]: I1125 12:10:03.800883 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ff941eb-20b9-484d-9ba2-f28f83f78a64-utilities\") pod \"redhat-operators-2b8wt\" (UID: \"2ff941eb-20b9-484d-9ba2-f28f83f78a64\") " pod="openshift-marketplace/redhat-operators-2b8wt" Nov 25 12:10:03 crc kubenswrapper[4776]: I1125 12:10:03.801104 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgt55\" (UniqueName: \"kubernetes.io/projected/2ff941eb-20b9-484d-9ba2-f28f83f78a64-kube-api-access-cgt55\") pod \"redhat-operators-2b8wt\" (UID: \"2ff941eb-20b9-484d-9ba2-f28f83f78a64\") " pod="openshift-marketplace/redhat-operators-2b8wt" Nov 25 12:10:03 crc kubenswrapper[4776]: I1125 12:10:03.805450 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ff941eb-20b9-484d-9ba2-f28f83f78a64-catalog-content\") pod \"redhat-operators-2b8wt\" (UID: \"2ff941eb-20b9-484d-9ba2-f28f83f78a64\") " pod="openshift-marketplace/redhat-operators-2b8wt" Nov 25 12:10:03 crc kubenswrapper[4776]: I1125 12:10:03.805589 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ff941eb-20b9-484d-9ba2-f28f83f78a64-utilities\") pod \"redhat-operators-2b8wt\" (UID: \"2ff941eb-20b9-484d-9ba2-f28f83f78a64\") " pod="openshift-marketplace/redhat-operators-2b8wt" Nov 25 12:10:03 crc kubenswrapper[4776]: I1125 12:10:03.847652 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgt55\" (UniqueName: \"kubernetes.io/projected/2ff941eb-20b9-484d-9ba2-f28f83f78a64-kube-api-access-cgt55\") pod \"redhat-operators-2b8wt\" (UID: \"2ff941eb-20b9-484d-9ba2-f28f83f78a64\") " pod="openshift-marketplace/redhat-operators-2b8wt" Nov 25 12:10:03 crc kubenswrapper[4776]: I1125 12:10:03.896421 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2b8wt" Nov 25 12:10:04 crc kubenswrapper[4776]: I1125 12:10:04.455982 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2b8wt"] Nov 25 12:10:05 crc kubenswrapper[4776]: I1125 12:10:05.248875 4776 generic.go:334] "Generic (PLEG): container finished" podID="2ff941eb-20b9-484d-9ba2-f28f83f78a64" containerID="18068ce64b784c62fdcaaa84e94072ef876075318fc0f338461786a26a3f2879" exitCode=0 Nov 25 12:10:05 crc kubenswrapper[4776]: I1125 12:10:05.249206 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2b8wt" event={"ID":"2ff941eb-20b9-484d-9ba2-f28f83f78a64","Type":"ContainerDied","Data":"18068ce64b784c62fdcaaa84e94072ef876075318fc0f338461786a26a3f2879"} Nov 25 12:10:05 crc kubenswrapper[4776]: I1125 12:10:05.249317 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2b8wt" event={"ID":"2ff941eb-20b9-484d-9ba2-f28f83f78a64","Type":"ContainerStarted","Data":"c6d2793379993068906bfff036de41f1ee33c8865fa1b289d79a71078ff70fa0"} Nov 25 12:10:07 crc kubenswrapper[4776]: I1125 12:10:07.271287 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2b8wt" event={"ID":"2ff941eb-20b9-484d-9ba2-f28f83f78a64","Type":"ContainerStarted","Data":"71eb808c6d7c1a54e55ddabb5b148caac3edd240820d8b230c7c97649866e4dd"} Nov 25 12:10:08 crc kubenswrapper[4776]: I1125 12:10:08.757569 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8fl5l"] Nov 25 12:10:08 crc kubenswrapper[4776]: I1125 12:10:08.764030 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8fl5l" Nov 25 12:10:08 crc kubenswrapper[4776]: I1125 12:10:08.778659 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8fl5l"] Nov 25 12:10:08 crc kubenswrapper[4776]: I1125 12:10:08.930642 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlj9n\" (UniqueName: \"kubernetes.io/projected/6389f6ea-bc9b-43ac-9776-951fa13aae51-kube-api-access-mlj9n\") pod \"community-operators-8fl5l\" (UID: \"6389f6ea-bc9b-43ac-9776-951fa13aae51\") " pod="openshift-marketplace/community-operators-8fl5l" Nov 25 12:10:08 crc kubenswrapper[4776]: I1125 12:10:08.930848 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6389f6ea-bc9b-43ac-9776-951fa13aae51-utilities\") pod \"community-operators-8fl5l\" (UID: \"6389f6ea-bc9b-43ac-9776-951fa13aae51\") " pod="openshift-marketplace/community-operators-8fl5l" Nov 25 12:10:08 crc kubenswrapper[4776]: I1125 12:10:08.930911 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6389f6ea-bc9b-43ac-9776-951fa13aae51-catalog-content\") pod \"community-operators-8fl5l\" (UID: \"6389f6ea-bc9b-43ac-9776-951fa13aae51\") " pod="openshift-marketplace/community-operators-8fl5l" Nov 25 12:10:09 crc kubenswrapper[4776]: I1125 12:10:09.032636 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6389f6ea-bc9b-43ac-9776-951fa13aae51-catalog-content\") pod \"community-operators-8fl5l\" (UID: \"6389f6ea-bc9b-43ac-9776-951fa13aae51\") " pod="openshift-marketplace/community-operators-8fl5l" Nov 25 12:10:09 crc kubenswrapper[4776]: I1125 12:10:09.032730 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlj9n\" (UniqueName: \"kubernetes.io/projected/6389f6ea-bc9b-43ac-9776-951fa13aae51-kube-api-access-mlj9n\") pod \"community-operators-8fl5l\" (UID: \"6389f6ea-bc9b-43ac-9776-951fa13aae51\") " pod="openshift-marketplace/community-operators-8fl5l" Nov 25 12:10:09 crc kubenswrapper[4776]: I1125 12:10:09.032863 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6389f6ea-bc9b-43ac-9776-951fa13aae51-utilities\") pod \"community-operators-8fl5l\" (UID: \"6389f6ea-bc9b-43ac-9776-951fa13aae51\") " pod="openshift-marketplace/community-operators-8fl5l" Nov 25 12:10:09 crc kubenswrapper[4776]: I1125 12:10:09.033501 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6389f6ea-bc9b-43ac-9776-951fa13aae51-utilities\") pod \"community-operators-8fl5l\" (UID: \"6389f6ea-bc9b-43ac-9776-951fa13aae51\") " pod="openshift-marketplace/community-operators-8fl5l" Nov 25 12:10:09 crc kubenswrapper[4776]: I1125 12:10:09.033625 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6389f6ea-bc9b-43ac-9776-951fa13aae51-catalog-content\") pod \"community-operators-8fl5l\" (UID: \"6389f6ea-bc9b-43ac-9776-951fa13aae51\") " pod="openshift-marketplace/community-operators-8fl5l" Nov 25 12:10:09 crc kubenswrapper[4776]: I1125 12:10:09.089563 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlj9n\" (UniqueName: \"kubernetes.io/projected/6389f6ea-bc9b-43ac-9776-951fa13aae51-kube-api-access-mlj9n\") pod \"community-operators-8fl5l\" (UID: \"6389f6ea-bc9b-43ac-9776-951fa13aae51\") " pod="openshift-marketplace/community-operators-8fl5l" Nov 25 12:10:09 crc kubenswrapper[4776]: I1125 12:10:09.141779 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8fl5l" Nov 25 12:10:09 crc kubenswrapper[4776]: I1125 12:10:09.782090 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8fl5l"] Nov 25 12:10:10 crc kubenswrapper[4776]: I1125 12:10:10.311865 4776 generic.go:334] "Generic (PLEG): container finished" podID="6389f6ea-bc9b-43ac-9776-951fa13aae51" containerID="16ce82dded4127672e717dbdf4d54ca5f0f552274e50da5bf36dbffc5ebdce79" exitCode=0 Nov 25 12:10:10 crc kubenswrapper[4776]: I1125 12:10:10.311957 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8fl5l" event={"ID":"6389f6ea-bc9b-43ac-9776-951fa13aae51","Type":"ContainerDied","Data":"16ce82dded4127672e717dbdf4d54ca5f0f552274e50da5bf36dbffc5ebdce79"} Nov 25 12:10:10 crc kubenswrapper[4776]: I1125 12:10:10.312276 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8fl5l" event={"ID":"6389f6ea-bc9b-43ac-9776-951fa13aae51","Type":"ContainerStarted","Data":"21823353093b00093800be49513ab26dad5b8308d7b31a961c6ccf3e9a4dc412"} Nov 25 12:10:11 crc kubenswrapper[4776]: I1125 12:10:11.335614 4776 generic.go:334] "Generic (PLEG): container finished" podID="2ff941eb-20b9-484d-9ba2-f28f83f78a64" containerID="71eb808c6d7c1a54e55ddabb5b148caac3edd240820d8b230c7c97649866e4dd" exitCode=0 Nov 25 12:10:11 crc kubenswrapper[4776]: I1125 12:10:11.335725 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2b8wt" event={"ID":"2ff941eb-20b9-484d-9ba2-f28f83f78a64","Type":"ContainerDied","Data":"71eb808c6d7c1a54e55ddabb5b148caac3edd240820d8b230c7c97649866e4dd"} Nov 25 12:10:12 crc kubenswrapper[4776]: I1125 12:10:12.349737 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8fl5l" event={"ID":"6389f6ea-bc9b-43ac-9776-951fa13aae51","Type":"ContainerStarted","Data":"6948579f170d9044c99736373dddba0666920268685434c069836691560f6f8b"} Nov 25 12:10:14 crc kubenswrapper[4776]: I1125 12:10:14.405299 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2b8wt" event={"ID":"2ff941eb-20b9-484d-9ba2-f28f83f78a64","Type":"ContainerStarted","Data":"26958d41c1cc9fed1ab7cde4c19375f83c02512d32ba42c7fe2402f0220e84bc"} Nov 25 12:10:14 crc kubenswrapper[4776]: I1125 12:10:14.430647 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2b8wt" podStartSLOduration=3.433994641 podStartE2EDuration="11.430620893s" podCreationTimestamp="2025-11-25 12:10:03 +0000 UTC" firstStartedPulling="2025-11-25 12:10:05.25475865 +0000 UTC m=+9950.295818203" lastFinishedPulling="2025-11-25 12:10:13.251384902 +0000 UTC m=+9958.292444455" observedRunningTime="2025-11-25 12:10:14.428202712 +0000 UTC m=+9959.469262275" watchObservedRunningTime="2025-11-25 12:10:14.430620893 +0000 UTC m=+9959.471680446" Nov 25 12:10:15 crc kubenswrapper[4776]: I1125 12:10:15.421034 4776 generic.go:334] "Generic (PLEG): container finished" podID="6389f6ea-bc9b-43ac-9776-951fa13aae51" containerID="6948579f170d9044c99736373dddba0666920268685434c069836691560f6f8b" exitCode=0 Nov 25 12:10:15 crc kubenswrapper[4776]: I1125 12:10:15.421127 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8fl5l" event={"ID":"6389f6ea-bc9b-43ac-9776-951fa13aae51","Type":"ContainerDied","Data":"6948579f170d9044c99736373dddba0666920268685434c069836691560f6f8b"} Nov 25 12:10:16 crc kubenswrapper[4776]: I1125 12:10:16.457300 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8fl5l" event={"ID":"6389f6ea-bc9b-43ac-9776-951fa13aae51","Type":"ContainerStarted","Data":"ba8d6374a435ba283c388a4709bc6ba06e6f86aa034c00f650e697b7114f24a4"} Nov 25 12:10:16 crc kubenswrapper[4776]: I1125 12:10:16.492734 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8fl5l" podStartSLOduration=2.9592909450000002 podStartE2EDuration="8.492707038s" podCreationTimestamp="2025-11-25 12:10:08 +0000 UTC" firstStartedPulling="2025-11-25 12:10:10.315763177 +0000 UTC m=+9955.356822730" lastFinishedPulling="2025-11-25 12:10:15.84917927 +0000 UTC m=+9960.890238823" observedRunningTime="2025-11-25 12:10:16.488905683 +0000 UTC m=+9961.529965236" watchObservedRunningTime="2025-11-25 12:10:16.492707038 +0000 UTC m=+9961.533766591" Nov 25 12:10:19 crc kubenswrapper[4776]: I1125 12:10:19.142319 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8fl5l" Nov 25 12:10:19 crc kubenswrapper[4776]: I1125 12:10:19.143192 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8fl5l" Nov 25 12:10:19 crc kubenswrapper[4776]: I1125 12:10:19.271355 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8fl5l" Nov 25 12:10:22 crc kubenswrapper[4776]: I1125 12:10:22.868487 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-5c6d9c7954-pvk2n" podUID="0dc0a5c5-9aa1-4190-985f-047a9406c24a" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Nov 25 12:10:23 crc kubenswrapper[4776]: I1125 12:10:23.897483 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2b8wt" Nov 25 12:10:23 crc kubenswrapper[4776]: I1125 12:10:23.897970 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2b8wt" Nov 25 12:10:24 crc kubenswrapper[4776]: I1125 12:10:24.958627 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2b8wt" podUID="2ff941eb-20b9-484d-9ba2-f28f83f78a64" containerName="registry-server" probeResult="failure" output=< Nov 25 12:10:24 crc kubenswrapper[4776]: timeout: failed to connect service ":50051" within 1s Nov 25 12:10:24 crc kubenswrapper[4776]: > Nov 25 12:10:29 crc kubenswrapper[4776]: I1125 12:10:29.201840 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8fl5l" Nov 25 12:10:29 crc kubenswrapper[4776]: I1125 12:10:29.260998 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8fl5l"] Nov 25 12:10:29 crc kubenswrapper[4776]: I1125 12:10:29.609226 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8fl5l" podUID="6389f6ea-bc9b-43ac-9776-951fa13aae51" containerName="registry-server" containerID="cri-o://ba8d6374a435ba283c388a4709bc6ba06e6f86aa034c00f650e697b7114f24a4" gracePeriod=2 Nov 25 12:10:30 crc kubenswrapper[4776]: I1125 12:10:30.177788 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8fl5l" Nov 25 12:10:30 crc kubenswrapper[4776]: I1125 12:10:30.353213 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlj9n\" (UniqueName: \"kubernetes.io/projected/6389f6ea-bc9b-43ac-9776-951fa13aae51-kube-api-access-mlj9n\") pod \"6389f6ea-bc9b-43ac-9776-951fa13aae51\" (UID: \"6389f6ea-bc9b-43ac-9776-951fa13aae51\") " Nov 25 12:10:30 crc kubenswrapper[4776]: I1125 12:10:30.353443 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6389f6ea-bc9b-43ac-9776-951fa13aae51-catalog-content\") pod \"6389f6ea-bc9b-43ac-9776-951fa13aae51\" (UID: \"6389f6ea-bc9b-43ac-9776-951fa13aae51\") " Nov 25 12:10:30 crc kubenswrapper[4776]: I1125 12:10:30.353539 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6389f6ea-bc9b-43ac-9776-951fa13aae51-utilities\") pod \"6389f6ea-bc9b-43ac-9776-951fa13aae51\" (UID: \"6389f6ea-bc9b-43ac-9776-951fa13aae51\") " Nov 25 12:10:30 crc kubenswrapper[4776]: I1125 12:10:30.354634 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6389f6ea-bc9b-43ac-9776-951fa13aae51-utilities" (OuterVolumeSpecName: "utilities") pod "6389f6ea-bc9b-43ac-9776-951fa13aae51" (UID: "6389f6ea-bc9b-43ac-9776-951fa13aae51"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:10:30 crc kubenswrapper[4776]: I1125 12:10:30.361650 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6389f6ea-bc9b-43ac-9776-951fa13aae51-kube-api-access-mlj9n" (OuterVolumeSpecName: "kube-api-access-mlj9n") pod "6389f6ea-bc9b-43ac-9776-951fa13aae51" (UID: "6389f6ea-bc9b-43ac-9776-951fa13aae51"). InnerVolumeSpecName "kube-api-access-mlj9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:10:30 crc kubenswrapper[4776]: I1125 12:10:30.430973 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6389f6ea-bc9b-43ac-9776-951fa13aae51-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6389f6ea-bc9b-43ac-9776-951fa13aae51" (UID: "6389f6ea-bc9b-43ac-9776-951fa13aae51"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:10:30 crc kubenswrapper[4776]: I1125 12:10:30.457735 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6389f6ea-bc9b-43ac-9776-951fa13aae51-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:10:30 crc kubenswrapper[4776]: I1125 12:10:30.458118 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6389f6ea-bc9b-43ac-9776-951fa13aae51-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:10:30 crc kubenswrapper[4776]: I1125 12:10:30.458192 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlj9n\" (UniqueName: \"kubernetes.io/projected/6389f6ea-bc9b-43ac-9776-951fa13aae51-kube-api-access-mlj9n\") on node \"crc\" DevicePath \"\"" Nov 25 12:10:30 crc kubenswrapper[4776]: I1125 12:10:30.669599 4776 generic.go:334] "Generic (PLEG): container finished" podID="6389f6ea-bc9b-43ac-9776-951fa13aae51" containerID="ba8d6374a435ba283c388a4709bc6ba06e6f86aa034c00f650e697b7114f24a4" exitCode=0 Nov 25 12:10:30 crc kubenswrapper[4776]: I1125 12:10:30.669675 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8fl5l" event={"ID":"6389f6ea-bc9b-43ac-9776-951fa13aae51","Type":"ContainerDied","Data":"ba8d6374a435ba283c388a4709bc6ba06e6f86aa034c00f650e697b7114f24a4"} Nov 25 12:10:30 crc kubenswrapper[4776]: I1125 12:10:30.669724 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8fl5l" event={"ID":"6389f6ea-bc9b-43ac-9776-951fa13aae51","Type":"ContainerDied","Data":"21823353093b00093800be49513ab26dad5b8308d7b31a961c6ccf3e9a4dc412"} Nov 25 12:10:30 crc kubenswrapper[4776]: I1125 12:10:30.669757 4776 scope.go:117] "RemoveContainer" containerID="ba8d6374a435ba283c388a4709bc6ba06e6f86aa034c00f650e697b7114f24a4" Nov 25 12:10:30 crc kubenswrapper[4776]: I1125 12:10:30.670118 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8fl5l" Nov 25 12:10:30 crc kubenswrapper[4776]: I1125 12:10:30.697315 4776 scope.go:117] "RemoveContainer" containerID="6948579f170d9044c99736373dddba0666920268685434c069836691560f6f8b" Nov 25 12:10:30 crc kubenswrapper[4776]: I1125 12:10:30.721451 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8fl5l"] Nov 25 12:10:30 crc kubenswrapper[4776]: I1125 12:10:30.730039 4776 scope.go:117] "RemoveContainer" containerID="16ce82dded4127672e717dbdf4d54ca5f0f552274e50da5bf36dbffc5ebdce79" Nov 25 12:10:30 crc kubenswrapper[4776]: I1125 12:10:30.732164 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8fl5l"] Nov 25 12:10:30 crc kubenswrapper[4776]: I1125 12:10:30.801056 4776 scope.go:117] "RemoveContainer" containerID="ba8d6374a435ba283c388a4709bc6ba06e6f86aa034c00f650e697b7114f24a4" Nov 25 12:10:30 crc kubenswrapper[4776]: E1125 12:10:30.802949 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba8d6374a435ba283c388a4709bc6ba06e6f86aa034c00f650e697b7114f24a4\": container with ID starting with ba8d6374a435ba283c388a4709bc6ba06e6f86aa034c00f650e697b7114f24a4 not found: ID does not exist" containerID="ba8d6374a435ba283c388a4709bc6ba06e6f86aa034c00f650e697b7114f24a4" Nov 25 12:10:30 crc kubenswrapper[4776]: I1125 12:10:30.802986 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba8d6374a435ba283c388a4709bc6ba06e6f86aa034c00f650e697b7114f24a4"} err="failed to get container status \"ba8d6374a435ba283c388a4709bc6ba06e6f86aa034c00f650e697b7114f24a4\": rpc error: code = NotFound desc = could not find container \"ba8d6374a435ba283c388a4709bc6ba06e6f86aa034c00f650e697b7114f24a4\": container with ID starting with ba8d6374a435ba283c388a4709bc6ba06e6f86aa034c00f650e697b7114f24a4 not found: ID does not exist" Nov 25 12:10:30 crc kubenswrapper[4776]: I1125 12:10:30.803009 4776 scope.go:117] "RemoveContainer" containerID="6948579f170d9044c99736373dddba0666920268685434c069836691560f6f8b" Nov 25 12:10:30 crc kubenswrapper[4776]: E1125 12:10:30.811122 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6948579f170d9044c99736373dddba0666920268685434c069836691560f6f8b\": container with ID starting with 6948579f170d9044c99736373dddba0666920268685434c069836691560f6f8b not found: ID does not exist" containerID="6948579f170d9044c99736373dddba0666920268685434c069836691560f6f8b" Nov 25 12:10:30 crc kubenswrapper[4776]: I1125 12:10:30.811187 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6948579f170d9044c99736373dddba0666920268685434c069836691560f6f8b"} err="failed to get container status \"6948579f170d9044c99736373dddba0666920268685434c069836691560f6f8b\": rpc error: code = NotFound desc = could not find container \"6948579f170d9044c99736373dddba0666920268685434c069836691560f6f8b\": container with ID starting with 6948579f170d9044c99736373dddba0666920268685434c069836691560f6f8b not found: ID does not exist" Nov 25 12:10:30 crc kubenswrapper[4776]: I1125 12:10:30.811220 4776 scope.go:117] "RemoveContainer" containerID="16ce82dded4127672e717dbdf4d54ca5f0f552274e50da5bf36dbffc5ebdce79" Nov 25 12:10:30 crc kubenswrapper[4776]: E1125 12:10:30.811964 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16ce82dded4127672e717dbdf4d54ca5f0f552274e50da5bf36dbffc5ebdce79\": container with ID starting with 16ce82dded4127672e717dbdf4d54ca5f0f552274e50da5bf36dbffc5ebdce79 not found: ID does not exist" containerID="16ce82dded4127672e717dbdf4d54ca5f0f552274e50da5bf36dbffc5ebdce79" Nov 25 12:10:30 crc kubenswrapper[4776]: I1125 12:10:30.812049 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16ce82dded4127672e717dbdf4d54ca5f0f552274e50da5bf36dbffc5ebdce79"} err="failed to get container status \"16ce82dded4127672e717dbdf4d54ca5f0f552274e50da5bf36dbffc5ebdce79\": rpc error: code = NotFound desc = could not find container \"16ce82dded4127672e717dbdf4d54ca5f0f552274e50da5bf36dbffc5ebdce79\": container with ID starting with 16ce82dded4127672e717dbdf4d54ca5f0f552274e50da5bf36dbffc5ebdce79 not found: ID does not exist" Nov 25 12:10:31 crc kubenswrapper[4776]: I1125 12:10:31.682898 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6389f6ea-bc9b-43ac-9776-951fa13aae51" path="/var/lib/kubelet/pods/6389f6ea-bc9b-43ac-9776-951fa13aae51/volumes" Nov 25 12:10:33 crc kubenswrapper[4776]: I1125 12:10:33.962637 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2b8wt" Nov 25 12:10:34 crc kubenswrapper[4776]: I1125 12:10:34.030517 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2b8wt" Nov 25 12:10:34 crc kubenswrapper[4776]: I1125 12:10:34.843238 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2b8wt"] Nov 25 12:10:35 crc kubenswrapper[4776]: I1125 12:10:35.730863 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2b8wt" podUID="2ff941eb-20b9-484d-9ba2-f28f83f78a64" containerName="registry-server" containerID="cri-o://26958d41c1cc9fed1ab7cde4c19375f83c02512d32ba42c7fe2402f0220e84bc" gracePeriod=2 Nov 25 12:10:36 crc kubenswrapper[4776]: I1125 12:10:36.232118 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2b8wt" Nov 25 12:10:36 crc kubenswrapper[4776]: I1125 12:10:36.256711 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgt55\" (UniqueName: \"kubernetes.io/projected/2ff941eb-20b9-484d-9ba2-f28f83f78a64-kube-api-access-cgt55\") pod \"2ff941eb-20b9-484d-9ba2-f28f83f78a64\" (UID: \"2ff941eb-20b9-484d-9ba2-f28f83f78a64\") " Nov 25 12:10:36 crc kubenswrapper[4776]: I1125 12:10:36.256919 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ff941eb-20b9-484d-9ba2-f28f83f78a64-utilities\") pod \"2ff941eb-20b9-484d-9ba2-f28f83f78a64\" (UID: \"2ff941eb-20b9-484d-9ba2-f28f83f78a64\") " Nov 25 12:10:36 crc kubenswrapper[4776]: I1125 12:10:36.256995 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ff941eb-20b9-484d-9ba2-f28f83f78a64-catalog-content\") pod \"2ff941eb-20b9-484d-9ba2-f28f83f78a64\" (UID: \"2ff941eb-20b9-484d-9ba2-f28f83f78a64\") " Nov 25 12:10:36 crc kubenswrapper[4776]: I1125 12:10:36.257746 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ff941eb-20b9-484d-9ba2-f28f83f78a64-utilities" (OuterVolumeSpecName: "utilities") pod "2ff941eb-20b9-484d-9ba2-f28f83f78a64" (UID: "2ff941eb-20b9-484d-9ba2-f28f83f78a64"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:10:36 crc kubenswrapper[4776]: I1125 12:10:36.359146 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ff941eb-20b9-484d-9ba2-f28f83f78a64-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:10:36 crc kubenswrapper[4776]: I1125 12:10:36.367484 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ff941eb-20b9-484d-9ba2-f28f83f78a64-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2ff941eb-20b9-484d-9ba2-f28f83f78a64" (UID: "2ff941eb-20b9-484d-9ba2-f28f83f78a64"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:10:36 crc kubenswrapper[4776]: I1125 12:10:36.461346 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ff941eb-20b9-484d-9ba2-f28f83f78a64-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:10:36 crc kubenswrapper[4776]: I1125 12:10:36.744664 4776 generic.go:334] "Generic (PLEG): container finished" podID="2ff941eb-20b9-484d-9ba2-f28f83f78a64" containerID="26958d41c1cc9fed1ab7cde4c19375f83c02512d32ba42c7fe2402f0220e84bc" exitCode=0 Nov 25 12:10:36 crc kubenswrapper[4776]: I1125 12:10:36.744712 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2b8wt" event={"ID":"2ff941eb-20b9-484d-9ba2-f28f83f78a64","Type":"ContainerDied","Data":"26958d41c1cc9fed1ab7cde4c19375f83c02512d32ba42c7fe2402f0220e84bc"} Nov 25 12:10:36 crc kubenswrapper[4776]: I1125 12:10:36.744794 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2b8wt" event={"ID":"2ff941eb-20b9-484d-9ba2-f28f83f78a64","Type":"ContainerDied","Data":"c6d2793379993068906bfff036de41f1ee33c8865fa1b289d79a71078ff70fa0"} Nov 25 12:10:36 crc kubenswrapper[4776]: I1125 12:10:36.744821 4776 scope.go:117] "RemoveContainer" containerID="26958d41c1cc9fed1ab7cde4c19375f83c02512d32ba42c7fe2402f0220e84bc" Nov 25 12:10:36 crc kubenswrapper[4776]: I1125 12:10:36.744746 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2b8wt" Nov 25 12:10:36 crc kubenswrapper[4776]: I1125 12:10:36.793332 4776 scope.go:117] "RemoveContainer" containerID="71eb808c6d7c1a54e55ddabb5b148caac3edd240820d8b230c7c97649866e4dd" Nov 25 12:10:36 crc kubenswrapper[4776]: I1125 12:10:36.973543 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ff941eb-20b9-484d-9ba2-f28f83f78a64-kube-api-access-cgt55" (OuterVolumeSpecName: "kube-api-access-cgt55") pod "2ff941eb-20b9-484d-9ba2-f28f83f78a64" (UID: "2ff941eb-20b9-484d-9ba2-f28f83f78a64"). InnerVolumeSpecName "kube-api-access-cgt55". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:10:36 crc kubenswrapper[4776]: I1125 12:10:36.976421 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgt55\" (UniqueName: \"kubernetes.io/projected/2ff941eb-20b9-484d-9ba2-f28f83f78a64-kube-api-access-cgt55\") on node \"crc\" DevicePath \"\"" Nov 25 12:10:37 crc kubenswrapper[4776]: I1125 12:10:37.001387 4776 scope.go:117] "RemoveContainer" containerID="18068ce64b784c62fdcaaa84e94072ef876075318fc0f338461786a26a3f2879" Nov 25 12:10:37 crc kubenswrapper[4776]: I1125 12:10:37.106887 4776 scope.go:117] "RemoveContainer" containerID="26958d41c1cc9fed1ab7cde4c19375f83c02512d32ba42c7fe2402f0220e84bc" Nov 25 12:10:37 crc kubenswrapper[4776]: E1125 12:10:37.107752 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26958d41c1cc9fed1ab7cde4c19375f83c02512d32ba42c7fe2402f0220e84bc\": container with ID starting with 26958d41c1cc9fed1ab7cde4c19375f83c02512d32ba42c7fe2402f0220e84bc not found: ID does not exist" containerID="26958d41c1cc9fed1ab7cde4c19375f83c02512d32ba42c7fe2402f0220e84bc" Nov 25 12:10:37 crc kubenswrapper[4776]: I1125 12:10:37.107806 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26958d41c1cc9fed1ab7cde4c19375f83c02512d32ba42c7fe2402f0220e84bc"} err="failed to get container status \"26958d41c1cc9fed1ab7cde4c19375f83c02512d32ba42c7fe2402f0220e84bc\": rpc error: code = NotFound desc = could not find container \"26958d41c1cc9fed1ab7cde4c19375f83c02512d32ba42c7fe2402f0220e84bc\": container with ID starting with 26958d41c1cc9fed1ab7cde4c19375f83c02512d32ba42c7fe2402f0220e84bc not found: ID does not exist" Nov 25 12:10:37 crc kubenswrapper[4776]: I1125 12:10:37.107837 4776 scope.go:117] "RemoveContainer" containerID="71eb808c6d7c1a54e55ddabb5b148caac3edd240820d8b230c7c97649866e4dd" Nov 25 12:10:37 crc kubenswrapper[4776]: E1125 12:10:37.108172 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71eb808c6d7c1a54e55ddabb5b148caac3edd240820d8b230c7c97649866e4dd\": container with ID starting with 71eb808c6d7c1a54e55ddabb5b148caac3edd240820d8b230c7c97649866e4dd not found: ID does not exist" containerID="71eb808c6d7c1a54e55ddabb5b148caac3edd240820d8b230c7c97649866e4dd" Nov 25 12:10:37 crc kubenswrapper[4776]: I1125 12:10:37.108207 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71eb808c6d7c1a54e55ddabb5b148caac3edd240820d8b230c7c97649866e4dd"} err="failed to get container status \"71eb808c6d7c1a54e55ddabb5b148caac3edd240820d8b230c7c97649866e4dd\": rpc error: code = NotFound desc = could not find container \"71eb808c6d7c1a54e55ddabb5b148caac3edd240820d8b230c7c97649866e4dd\": container with ID starting with 71eb808c6d7c1a54e55ddabb5b148caac3edd240820d8b230c7c97649866e4dd not found: ID does not exist" Nov 25 12:10:37 crc kubenswrapper[4776]: I1125 12:10:37.108244 4776 scope.go:117] "RemoveContainer" containerID="18068ce64b784c62fdcaaa84e94072ef876075318fc0f338461786a26a3f2879" Nov 25 12:10:37 crc kubenswrapper[4776]: E1125 12:10:37.108478 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18068ce64b784c62fdcaaa84e94072ef876075318fc0f338461786a26a3f2879\": container with ID starting with 18068ce64b784c62fdcaaa84e94072ef876075318fc0f338461786a26a3f2879 not found: ID does not exist" containerID="18068ce64b784c62fdcaaa84e94072ef876075318fc0f338461786a26a3f2879" Nov 25 12:10:37 crc kubenswrapper[4776]: I1125 12:10:37.108508 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18068ce64b784c62fdcaaa84e94072ef876075318fc0f338461786a26a3f2879"} err="failed to get container status \"18068ce64b784c62fdcaaa84e94072ef876075318fc0f338461786a26a3f2879\": rpc error: code = NotFound desc = could not find container \"18068ce64b784c62fdcaaa84e94072ef876075318fc0f338461786a26a3f2879\": container with ID starting with 18068ce64b784c62fdcaaa84e94072ef876075318fc0f338461786a26a3f2879 not found: ID does not exist" Nov 25 12:10:37 crc kubenswrapper[4776]: I1125 12:10:37.115664 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2b8wt"] Nov 25 12:10:37 crc kubenswrapper[4776]: I1125 12:10:37.126456 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2b8wt"] Nov 25 12:10:37 crc kubenswrapper[4776]: I1125 12:10:37.674453 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ff941eb-20b9-484d-9ba2-f28f83f78a64" path="/var/lib/kubelet/pods/2ff941eb-20b9-484d-9ba2-f28f83f78a64/volumes" Nov 25 12:11:07 crc kubenswrapper[4776]: I1125 12:11:07.510430 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jkpql"] Nov 25 12:11:07 crc kubenswrapper[4776]: E1125 12:11:07.511597 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6389f6ea-bc9b-43ac-9776-951fa13aae51" containerName="registry-server" Nov 25 12:11:07 crc kubenswrapper[4776]: I1125 12:11:07.511616 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="6389f6ea-bc9b-43ac-9776-951fa13aae51" containerName="registry-server" Nov 25 12:11:07 crc kubenswrapper[4776]: E1125 12:11:07.511650 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff941eb-20b9-484d-9ba2-f28f83f78a64" containerName="registry-server" Nov 25 12:11:07 crc kubenswrapper[4776]: I1125 12:11:07.511656 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff941eb-20b9-484d-9ba2-f28f83f78a64" containerName="registry-server" Nov 25 12:11:07 crc kubenswrapper[4776]: E1125 12:11:07.511664 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff941eb-20b9-484d-9ba2-f28f83f78a64" containerName="extract-utilities" Nov 25 12:11:07 crc kubenswrapper[4776]: I1125 12:11:07.511670 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff941eb-20b9-484d-9ba2-f28f83f78a64" containerName="extract-utilities" Nov 25 12:11:07 crc kubenswrapper[4776]: E1125 12:11:07.511691 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6389f6ea-bc9b-43ac-9776-951fa13aae51" containerName="extract-content" Nov 25 12:11:07 crc kubenswrapper[4776]: I1125 12:11:07.511697 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="6389f6ea-bc9b-43ac-9776-951fa13aae51" containerName="extract-content" Nov 25 12:11:07 crc kubenswrapper[4776]: E1125 12:11:07.511707 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ff941eb-20b9-484d-9ba2-f28f83f78a64" containerName="extract-content" Nov 25 12:11:07 crc kubenswrapper[4776]: I1125 12:11:07.511714 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ff941eb-20b9-484d-9ba2-f28f83f78a64" containerName="extract-content" Nov 25 12:11:07 crc kubenswrapper[4776]: E1125 12:11:07.511723 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6389f6ea-bc9b-43ac-9776-951fa13aae51" containerName="extract-utilities" Nov 25 12:11:07 crc kubenswrapper[4776]: I1125 12:11:07.511729 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="6389f6ea-bc9b-43ac-9776-951fa13aae51" containerName="extract-utilities" Nov 25 12:11:07 crc kubenswrapper[4776]: I1125 12:11:07.511973 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="6389f6ea-bc9b-43ac-9776-951fa13aae51" containerName="registry-server" Nov 25 12:11:07 crc kubenswrapper[4776]: I1125 12:11:07.511993 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ff941eb-20b9-484d-9ba2-f28f83f78a64" containerName="registry-server" Nov 25 12:11:07 crc kubenswrapper[4776]: I1125 12:11:07.513814 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jkpql" Nov 25 12:11:07 crc kubenswrapper[4776]: I1125 12:11:07.523804 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jkpql"] Nov 25 12:11:07 crc kubenswrapper[4776]: I1125 12:11:07.700735 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b-catalog-content\") pod \"certified-operators-jkpql\" (UID: \"5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b\") " pod="openshift-marketplace/certified-operators-jkpql" Nov 25 12:11:07 crc kubenswrapper[4776]: I1125 12:11:07.700870 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42zgq\" (UniqueName: \"kubernetes.io/projected/5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b-kube-api-access-42zgq\") pod \"certified-operators-jkpql\" (UID: \"5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b\") " pod="openshift-marketplace/certified-operators-jkpql" Nov 25 12:11:07 crc kubenswrapper[4776]: I1125 12:11:07.701160 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b-utilities\") pod \"certified-operators-jkpql\" (UID: \"5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b\") " pod="openshift-marketplace/certified-operators-jkpql" Nov 25 12:11:07 crc kubenswrapper[4776]: I1125 12:11:07.804211 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b-utilities\") pod \"certified-operators-jkpql\" (UID: \"5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b\") " pod="openshift-marketplace/certified-operators-jkpql" Nov 25 12:11:07 crc kubenswrapper[4776]: I1125 12:11:07.804880 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b-catalog-content\") pod \"certified-operators-jkpql\" (UID: \"5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b\") " pod="openshift-marketplace/certified-operators-jkpql" Nov 25 12:11:07 crc kubenswrapper[4776]: I1125 12:11:07.805148 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42zgq\" (UniqueName: \"kubernetes.io/projected/5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b-kube-api-access-42zgq\") pod \"certified-operators-jkpql\" (UID: \"5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b\") " pod="openshift-marketplace/certified-operators-jkpql" Nov 25 12:11:07 crc kubenswrapper[4776]: I1125 12:11:07.805318 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b-utilities\") pod \"certified-operators-jkpql\" (UID: \"5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b\") " pod="openshift-marketplace/certified-operators-jkpql" Nov 25 12:11:07 crc kubenswrapper[4776]: I1125 12:11:07.805448 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b-catalog-content\") pod \"certified-operators-jkpql\" (UID: \"5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b\") " pod="openshift-marketplace/certified-operators-jkpql" Nov 25 12:11:07 crc kubenswrapper[4776]: I1125 12:11:07.831732 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42zgq\" (UniqueName: \"kubernetes.io/projected/5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b-kube-api-access-42zgq\") pod \"certified-operators-jkpql\" (UID: \"5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b\") " pod="openshift-marketplace/certified-operators-jkpql" Nov 25 12:11:07 crc kubenswrapper[4776]: I1125 12:11:07.842037 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jkpql" Nov 25 12:11:08 crc kubenswrapper[4776]: I1125 12:11:08.453819 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jkpql"] Nov 25 12:11:09 crc kubenswrapper[4776]: I1125 12:11:09.171376 4776 generic.go:334] "Generic (PLEG): container finished" podID="5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b" containerID="219df6df3d105e05f5883b2877ffb0cba4349ee0549906878c7afd17429d331b" exitCode=0 Nov 25 12:11:09 crc kubenswrapper[4776]: I1125 12:11:09.171525 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jkpql" event={"ID":"5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b","Type":"ContainerDied","Data":"219df6df3d105e05f5883b2877ffb0cba4349ee0549906878c7afd17429d331b"} Nov 25 12:11:09 crc kubenswrapper[4776]: I1125 12:11:09.171748 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jkpql" event={"ID":"5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b","Type":"ContainerStarted","Data":"7f5a0447c23325389875321136ead8d7c32406a128567df06c1f05f976a3d9d7"} Nov 25 12:11:10 crc kubenswrapper[4776]: I1125 12:11:10.187513 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jkpql" event={"ID":"5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b","Type":"ContainerStarted","Data":"9e019e09df35fb343a8c0fbb4dbe51d31eec3f6c9a0e94a718a6beadb01fae3d"} Nov 25 12:11:11 crc kubenswrapper[4776]: I1125 12:11:11.202210 4776 generic.go:334] "Generic (PLEG): container finished" podID="5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b" containerID="9e019e09df35fb343a8c0fbb4dbe51d31eec3f6c9a0e94a718a6beadb01fae3d" exitCode=0 Nov 25 12:11:11 crc kubenswrapper[4776]: I1125 12:11:11.202321 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jkpql" event={"ID":"5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b","Type":"ContainerDied","Data":"9e019e09df35fb343a8c0fbb4dbe51d31eec3f6c9a0e94a718a6beadb01fae3d"} Nov 25 12:11:12 crc kubenswrapper[4776]: I1125 12:11:12.218853 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jkpql" event={"ID":"5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b","Type":"ContainerStarted","Data":"f28154c3900f93c38eefd83f2617bd75e83092fccd384c4b13d8092dd9f11b5b"} Nov 25 12:11:12 crc kubenswrapper[4776]: I1125 12:11:12.241375 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jkpql" podStartSLOduration=2.761350923 podStartE2EDuration="5.241346724s" podCreationTimestamp="2025-11-25 12:11:07 +0000 UTC" firstStartedPulling="2025-11-25 12:11:09.175925263 +0000 UTC m=+10014.216984806" lastFinishedPulling="2025-11-25 12:11:11.655921054 +0000 UTC m=+10016.696980607" observedRunningTime="2025-11-25 12:11:12.238516383 +0000 UTC m=+10017.279575926" watchObservedRunningTime="2025-11-25 12:11:12.241346724 +0000 UTC m=+10017.282406307" Nov 25 12:11:17 crc kubenswrapper[4776]: I1125 12:11:17.844343 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jkpql" Nov 25 12:11:17 crc kubenswrapper[4776]: I1125 12:11:17.845278 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jkpql" Nov 25 12:11:17 crc kubenswrapper[4776]: I1125 12:11:17.930036 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jkpql" Nov 25 12:11:18 crc kubenswrapper[4776]: I1125 12:11:18.344797 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jkpql" Nov 25 12:11:18 crc kubenswrapper[4776]: I1125 12:11:18.416726 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jkpql"] Nov 25 12:11:20 crc kubenswrapper[4776]: I1125 12:11:20.314309 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jkpql" podUID="5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b" containerName="registry-server" containerID="cri-o://f28154c3900f93c38eefd83f2617bd75e83092fccd384c4b13d8092dd9f11b5b" gracePeriod=2 Nov 25 12:11:20 crc kubenswrapper[4776]: I1125 12:11:20.810520 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jkpql" Nov 25 12:11:20 crc kubenswrapper[4776]: I1125 12:11:20.958973 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b-catalog-content\") pod \"5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b\" (UID: \"5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b\") " Nov 25 12:11:20 crc kubenswrapper[4776]: I1125 12:11:20.959287 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b-utilities\") pod \"5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b\" (UID: \"5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b\") " Nov 25 12:11:20 crc kubenswrapper[4776]: I1125 12:11:20.959442 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42zgq\" (UniqueName: \"kubernetes.io/projected/5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b-kube-api-access-42zgq\") pod \"5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b\" (UID: \"5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b\") " Nov 25 12:11:20 crc kubenswrapper[4776]: I1125 12:11:20.960657 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b-utilities" (OuterVolumeSpecName: "utilities") pod "5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b" (UID: "5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:11:20 crc kubenswrapper[4776]: I1125 12:11:20.969753 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b-kube-api-access-42zgq" (OuterVolumeSpecName: "kube-api-access-42zgq") pod "5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b" (UID: "5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b"). InnerVolumeSpecName "kube-api-access-42zgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:11:21 crc kubenswrapper[4776]: I1125 12:11:21.062579 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42zgq\" (UniqueName: \"kubernetes.io/projected/5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b-kube-api-access-42zgq\") on node \"crc\" DevicePath \"\"" Nov 25 12:11:21 crc kubenswrapper[4776]: I1125 12:11:21.062653 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:11:21 crc kubenswrapper[4776]: I1125 12:11:21.262348 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b" (UID: "5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:11:21 crc kubenswrapper[4776]: I1125 12:11:21.268839 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:11:21 crc kubenswrapper[4776]: I1125 12:11:21.329355 4776 generic.go:334] "Generic (PLEG): container finished" podID="5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b" containerID="f28154c3900f93c38eefd83f2617bd75e83092fccd384c4b13d8092dd9f11b5b" exitCode=0 Nov 25 12:11:21 crc kubenswrapper[4776]: I1125 12:11:21.329407 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jkpql" event={"ID":"5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b","Type":"ContainerDied","Data":"f28154c3900f93c38eefd83f2617bd75e83092fccd384c4b13d8092dd9f11b5b"} Nov 25 12:11:21 crc kubenswrapper[4776]: I1125 12:11:21.329440 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jkpql" event={"ID":"5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b","Type":"ContainerDied","Data":"7f5a0447c23325389875321136ead8d7c32406a128567df06c1f05f976a3d9d7"} Nov 25 12:11:21 crc kubenswrapper[4776]: I1125 12:11:21.329461 4776 scope.go:117] "RemoveContainer" containerID="f28154c3900f93c38eefd83f2617bd75e83092fccd384c4b13d8092dd9f11b5b" Nov 25 12:11:21 crc kubenswrapper[4776]: I1125 12:11:21.329630 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jkpql" Nov 25 12:11:21 crc kubenswrapper[4776]: I1125 12:11:21.372813 4776 scope.go:117] "RemoveContainer" containerID="9e019e09df35fb343a8c0fbb4dbe51d31eec3f6c9a0e94a718a6beadb01fae3d" Nov 25 12:11:21 crc kubenswrapper[4776]: I1125 12:11:21.378079 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jkpql"] Nov 25 12:11:21 crc kubenswrapper[4776]: I1125 12:11:21.395851 4776 scope.go:117] "RemoveContainer" containerID="219df6df3d105e05f5883b2877ffb0cba4349ee0549906878c7afd17429d331b" Nov 25 12:11:21 crc kubenswrapper[4776]: I1125 12:11:21.396024 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jkpql"] Nov 25 12:11:21 crc kubenswrapper[4776]: I1125 12:11:21.453144 4776 scope.go:117] "RemoveContainer" containerID="f28154c3900f93c38eefd83f2617bd75e83092fccd384c4b13d8092dd9f11b5b" Nov 25 12:11:21 crc kubenswrapper[4776]: E1125 12:11:21.454028 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f28154c3900f93c38eefd83f2617bd75e83092fccd384c4b13d8092dd9f11b5b\": container with ID starting with f28154c3900f93c38eefd83f2617bd75e83092fccd384c4b13d8092dd9f11b5b not found: ID does not exist" containerID="f28154c3900f93c38eefd83f2617bd75e83092fccd384c4b13d8092dd9f11b5b" Nov 25 12:11:21 crc kubenswrapper[4776]: I1125 12:11:21.454106 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f28154c3900f93c38eefd83f2617bd75e83092fccd384c4b13d8092dd9f11b5b"} err="failed to get container status \"f28154c3900f93c38eefd83f2617bd75e83092fccd384c4b13d8092dd9f11b5b\": rpc error: code = NotFound desc = could not find container \"f28154c3900f93c38eefd83f2617bd75e83092fccd384c4b13d8092dd9f11b5b\": container with ID starting with f28154c3900f93c38eefd83f2617bd75e83092fccd384c4b13d8092dd9f11b5b not found: ID does not exist" Nov 25 12:11:21 crc kubenswrapper[4776]: I1125 12:11:21.454157 4776 scope.go:117] "RemoveContainer" containerID="9e019e09df35fb343a8c0fbb4dbe51d31eec3f6c9a0e94a718a6beadb01fae3d" Nov 25 12:11:21 crc kubenswrapper[4776]: E1125 12:11:21.454744 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e019e09df35fb343a8c0fbb4dbe51d31eec3f6c9a0e94a718a6beadb01fae3d\": container with ID starting with 9e019e09df35fb343a8c0fbb4dbe51d31eec3f6c9a0e94a718a6beadb01fae3d not found: ID does not exist" containerID="9e019e09df35fb343a8c0fbb4dbe51d31eec3f6c9a0e94a718a6beadb01fae3d" Nov 25 12:11:21 crc kubenswrapper[4776]: I1125 12:11:21.454810 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e019e09df35fb343a8c0fbb4dbe51d31eec3f6c9a0e94a718a6beadb01fae3d"} err="failed to get container status \"9e019e09df35fb343a8c0fbb4dbe51d31eec3f6c9a0e94a718a6beadb01fae3d\": rpc error: code = NotFound desc = could not find container \"9e019e09df35fb343a8c0fbb4dbe51d31eec3f6c9a0e94a718a6beadb01fae3d\": container with ID starting with 9e019e09df35fb343a8c0fbb4dbe51d31eec3f6c9a0e94a718a6beadb01fae3d not found: ID does not exist" Nov 25 12:11:21 crc kubenswrapper[4776]: I1125 12:11:21.454847 4776 scope.go:117] "RemoveContainer" containerID="219df6df3d105e05f5883b2877ffb0cba4349ee0549906878c7afd17429d331b" Nov 25 12:11:21 crc kubenswrapper[4776]: E1125 12:11:21.455295 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"219df6df3d105e05f5883b2877ffb0cba4349ee0549906878c7afd17429d331b\": container with ID starting with 219df6df3d105e05f5883b2877ffb0cba4349ee0549906878c7afd17429d331b not found: ID does not exist" containerID="219df6df3d105e05f5883b2877ffb0cba4349ee0549906878c7afd17429d331b" Nov 25 12:11:21 crc kubenswrapper[4776]: I1125 12:11:21.455340 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"219df6df3d105e05f5883b2877ffb0cba4349ee0549906878c7afd17429d331b"} err="failed to get container status \"219df6df3d105e05f5883b2877ffb0cba4349ee0549906878c7afd17429d331b\": rpc error: code = NotFound desc = could not find container \"219df6df3d105e05f5883b2877ffb0cba4349ee0549906878c7afd17429d331b\": container with ID starting with 219df6df3d105e05f5883b2877ffb0cba4349ee0549906878c7afd17429d331b not found: ID does not exist" Nov 25 12:11:21 crc kubenswrapper[4776]: I1125 12:11:21.679012 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b" path="/var/lib/kubelet/pods/5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b/volumes" Nov 25 12:11:47 crc kubenswrapper[4776]: I1125 12:11:47.818642 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:11:47 crc kubenswrapper[4776]: I1125 12:11:47.819309 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:12:17 crc kubenswrapper[4776]: I1125 12:12:17.818037 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:12:17 crc kubenswrapper[4776]: I1125 12:12:17.818968 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:12:47 crc kubenswrapper[4776]: I1125 12:12:47.818995 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:12:47 crc kubenswrapper[4776]: I1125 12:12:47.819898 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:12:47 crc kubenswrapper[4776]: I1125 12:12:47.819954 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 12:12:47 crc kubenswrapper[4776]: I1125 12:12:47.820928 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5b13db60744f199580371de408731797095f5765187d672f3989cfe03fd42a7f"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 12:12:47 crc kubenswrapper[4776]: I1125 12:12:47.820985 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://5b13db60744f199580371de408731797095f5765187d672f3989cfe03fd42a7f" gracePeriod=600 Nov 25 12:12:48 crc kubenswrapper[4776]: I1125 12:12:48.378033 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="5b13db60744f199580371de408731797095f5765187d672f3989cfe03fd42a7f" exitCode=0 Nov 25 12:12:48 crc kubenswrapper[4776]: I1125 12:12:48.378097 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"5b13db60744f199580371de408731797095f5765187d672f3989cfe03fd42a7f"} Nov 25 12:12:48 crc kubenswrapper[4776]: I1125 12:12:48.378550 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563"} Nov 25 12:12:48 crc kubenswrapper[4776]: I1125 12:12:48.378577 4776 scope.go:117] "RemoveContainer" containerID="b3286294cf43f7ec564e0aa3e67e4b73dfa9f3a80af5ab29bfd928bdfd26af0d" Nov 25 12:13:43 crc kubenswrapper[4776]: I1125 12:13:43.536135 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_464f639d-26e5-4d21-9d9b-c84585fbd8e2/init-config-reloader/0.log" Nov 25 12:13:43 crc kubenswrapper[4776]: I1125 12:13:43.854901 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_464f639d-26e5-4d21-9d9b-c84585fbd8e2/init-config-reloader/0.log" Nov 25 12:13:43 crc kubenswrapper[4776]: I1125 12:13:43.893368 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_464f639d-26e5-4d21-9d9b-c84585fbd8e2/config-reloader/0.log" Nov 25 12:13:43 crc kubenswrapper[4776]: I1125 12:13:43.935767 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_464f639d-26e5-4d21-9d9b-c84585fbd8e2/alertmanager/0.log" Nov 25 12:13:44 crc kubenswrapper[4776]: I1125 12:13:44.142378 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_81d6f945-72ee-4286-a46b-e6452508c428/aodh-api/0.log" Nov 25 12:13:44 crc kubenswrapper[4776]: I1125 12:13:44.183761 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_81d6f945-72ee-4286-a46b-e6452508c428/aodh-evaluator/0.log" Nov 25 12:13:44 crc kubenswrapper[4776]: I1125 12:13:44.405255 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_81d6f945-72ee-4286-a46b-e6452508c428/aodh-notifier/0.log" Nov 25 12:13:44 crc kubenswrapper[4776]: I1125 12:13:44.437051 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_81d6f945-72ee-4286-a46b-e6452508c428/aodh-listener/0.log" Nov 25 12:13:44 crc kubenswrapper[4776]: I1125 12:13:44.452046 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-645d68dd56-v4nsw_90cec8c4-dd53-4430-839e-ed26c40b10ed/barbican-api/0.log" Nov 25 12:13:44 crc kubenswrapper[4776]: I1125 12:13:44.655628 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-645d68dd56-v4nsw_90cec8c4-dd53-4430-839e-ed26c40b10ed/barbican-api-log/0.log" Nov 25 12:13:44 crc kubenswrapper[4776]: I1125 12:13:44.784021 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5d87dd74f4-rzv2q_cd7d445b-c6b1-4479-8158-1be3fd658211/barbican-keystone-listener-log/0.log" Nov 25 12:13:44 crc kubenswrapper[4776]: I1125 12:13:44.805686 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5d87dd74f4-rzv2q_cd7d445b-c6b1-4479-8158-1be3fd658211/barbican-keystone-listener/0.log" Nov 25 12:13:45 crc kubenswrapper[4776]: I1125 12:13:45.029664 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5647554789-tq4dw_ea1d5f90-b587-4071-a507-c5869e7fc650/barbican-worker/0.log" Nov 25 12:13:45 crc kubenswrapper[4776]: I1125 12:13:45.039346 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5647554789-tq4dw_ea1d5f90-b587-4071-a507-c5869e7fc650/barbican-worker-log/0.log" Nov 25 12:13:45 crc kubenswrapper[4776]: I1125 12:13:45.344765 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-cell1-pzvxn_b771a78c-c9b8-4c94-96b3-caa28c5cacc4/bootstrap-openstack-openstack-cell1/0.log" Nov 25 12:13:45 crc kubenswrapper[4776]: I1125 12:13:45.447336 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b11bc2ae-de8a-45a7-be5a-87a8d81180a2/ceilometer-central-agent/0.log" Nov 25 12:13:45 crc kubenswrapper[4776]: I1125 12:13:45.466052 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b11bc2ae-de8a-45a7-be5a-87a8d81180a2/ceilometer-notification-agent/0.log" Nov 25 12:13:45 crc kubenswrapper[4776]: I1125 12:13:45.666414 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b11bc2ae-de8a-45a7-be5a-87a8d81180a2/proxy-httpd/0.log" Nov 25 12:13:45 crc kubenswrapper[4776]: I1125 12:13:45.734490 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_b11bc2ae-de8a-45a7-be5a-87a8d81180a2/sg-core/0.log" Nov 25 12:13:45 crc kubenswrapper[4776]: I1125 12:13:45.880294 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_6d076c50-21b1-4e6d-b345-a8c46bde8e8f/cinder-api/0.log" Nov 25 12:13:45 crc kubenswrapper[4776]: I1125 12:13:45.989380 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_6d076c50-21b1-4e6d-b345-a8c46bde8e8f/cinder-api-log/0.log" Nov 25 12:13:46 crc kubenswrapper[4776]: I1125 12:13:46.184928 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_18d53345-e467-4604-8446-450375280e28/probe/0.log" Nov 25 12:13:46 crc kubenswrapper[4776]: I1125 12:13:46.198371 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_18d53345-e467-4604-8446-450375280e28/cinder-scheduler/0.log" Nov 25 12:13:46 crc kubenswrapper[4776]: I1125 12:13:46.504415 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-lwvwk_7ed85cc7-157f-4a26-85aa-5266b68246f7/configure-os-openstack-openstack-cell1/0.log" Nov 25 12:13:46 crc kubenswrapper[4776]: I1125 12:13:46.511282 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-cell1-75kbb_5cb01ee8-be4e-4cb2-9d08-0ee2509afac6/configure-network-openstack-openstack-cell1/0.log" Nov 25 12:13:46 crc kubenswrapper[4776]: I1125 12:13:46.703408 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-568cc6cf7c-6mkrs_82c0a51a-57e4-4ab3-98b6-e2052e449bf0/init/0.log" Nov 25 12:13:46 crc kubenswrapper[4776]: I1125 12:13:46.969559 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-cell1-nhxsr_9eccbe5b-80cd-4633-829c-0f79f228bf48/download-cache-openstack-openstack-cell1/0.log" Nov 25 12:13:47 crc kubenswrapper[4776]: I1125 12:13:47.013607 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-568cc6cf7c-6mkrs_82c0a51a-57e4-4ab3-98b6-e2052e449bf0/dnsmasq-dns/0.log" Nov 25 12:13:47 crc kubenswrapper[4776]: I1125 12:13:47.091199 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-568cc6cf7c-6mkrs_82c0a51a-57e4-4ab3-98b6-e2052e449bf0/init/0.log" Nov 25 12:13:47 crc kubenswrapper[4776]: I1125 12:13:47.283950 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54/glance-httpd/0.log" Nov 25 12:13:47 crc kubenswrapper[4776]: I1125 12:13:47.365756 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_6a43c9d6-0d1b-48a5-8fe6-07abcf9c2b54/glance-log/0.log" Nov 25 12:13:47 crc kubenswrapper[4776]: I1125 12:13:47.798328 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_ea2660a8-255a-45db-a684-d089438652e8/glance-httpd/0.log" Nov 25 12:13:47 crc kubenswrapper[4776]: I1125 12:13:47.920352 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_ea2660a8-255a-45db-a684-d089438652e8/glance-log/0.log" Nov 25 12:13:48 crc kubenswrapper[4776]: I1125 12:13:48.408287 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-75966d5674-cmtbw_e7441fab-e3ef-4668-939a-b8d012872bdb/heat-api/0.log" Nov 25 12:13:48 crc kubenswrapper[4776]: I1125 12:13:48.449041 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-5954f4557-4hz28_4291e940-29c4-4c08-98ea-227e96cd687f/heat-engine/0.log" Nov 25 12:13:48 crc kubenswrapper[4776]: I1125 12:13:48.837514 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-f4f475bb-4hkzw_029b9a31-76a8-446d-be2c-53760cb3980a/horizon/0.log" Nov 25 12:13:48 crc kubenswrapper[4776]: I1125 12:13:48.885719 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-cell1-bhxjr_4af7b0e8-aca0-4803-9d2c-0502406152e0/install-certs-openstack-openstack-cell1/0.log" Nov 25 12:13:49 crc kubenswrapper[4776]: I1125 12:13:49.016657 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-58c77f6f47-2zhwf_5aa7be27-73be-4401-a206-69b4d6a47a12/heat-cfnapi/0.log" Nov 25 12:13:49 crc kubenswrapper[4776]: I1125 12:13:49.187849 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-cell1-lxhrp_8ac69e36-fa36-4a86-843d-4deffacc0f1f/install-os-openstack-openstack-cell1/0.log" Nov 25 12:13:49 crc kubenswrapper[4776]: I1125 12:13:49.321791 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-f4f475bb-4hkzw_029b9a31-76a8-446d-be2c-53760cb3980a/horizon-log/0.log" Nov 25 12:13:49 crc kubenswrapper[4776]: I1125 12:13:49.406051 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29401141-9cqm2_ad73ef14-feaa-40da-98f8-4ac5687bf72b/keystone-cron/0.log" Nov 25 12:13:49 crc kubenswrapper[4776]: I1125 12:13:49.498802 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-c5d66c966-5c6vj_974cf029-f130-4748-81c3-1dca65481a17/keystone-api/0.log" Nov 25 12:13:49 crc kubenswrapper[4776]: I1125 12:13:49.579275 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29401201-8vgn6_841c9ff3-7f79-424b-9da0-4bca60584e2a/keystone-cron/0.log" Nov 25 12:13:49 crc kubenswrapper[4776]: I1125 12:13:49.720470 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_a25bfd2c-28ed-4856-9805-3242c9b7492b/kube-state-metrics/0.log" Nov 25 12:13:49 crc kubenswrapper[4776]: I1125 12:13:49.857004 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-openstack-openstack-cell1-c2plb_bc20c5f3-b0e5-47e2-8878-015b66dad6f0/libvirt-openstack-openstack-cell1/0.log" Nov 25 12:13:51 crc kubenswrapper[4776]: I1125 12:13:51.018823 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7db9d4c697-g5n67_79fed1e6-a706-4827-85e1-1d46cd4fe514/neutron-httpd/0.log" Nov 25 12:13:51 crc kubenswrapper[4776]: I1125 12:13:51.159984 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7db9d4c697-g5n67_79fed1e6-a706-4827-85e1-1d46cd4fe514/neutron-api/0.log" Nov 25 12:13:51 crc kubenswrapper[4776]: I1125 12:13:51.460888 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-cell1-nzsns_795a3ef2-55b2-4550-a681-12b17cb26dad/neutron-metadata-openstack-openstack-cell1/0.log" Nov 25 12:13:51 crc kubenswrapper[4776]: I1125 12:13:51.483125 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dhcp-openstack-openstack-cell1-k4kxp_1a0bc66d-b624-4636-baec-30c555632825/neutron-dhcp-openstack-openstack-cell1/0.log" Nov 25 12:13:51 crc kubenswrapper[4776]: I1125 12:13:51.739612 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-sriov-openstack-openstack-cell1-x7ctw_c622a8cc-2543-4ba5-b53f-32d680331106/neutron-sriov-openstack-openstack-cell1/0.log" Nov 25 12:13:51 crc kubenswrapper[4776]: I1125 12:13:51.943276 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_4ae65eee-ea73-403b-a154-e6ed7865d548/nova-api-api/0.log" Nov 25 12:13:52 crc kubenswrapper[4776]: I1125 12:13:52.086866 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_4ae65eee-ea73-403b-a154-e6ed7865d548/nova-api-log/0.log" Nov 25 12:13:52 crc kubenswrapper[4776]: I1125 12:13:52.270055 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_0945270e-80b0-44ca-9c2b-642abacd57bf/nova-cell0-conductor-conductor/0.log" Nov 25 12:13:52 crc kubenswrapper[4776]: I1125 12:13:52.544222 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_8a2e68e3-1379-4a66-9b36-bd74718b49ab/nova-cell1-conductor-conductor/0.log" Nov 25 12:13:52 crc kubenswrapper[4776]: I1125 12:13:52.716644 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_a40b8b0e-76a4-402b-962d-46e22f5bbaef/nova-cell1-novncproxy-novncproxy/0.log" Nov 25 12:13:52 crc kubenswrapper[4776]: I1125 12:13:52.985664 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellcfgzk_1db10c36-1c34-47f5-a872-a82d15d88299/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Nov 25 12:13:53 crc kubenswrapper[4776]: I1125 12:13:53.623882 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_39189bd1-98eb-48a8-b087-ac932aa0dace/nova-metadata-log/0.log" Nov 25 12:13:53 crc kubenswrapper[4776]: I1125 12:13:53.820664 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-openstack-cell1-bh42p_2aba01e8-085c-4bef-b6bc-1ae855ced88f/nova-cell1-openstack-openstack-cell1/0.log" Nov 25 12:13:54 crc kubenswrapper[4776]: I1125 12:13:54.316324 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_47179497-a4ae-4629-8a62-f1875880b8df/nova-scheduler-scheduler/0.log" Nov 25 12:13:54 crc kubenswrapper[4776]: I1125 12:13:54.473346 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_39189bd1-98eb-48a8-b087-ac932aa0dace/nova-metadata-metadata/0.log" Nov 25 12:13:54 crc kubenswrapper[4776]: I1125 12:13:54.547594 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-c694cd6fb-994xl_21522877-e085-4ff3-857d-edb4989664fd/init/0.log" Nov 25 12:13:54 crc kubenswrapper[4776]: I1125 12:13:54.787563 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-c694cd6fb-994xl_21522877-e085-4ff3-857d-edb4989664fd/octavia-api-provider-agent/0.log" Nov 25 12:13:54 crc kubenswrapper[4776]: I1125 12:13:54.812808 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-c694cd6fb-994xl_21522877-e085-4ff3-857d-edb4989664fd/init/0.log" Nov 25 12:13:55 crc kubenswrapper[4776]: I1125 12:13:55.112776 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-c694cd6fb-994xl_21522877-e085-4ff3-857d-edb4989664fd/octavia-api/0.log" Nov 25 12:13:55 crc kubenswrapper[4776]: I1125 12:13:55.372104 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-z8qlp_d017f1f2-3baf-46a7-9dcd-d361a86c51f6/init/0.log" Nov 25 12:13:55 crc kubenswrapper[4776]: I1125 12:13:55.664625 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-z8qlp_d017f1f2-3baf-46a7-9dcd-d361a86c51f6/init/0.log" Nov 25 12:13:55 crc kubenswrapper[4776]: I1125 12:13:55.716105 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-z8qlp_d017f1f2-3baf-46a7-9dcd-d361a86c51f6/octavia-healthmanager/0.log" Nov 25 12:13:55 crc kubenswrapper[4776]: I1125 12:13:55.733221 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-bd8xp_de8204cc-520a-41de-8079-2b25831ae0ae/init/0.log" Nov 25 12:13:55 crc kubenswrapper[4776]: I1125 12:13:55.976724 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-bd8xp_de8204cc-520a-41de-8079-2b25831ae0ae/octavia-housekeeping/0.log" Nov 25 12:13:55 crc kubenswrapper[4776]: I1125 12:13:55.984101 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-bd8xp_de8204cc-520a-41de-8079-2b25831ae0ae/init/0.log" Nov 25 12:13:56 crc kubenswrapper[4776]: I1125 12:13:56.066716 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-nbkx7_2e7c13e5-cbf7-408b-8a3c-edfc1d208c12/init/0.log" Nov 25 12:13:56 crc kubenswrapper[4776]: I1125 12:13:56.344619 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-nbkx7_2e7c13e5-cbf7-408b-8a3c-edfc1d208c12/init/0.log" Nov 25 12:13:56 crc kubenswrapper[4776]: I1125 12:13:56.414474 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-jzw7v_eb21722f-848b-44ff-8658-a5a890bf0855/init/0.log" Nov 25 12:13:56 crc kubenswrapper[4776]: I1125 12:13:56.419917 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-nbkx7_2e7c13e5-cbf7-408b-8a3c-edfc1d208c12/octavia-rsyslog/0.log" Nov 25 12:13:56 crc kubenswrapper[4776]: I1125 12:13:56.727211 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-jzw7v_eb21722f-848b-44ff-8658-a5a890bf0855/init/0.log" Nov 25 12:13:56 crc kubenswrapper[4776]: I1125 12:13:56.885799 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_8c3aa288-f360-4925-8229-5a3593b31be7/mysql-bootstrap/0.log" Nov 25 12:13:57 crc kubenswrapper[4776]: I1125 12:13:57.049913 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_8c3aa288-f360-4925-8229-5a3593b31be7/mysql-bootstrap/0.log" Nov 25 12:13:57 crc kubenswrapper[4776]: I1125 12:13:57.063806 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-jzw7v_eb21722f-848b-44ff-8658-a5a890bf0855/octavia-worker/0.log" Nov 25 12:13:57 crc kubenswrapper[4776]: I1125 12:13:57.100934 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_8c3aa288-f360-4925-8229-5a3593b31be7/galera/0.log" Nov 25 12:13:57 crc kubenswrapper[4776]: I1125 12:13:57.313999 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ed1eddec-2d91-434e-803e-69cf7501f99d/mysql-bootstrap/0.log" Nov 25 12:13:57 crc kubenswrapper[4776]: I1125 12:13:57.520838 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ed1eddec-2d91-434e-803e-69cf7501f99d/mysql-bootstrap/0.log" Nov 25 12:13:57 crc kubenswrapper[4776]: I1125 12:13:57.538528 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ed1eddec-2d91-434e-803e-69cf7501f99d/galera/0.log" Nov 25 12:13:57 crc kubenswrapper[4776]: I1125 12:13:57.582517 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_96898351-caa6-4edc-8080-898f9fff0f56/openstackclient/0.log" Nov 25 12:13:57 crc kubenswrapper[4776]: I1125 12:13:57.790813 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-85l79_a6e2d225-3db6-4189-a956-fd709c242387/ovn-controller/0.log" Nov 25 12:13:57 crc kubenswrapper[4776]: I1125 12:13:57.970130 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-jgqps_5dcfc2e2-07bd-4f08-b20f-604e1a6a49e6/openstack-network-exporter/0.log" Nov 25 12:13:58 crc kubenswrapper[4776]: I1125 12:13:58.148501 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-wlphl_845d70c5-c0ad-4143-8778-3bccb4724aa4/ovsdb-server-init/0.log" Nov 25 12:13:58 crc kubenswrapper[4776]: I1125 12:13:58.513967 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-wlphl_845d70c5-c0ad-4143-8778-3bccb4724aa4/ovsdb-server-init/0.log" Nov 25 12:13:58 crc kubenswrapper[4776]: I1125 12:13:58.516189 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-wlphl_845d70c5-c0ad-4143-8778-3bccb4724aa4/ovsdb-server/0.log" Nov 25 12:13:58 crc kubenswrapper[4776]: I1125 12:13:58.533739 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-wlphl_845d70c5-c0ad-4143-8778-3bccb4724aa4/ovs-vswitchd/0.log" Nov 25 12:13:58 crc kubenswrapper[4776]: I1125 12:13:58.734959 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_7f396897-96fa-422c-a2fd-92faadaff6e3/ovn-northd/0.log" Nov 25 12:13:58 crc kubenswrapper[4776]: I1125 12:13:58.787809 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_7f396897-96fa-422c-a2fd-92faadaff6e3/openstack-network-exporter/0.log" Nov 25 12:13:58 crc kubenswrapper[4776]: I1125 12:13:58.981322 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-cell1-jppkj_458d4eb8-3fdd-4a3c-9457-83af2f4dd0c4/ovn-openstack-openstack-cell1/0.log" Nov 25 12:13:59 crc kubenswrapper[4776]: I1125 12:13:59.383586 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_3d16f6ea-879b-4b44-a9aa-187c59681215/openstack-network-exporter/0.log" Nov 25 12:13:59 crc kubenswrapper[4776]: I1125 12:13:59.558038 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_3d16f6ea-879b-4b44-a9aa-187c59681215/ovsdbserver-nb/0.log" Nov 25 12:13:59 crc kubenswrapper[4776]: I1125 12:13:59.618526 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_bce8180f-fb70-4f9c-bc00-3a39e62da7a3/openstack-network-exporter/0.log" Nov 25 12:13:59 crc kubenswrapper[4776]: I1125 12:13:59.811206 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_bce8180f-fb70-4f9c-bc00-3a39e62da7a3/ovsdbserver-nb/0.log" Nov 25 12:13:59 crc kubenswrapper[4776]: I1125 12:13:59.876251 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_f5b7e9cb-0aa9-434f-b75c-783829b9ba56/openstack-network-exporter/0.log" Nov 25 12:14:00 crc kubenswrapper[4776]: I1125 12:14:00.071974 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_f5b7e9cb-0aa9-434f-b75c-783829b9ba56/ovsdbserver-nb/0.log" Nov 25 12:14:00 crc kubenswrapper[4776]: I1125 12:14:00.155821 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ce32cbf1-d6a1-4361-8921-ce0abcda4667/openstack-network-exporter/0.log" Nov 25 12:14:00 crc kubenswrapper[4776]: I1125 12:14:00.173711 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ce32cbf1-d6a1-4361-8921-ce0abcda4667/ovsdbserver-sb/0.log" Nov 25 12:14:00 crc kubenswrapper[4776]: I1125 12:14:00.389364 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_0d2f39ca-cd42-42c2-9d0f-51e31961a349/openstack-network-exporter/0.log" Nov 25 12:14:00 crc kubenswrapper[4776]: I1125 12:14:00.435406 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_0d2f39ca-cd42-42c2-9d0f-51e31961a349/ovsdbserver-sb/0.log" Nov 25 12:14:00 crc kubenswrapper[4776]: I1125 12:14:00.705314 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_676f4247-9871-4bee-ad0f-3b42889be224/openstack-network-exporter/0.log" Nov 25 12:14:00 crc kubenswrapper[4776]: I1125 12:14:00.721491 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_676f4247-9871-4bee-ad0f-3b42889be224/ovsdbserver-sb/0.log" Nov 25 12:14:01 crc kubenswrapper[4776]: I1125 12:14:01.012660 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5488dbdb98-xmdn2_5720535b-1e84-4ca1-8a09-67ad129337f1/placement-api/0.log" Nov 25 12:14:01 crc kubenswrapper[4776]: I1125 12:14:01.159659 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5488dbdb98-xmdn2_5720535b-1e84-4ca1-8a09-67ad129337f1/placement-log/0.log" Nov 25 12:14:01 crc kubenswrapper[4776]: I1125 12:14:01.564082 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_31358dae-00d6-47d3-b9aa-151088320199/init-config-reloader/0.log" Nov 25 12:14:01 crc kubenswrapper[4776]: I1125 12:14:01.579051 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-c8ngsm_de94e490-fe33-4e32-ac8f-5dd704a298f1/pre-adoption-validation-openstack-pre-adoption-openstack-cell1/0.log" Nov 25 12:14:01 crc kubenswrapper[4776]: I1125 12:14:01.817021 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_31358dae-00d6-47d3-b9aa-151088320199/prometheus/0.log" Nov 25 12:14:01 crc kubenswrapper[4776]: I1125 12:14:01.903960 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_31358dae-00d6-47d3-b9aa-151088320199/init-config-reloader/0.log" Nov 25 12:14:01 crc kubenswrapper[4776]: I1125 12:14:01.969259 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_31358dae-00d6-47d3-b9aa-151088320199/config-reloader/0.log" Nov 25 12:14:02 crc kubenswrapper[4776]: I1125 12:14:02.013241 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_31358dae-00d6-47d3-b9aa-151088320199/thanos-sidecar/0.log" Nov 25 12:14:02 crc kubenswrapper[4776]: I1125 12:14:02.213015 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_147a21c0-308e-4792-ae3e-bfc852327d0b/setup-container/0.log" Nov 25 12:14:02 crc kubenswrapper[4776]: I1125 12:14:02.452717 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_147a21c0-308e-4792-ae3e-bfc852327d0b/setup-container/0.log" Nov 25 12:14:02 crc kubenswrapper[4776]: I1125 12:14:02.524491 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_188d5ae0-1ff7-44fd-b0db-5500b52f2b63/setup-container/0.log" Nov 25 12:14:02 crc kubenswrapper[4776]: I1125 12:14:02.545954 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_147a21c0-308e-4792-ae3e-bfc852327d0b/rabbitmq/0.log" Nov 25 12:14:02 crc kubenswrapper[4776]: I1125 12:14:02.747984 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_188d5ae0-1ff7-44fd-b0db-5500b52f2b63/setup-container/0.log" Nov 25 12:14:02 crc kubenswrapper[4776]: I1125 12:14:02.821675 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_188d5ae0-1ff7-44fd-b0db-5500b52f2b63/rabbitmq/0.log" Nov 25 12:14:02 crc kubenswrapper[4776]: I1125 12:14:02.912956 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-cell1-jrtwp_4404e0f3-9f58-4639-9045-afdcc964e742/reboot-os-openstack-openstack-cell1/0.log" Nov 25 12:14:03 crc kubenswrapper[4776]: I1125 12:14:03.140466 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-cell1-q69dh_d9ec86f2-22e3-48aa-b7c7-00d8ef5f2857/run-os-openstack-openstack-cell1/0.log" Nov 25 12:14:03 crc kubenswrapper[4776]: I1125 12:14:03.503414 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-openstack-h7v7g_087b7d2d-0326-4d89-9262-902367375a92/ssh-known-hosts-openstack/0.log" Nov 25 12:14:04 crc kubenswrapper[4776]: I1125 12:14:04.032017 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5c6d9c7954-pvk2n_0dc0a5c5-9aa1-4190-985f-047a9406c24a/proxy-httpd/0.log" Nov 25 12:14:04 crc kubenswrapper[4776]: I1125 12:14:04.272995 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_4f2267d1-b0d6-4a66-8dde-beb9d280b4e1/memcached/0.log" Nov 25 12:14:04 crc kubenswrapper[4776]: I1125 12:14:04.294352 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5c6d9c7954-pvk2n_0dc0a5c5-9aa1-4190-985f-047a9406c24a/proxy-server/0.log" Nov 25 12:14:04 crc kubenswrapper[4776]: I1125 12:14:04.346366 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-7wxbq_8df08eab-1375-42cd-b1dc-31a953e2bca6/swift-ring-rebalance/0.log" Nov 25 12:14:04 crc kubenswrapper[4776]: I1125 12:14:04.548733 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-openstack-openstack-cell1-vdhqb_13669d03-2d4a-4bed-8386-cfcae6272ffb/telemetry-openstack-openstack-cell1/0.log" Nov 25 12:14:04 crc kubenswrapper[4776]: I1125 12:14:04.681310 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-cell1-k5b8t_0ec16c9c-ee71-4edb-b4e8-6570328756d8/validate-network-openstack-openstack-cell1/0.log" Nov 25 12:14:04 crc kubenswrapper[4776]: I1125 12:14:04.693709 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-cell1-5gh8s_4ffa64e0-2cd8-4fe7-90d1-0e4ce4a88ba8/tripleo-cleanup-tripleo-cleanup-openstack-cell1/0.log" Nov 25 12:14:34 crc kubenswrapper[4776]: I1125 12:14:34.382455 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-86dc4d89c8-58r67_5449efd1-7801-454a-bc48-a5e9e91aba7d/kube-rbac-proxy/0.log" Nov 25 12:14:34 crc kubenswrapper[4776]: I1125 12:14:34.533137 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-86dc4d89c8-58r67_5449efd1-7801-454a-bc48-a5e9e91aba7d/manager/0.log" Nov 25 12:14:34 crc kubenswrapper[4776]: I1125 12:14:34.622813 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv_a13d792f-8c2b-4dce-ba68-c83effbbbd12/util/0.log" Nov 25 12:14:34 crc kubenswrapper[4776]: I1125 12:14:34.864786 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv_a13d792f-8c2b-4dce-ba68-c83effbbbd12/pull/0.log" Nov 25 12:14:34 crc kubenswrapper[4776]: I1125 12:14:34.880261 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv_a13d792f-8c2b-4dce-ba68-c83effbbbd12/util/0.log" Nov 25 12:14:34 crc kubenswrapper[4776]: I1125 12:14:34.925720 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv_a13d792f-8c2b-4dce-ba68-c83effbbbd12/pull/0.log" Nov 25 12:14:35 crc kubenswrapper[4776]: I1125 12:14:35.102719 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv_a13d792f-8c2b-4dce-ba68-c83effbbbd12/util/0.log" Nov 25 12:14:35 crc kubenswrapper[4776]: I1125 12:14:35.127659 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv_a13d792f-8c2b-4dce-ba68-c83effbbbd12/pull/0.log" Nov 25 12:14:35 crc kubenswrapper[4776]: I1125 12:14:35.138050 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15fz2cvv_a13d792f-8c2b-4dce-ba68-c83effbbbd12/extract/0.log" Nov 25 12:14:35 crc kubenswrapper[4776]: I1125 12:14:35.394281 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-rnrpd_2d511feb-5cd8-4ba3-b5af-5d9d01086a82/kube-rbac-proxy/0.log" Nov 25 12:14:35 crc kubenswrapper[4776]: I1125 12:14:35.434446 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-rnrpd_2d511feb-5cd8-4ba3-b5af-5d9d01086a82/manager/0.log" Nov 25 12:14:35 crc kubenswrapper[4776]: I1125 12:14:35.475300 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-zwk9j_a590fe89-93f7-4c94-bbdd-82531eedf52f/kube-rbac-proxy/0.log" Nov 25 12:14:36 crc kubenswrapper[4776]: I1125 12:14:36.181992 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-zwk9j_a590fe89-93f7-4c94-bbdd-82531eedf52f/manager/0.log" Nov 25 12:14:36 crc kubenswrapper[4776]: I1125 12:14:36.204181 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-68b95954c9-h6k9v_7a21bb84-b9c5-4e6f-8fbb-3c2bc65dd409/kube-rbac-proxy/0.log" Nov 25 12:14:36 crc kubenswrapper[4776]: I1125 12:14:36.371277 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-68b95954c9-h6k9v_7a21bb84-b9c5-4e6f-8fbb-3c2bc65dd409/manager/0.log" Nov 25 12:14:36 crc kubenswrapper[4776]: I1125 12:14:36.541212 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-774b86978c-gmcp9_61810388-3bac-4a06-88ea-98cc0337fdee/kube-rbac-proxy/0.log" Nov 25 12:14:36 crc kubenswrapper[4776]: I1125 12:14:36.657768 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-774b86978c-gmcp9_61810388-3bac-4a06-88ea-98cc0337fdee/manager/0.log" Nov 25 12:14:36 crc kubenswrapper[4776]: I1125 12:14:36.873045 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-khxdw_187d1384-3de2-4135-abac-45f15155b942/kube-rbac-proxy/0.log" Nov 25 12:14:36 crc kubenswrapper[4776]: I1125 12:14:36.880130 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-khxdw_187d1384-3de2-4135-abac-45f15155b942/manager/0.log" Nov 25 12:14:36 crc kubenswrapper[4776]: I1125 12:14:36.940851 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-d5cc86f4b-clnxq_bf227b94-0dfc-4055-b688-0f73585af089/kube-rbac-proxy/0.log" Nov 25 12:14:37 crc kubenswrapper[4776]: I1125 12:14:37.136087 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-cc9wf_ebac9e21-7ae6-45e4-a6a8-9728b3670733/kube-rbac-proxy/0.log" Nov 25 12:14:37 crc kubenswrapper[4776]: I1125 12:14:37.253473 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-cc9wf_ebac9e21-7ae6-45e4-a6a8-9728b3670733/manager/0.log" Nov 25 12:14:37 crc kubenswrapper[4776]: I1125 12:14:37.447991 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-d5cc86f4b-clnxq_bf227b94-0dfc-4055-b688-0f73585af089/manager/0.log" Nov 25 12:14:37 crc kubenswrapper[4776]: I1125 12:14:37.487955 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-pckfv_63e1e612-6ab8-4144-b2cc-c8354bd0894a/kube-rbac-proxy/0.log" Nov 25 12:14:38 crc kubenswrapper[4776]: I1125 12:14:38.003682 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-pckfv_63e1e612-6ab8-4144-b2cc-c8354bd0894a/manager/0.log" Nov 25 12:14:38 crc kubenswrapper[4776]: I1125 12:14:38.754212 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58bb8d67cc-h57b7_abb2c6a9-56ce-45ff-902f-ac49a249fd81/manager/0.log" Nov 25 12:14:38 crc kubenswrapper[4776]: I1125 12:14:38.810557 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58bb8d67cc-h57b7_abb2c6a9-56ce-45ff-902f-ac49a249fd81/kube-rbac-proxy/0.log" Nov 25 12:14:38 crc kubenswrapper[4776]: I1125 12:14:38.871351 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-cb6c4fdb7-jr7jl_7af82806-526b-4fe2-8314-65298d1a0539/kube-rbac-proxy/0.log" Nov 25 12:14:38 crc kubenswrapper[4776]: I1125 12:14:38.902810 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-cb6c4fdb7-jr7jl_7af82806-526b-4fe2-8314-65298d1a0539/manager/0.log" Nov 25 12:14:39 crc kubenswrapper[4776]: I1125 12:14:39.027347 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7c57c8bbc4-gwj5b_88b8f0f1-da6d-43fe-940e-23c48e9248d2/kube-rbac-proxy/0.log" Nov 25 12:14:39 crc kubenswrapper[4776]: I1125 12:14:39.135495 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-sddh8_5c262b40-b415-4cca-b1e8-635a9c153d81/kube-rbac-proxy/0.log" Nov 25 12:14:39 crc kubenswrapper[4776]: I1125 12:14:39.167127 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7c57c8bbc4-gwj5b_88b8f0f1-da6d-43fe-940e-23c48e9248d2/manager/0.log" Nov 25 12:14:39 crc kubenswrapper[4776]: I1125 12:14:39.411867 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-h8s6x_68f49316-eb77-464a-aebb-189800786fa5/kube-rbac-proxy/0.log" Nov 25 12:14:39 crc kubenswrapper[4776]: I1125 12:14:39.505974 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-sddh8_5c262b40-b415-4cca-b1e8-635a9c153d81/manager/0.log" Nov 25 12:14:39 crc kubenswrapper[4776]: I1125 12:14:39.565650 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-h8s6x_68f49316-eb77-464a-aebb-189800786fa5/manager/0.log" Nov 25 12:14:39 crc kubenswrapper[4776]: I1125 12:14:39.641504 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-b58f89467-t9s4l_ca7c46cd-e5ff-4732-9718-5c2f08b75221/kube-rbac-proxy/0.log" Nov 25 12:14:39 crc kubenswrapper[4776]: I1125 12:14:39.701170 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-b58f89467-t9s4l_ca7c46cd-e5ff-4732-9718-5c2f08b75221/manager/0.log" Nov 25 12:14:40 crc kubenswrapper[4776]: I1125 12:14:40.172464 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7b567956b5-q468k_1a1c5930-5dee-4eeb-a09a-842791a33f88/operator/0.log" Nov 25 12:14:40 crc kubenswrapper[4776]: I1125 12:14:40.259911 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-66cf5c67ff-hv6qx_400f3b2c-a118-4fab-8c8e-2f4d29f5c354/kube-rbac-proxy/0.log" Nov 25 12:14:40 crc kubenswrapper[4776]: I1125 12:14:40.409187 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-k6vg9_96957ec4-0d5b-42a9-8ddb-976e00abe710/registry-server/0.log" Nov 25 12:14:40 crc kubenswrapper[4776]: I1125 12:14:40.538776 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-kbcml_dac570d6-3425-4117-be42-1c6d30785a4c/kube-rbac-proxy/0.log" Nov 25 12:14:40 crc kubenswrapper[4776]: I1125 12:14:40.660199 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-66cf5c67ff-hv6qx_400f3b2c-a118-4fab-8c8e-2f4d29f5c354/manager/0.log" Nov 25 12:14:40 crc kubenswrapper[4776]: I1125 12:14:40.759303 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-kbcml_dac570d6-3425-4117-be42-1c6d30785a4c/manager/0.log" Nov 25 12:14:40 crc kubenswrapper[4776]: I1125 12:14:40.878727 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-npwhb_8f2c9384-1b11-4925-8338-babf955143b1/operator/0.log" Nov 25 12:14:41 crc kubenswrapper[4776]: I1125 12:14:41.011463 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-mtqhm_9018dbd6-ab49-4231-a172-33f9a0b56d90/kube-rbac-proxy/0.log" Nov 25 12:14:41 crc kubenswrapper[4776]: I1125 12:14:41.120236 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-mtqhm_9018dbd6-ab49-4231-a172-33f9a0b56d90/manager/0.log" Nov 25 12:14:41 crc kubenswrapper[4776]: I1125 12:14:41.395395 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-567f98c9d-ppmfr_3ebf4825-05f2-4fdd-bc60-d35e80e1b294/kube-rbac-proxy/0.log" Nov 25 12:14:41 crc kubenswrapper[4776]: I1125 12:14:41.584697 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cb74df96-v4xgj_c2be0972-c768-499c-95e3-4cb180ad510c/kube-rbac-proxy/0.log" Nov 25 12:14:41 crc kubenswrapper[4776]: I1125 12:14:41.619931 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cb74df96-v4xgj_c2be0972-c768-499c-95e3-4cb180ad510c/manager/0.log" Nov 25 12:14:41 crc kubenswrapper[4776]: I1125 12:14:41.745301 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-567f98c9d-ppmfr_3ebf4825-05f2-4fdd-bc60-d35e80e1b294/manager/0.log" Nov 25 12:14:41 crc kubenswrapper[4776]: I1125 12:14:41.812129 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-864885998-sxmf5_27834511-7fe9-454d-b318-efa218524c0d/manager/0.log" Nov 25 12:14:41 crc kubenswrapper[4776]: I1125 12:14:41.845745 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-864885998-sxmf5_27834511-7fe9-454d-b318-efa218524c0d/kube-rbac-proxy/0.log" Nov 25 12:14:43 crc kubenswrapper[4776]: I1125 12:14:43.393398 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7cd5954d9-mnd72_f835b8f9-faca-47e0-99ef-8a11a20c5a56/manager/0.log" Nov 25 12:15:00 crc kubenswrapper[4776]: I1125 12:15:00.184102 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401215-zrwkj"] Nov 25 12:15:00 crc kubenswrapper[4776]: E1125 12:15:00.185446 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b" containerName="extract-utilities" Nov 25 12:15:00 crc kubenswrapper[4776]: I1125 12:15:00.185463 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b" containerName="extract-utilities" Nov 25 12:15:00 crc kubenswrapper[4776]: E1125 12:15:00.185489 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b" containerName="extract-content" Nov 25 12:15:00 crc kubenswrapper[4776]: I1125 12:15:00.185497 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b" containerName="extract-content" Nov 25 12:15:00 crc kubenswrapper[4776]: E1125 12:15:00.185518 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b" containerName="registry-server" Nov 25 12:15:00 crc kubenswrapper[4776]: I1125 12:15:00.185530 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b" containerName="registry-server" Nov 25 12:15:00 crc kubenswrapper[4776]: I1125 12:15:00.185757 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cc887c5-a4d8-4e38-ae8c-fcd8a5fb532b" containerName="registry-server" Nov 25 12:15:00 crc kubenswrapper[4776]: I1125 12:15:00.186673 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-zrwkj" Nov 25 12:15:00 crc kubenswrapper[4776]: I1125 12:15:00.197788 4776 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 12:15:00 crc kubenswrapper[4776]: I1125 12:15:00.197813 4776 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 12:15:00 crc kubenswrapper[4776]: I1125 12:15:00.202209 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401215-zrwkj"] Nov 25 12:15:00 crc kubenswrapper[4776]: I1125 12:15:00.297256 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwwgf\" (UniqueName: \"kubernetes.io/projected/fa9774f8-e062-493f-9480-c71c2b6ff403-kube-api-access-dwwgf\") pod \"collect-profiles-29401215-zrwkj\" (UID: \"fa9774f8-e062-493f-9480-c71c2b6ff403\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-zrwkj" Nov 25 12:15:00 crc kubenswrapper[4776]: I1125 12:15:00.297952 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fa9774f8-e062-493f-9480-c71c2b6ff403-config-volume\") pod \"collect-profiles-29401215-zrwkj\" (UID: \"fa9774f8-e062-493f-9480-c71c2b6ff403\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-zrwkj" Nov 25 12:15:00 crc kubenswrapper[4776]: I1125 12:15:00.298270 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fa9774f8-e062-493f-9480-c71c2b6ff403-secret-volume\") pod \"collect-profiles-29401215-zrwkj\" (UID: \"fa9774f8-e062-493f-9480-c71c2b6ff403\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-zrwkj" Nov 25 12:15:00 crc kubenswrapper[4776]: I1125 12:15:00.400385 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwwgf\" (UniqueName: \"kubernetes.io/projected/fa9774f8-e062-493f-9480-c71c2b6ff403-kube-api-access-dwwgf\") pod \"collect-profiles-29401215-zrwkj\" (UID: \"fa9774f8-e062-493f-9480-c71c2b6ff403\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-zrwkj" Nov 25 12:15:00 crc kubenswrapper[4776]: I1125 12:15:00.400469 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fa9774f8-e062-493f-9480-c71c2b6ff403-config-volume\") pod \"collect-profiles-29401215-zrwkj\" (UID: \"fa9774f8-e062-493f-9480-c71c2b6ff403\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-zrwkj" Nov 25 12:15:00 crc kubenswrapper[4776]: I1125 12:15:00.400584 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fa9774f8-e062-493f-9480-c71c2b6ff403-secret-volume\") pod \"collect-profiles-29401215-zrwkj\" (UID: \"fa9774f8-e062-493f-9480-c71c2b6ff403\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-zrwkj" Nov 25 12:15:00 crc kubenswrapper[4776]: I1125 12:15:00.403141 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fa9774f8-e062-493f-9480-c71c2b6ff403-config-volume\") pod \"collect-profiles-29401215-zrwkj\" (UID: \"fa9774f8-e062-493f-9480-c71c2b6ff403\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-zrwkj" Nov 25 12:15:00 crc kubenswrapper[4776]: I1125 12:15:00.414592 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fa9774f8-e062-493f-9480-c71c2b6ff403-secret-volume\") pod \"collect-profiles-29401215-zrwkj\" (UID: \"fa9774f8-e062-493f-9480-c71c2b6ff403\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-zrwkj" Nov 25 12:15:00 crc kubenswrapper[4776]: I1125 12:15:00.428770 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwwgf\" (UniqueName: \"kubernetes.io/projected/fa9774f8-e062-493f-9480-c71c2b6ff403-kube-api-access-dwwgf\") pod \"collect-profiles-29401215-zrwkj\" (UID: \"fa9774f8-e062-493f-9480-c71c2b6ff403\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-zrwkj" Nov 25 12:15:00 crc kubenswrapper[4776]: I1125 12:15:00.522236 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-zrwkj" Nov 25 12:15:01 crc kubenswrapper[4776]: I1125 12:15:01.048425 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401215-zrwkj"] Nov 25 12:15:01 crc kubenswrapper[4776]: I1125 12:15:01.935837 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-zrwkj" event={"ID":"fa9774f8-e062-493f-9480-c71c2b6ff403","Type":"ContainerStarted","Data":"172ad9574055df6b5941de65c56977552ee2e90a8edf13a4ec64dd86b2c2bd53"} Nov 25 12:15:01 crc kubenswrapper[4776]: I1125 12:15:01.936379 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-zrwkj" event={"ID":"fa9774f8-e062-493f-9480-c71c2b6ff403","Type":"ContainerStarted","Data":"1ecda469e4435dd838d976119d5a55a1deab91db74d4e2f7f5af61ead326c2b5"} Nov 25 12:15:01 crc kubenswrapper[4776]: I1125 12:15:01.970620 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-zrwkj" podStartSLOduration=1.970585348 podStartE2EDuration="1.970585348s" podCreationTimestamp="2025-11-25 12:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 12:15:01.96030489 +0000 UTC m=+10247.001364433" watchObservedRunningTime="2025-11-25 12:15:01.970585348 +0000 UTC m=+10247.011644901" Nov 25 12:15:02 crc kubenswrapper[4776]: I1125 12:15:02.957004 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa9774f8-e062-493f-9480-c71c2b6ff403" containerID="172ad9574055df6b5941de65c56977552ee2e90a8edf13a4ec64dd86b2c2bd53" exitCode=0 Nov 25 12:15:02 crc kubenswrapper[4776]: I1125 12:15:02.958158 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-zrwkj" event={"ID":"fa9774f8-e062-493f-9480-c71c2b6ff403","Type":"ContainerDied","Data":"172ad9574055df6b5941de65c56977552ee2e90a8edf13a4ec64dd86b2c2bd53"} Nov 25 12:15:03 crc kubenswrapper[4776]: I1125 12:15:03.827130 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-rbfl6_a229c7a5-8a0d-451b-9b28-da9c8762044a/control-plane-machine-set-operator/0.log" Nov 25 12:15:04 crc kubenswrapper[4776]: I1125 12:15:04.018083 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-fjr5v_fffe85b6-294d-4f43-ad27-65a5d093c076/machine-api-operator/0.log" Nov 25 12:15:04 crc kubenswrapper[4776]: I1125 12:15:04.074055 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-fjr5v_fffe85b6-294d-4f43-ad27-65a5d093c076/kube-rbac-proxy/0.log" Nov 25 12:15:04 crc kubenswrapper[4776]: I1125 12:15:04.420897 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-zrwkj" Nov 25 12:15:04 crc kubenswrapper[4776]: I1125 12:15:04.525889 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwwgf\" (UniqueName: \"kubernetes.io/projected/fa9774f8-e062-493f-9480-c71c2b6ff403-kube-api-access-dwwgf\") pod \"fa9774f8-e062-493f-9480-c71c2b6ff403\" (UID: \"fa9774f8-e062-493f-9480-c71c2b6ff403\") " Nov 25 12:15:04 crc kubenswrapper[4776]: I1125 12:15:04.526177 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fa9774f8-e062-493f-9480-c71c2b6ff403-secret-volume\") pod \"fa9774f8-e062-493f-9480-c71c2b6ff403\" (UID: \"fa9774f8-e062-493f-9480-c71c2b6ff403\") " Nov 25 12:15:04 crc kubenswrapper[4776]: I1125 12:15:04.526300 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fa9774f8-e062-493f-9480-c71c2b6ff403-config-volume\") pod \"fa9774f8-e062-493f-9480-c71c2b6ff403\" (UID: \"fa9774f8-e062-493f-9480-c71c2b6ff403\") " Nov 25 12:15:04 crc kubenswrapper[4776]: I1125 12:15:04.527136 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa9774f8-e062-493f-9480-c71c2b6ff403-config-volume" (OuterVolumeSpecName: "config-volume") pod "fa9774f8-e062-493f-9480-c71c2b6ff403" (UID: "fa9774f8-e062-493f-9480-c71c2b6ff403"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 12:15:04 crc kubenswrapper[4776]: I1125 12:15:04.538446 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa9774f8-e062-493f-9480-c71c2b6ff403-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "fa9774f8-e062-493f-9480-c71c2b6ff403" (UID: "fa9774f8-e062-493f-9480-c71c2b6ff403"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 12:15:04 crc kubenswrapper[4776]: I1125 12:15:04.539077 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa9774f8-e062-493f-9480-c71c2b6ff403-kube-api-access-dwwgf" (OuterVolumeSpecName: "kube-api-access-dwwgf") pod "fa9774f8-e062-493f-9480-c71c2b6ff403" (UID: "fa9774f8-e062-493f-9480-c71c2b6ff403"). InnerVolumeSpecName "kube-api-access-dwwgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:15:04 crc kubenswrapper[4776]: I1125 12:15:04.629079 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwwgf\" (UniqueName: \"kubernetes.io/projected/fa9774f8-e062-493f-9480-c71c2b6ff403-kube-api-access-dwwgf\") on node \"crc\" DevicePath \"\"" Nov 25 12:15:04 crc kubenswrapper[4776]: I1125 12:15:04.629122 4776 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fa9774f8-e062-493f-9480-c71c2b6ff403-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 12:15:04 crc kubenswrapper[4776]: I1125 12:15:04.629132 4776 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fa9774f8-e062-493f-9480-c71c2b6ff403-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 12:15:04 crc kubenswrapper[4776]: I1125 12:15:04.986448 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-zrwkj" event={"ID":"fa9774f8-e062-493f-9480-c71c2b6ff403","Type":"ContainerDied","Data":"1ecda469e4435dd838d976119d5a55a1deab91db74d4e2f7f5af61ead326c2b5"} Nov 25 12:15:04 crc kubenswrapper[4776]: I1125 12:15:04.986499 4776 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ecda469e4435dd838d976119d5a55a1deab91db74d4e2f7f5af61ead326c2b5" Nov 25 12:15:04 crc kubenswrapper[4776]: I1125 12:15:04.986573 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401215-zrwkj" Nov 25 12:15:05 crc kubenswrapper[4776]: I1125 12:15:05.060851 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401170-t572v"] Nov 25 12:15:05 crc kubenswrapper[4776]: I1125 12:15:05.073434 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401170-t572v"] Nov 25 12:15:05 crc kubenswrapper[4776]: I1125 12:15:05.682482 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe" path="/var/lib/kubelet/pods/a38ad9f5-2ea1-4da3-93ce-d1c0033a3fbe/volumes" Nov 25 12:15:17 crc kubenswrapper[4776]: I1125 12:15:17.214234 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-qzgc4_24ae1849-101b-4e31-8189-eb4db03d8c73/cert-manager-controller/0.log" Nov 25 12:15:17 crc kubenswrapper[4776]: I1125 12:15:17.433798 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-d9hk5_c688fa59-c409-4344-bd47-65cc299fc20e/cert-manager-cainjector/0.log" Nov 25 12:15:17 crc kubenswrapper[4776]: I1125 12:15:17.466733 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-pt9xh_052e47c1-028a-4f6f-bc78-67b4ad351ffe/cert-manager-webhook/0.log" Nov 25 12:15:17 crc kubenswrapper[4776]: I1125 12:15:17.817913 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:15:17 crc kubenswrapper[4776]: I1125 12:15:17.817976 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:15:31 crc kubenswrapper[4776]: I1125 12:15:31.276908 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-p5grn_a3994dc8-916f-4efb-b723-03650a70411d/nmstate-console-plugin/0.log" Nov 25 12:15:31 crc kubenswrapper[4776]: I1125 12:15:31.478780 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-fv9bw_a4451cf3-5459-4208-bb86-49d3d60c180c/nmstate-handler/0.log" Nov 25 12:15:31 crc kubenswrapper[4776]: I1125 12:15:31.503365 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-8bncp_d070507a-81b9-4c71-b82e-5023abea5613/kube-rbac-proxy/0.log" Nov 25 12:15:31 crc kubenswrapper[4776]: I1125 12:15:31.567800 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-8bncp_d070507a-81b9-4c71-b82e-5023abea5613/nmstate-metrics/0.log" Nov 25 12:15:31 crc kubenswrapper[4776]: I1125 12:15:31.718887 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-klsvw_87dd97b9-3af4-41c6-8c24-db941e082eb6/nmstate-operator/0.log" Nov 25 12:15:31 crc kubenswrapper[4776]: I1125 12:15:31.807861 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-gdccx_cb651a64-745e-4587-b786-9e00604a5a77/nmstate-webhook/0.log" Nov 25 12:15:43 crc kubenswrapper[4776]: I1125 12:15:43.513981 4776 scope.go:117] "RemoveContainer" containerID="7b381a199dd0fbf4c2df793c6a00aa1f2b429bf2974992eb737c590f24e9c3be" Nov 25 12:15:47 crc kubenswrapper[4776]: I1125 12:15:47.818615 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:15:47 crc kubenswrapper[4776]: I1125 12:15:47.819479 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:15:48 crc kubenswrapper[4776]: I1125 12:15:48.000020 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-jvm2p_d74cede2-6a31-4173-a6af-87669c0804bf/kube-rbac-proxy/0.log" Nov 25 12:15:48 crc kubenswrapper[4776]: I1125 12:15:48.247349 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj78r_6bd80dc6-3a88-459e-bfea-f2fc22aed477/cp-frr-files/0.log" Nov 25 12:15:48 crc kubenswrapper[4776]: I1125 12:15:48.485657 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-jvm2p_d74cede2-6a31-4173-a6af-87669c0804bf/controller/0.log" Nov 25 12:15:48 crc kubenswrapper[4776]: I1125 12:15:48.511389 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj78r_6bd80dc6-3a88-459e-bfea-f2fc22aed477/cp-frr-files/0.log" Nov 25 12:15:48 crc kubenswrapper[4776]: I1125 12:15:48.530676 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj78r_6bd80dc6-3a88-459e-bfea-f2fc22aed477/cp-metrics/0.log" Nov 25 12:15:48 crc kubenswrapper[4776]: I1125 12:15:48.558537 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj78r_6bd80dc6-3a88-459e-bfea-f2fc22aed477/cp-reloader/0.log" Nov 25 12:15:48 crc kubenswrapper[4776]: I1125 12:15:48.706567 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj78r_6bd80dc6-3a88-459e-bfea-f2fc22aed477/cp-reloader/0.log" Nov 25 12:15:48 crc kubenswrapper[4776]: I1125 12:15:48.949813 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj78r_6bd80dc6-3a88-459e-bfea-f2fc22aed477/cp-frr-files/0.log" Nov 25 12:15:48 crc kubenswrapper[4776]: I1125 12:15:48.957239 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj78r_6bd80dc6-3a88-459e-bfea-f2fc22aed477/cp-metrics/0.log" Nov 25 12:15:49 crc kubenswrapper[4776]: I1125 12:15:49.010691 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj78r_6bd80dc6-3a88-459e-bfea-f2fc22aed477/cp-metrics/0.log" Nov 25 12:15:49 crc kubenswrapper[4776]: I1125 12:15:49.014384 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj78r_6bd80dc6-3a88-459e-bfea-f2fc22aed477/cp-reloader/0.log" Nov 25 12:15:49 crc kubenswrapper[4776]: I1125 12:15:49.171942 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj78r_6bd80dc6-3a88-459e-bfea-f2fc22aed477/cp-frr-files/0.log" Nov 25 12:15:49 crc kubenswrapper[4776]: I1125 12:15:49.210310 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj78r_6bd80dc6-3a88-459e-bfea-f2fc22aed477/cp-reloader/0.log" Nov 25 12:15:49 crc kubenswrapper[4776]: I1125 12:15:49.256903 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj78r_6bd80dc6-3a88-459e-bfea-f2fc22aed477/cp-metrics/0.log" Nov 25 12:15:49 crc kubenswrapper[4776]: I1125 12:15:49.301533 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj78r_6bd80dc6-3a88-459e-bfea-f2fc22aed477/controller/0.log" Nov 25 12:15:49 crc kubenswrapper[4776]: I1125 12:15:49.437227 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj78r_6bd80dc6-3a88-459e-bfea-f2fc22aed477/frr-metrics/0.log" Nov 25 12:15:49 crc kubenswrapper[4776]: I1125 12:15:49.505457 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj78r_6bd80dc6-3a88-459e-bfea-f2fc22aed477/kube-rbac-proxy/0.log" Nov 25 12:15:49 crc kubenswrapper[4776]: I1125 12:15:49.577042 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj78r_6bd80dc6-3a88-459e-bfea-f2fc22aed477/kube-rbac-proxy-frr/0.log" Nov 25 12:15:49 crc kubenswrapper[4776]: I1125 12:15:49.730927 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj78r_6bd80dc6-3a88-459e-bfea-f2fc22aed477/reloader/0.log" Nov 25 12:15:49 crc kubenswrapper[4776]: I1125 12:15:49.875493 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-dfbc5_14a3e0bf-32be-4251-9d91-9e41b841d9e2/frr-k8s-webhook-server/0.log" Nov 25 12:15:50 crc kubenswrapper[4776]: I1125 12:15:50.046209 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6bc7ff74-znqxl_3700db74-6cd7-4370-8b44-fea7c84052af/manager/0.log" Nov 25 12:15:50 crc kubenswrapper[4776]: I1125 12:15:50.316904 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5c67c9cb54-vgffr_12f6e05c-cc4d-4e0c-8a1a-bfe459e58d70/webhook-server/0.log" Nov 25 12:15:50 crc kubenswrapper[4776]: I1125 12:15:50.469881 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-p9fdc_294e21df-2814-43a8-975c-e0beb19b3e03/kube-rbac-proxy/0.log" Nov 25 12:15:51 crc kubenswrapper[4776]: I1125 12:15:51.560671 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-p9fdc_294e21df-2814-43a8-975c-e0beb19b3e03/speaker/0.log" Nov 25 12:15:53 crc kubenswrapper[4776]: I1125 12:15:53.339586 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj78r_6bd80dc6-3a88-459e-bfea-f2fc22aed477/frr/0.log" Nov 25 12:16:08 crc kubenswrapper[4776]: I1125 12:16:08.474817 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s_ee805461-014c-4400-9310-f2862d366911/util/0.log" Nov 25 12:16:08 crc kubenswrapper[4776]: I1125 12:16:08.721497 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s_ee805461-014c-4400-9310-f2862d366911/pull/0.log" Nov 25 12:16:08 crc kubenswrapper[4776]: I1125 12:16:08.727090 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s_ee805461-014c-4400-9310-f2862d366911/pull/0.log" Nov 25 12:16:08 crc kubenswrapper[4776]: I1125 12:16:08.897744 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s_ee805461-014c-4400-9310-f2862d366911/util/0.log" Nov 25 12:16:08 crc kubenswrapper[4776]: I1125 12:16:08.998998 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s_ee805461-014c-4400-9310-f2862d366911/pull/0.log" Nov 25 12:16:09 crc kubenswrapper[4776]: I1125 12:16:09.036537 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s_ee805461-014c-4400-9310-f2862d366911/util/0.log" Nov 25 12:16:09 crc kubenswrapper[4776]: I1125 12:16:09.078834 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931abhm7s_ee805461-014c-4400-9310-f2862d366911/extract/0.log" Nov 25 12:16:09 crc kubenswrapper[4776]: I1125 12:16:09.271814 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m_b55a5ed7-b167-447e-b9ae-16ea2c904b39/util/0.log" Nov 25 12:16:09 crc kubenswrapper[4776]: I1125 12:16:09.407090 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m_b55a5ed7-b167-447e-b9ae-16ea2c904b39/util/0.log" Nov 25 12:16:09 crc kubenswrapper[4776]: I1125 12:16:09.421802 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m_b55a5ed7-b167-447e-b9ae-16ea2c904b39/pull/0.log" Nov 25 12:16:09 crc kubenswrapper[4776]: I1125 12:16:09.426579 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m_b55a5ed7-b167-447e-b9ae-16ea2c904b39/pull/0.log" Nov 25 12:16:09 crc kubenswrapper[4776]: I1125 12:16:09.636058 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m_b55a5ed7-b167-447e-b9ae-16ea2c904b39/extract/0.log" Nov 25 12:16:09 crc kubenswrapper[4776]: I1125 12:16:09.637661 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m_b55a5ed7-b167-447e-b9ae-16ea2c904b39/pull/0.log" Nov 25 12:16:09 crc kubenswrapper[4776]: I1125 12:16:09.684467 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ehhm7m_b55a5ed7-b167-447e-b9ae-16ea2c904b39/util/0.log" Nov 25 12:16:09 crc kubenswrapper[4776]: I1125 12:16:09.870707 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz_dbc20081-c609-4ddf-a321-2ae092fd67e9/util/0.log" Nov 25 12:16:09 crc kubenswrapper[4776]: I1125 12:16:09.975724 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz_dbc20081-c609-4ddf-a321-2ae092fd67e9/util/0.log" Nov 25 12:16:10 crc kubenswrapper[4776]: I1125 12:16:10.033015 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz_dbc20081-c609-4ddf-a321-2ae092fd67e9/pull/0.log" Nov 25 12:16:11 crc kubenswrapper[4776]: I1125 12:16:11.152827 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz_dbc20081-c609-4ddf-a321-2ae092fd67e9/pull/0.log" Nov 25 12:16:11 crc kubenswrapper[4776]: I1125 12:16:11.325254 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz_dbc20081-c609-4ddf-a321-2ae092fd67e9/util/0.log" Nov 25 12:16:11 crc kubenswrapper[4776]: I1125 12:16:11.364831 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz_dbc20081-c609-4ddf-a321-2ae092fd67e9/extract/0.log" Nov 25 12:16:11 crc kubenswrapper[4776]: I1125 12:16:11.371002 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q5dtz_dbc20081-c609-4ddf-a321-2ae092fd67e9/pull/0.log" Nov 25 12:16:11 crc kubenswrapper[4776]: I1125 12:16:11.493623 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rzzlh_de4c5588-c756-4f79-b6a3-3d533fce4008/extract-utilities/0.log" Nov 25 12:16:11 crc kubenswrapper[4776]: I1125 12:16:11.706895 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rzzlh_de4c5588-c756-4f79-b6a3-3d533fce4008/extract-utilities/0.log" Nov 25 12:16:11 crc kubenswrapper[4776]: I1125 12:16:11.719369 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rzzlh_de4c5588-c756-4f79-b6a3-3d533fce4008/extract-content/0.log" Nov 25 12:16:11 crc kubenswrapper[4776]: I1125 12:16:11.726029 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rzzlh_de4c5588-c756-4f79-b6a3-3d533fce4008/extract-content/0.log" Nov 25 12:16:11 crc kubenswrapper[4776]: I1125 12:16:11.968336 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rzzlh_de4c5588-c756-4f79-b6a3-3d533fce4008/extract-utilities/0.log" Nov 25 12:16:12 crc kubenswrapper[4776]: I1125 12:16:12.249044 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rzzlh_de4c5588-c756-4f79-b6a3-3d533fce4008/extract-content/0.log" Nov 25 12:16:12 crc kubenswrapper[4776]: I1125 12:16:12.256860 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qtfp5_803c71a8-0f91-4fa4-949f-5995ae8af48c/extract-utilities/0.log" Nov 25 12:16:12 crc kubenswrapper[4776]: I1125 12:16:12.424023 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qtfp5_803c71a8-0f91-4fa4-949f-5995ae8af48c/extract-content/0.log" Nov 25 12:16:12 crc kubenswrapper[4776]: I1125 12:16:12.447920 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qtfp5_803c71a8-0f91-4fa4-949f-5995ae8af48c/extract-content/0.log" Nov 25 12:16:12 crc kubenswrapper[4776]: I1125 12:16:12.450474 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qtfp5_803c71a8-0f91-4fa4-949f-5995ae8af48c/extract-utilities/0.log" Nov 25 12:16:12 crc kubenswrapper[4776]: I1125 12:16:12.802734 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qtfp5_803c71a8-0f91-4fa4-949f-5995ae8af48c/extract-utilities/0.log" Nov 25 12:16:12 crc kubenswrapper[4776]: I1125 12:16:12.896450 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qtfp5_803c71a8-0f91-4fa4-949f-5995ae8af48c/extract-content/0.log" Nov 25 12:16:13 crc kubenswrapper[4776]: I1125 12:16:13.103727 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq_c7cdbd4e-245e-4b7a-890d-c7695d445cb1/util/0.log" Nov 25 12:16:13 crc kubenswrapper[4776]: I1125 12:16:13.322922 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq_c7cdbd4e-245e-4b7a-890d-c7695d445cb1/pull/0.log" Nov 25 12:16:13 crc kubenswrapper[4776]: I1125 12:16:13.386315 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq_c7cdbd4e-245e-4b7a-890d-c7695d445cb1/pull/0.log" Nov 25 12:16:13 crc kubenswrapper[4776]: I1125 12:16:13.449465 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq_c7cdbd4e-245e-4b7a-890d-c7695d445cb1/util/0.log" Nov 25 12:16:13 crc kubenswrapper[4776]: I1125 12:16:13.532996 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq_c7cdbd4e-245e-4b7a-890d-c7695d445cb1/util/0.log" Nov 25 12:16:13 crc kubenswrapper[4776]: I1125 12:16:13.593786 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rzzlh_de4c5588-c756-4f79-b6a3-3d533fce4008/registry-server/0.log" Nov 25 12:16:13 crc kubenswrapper[4776]: I1125 12:16:13.695270 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq_c7cdbd4e-245e-4b7a-890d-c7695d445cb1/pull/0.log" Nov 25 12:16:13 crc kubenswrapper[4776]: I1125 12:16:13.874204 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6ltgpq_c7cdbd4e-245e-4b7a-890d-c7695d445cb1/extract/0.log" Nov 25 12:16:13 crc kubenswrapper[4776]: I1125 12:16:13.982482 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-xcd5g_c7a039e1-a9b6-4fc9-b297-6cb3eb6618f2/marketplace-operator/0.log" Nov 25 12:16:14 crc kubenswrapper[4776]: I1125 12:16:14.080655 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sz9dx_cb60184b-c3b3-4ddb-90da-01c62fb183d7/extract-utilities/0.log" Nov 25 12:16:14 crc kubenswrapper[4776]: I1125 12:16:14.245039 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sz9dx_cb60184b-c3b3-4ddb-90da-01c62fb183d7/extract-utilities/0.log" Nov 25 12:16:14 crc kubenswrapper[4776]: I1125 12:16:14.253683 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sz9dx_cb60184b-c3b3-4ddb-90da-01c62fb183d7/extract-content/0.log" Nov 25 12:16:14 crc kubenswrapper[4776]: I1125 12:16:14.260594 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sz9dx_cb60184b-c3b3-4ddb-90da-01c62fb183d7/extract-content/0.log" Nov 25 12:16:14 crc kubenswrapper[4776]: I1125 12:16:14.544483 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sz9dx_cb60184b-c3b3-4ddb-90da-01c62fb183d7/extract-content/0.log" Nov 25 12:16:14 crc kubenswrapper[4776]: I1125 12:16:14.577592 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sz9dx_cb60184b-c3b3-4ddb-90da-01c62fb183d7/extract-utilities/0.log" Nov 25 12:16:14 crc kubenswrapper[4776]: I1125 12:16:14.859083 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c2d7k_5d46339c-f862-49cf-b4e9-fde20e8688de/extract-utilities/0.log" Nov 25 12:16:14 crc kubenswrapper[4776]: I1125 12:16:14.993388 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c2d7k_5d46339c-f862-49cf-b4e9-fde20e8688de/extract-utilities/0.log" Nov 25 12:16:14 crc kubenswrapper[4776]: I1125 12:16:14.993425 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c2d7k_5d46339c-f862-49cf-b4e9-fde20e8688de/extract-content/0.log" Nov 25 12:16:15 crc kubenswrapper[4776]: I1125 12:16:15.090904 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c2d7k_5d46339c-f862-49cf-b4e9-fde20e8688de/extract-content/0.log" Nov 25 12:16:15 crc kubenswrapper[4776]: I1125 12:16:15.254454 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-sz9dx_cb60184b-c3b3-4ddb-90da-01c62fb183d7/registry-server/0.log" Nov 25 12:16:15 crc kubenswrapper[4776]: I1125 12:16:15.339955 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c2d7k_5d46339c-f862-49cf-b4e9-fde20e8688de/extract-content/0.log" Nov 25 12:16:15 crc kubenswrapper[4776]: I1125 12:16:15.406555 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c2d7k_5d46339c-f862-49cf-b4e9-fde20e8688de/extract-utilities/0.log" Nov 25 12:16:15 crc kubenswrapper[4776]: I1125 12:16:15.930194 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-c2d7k_5d46339c-f862-49cf-b4e9-fde20e8688de/registry-server/0.log" Nov 25 12:16:16 crc kubenswrapper[4776]: I1125 12:16:16.625516 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qtfp5_803c71a8-0f91-4fa4-949f-5995ae8af48c/registry-server/0.log" Nov 25 12:16:17 crc kubenswrapper[4776]: I1125 12:16:17.818012 4776 patch_prober.go:28] interesting pod/machine-config-daemon-84dqb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 12:16:17 crc kubenswrapper[4776]: I1125 12:16:17.819592 4776 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 12:16:17 crc kubenswrapper[4776]: I1125 12:16:17.819743 4776 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" Nov 25 12:16:17 crc kubenswrapper[4776]: I1125 12:16:17.820839 4776 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563"} pod="openshift-machine-config-operator/machine-config-daemon-84dqb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 12:16:17 crc kubenswrapper[4776]: I1125 12:16:17.821021 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerName="machine-config-daemon" containerID="cri-o://d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563" gracePeriod=600 Nov 25 12:16:18 crc kubenswrapper[4776]: E1125 12:16:18.504839 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:16:18 crc kubenswrapper[4776]: I1125 12:16:18.877317 4776 generic.go:334] "Generic (PLEG): container finished" podID="fa47ebcc-a95e-4693-876d-7284c28c3ade" containerID="d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563" exitCode=0 Nov 25 12:16:18 crc kubenswrapper[4776]: I1125 12:16:18.877363 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerDied","Data":"d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563"} Nov 25 12:16:18 crc kubenswrapper[4776]: I1125 12:16:18.877396 4776 scope.go:117] "RemoveContainer" containerID="5b13db60744f199580371de408731797095f5765187d672f3989cfe03fd42a7f" Nov 25 12:16:18 crc kubenswrapper[4776]: I1125 12:16:18.878196 4776 scope.go:117] "RemoveContainer" containerID="d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563" Nov 25 12:16:18 crc kubenswrapper[4776]: E1125 12:16:18.878565 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:16:31 crc kubenswrapper[4776]: I1125 12:16:31.227942 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-n9rjm_05d15308-0620-46cf-838e-1c8da434b69f/prometheus-operator/0.log" Nov 25 12:16:31 crc kubenswrapper[4776]: I1125 12:16:31.423813 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-544ff7697f-mdz99_981f87ca-c762-4f02-b2bf-b22732edfd22/prometheus-operator-admission-webhook/0.log" Nov 25 12:16:31 crc kubenswrapper[4776]: I1125 12:16:31.429823 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-544ff7697f-l2d6r_404823ba-8531-4881-a133-e3900d9bd6c8/prometheus-operator-admission-webhook/0.log" Nov 25 12:16:31 crc kubenswrapper[4776]: I1125 12:16:31.612133 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-6hnbp_79fbf35b-52fb-4f94-89b9-d5b257b86de6/operator/0.log" Nov 25 12:16:31 crc kubenswrapper[4776]: I1125 12:16:31.657937 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-zr5h9_06bd1f63-0261-44b4-91ca-483b379c21d8/perses-operator/0.log" Nov 25 12:16:31 crc kubenswrapper[4776]: I1125 12:16:31.663084 4776 scope.go:117] "RemoveContainer" containerID="d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563" Nov 25 12:16:31 crc kubenswrapper[4776]: E1125 12:16:31.663387 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:16:38 crc kubenswrapper[4776]: E1125 12:16:38.919485 4776 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.193:45244->38.102.83.193:36915: read tcp 38.102.83.193:45244->38.102.83.193:36915: read: connection reset by peer Nov 25 12:16:46 crc kubenswrapper[4776]: I1125 12:16:46.662794 4776 scope.go:117] "RemoveContainer" containerID="d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563" Nov 25 12:16:46 crc kubenswrapper[4776]: E1125 12:16:46.663829 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:16:59 crc kubenswrapper[4776]: I1125 12:16:59.663346 4776 scope.go:117] "RemoveContainer" containerID="d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563" Nov 25 12:16:59 crc kubenswrapper[4776]: E1125 12:16:59.664524 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:17:04 crc kubenswrapper[4776]: E1125 12:17:04.041719 4776 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.193:51938->38.102.83.193:36915: write tcp 38.102.83.193:51938->38.102.83.193:36915: write: connection reset by peer Nov 25 12:17:12 crc kubenswrapper[4776]: I1125 12:17:12.664014 4776 scope.go:117] "RemoveContainer" containerID="d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563" Nov 25 12:17:12 crc kubenswrapper[4776]: E1125 12:17:12.664910 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:17:26 crc kubenswrapper[4776]: I1125 12:17:26.663047 4776 scope.go:117] "RemoveContainer" containerID="d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563" Nov 25 12:17:26 crc kubenswrapper[4776]: E1125 12:17:26.664128 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:17:41 crc kubenswrapper[4776]: I1125 12:17:41.664509 4776 scope.go:117] "RemoveContainer" containerID="d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563" Nov 25 12:17:41 crc kubenswrapper[4776]: E1125 12:17:41.668937 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:17:55 crc kubenswrapper[4776]: I1125 12:17:55.676538 4776 scope.go:117] "RemoveContainer" containerID="d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563" Nov 25 12:17:55 crc kubenswrapper[4776]: E1125 12:17:55.677749 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:18:09 crc kubenswrapper[4776]: I1125 12:18:09.662619 4776 scope.go:117] "RemoveContainer" containerID="d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563" Nov 25 12:18:09 crc kubenswrapper[4776]: E1125 12:18:09.663647 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:18:09 crc kubenswrapper[4776]: I1125 12:18:09.864604 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x44pz"] Nov 25 12:18:09 crc kubenswrapper[4776]: E1125 12:18:09.865192 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa9774f8-e062-493f-9480-c71c2b6ff403" containerName="collect-profiles" Nov 25 12:18:09 crc kubenswrapper[4776]: I1125 12:18:09.865218 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa9774f8-e062-493f-9480-c71c2b6ff403" containerName="collect-profiles" Nov 25 12:18:09 crc kubenswrapper[4776]: I1125 12:18:09.865496 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa9774f8-e062-493f-9480-c71c2b6ff403" containerName="collect-profiles" Nov 25 12:18:09 crc kubenswrapper[4776]: I1125 12:18:09.867110 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x44pz" Nov 25 12:18:09 crc kubenswrapper[4776]: I1125 12:18:09.876468 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x44pz"] Nov 25 12:18:09 crc kubenswrapper[4776]: I1125 12:18:09.997731 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pr4w\" (UniqueName: \"kubernetes.io/projected/9f4f4be9-19a4-4872-961e-aa545c563aad-kube-api-access-7pr4w\") pod \"redhat-marketplace-x44pz\" (UID: \"9f4f4be9-19a4-4872-961e-aa545c563aad\") " pod="openshift-marketplace/redhat-marketplace-x44pz" Nov 25 12:18:09 crc kubenswrapper[4776]: I1125 12:18:09.997999 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f4f4be9-19a4-4872-961e-aa545c563aad-utilities\") pod \"redhat-marketplace-x44pz\" (UID: \"9f4f4be9-19a4-4872-961e-aa545c563aad\") " pod="openshift-marketplace/redhat-marketplace-x44pz" Nov 25 12:18:09 crc kubenswrapper[4776]: I1125 12:18:09.998092 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f4f4be9-19a4-4872-961e-aa545c563aad-catalog-content\") pod \"redhat-marketplace-x44pz\" (UID: \"9f4f4be9-19a4-4872-961e-aa545c563aad\") " pod="openshift-marketplace/redhat-marketplace-x44pz" Nov 25 12:18:10 crc kubenswrapper[4776]: I1125 12:18:10.100029 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f4f4be9-19a4-4872-961e-aa545c563aad-utilities\") pod \"redhat-marketplace-x44pz\" (UID: \"9f4f4be9-19a4-4872-961e-aa545c563aad\") " pod="openshift-marketplace/redhat-marketplace-x44pz" Nov 25 12:18:10 crc kubenswrapper[4776]: I1125 12:18:10.100120 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f4f4be9-19a4-4872-961e-aa545c563aad-catalog-content\") pod \"redhat-marketplace-x44pz\" (UID: \"9f4f4be9-19a4-4872-961e-aa545c563aad\") " pod="openshift-marketplace/redhat-marketplace-x44pz" Nov 25 12:18:10 crc kubenswrapper[4776]: I1125 12:18:10.100221 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pr4w\" (UniqueName: \"kubernetes.io/projected/9f4f4be9-19a4-4872-961e-aa545c563aad-kube-api-access-7pr4w\") pod \"redhat-marketplace-x44pz\" (UID: \"9f4f4be9-19a4-4872-961e-aa545c563aad\") " pod="openshift-marketplace/redhat-marketplace-x44pz" Nov 25 12:18:10 crc kubenswrapper[4776]: I1125 12:18:10.100595 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f4f4be9-19a4-4872-961e-aa545c563aad-utilities\") pod \"redhat-marketplace-x44pz\" (UID: \"9f4f4be9-19a4-4872-961e-aa545c563aad\") " pod="openshift-marketplace/redhat-marketplace-x44pz" Nov 25 12:18:10 crc kubenswrapper[4776]: I1125 12:18:10.100634 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f4f4be9-19a4-4872-961e-aa545c563aad-catalog-content\") pod \"redhat-marketplace-x44pz\" (UID: \"9f4f4be9-19a4-4872-961e-aa545c563aad\") " pod="openshift-marketplace/redhat-marketplace-x44pz" Nov 25 12:18:10 crc kubenswrapper[4776]: I1125 12:18:10.128318 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pr4w\" (UniqueName: \"kubernetes.io/projected/9f4f4be9-19a4-4872-961e-aa545c563aad-kube-api-access-7pr4w\") pod \"redhat-marketplace-x44pz\" (UID: \"9f4f4be9-19a4-4872-961e-aa545c563aad\") " pod="openshift-marketplace/redhat-marketplace-x44pz" Nov 25 12:18:10 crc kubenswrapper[4776]: I1125 12:18:10.196427 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x44pz" Nov 25 12:18:10 crc kubenswrapper[4776]: I1125 12:18:10.705883 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x44pz"] Nov 25 12:18:11 crc kubenswrapper[4776]: I1125 12:18:11.187158 4776 generic.go:334] "Generic (PLEG): container finished" podID="9f4f4be9-19a4-4872-961e-aa545c563aad" containerID="817faa7aee46a29b0145616eedde10d723403c57e194bbf1709f68a8b85aed9a" exitCode=0 Nov 25 12:18:11 crc kubenswrapper[4776]: I1125 12:18:11.187402 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x44pz" event={"ID":"9f4f4be9-19a4-4872-961e-aa545c563aad","Type":"ContainerDied","Data":"817faa7aee46a29b0145616eedde10d723403c57e194bbf1709f68a8b85aed9a"} Nov 25 12:18:11 crc kubenswrapper[4776]: I1125 12:18:11.187507 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x44pz" event={"ID":"9f4f4be9-19a4-4872-961e-aa545c563aad","Type":"ContainerStarted","Data":"95cef06bbec86674b6e15cffc16ec52192b476ba4f3f7e02d5922ed06980ba64"} Nov 25 12:18:11 crc kubenswrapper[4776]: I1125 12:18:11.192716 4776 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 12:18:12 crc kubenswrapper[4776]: I1125 12:18:12.202221 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x44pz" event={"ID":"9f4f4be9-19a4-4872-961e-aa545c563aad","Type":"ContainerStarted","Data":"4660b9b162d08cede6f20d98ab510bcb4cb61f68a906181eaf43e9ec38c2c4a6"} Nov 25 12:18:13 crc kubenswrapper[4776]: I1125 12:18:13.213727 4776 generic.go:334] "Generic (PLEG): container finished" podID="9f4f4be9-19a4-4872-961e-aa545c563aad" containerID="4660b9b162d08cede6f20d98ab510bcb4cb61f68a906181eaf43e9ec38c2c4a6" exitCode=0 Nov 25 12:18:13 crc kubenswrapper[4776]: I1125 12:18:13.213773 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x44pz" event={"ID":"9f4f4be9-19a4-4872-961e-aa545c563aad","Type":"ContainerDied","Data":"4660b9b162d08cede6f20d98ab510bcb4cb61f68a906181eaf43e9ec38c2c4a6"} Nov 25 12:18:14 crc kubenswrapper[4776]: I1125 12:18:14.225800 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x44pz" event={"ID":"9f4f4be9-19a4-4872-961e-aa545c563aad","Type":"ContainerStarted","Data":"99781065f106949cb1a47961e65b7cc950a753562eca2343797c3725929af90a"} Nov 25 12:18:14 crc kubenswrapper[4776]: I1125 12:18:14.252850 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x44pz" podStartSLOduration=2.696540797 podStartE2EDuration="5.252817912s" podCreationTimestamp="2025-11-25 12:18:09 +0000 UTC" firstStartedPulling="2025-11-25 12:18:11.192439958 +0000 UTC m=+10436.233499511" lastFinishedPulling="2025-11-25 12:18:13.748717063 +0000 UTC m=+10438.789776626" observedRunningTime="2025-11-25 12:18:14.24595531 +0000 UTC m=+10439.287014873" watchObservedRunningTime="2025-11-25 12:18:14.252817912 +0000 UTC m=+10439.293877465" Nov 25 12:18:20 crc kubenswrapper[4776]: I1125 12:18:20.201608 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x44pz" Nov 25 12:18:20 crc kubenswrapper[4776]: I1125 12:18:20.202368 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x44pz" Nov 25 12:18:20 crc kubenswrapper[4776]: I1125 12:18:20.271391 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x44pz" Nov 25 12:18:20 crc kubenswrapper[4776]: I1125 12:18:20.348281 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x44pz" Nov 25 12:18:20 crc kubenswrapper[4776]: I1125 12:18:20.518104 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x44pz"] Nov 25 12:18:22 crc kubenswrapper[4776]: I1125 12:18:22.308773 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-x44pz" podUID="9f4f4be9-19a4-4872-961e-aa545c563aad" containerName="registry-server" containerID="cri-o://99781065f106949cb1a47961e65b7cc950a753562eca2343797c3725929af90a" gracePeriod=2 Nov 25 12:18:22 crc kubenswrapper[4776]: I1125 12:18:22.846169 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x44pz" Nov 25 12:18:22 crc kubenswrapper[4776]: I1125 12:18:22.912743 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f4f4be9-19a4-4872-961e-aa545c563aad-catalog-content\") pod \"9f4f4be9-19a4-4872-961e-aa545c563aad\" (UID: \"9f4f4be9-19a4-4872-961e-aa545c563aad\") " Nov 25 12:18:22 crc kubenswrapper[4776]: I1125 12:18:22.912806 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pr4w\" (UniqueName: \"kubernetes.io/projected/9f4f4be9-19a4-4872-961e-aa545c563aad-kube-api-access-7pr4w\") pod \"9f4f4be9-19a4-4872-961e-aa545c563aad\" (UID: \"9f4f4be9-19a4-4872-961e-aa545c563aad\") " Nov 25 12:18:22 crc kubenswrapper[4776]: I1125 12:18:22.912861 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f4f4be9-19a4-4872-961e-aa545c563aad-utilities\") pod \"9f4f4be9-19a4-4872-961e-aa545c563aad\" (UID: \"9f4f4be9-19a4-4872-961e-aa545c563aad\") " Nov 25 12:18:22 crc kubenswrapper[4776]: I1125 12:18:22.914394 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f4f4be9-19a4-4872-961e-aa545c563aad-utilities" (OuterVolumeSpecName: "utilities") pod "9f4f4be9-19a4-4872-961e-aa545c563aad" (UID: "9f4f4be9-19a4-4872-961e-aa545c563aad"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:18:22 crc kubenswrapper[4776]: I1125 12:18:22.933209 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f4f4be9-19a4-4872-961e-aa545c563aad-kube-api-access-7pr4w" (OuterVolumeSpecName: "kube-api-access-7pr4w") pod "9f4f4be9-19a4-4872-961e-aa545c563aad" (UID: "9f4f4be9-19a4-4872-961e-aa545c563aad"). InnerVolumeSpecName "kube-api-access-7pr4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:18:22 crc kubenswrapper[4776]: I1125 12:18:22.933453 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f4f4be9-19a4-4872-961e-aa545c563aad-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9f4f4be9-19a4-4872-961e-aa545c563aad" (UID: "9f4f4be9-19a4-4872-961e-aa545c563aad"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:18:23 crc kubenswrapper[4776]: I1125 12:18:23.014840 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f4f4be9-19a4-4872-961e-aa545c563aad-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:18:23 crc kubenswrapper[4776]: I1125 12:18:23.015178 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pr4w\" (UniqueName: \"kubernetes.io/projected/9f4f4be9-19a4-4872-961e-aa545c563aad-kube-api-access-7pr4w\") on node \"crc\" DevicePath \"\"" Nov 25 12:18:23 crc kubenswrapper[4776]: I1125 12:18:23.015278 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f4f4be9-19a4-4872-961e-aa545c563aad-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:18:23 crc kubenswrapper[4776]: I1125 12:18:23.326662 4776 generic.go:334] "Generic (PLEG): container finished" podID="9f4f4be9-19a4-4872-961e-aa545c563aad" containerID="99781065f106949cb1a47961e65b7cc950a753562eca2343797c3725929af90a" exitCode=0 Nov 25 12:18:23 crc kubenswrapper[4776]: I1125 12:18:23.327114 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x44pz" event={"ID":"9f4f4be9-19a4-4872-961e-aa545c563aad","Type":"ContainerDied","Data":"99781065f106949cb1a47961e65b7cc950a753562eca2343797c3725929af90a"} Nov 25 12:18:23 crc kubenswrapper[4776]: I1125 12:18:23.327141 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x44pz" event={"ID":"9f4f4be9-19a4-4872-961e-aa545c563aad","Type":"ContainerDied","Data":"95cef06bbec86674b6e15cffc16ec52192b476ba4f3f7e02d5922ed06980ba64"} Nov 25 12:18:23 crc kubenswrapper[4776]: I1125 12:18:23.327158 4776 scope.go:117] "RemoveContainer" containerID="99781065f106949cb1a47961e65b7cc950a753562eca2343797c3725929af90a" Nov 25 12:18:23 crc kubenswrapper[4776]: I1125 12:18:23.327331 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x44pz" Nov 25 12:18:23 crc kubenswrapper[4776]: I1125 12:18:23.353719 4776 scope.go:117] "RemoveContainer" containerID="4660b9b162d08cede6f20d98ab510bcb4cb61f68a906181eaf43e9ec38c2c4a6" Nov 25 12:18:23 crc kubenswrapper[4776]: I1125 12:18:23.381697 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x44pz"] Nov 25 12:18:23 crc kubenswrapper[4776]: I1125 12:18:23.391585 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-x44pz"] Nov 25 12:18:23 crc kubenswrapper[4776]: I1125 12:18:23.396152 4776 scope.go:117] "RemoveContainer" containerID="817faa7aee46a29b0145616eedde10d723403c57e194bbf1709f68a8b85aed9a" Nov 25 12:18:23 crc kubenswrapper[4776]: I1125 12:18:23.442437 4776 scope.go:117] "RemoveContainer" containerID="99781065f106949cb1a47961e65b7cc950a753562eca2343797c3725929af90a" Nov 25 12:18:23 crc kubenswrapper[4776]: E1125 12:18:23.442918 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99781065f106949cb1a47961e65b7cc950a753562eca2343797c3725929af90a\": container with ID starting with 99781065f106949cb1a47961e65b7cc950a753562eca2343797c3725929af90a not found: ID does not exist" containerID="99781065f106949cb1a47961e65b7cc950a753562eca2343797c3725929af90a" Nov 25 12:18:23 crc kubenswrapper[4776]: I1125 12:18:23.442953 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99781065f106949cb1a47961e65b7cc950a753562eca2343797c3725929af90a"} err="failed to get container status \"99781065f106949cb1a47961e65b7cc950a753562eca2343797c3725929af90a\": rpc error: code = NotFound desc = could not find container \"99781065f106949cb1a47961e65b7cc950a753562eca2343797c3725929af90a\": container with ID starting with 99781065f106949cb1a47961e65b7cc950a753562eca2343797c3725929af90a not found: ID does not exist" Nov 25 12:18:23 crc kubenswrapper[4776]: I1125 12:18:23.442981 4776 scope.go:117] "RemoveContainer" containerID="4660b9b162d08cede6f20d98ab510bcb4cb61f68a906181eaf43e9ec38c2c4a6" Nov 25 12:18:23 crc kubenswrapper[4776]: E1125 12:18:23.443233 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4660b9b162d08cede6f20d98ab510bcb4cb61f68a906181eaf43e9ec38c2c4a6\": container with ID starting with 4660b9b162d08cede6f20d98ab510bcb4cb61f68a906181eaf43e9ec38c2c4a6 not found: ID does not exist" containerID="4660b9b162d08cede6f20d98ab510bcb4cb61f68a906181eaf43e9ec38c2c4a6" Nov 25 12:18:23 crc kubenswrapper[4776]: I1125 12:18:23.443285 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4660b9b162d08cede6f20d98ab510bcb4cb61f68a906181eaf43e9ec38c2c4a6"} err="failed to get container status \"4660b9b162d08cede6f20d98ab510bcb4cb61f68a906181eaf43e9ec38c2c4a6\": rpc error: code = NotFound desc = could not find container \"4660b9b162d08cede6f20d98ab510bcb4cb61f68a906181eaf43e9ec38c2c4a6\": container with ID starting with 4660b9b162d08cede6f20d98ab510bcb4cb61f68a906181eaf43e9ec38c2c4a6 not found: ID does not exist" Nov 25 12:18:23 crc kubenswrapper[4776]: I1125 12:18:23.443305 4776 scope.go:117] "RemoveContainer" containerID="817faa7aee46a29b0145616eedde10d723403c57e194bbf1709f68a8b85aed9a" Nov 25 12:18:23 crc kubenswrapper[4776]: E1125 12:18:23.443499 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"817faa7aee46a29b0145616eedde10d723403c57e194bbf1709f68a8b85aed9a\": container with ID starting with 817faa7aee46a29b0145616eedde10d723403c57e194bbf1709f68a8b85aed9a not found: ID does not exist" containerID="817faa7aee46a29b0145616eedde10d723403c57e194bbf1709f68a8b85aed9a" Nov 25 12:18:23 crc kubenswrapper[4776]: I1125 12:18:23.443520 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"817faa7aee46a29b0145616eedde10d723403c57e194bbf1709f68a8b85aed9a"} err="failed to get container status \"817faa7aee46a29b0145616eedde10d723403c57e194bbf1709f68a8b85aed9a\": rpc error: code = NotFound desc = could not find container \"817faa7aee46a29b0145616eedde10d723403c57e194bbf1709f68a8b85aed9a\": container with ID starting with 817faa7aee46a29b0145616eedde10d723403c57e194bbf1709f68a8b85aed9a not found: ID does not exist" Nov 25 12:18:23 crc kubenswrapper[4776]: I1125 12:18:23.662984 4776 scope.go:117] "RemoveContainer" containerID="d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563" Nov 25 12:18:23 crc kubenswrapper[4776]: E1125 12:18:23.663410 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:18:23 crc kubenswrapper[4776]: I1125 12:18:23.675124 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f4f4be9-19a4-4872-961e-aa545c563aad" path="/var/lib/kubelet/pods/9f4f4be9-19a4-4872-961e-aa545c563aad/volumes" Nov 25 12:18:34 crc kubenswrapper[4776]: I1125 12:18:34.662959 4776 scope.go:117] "RemoveContainer" containerID="d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563" Nov 25 12:18:34 crc kubenswrapper[4776]: E1125 12:18:34.664291 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:18:48 crc kubenswrapper[4776]: I1125 12:18:48.662424 4776 scope.go:117] "RemoveContainer" containerID="d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563" Nov 25 12:18:48 crc kubenswrapper[4776]: E1125 12:18:48.663631 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:19:03 crc kubenswrapper[4776]: I1125 12:19:03.662953 4776 scope.go:117] "RemoveContainer" containerID="d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563" Nov 25 12:19:03 crc kubenswrapper[4776]: E1125 12:19:03.665036 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:19:10 crc kubenswrapper[4776]: I1125 12:19:10.903990 4776 generic.go:334] "Generic (PLEG): container finished" podID="7256708a-6fd0-414d-902c-105d6efc08e8" containerID="5a1cce6bea9b448d7b7c9493aa43fb0c00de36186e83fd5eb96b09b4ba3a735d" exitCode=0 Nov 25 12:19:10 crc kubenswrapper[4776]: I1125 12:19:10.904747 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7b7jq/must-gather-wmmkf" event={"ID":"7256708a-6fd0-414d-902c-105d6efc08e8","Type":"ContainerDied","Data":"5a1cce6bea9b448d7b7c9493aa43fb0c00de36186e83fd5eb96b09b4ba3a735d"} Nov 25 12:19:10 crc kubenswrapper[4776]: I1125 12:19:10.905927 4776 scope.go:117] "RemoveContainer" containerID="5a1cce6bea9b448d7b7c9493aa43fb0c00de36186e83fd5eb96b09b4ba3a735d" Nov 25 12:19:11 crc kubenswrapper[4776]: I1125 12:19:11.037377 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7b7jq_must-gather-wmmkf_7256708a-6fd0-414d-902c-105d6efc08e8/gather/0.log" Nov 25 12:19:18 crc kubenswrapper[4776]: I1125 12:19:18.663701 4776 scope.go:117] "RemoveContainer" containerID="d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563" Nov 25 12:19:18 crc kubenswrapper[4776]: E1125 12:19:18.664858 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:19:19 crc kubenswrapper[4776]: I1125 12:19:19.129329 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-7b7jq/must-gather-wmmkf"] Nov 25 12:19:19 crc kubenswrapper[4776]: I1125 12:19:19.129691 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-7b7jq/must-gather-wmmkf" podUID="7256708a-6fd0-414d-902c-105d6efc08e8" containerName="copy" containerID="cri-o://a6559546a505d58e7870ea9ef3612997d38c54df3b2a8d5b62720da5a4dd70a6" gracePeriod=2 Nov 25 12:19:19 crc kubenswrapper[4776]: I1125 12:19:19.146615 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-7b7jq/must-gather-wmmkf"] Nov 25 12:19:19 crc kubenswrapper[4776]: I1125 12:19:19.758305 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7b7jq_must-gather-wmmkf_7256708a-6fd0-414d-902c-105d6efc08e8/copy/0.log" Nov 25 12:19:19 crc kubenswrapper[4776]: I1125 12:19:19.759423 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b7jq/must-gather-wmmkf" Nov 25 12:19:19 crc kubenswrapper[4776]: I1125 12:19:19.853837 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7256708a-6fd0-414d-902c-105d6efc08e8-must-gather-output\") pod \"7256708a-6fd0-414d-902c-105d6efc08e8\" (UID: \"7256708a-6fd0-414d-902c-105d6efc08e8\") " Nov 25 12:19:19 crc kubenswrapper[4776]: I1125 12:19:19.853943 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tl455\" (UniqueName: \"kubernetes.io/projected/7256708a-6fd0-414d-902c-105d6efc08e8-kube-api-access-tl455\") pod \"7256708a-6fd0-414d-902c-105d6efc08e8\" (UID: \"7256708a-6fd0-414d-902c-105d6efc08e8\") " Nov 25 12:19:19 crc kubenswrapper[4776]: I1125 12:19:19.866438 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7256708a-6fd0-414d-902c-105d6efc08e8-kube-api-access-tl455" (OuterVolumeSpecName: "kube-api-access-tl455") pod "7256708a-6fd0-414d-902c-105d6efc08e8" (UID: "7256708a-6fd0-414d-902c-105d6efc08e8"). InnerVolumeSpecName "kube-api-access-tl455". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:19:19 crc kubenswrapper[4776]: I1125 12:19:19.956571 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tl455\" (UniqueName: \"kubernetes.io/projected/7256708a-6fd0-414d-902c-105d6efc08e8-kube-api-access-tl455\") on node \"crc\" DevicePath \"\"" Nov 25 12:19:20 crc kubenswrapper[4776]: I1125 12:19:20.056368 4776 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-7b7jq_must-gather-wmmkf_7256708a-6fd0-414d-902c-105d6efc08e8/copy/0.log" Nov 25 12:19:20 crc kubenswrapper[4776]: I1125 12:19:20.058227 4776 generic.go:334] "Generic (PLEG): container finished" podID="7256708a-6fd0-414d-902c-105d6efc08e8" containerID="a6559546a505d58e7870ea9ef3612997d38c54df3b2a8d5b62720da5a4dd70a6" exitCode=143 Nov 25 12:19:20 crc kubenswrapper[4776]: I1125 12:19:20.058308 4776 scope.go:117] "RemoveContainer" containerID="a6559546a505d58e7870ea9ef3612997d38c54df3b2a8d5b62720da5a4dd70a6" Nov 25 12:19:20 crc kubenswrapper[4776]: I1125 12:19:20.058501 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7b7jq/must-gather-wmmkf" Nov 25 12:19:20 crc kubenswrapper[4776]: I1125 12:19:20.084894 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7256708a-6fd0-414d-902c-105d6efc08e8-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "7256708a-6fd0-414d-902c-105d6efc08e8" (UID: "7256708a-6fd0-414d-902c-105d6efc08e8"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:19:20 crc kubenswrapper[4776]: I1125 12:19:20.105086 4776 scope.go:117] "RemoveContainer" containerID="5a1cce6bea9b448d7b7c9493aa43fb0c00de36186e83fd5eb96b09b4ba3a735d" Nov 25 12:19:20 crc kubenswrapper[4776]: I1125 12:19:20.159384 4776 scope.go:117] "RemoveContainer" containerID="a6559546a505d58e7870ea9ef3612997d38c54df3b2a8d5b62720da5a4dd70a6" Nov 25 12:19:20 crc kubenswrapper[4776]: E1125 12:19:20.160013 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6559546a505d58e7870ea9ef3612997d38c54df3b2a8d5b62720da5a4dd70a6\": container with ID starting with a6559546a505d58e7870ea9ef3612997d38c54df3b2a8d5b62720da5a4dd70a6 not found: ID does not exist" containerID="a6559546a505d58e7870ea9ef3612997d38c54df3b2a8d5b62720da5a4dd70a6" Nov 25 12:19:20 crc kubenswrapper[4776]: I1125 12:19:20.160047 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6559546a505d58e7870ea9ef3612997d38c54df3b2a8d5b62720da5a4dd70a6"} err="failed to get container status \"a6559546a505d58e7870ea9ef3612997d38c54df3b2a8d5b62720da5a4dd70a6\": rpc error: code = NotFound desc = could not find container \"a6559546a505d58e7870ea9ef3612997d38c54df3b2a8d5b62720da5a4dd70a6\": container with ID starting with a6559546a505d58e7870ea9ef3612997d38c54df3b2a8d5b62720da5a4dd70a6 not found: ID does not exist" Nov 25 12:19:20 crc kubenswrapper[4776]: I1125 12:19:20.160551 4776 scope.go:117] "RemoveContainer" containerID="5a1cce6bea9b448d7b7c9493aa43fb0c00de36186e83fd5eb96b09b4ba3a735d" Nov 25 12:19:20 crc kubenswrapper[4776]: E1125 12:19:20.161487 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a1cce6bea9b448d7b7c9493aa43fb0c00de36186e83fd5eb96b09b4ba3a735d\": container with ID starting with 5a1cce6bea9b448d7b7c9493aa43fb0c00de36186e83fd5eb96b09b4ba3a735d not found: ID does not exist" containerID="5a1cce6bea9b448d7b7c9493aa43fb0c00de36186e83fd5eb96b09b4ba3a735d" Nov 25 12:19:20 crc kubenswrapper[4776]: I1125 12:19:20.161517 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a1cce6bea9b448d7b7c9493aa43fb0c00de36186e83fd5eb96b09b4ba3a735d"} err="failed to get container status \"5a1cce6bea9b448d7b7c9493aa43fb0c00de36186e83fd5eb96b09b4ba3a735d\": rpc error: code = NotFound desc = could not find container \"5a1cce6bea9b448d7b7c9493aa43fb0c00de36186e83fd5eb96b09b4ba3a735d\": container with ID starting with 5a1cce6bea9b448d7b7c9493aa43fb0c00de36186e83fd5eb96b09b4ba3a735d not found: ID does not exist" Nov 25 12:19:20 crc kubenswrapper[4776]: I1125 12:19:20.165342 4776 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7256708a-6fd0-414d-902c-105d6efc08e8-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 25 12:19:21 crc kubenswrapper[4776]: I1125 12:19:21.675116 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7256708a-6fd0-414d-902c-105d6efc08e8" path="/var/lib/kubelet/pods/7256708a-6fd0-414d-902c-105d6efc08e8/volumes" Nov 25 12:19:32 crc kubenswrapper[4776]: I1125 12:19:32.662115 4776 scope.go:117] "RemoveContainer" containerID="d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563" Nov 25 12:19:32 crc kubenswrapper[4776]: E1125 12:19:32.663022 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:19:43 crc kubenswrapper[4776]: I1125 12:19:43.662012 4776 scope.go:117] "RemoveContainer" containerID="d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563" Nov 25 12:19:43 crc kubenswrapper[4776]: E1125 12:19:43.663042 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:19:58 crc kubenswrapper[4776]: I1125 12:19:58.662365 4776 scope.go:117] "RemoveContainer" containerID="d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563" Nov 25 12:19:58 crc kubenswrapper[4776]: E1125 12:19:58.663437 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:20:09 crc kubenswrapper[4776]: I1125 12:20:09.662381 4776 scope.go:117] "RemoveContainer" containerID="d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563" Nov 25 12:20:09 crc kubenswrapper[4776]: E1125 12:20:09.665026 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:20:21 crc kubenswrapper[4776]: I1125 12:20:21.662591 4776 scope.go:117] "RemoveContainer" containerID="d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563" Nov 25 12:20:21 crc kubenswrapper[4776]: E1125 12:20:21.663556 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:20:32 crc kubenswrapper[4776]: I1125 12:20:32.662769 4776 scope.go:117] "RemoveContainer" containerID="d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563" Nov 25 12:20:32 crc kubenswrapper[4776]: E1125 12:20:32.664174 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:20:46 crc kubenswrapper[4776]: I1125 12:20:46.663047 4776 scope.go:117] "RemoveContainer" containerID="d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563" Nov 25 12:20:46 crc kubenswrapper[4776]: E1125 12:20:46.663910 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:20:51 crc kubenswrapper[4776]: I1125 12:20:51.390311 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9nzgr"] Nov 25 12:20:51 crc kubenswrapper[4776]: E1125 12:20:51.393325 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7256708a-6fd0-414d-902c-105d6efc08e8" containerName="copy" Nov 25 12:20:51 crc kubenswrapper[4776]: I1125 12:20:51.393532 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="7256708a-6fd0-414d-902c-105d6efc08e8" containerName="copy" Nov 25 12:20:51 crc kubenswrapper[4776]: E1125 12:20:51.393650 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f4f4be9-19a4-4872-961e-aa545c563aad" containerName="registry-server" Nov 25 12:20:51 crc kubenswrapper[4776]: I1125 12:20:51.393735 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f4f4be9-19a4-4872-961e-aa545c563aad" containerName="registry-server" Nov 25 12:20:51 crc kubenswrapper[4776]: E1125 12:20:51.393824 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f4f4be9-19a4-4872-961e-aa545c563aad" containerName="extract-content" Nov 25 12:20:51 crc kubenswrapper[4776]: I1125 12:20:51.393909 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f4f4be9-19a4-4872-961e-aa545c563aad" containerName="extract-content" Nov 25 12:20:51 crc kubenswrapper[4776]: E1125 12:20:51.394008 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7256708a-6fd0-414d-902c-105d6efc08e8" containerName="gather" Nov 25 12:20:51 crc kubenswrapper[4776]: I1125 12:20:51.394103 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="7256708a-6fd0-414d-902c-105d6efc08e8" containerName="gather" Nov 25 12:20:51 crc kubenswrapper[4776]: E1125 12:20:51.394201 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f4f4be9-19a4-4872-961e-aa545c563aad" containerName="extract-utilities" Nov 25 12:20:51 crc kubenswrapper[4776]: I1125 12:20:51.394286 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f4f4be9-19a4-4872-961e-aa545c563aad" containerName="extract-utilities" Nov 25 12:20:51 crc kubenswrapper[4776]: I1125 12:20:51.394628 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="7256708a-6fd0-414d-902c-105d6efc08e8" containerName="gather" Nov 25 12:20:51 crc kubenswrapper[4776]: I1125 12:20:51.394744 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f4f4be9-19a4-4872-961e-aa545c563aad" containerName="registry-server" Nov 25 12:20:51 crc kubenswrapper[4776]: I1125 12:20:51.394832 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="7256708a-6fd0-414d-902c-105d6efc08e8" containerName="copy" Nov 25 12:20:51 crc kubenswrapper[4776]: I1125 12:20:51.397107 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9nzgr" Nov 25 12:20:51 crc kubenswrapper[4776]: I1125 12:20:51.415520 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9nzgr"] Nov 25 12:20:51 crc kubenswrapper[4776]: I1125 12:20:51.481664 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff0146ba-9ffb-4d4a-ab66-541aea2c1065-catalog-content\") pod \"redhat-operators-9nzgr\" (UID: \"ff0146ba-9ffb-4d4a-ab66-541aea2c1065\") " pod="openshift-marketplace/redhat-operators-9nzgr" Nov 25 12:20:51 crc kubenswrapper[4776]: I1125 12:20:51.481831 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qvz4\" (UniqueName: \"kubernetes.io/projected/ff0146ba-9ffb-4d4a-ab66-541aea2c1065-kube-api-access-9qvz4\") pod \"redhat-operators-9nzgr\" (UID: \"ff0146ba-9ffb-4d4a-ab66-541aea2c1065\") " pod="openshift-marketplace/redhat-operators-9nzgr" Nov 25 12:20:51 crc kubenswrapper[4776]: I1125 12:20:51.481920 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff0146ba-9ffb-4d4a-ab66-541aea2c1065-utilities\") pod \"redhat-operators-9nzgr\" (UID: \"ff0146ba-9ffb-4d4a-ab66-541aea2c1065\") " pod="openshift-marketplace/redhat-operators-9nzgr" Nov 25 12:20:51 crc kubenswrapper[4776]: I1125 12:20:51.584220 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qvz4\" (UniqueName: \"kubernetes.io/projected/ff0146ba-9ffb-4d4a-ab66-541aea2c1065-kube-api-access-9qvz4\") pod \"redhat-operators-9nzgr\" (UID: \"ff0146ba-9ffb-4d4a-ab66-541aea2c1065\") " pod="openshift-marketplace/redhat-operators-9nzgr" Nov 25 12:20:51 crc kubenswrapper[4776]: I1125 12:20:51.584348 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff0146ba-9ffb-4d4a-ab66-541aea2c1065-utilities\") pod \"redhat-operators-9nzgr\" (UID: \"ff0146ba-9ffb-4d4a-ab66-541aea2c1065\") " pod="openshift-marketplace/redhat-operators-9nzgr" Nov 25 12:20:51 crc kubenswrapper[4776]: I1125 12:20:51.584427 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff0146ba-9ffb-4d4a-ab66-541aea2c1065-catalog-content\") pod \"redhat-operators-9nzgr\" (UID: \"ff0146ba-9ffb-4d4a-ab66-541aea2c1065\") " pod="openshift-marketplace/redhat-operators-9nzgr" Nov 25 12:20:51 crc kubenswrapper[4776]: I1125 12:20:51.585138 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff0146ba-9ffb-4d4a-ab66-541aea2c1065-utilities\") pod \"redhat-operators-9nzgr\" (UID: \"ff0146ba-9ffb-4d4a-ab66-541aea2c1065\") " pod="openshift-marketplace/redhat-operators-9nzgr" Nov 25 12:20:51 crc kubenswrapper[4776]: I1125 12:20:51.585156 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff0146ba-9ffb-4d4a-ab66-541aea2c1065-catalog-content\") pod \"redhat-operators-9nzgr\" (UID: \"ff0146ba-9ffb-4d4a-ab66-541aea2c1065\") " pod="openshift-marketplace/redhat-operators-9nzgr" Nov 25 12:20:51 crc kubenswrapper[4776]: I1125 12:20:51.613553 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qvz4\" (UniqueName: \"kubernetes.io/projected/ff0146ba-9ffb-4d4a-ab66-541aea2c1065-kube-api-access-9qvz4\") pod \"redhat-operators-9nzgr\" (UID: \"ff0146ba-9ffb-4d4a-ab66-541aea2c1065\") " pod="openshift-marketplace/redhat-operators-9nzgr" Nov 25 12:20:51 crc kubenswrapper[4776]: I1125 12:20:51.725466 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9nzgr" Nov 25 12:20:52 crc kubenswrapper[4776]: I1125 12:20:52.279635 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9nzgr"] Nov 25 12:20:53 crc kubenswrapper[4776]: I1125 12:20:53.096728 4776 generic.go:334] "Generic (PLEG): container finished" podID="ff0146ba-9ffb-4d4a-ab66-541aea2c1065" containerID="3e431580800e54bffccb239b4ab4eaa33e74767460bb2d462577637ed8d14efe" exitCode=0 Nov 25 12:20:53 crc kubenswrapper[4776]: I1125 12:20:53.096820 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9nzgr" event={"ID":"ff0146ba-9ffb-4d4a-ab66-541aea2c1065","Type":"ContainerDied","Data":"3e431580800e54bffccb239b4ab4eaa33e74767460bb2d462577637ed8d14efe"} Nov 25 12:20:53 crc kubenswrapper[4776]: I1125 12:20:53.097274 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9nzgr" event={"ID":"ff0146ba-9ffb-4d4a-ab66-541aea2c1065","Type":"ContainerStarted","Data":"203e4ef8bd14d3bcaf9fadb4d8ff0795081f89ea3544a630d845cf5a26e30cd9"} Nov 25 12:20:54 crc kubenswrapper[4776]: I1125 12:20:54.109392 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9nzgr" event={"ID":"ff0146ba-9ffb-4d4a-ab66-541aea2c1065","Type":"ContainerStarted","Data":"33e9c5be8ad8dcf6b3b06c40de7b82fd33d64a90b1290946b41dddbf3e863bea"} Nov 25 12:20:57 crc kubenswrapper[4776]: I1125 12:20:57.145300 4776 generic.go:334] "Generic (PLEG): container finished" podID="ff0146ba-9ffb-4d4a-ab66-541aea2c1065" containerID="33e9c5be8ad8dcf6b3b06c40de7b82fd33d64a90b1290946b41dddbf3e863bea" exitCode=0 Nov 25 12:20:57 crc kubenswrapper[4776]: I1125 12:20:57.145406 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9nzgr" event={"ID":"ff0146ba-9ffb-4d4a-ab66-541aea2c1065","Type":"ContainerDied","Data":"33e9c5be8ad8dcf6b3b06c40de7b82fd33d64a90b1290946b41dddbf3e863bea"} Nov 25 12:20:58 crc kubenswrapper[4776]: I1125 12:20:58.161380 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9nzgr" event={"ID":"ff0146ba-9ffb-4d4a-ab66-541aea2c1065","Type":"ContainerStarted","Data":"83c9dd26503b6772dddcf6d55212f6f9e5bacd92483555fa481435a17d6b63ee"} Nov 25 12:20:58 crc kubenswrapper[4776]: I1125 12:20:58.194812 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9nzgr" podStartSLOduration=2.652432118 podStartE2EDuration="7.19478066s" podCreationTimestamp="2025-11-25 12:20:51 +0000 UTC" firstStartedPulling="2025-11-25 12:20:53.101190375 +0000 UTC m=+10598.142249928" lastFinishedPulling="2025-11-25 12:20:57.643538917 +0000 UTC m=+10602.684598470" observedRunningTime="2025-11-25 12:20:58.181377113 +0000 UTC m=+10603.222436666" watchObservedRunningTime="2025-11-25 12:20:58.19478066 +0000 UTC m=+10603.235840213" Nov 25 12:21:00 crc kubenswrapper[4776]: I1125 12:21:00.662413 4776 scope.go:117] "RemoveContainer" containerID="d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563" Nov 25 12:21:00 crc kubenswrapper[4776]: E1125 12:21:00.663891 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:21:01 crc kubenswrapper[4776]: I1125 12:21:01.725628 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9nzgr" Nov 25 12:21:01 crc kubenswrapper[4776]: I1125 12:21:01.725715 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9nzgr" Nov 25 12:21:02 crc kubenswrapper[4776]: I1125 12:21:02.779100 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9nzgr" podUID="ff0146ba-9ffb-4d4a-ab66-541aea2c1065" containerName="registry-server" probeResult="failure" output=< Nov 25 12:21:02 crc kubenswrapper[4776]: timeout: failed to connect service ":50051" within 1s Nov 25 12:21:02 crc kubenswrapper[4776]: > Nov 25 12:21:11 crc kubenswrapper[4776]: I1125 12:21:11.773561 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9nzgr" Nov 25 12:21:11 crc kubenswrapper[4776]: I1125 12:21:11.829742 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9nzgr" Nov 25 12:21:12 crc kubenswrapper[4776]: I1125 12:21:12.018220 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9nzgr"] Nov 25 12:21:13 crc kubenswrapper[4776]: I1125 12:21:13.303618 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9nzgr" podUID="ff0146ba-9ffb-4d4a-ab66-541aea2c1065" containerName="registry-server" containerID="cri-o://83c9dd26503b6772dddcf6d55212f6f9e5bacd92483555fa481435a17d6b63ee" gracePeriod=2 Nov 25 12:21:13 crc kubenswrapper[4776]: I1125 12:21:13.854268 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9nzgr" Nov 25 12:21:14 crc kubenswrapper[4776]: I1125 12:21:14.039968 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff0146ba-9ffb-4d4a-ab66-541aea2c1065-utilities\") pod \"ff0146ba-9ffb-4d4a-ab66-541aea2c1065\" (UID: \"ff0146ba-9ffb-4d4a-ab66-541aea2c1065\") " Nov 25 12:21:14 crc kubenswrapper[4776]: I1125 12:21:14.040927 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff0146ba-9ffb-4d4a-ab66-541aea2c1065-utilities" (OuterVolumeSpecName: "utilities") pod "ff0146ba-9ffb-4d4a-ab66-541aea2c1065" (UID: "ff0146ba-9ffb-4d4a-ab66-541aea2c1065"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:21:14 crc kubenswrapper[4776]: I1125 12:21:14.042524 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qvz4\" (UniqueName: \"kubernetes.io/projected/ff0146ba-9ffb-4d4a-ab66-541aea2c1065-kube-api-access-9qvz4\") pod \"ff0146ba-9ffb-4d4a-ab66-541aea2c1065\" (UID: \"ff0146ba-9ffb-4d4a-ab66-541aea2c1065\") " Nov 25 12:21:14 crc kubenswrapper[4776]: I1125 12:21:14.042665 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff0146ba-9ffb-4d4a-ab66-541aea2c1065-catalog-content\") pod \"ff0146ba-9ffb-4d4a-ab66-541aea2c1065\" (UID: \"ff0146ba-9ffb-4d4a-ab66-541aea2c1065\") " Nov 25 12:21:14 crc kubenswrapper[4776]: I1125 12:21:14.053490 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff0146ba-9ffb-4d4a-ab66-541aea2c1065-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:21:14 crc kubenswrapper[4776]: I1125 12:21:14.053730 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff0146ba-9ffb-4d4a-ab66-541aea2c1065-kube-api-access-9qvz4" (OuterVolumeSpecName: "kube-api-access-9qvz4") pod "ff0146ba-9ffb-4d4a-ab66-541aea2c1065" (UID: "ff0146ba-9ffb-4d4a-ab66-541aea2c1065"). InnerVolumeSpecName "kube-api-access-9qvz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:21:14 crc kubenswrapper[4776]: I1125 12:21:14.136456 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff0146ba-9ffb-4d4a-ab66-541aea2c1065-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff0146ba-9ffb-4d4a-ab66-541aea2c1065" (UID: "ff0146ba-9ffb-4d4a-ab66-541aea2c1065"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:21:14 crc kubenswrapper[4776]: I1125 12:21:14.155893 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qvz4\" (UniqueName: \"kubernetes.io/projected/ff0146ba-9ffb-4d4a-ab66-541aea2c1065-kube-api-access-9qvz4\") on node \"crc\" DevicePath \"\"" Nov 25 12:21:14 crc kubenswrapper[4776]: I1125 12:21:14.155973 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff0146ba-9ffb-4d4a-ab66-541aea2c1065-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:21:14 crc kubenswrapper[4776]: I1125 12:21:14.316368 4776 generic.go:334] "Generic (PLEG): container finished" podID="ff0146ba-9ffb-4d4a-ab66-541aea2c1065" containerID="83c9dd26503b6772dddcf6d55212f6f9e5bacd92483555fa481435a17d6b63ee" exitCode=0 Nov 25 12:21:14 crc kubenswrapper[4776]: I1125 12:21:14.316418 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9nzgr" event={"ID":"ff0146ba-9ffb-4d4a-ab66-541aea2c1065","Type":"ContainerDied","Data":"83c9dd26503b6772dddcf6d55212f6f9e5bacd92483555fa481435a17d6b63ee"} Nov 25 12:21:14 crc kubenswrapper[4776]: I1125 12:21:14.316448 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9nzgr" event={"ID":"ff0146ba-9ffb-4d4a-ab66-541aea2c1065","Type":"ContainerDied","Data":"203e4ef8bd14d3bcaf9fadb4d8ff0795081f89ea3544a630d845cf5a26e30cd9"} Nov 25 12:21:14 crc kubenswrapper[4776]: I1125 12:21:14.316469 4776 scope.go:117] "RemoveContainer" containerID="83c9dd26503b6772dddcf6d55212f6f9e5bacd92483555fa481435a17d6b63ee" Nov 25 12:21:14 crc kubenswrapper[4776]: I1125 12:21:14.316623 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9nzgr" Nov 25 12:21:14 crc kubenswrapper[4776]: I1125 12:21:14.360123 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9nzgr"] Nov 25 12:21:14 crc kubenswrapper[4776]: I1125 12:21:14.370728 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9nzgr"] Nov 25 12:21:14 crc kubenswrapper[4776]: I1125 12:21:14.374905 4776 scope.go:117] "RemoveContainer" containerID="33e9c5be8ad8dcf6b3b06c40de7b82fd33d64a90b1290946b41dddbf3e863bea" Nov 25 12:21:14 crc kubenswrapper[4776]: I1125 12:21:14.395356 4776 scope.go:117] "RemoveContainer" containerID="3e431580800e54bffccb239b4ab4eaa33e74767460bb2d462577637ed8d14efe" Nov 25 12:21:14 crc kubenswrapper[4776]: I1125 12:21:14.450833 4776 scope.go:117] "RemoveContainer" containerID="83c9dd26503b6772dddcf6d55212f6f9e5bacd92483555fa481435a17d6b63ee" Nov 25 12:21:14 crc kubenswrapper[4776]: E1125 12:21:14.451475 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83c9dd26503b6772dddcf6d55212f6f9e5bacd92483555fa481435a17d6b63ee\": container with ID starting with 83c9dd26503b6772dddcf6d55212f6f9e5bacd92483555fa481435a17d6b63ee not found: ID does not exist" containerID="83c9dd26503b6772dddcf6d55212f6f9e5bacd92483555fa481435a17d6b63ee" Nov 25 12:21:14 crc kubenswrapper[4776]: I1125 12:21:14.451541 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83c9dd26503b6772dddcf6d55212f6f9e5bacd92483555fa481435a17d6b63ee"} err="failed to get container status \"83c9dd26503b6772dddcf6d55212f6f9e5bacd92483555fa481435a17d6b63ee\": rpc error: code = NotFound desc = could not find container \"83c9dd26503b6772dddcf6d55212f6f9e5bacd92483555fa481435a17d6b63ee\": container with ID starting with 83c9dd26503b6772dddcf6d55212f6f9e5bacd92483555fa481435a17d6b63ee not found: ID does not exist" Nov 25 12:21:14 crc kubenswrapper[4776]: I1125 12:21:14.451577 4776 scope.go:117] "RemoveContainer" containerID="33e9c5be8ad8dcf6b3b06c40de7b82fd33d64a90b1290946b41dddbf3e863bea" Nov 25 12:21:14 crc kubenswrapper[4776]: E1125 12:21:14.452007 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33e9c5be8ad8dcf6b3b06c40de7b82fd33d64a90b1290946b41dddbf3e863bea\": container with ID starting with 33e9c5be8ad8dcf6b3b06c40de7b82fd33d64a90b1290946b41dddbf3e863bea not found: ID does not exist" containerID="33e9c5be8ad8dcf6b3b06c40de7b82fd33d64a90b1290946b41dddbf3e863bea" Nov 25 12:21:14 crc kubenswrapper[4776]: I1125 12:21:14.452033 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33e9c5be8ad8dcf6b3b06c40de7b82fd33d64a90b1290946b41dddbf3e863bea"} err="failed to get container status \"33e9c5be8ad8dcf6b3b06c40de7b82fd33d64a90b1290946b41dddbf3e863bea\": rpc error: code = NotFound desc = could not find container \"33e9c5be8ad8dcf6b3b06c40de7b82fd33d64a90b1290946b41dddbf3e863bea\": container with ID starting with 33e9c5be8ad8dcf6b3b06c40de7b82fd33d64a90b1290946b41dddbf3e863bea not found: ID does not exist" Nov 25 12:21:14 crc kubenswrapper[4776]: I1125 12:21:14.452050 4776 scope.go:117] "RemoveContainer" containerID="3e431580800e54bffccb239b4ab4eaa33e74767460bb2d462577637ed8d14efe" Nov 25 12:21:14 crc kubenswrapper[4776]: E1125 12:21:14.452285 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e431580800e54bffccb239b4ab4eaa33e74767460bb2d462577637ed8d14efe\": container with ID starting with 3e431580800e54bffccb239b4ab4eaa33e74767460bb2d462577637ed8d14efe not found: ID does not exist" containerID="3e431580800e54bffccb239b4ab4eaa33e74767460bb2d462577637ed8d14efe" Nov 25 12:21:14 crc kubenswrapper[4776]: I1125 12:21:14.452316 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e431580800e54bffccb239b4ab4eaa33e74767460bb2d462577637ed8d14efe"} err="failed to get container status \"3e431580800e54bffccb239b4ab4eaa33e74767460bb2d462577637ed8d14efe\": rpc error: code = NotFound desc = could not find container \"3e431580800e54bffccb239b4ab4eaa33e74767460bb2d462577637ed8d14efe\": container with ID starting with 3e431580800e54bffccb239b4ab4eaa33e74767460bb2d462577637ed8d14efe not found: ID does not exist" Nov 25 12:21:14 crc kubenswrapper[4776]: I1125 12:21:14.663990 4776 scope.go:117] "RemoveContainer" containerID="d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563" Nov 25 12:21:14 crc kubenswrapper[4776]: E1125 12:21:14.664312 4776 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-84dqb_openshift-machine-config-operator(fa47ebcc-a95e-4693-876d-7284c28c3ade)\"" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" podUID="fa47ebcc-a95e-4693-876d-7284c28c3ade" Nov 25 12:21:15 crc kubenswrapper[4776]: I1125 12:21:15.677838 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff0146ba-9ffb-4d4a-ab66-541aea2c1065" path="/var/lib/kubelet/pods/ff0146ba-9ffb-4d4a-ab66-541aea2c1065/volumes" Nov 25 12:21:17 crc kubenswrapper[4776]: I1125 12:21:17.423389 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gr69n"] Nov 25 12:21:17 crc kubenswrapper[4776]: E1125 12:21:17.424249 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff0146ba-9ffb-4d4a-ab66-541aea2c1065" containerName="extract-utilities" Nov 25 12:21:17 crc kubenswrapper[4776]: I1125 12:21:17.424265 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff0146ba-9ffb-4d4a-ab66-541aea2c1065" containerName="extract-utilities" Nov 25 12:21:17 crc kubenswrapper[4776]: E1125 12:21:17.424289 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff0146ba-9ffb-4d4a-ab66-541aea2c1065" containerName="extract-content" Nov 25 12:21:17 crc kubenswrapper[4776]: I1125 12:21:17.424297 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff0146ba-9ffb-4d4a-ab66-541aea2c1065" containerName="extract-content" Nov 25 12:21:17 crc kubenswrapper[4776]: E1125 12:21:17.424310 4776 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff0146ba-9ffb-4d4a-ab66-541aea2c1065" containerName="registry-server" Nov 25 12:21:17 crc kubenswrapper[4776]: I1125 12:21:17.424318 4776 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff0146ba-9ffb-4d4a-ab66-541aea2c1065" containerName="registry-server" Nov 25 12:21:17 crc kubenswrapper[4776]: I1125 12:21:17.424609 4776 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff0146ba-9ffb-4d4a-ab66-541aea2c1065" containerName="registry-server" Nov 25 12:21:17 crc kubenswrapper[4776]: I1125 12:21:17.426510 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gr69n" Nov 25 12:21:17 crc kubenswrapper[4776]: I1125 12:21:17.437477 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14a73053-d4c1-4633-bbf0-bed6e1ef64da-catalog-content\") pod \"community-operators-gr69n\" (UID: \"14a73053-d4c1-4633-bbf0-bed6e1ef64da\") " pod="openshift-marketplace/community-operators-gr69n" Nov 25 12:21:17 crc kubenswrapper[4776]: I1125 12:21:17.438022 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14a73053-d4c1-4633-bbf0-bed6e1ef64da-utilities\") pod \"community-operators-gr69n\" (UID: \"14a73053-d4c1-4633-bbf0-bed6e1ef64da\") " pod="openshift-marketplace/community-operators-gr69n" Nov 25 12:21:17 crc kubenswrapper[4776]: I1125 12:21:17.438281 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hqr2\" (UniqueName: \"kubernetes.io/projected/14a73053-d4c1-4633-bbf0-bed6e1ef64da-kube-api-access-2hqr2\") pod \"community-operators-gr69n\" (UID: \"14a73053-d4c1-4633-bbf0-bed6e1ef64da\") " pod="openshift-marketplace/community-operators-gr69n" Nov 25 12:21:17 crc kubenswrapper[4776]: I1125 12:21:17.442617 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gr69n"] Nov 25 12:21:17 crc kubenswrapper[4776]: I1125 12:21:17.540332 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hqr2\" (UniqueName: \"kubernetes.io/projected/14a73053-d4c1-4633-bbf0-bed6e1ef64da-kube-api-access-2hqr2\") pod \"community-operators-gr69n\" (UID: \"14a73053-d4c1-4633-bbf0-bed6e1ef64da\") " pod="openshift-marketplace/community-operators-gr69n" Nov 25 12:21:17 crc kubenswrapper[4776]: I1125 12:21:17.540414 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14a73053-d4c1-4633-bbf0-bed6e1ef64da-catalog-content\") pod \"community-operators-gr69n\" (UID: \"14a73053-d4c1-4633-bbf0-bed6e1ef64da\") " pod="openshift-marketplace/community-operators-gr69n" Nov 25 12:21:17 crc kubenswrapper[4776]: I1125 12:21:17.540595 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14a73053-d4c1-4633-bbf0-bed6e1ef64da-utilities\") pod \"community-operators-gr69n\" (UID: \"14a73053-d4c1-4633-bbf0-bed6e1ef64da\") " pod="openshift-marketplace/community-operators-gr69n" Nov 25 12:21:17 crc kubenswrapper[4776]: I1125 12:21:17.541136 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14a73053-d4c1-4633-bbf0-bed6e1ef64da-catalog-content\") pod \"community-operators-gr69n\" (UID: \"14a73053-d4c1-4633-bbf0-bed6e1ef64da\") " pod="openshift-marketplace/community-operators-gr69n" Nov 25 12:21:17 crc kubenswrapper[4776]: I1125 12:21:17.541320 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14a73053-d4c1-4633-bbf0-bed6e1ef64da-utilities\") pod \"community-operators-gr69n\" (UID: \"14a73053-d4c1-4633-bbf0-bed6e1ef64da\") " pod="openshift-marketplace/community-operators-gr69n" Nov 25 12:21:17 crc kubenswrapper[4776]: I1125 12:21:17.560947 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hqr2\" (UniqueName: \"kubernetes.io/projected/14a73053-d4c1-4633-bbf0-bed6e1ef64da-kube-api-access-2hqr2\") pod \"community-operators-gr69n\" (UID: \"14a73053-d4c1-4633-bbf0-bed6e1ef64da\") " pod="openshift-marketplace/community-operators-gr69n" Nov 25 12:21:17 crc kubenswrapper[4776]: I1125 12:21:17.749587 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gr69n" Nov 25 12:21:18 crc kubenswrapper[4776]: I1125 12:21:18.437537 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gr69n"] Nov 25 12:21:19 crc kubenswrapper[4776]: I1125 12:21:19.383033 4776 generic.go:334] "Generic (PLEG): container finished" podID="14a73053-d4c1-4633-bbf0-bed6e1ef64da" containerID="17e70e93c6d1dbb45c455d36fe165d434c26be1243daf949393d272fb2e8ce85" exitCode=0 Nov 25 12:21:19 crc kubenswrapper[4776]: I1125 12:21:19.383174 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gr69n" event={"ID":"14a73053-d4c1-4633-bbf0-bed6e1ef64da","Type":"ContainerDied","Data":"17e70e93c6d1dbb45c455d36fe165d434c26be1243daf949393d272fb2e8ce85"} Nov 25 12:21:19 crc kubenswrapper[4776]: I1125 12:21:19.383610 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gr69n" event={"ID":"14a73053-d4c1-4633-bbf0-bed6e1ef64da","Type":"ContainerStarted","Data":"314f2b648f07718b7b046759c107ed369b7ec2477eb77806a580ce5fa231cdaf"} Nov 25 12:21:19 crc kubenswrapper[4776]: I1125 12:21:19.644693 4776 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2pj5g"] Nov 25 12:21:19 crc kubenswrapper[4776]: I1125 12:21:19.648086 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2pj5g" Nov 25 12:21:19 crc kubenswrapper[4776]: I1125 12:21:19.661556 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2pj5g"] Nov 25 12:21:19 crc kubenswrapper[4776]: I1125 12:21:19.801317 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/136e5670-c1db-4e24-a402-9163e0045408-catalog-content\") pod \"certified-operators-2pj5g\" (UID: \"136e5670-c1db-4e24-a402-9163e0045408\") " pod="openshift-marketplace/certified-operators-2pj5g" Nov 25 12:21:19 crc kubenswrapper[4776]: I1125 12:21:19.801409 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/136e5670-c1db-4e24-a402-9163e0045408-utilities\") pod \"certified-operators-2pj5g\" (UID: \"136e5670-c1db-4e24-a402-9163e0045408\") " pod="openshift-marketplace/certified-operators-2pj5g" Nov 25 12:21:19 crc kubenswrapper[4776]: I1125 12:21:19.801787 4776 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhglg\" (UniqueName: \"kubernetes.io/projected/136e5670-c1db-4e24-a402-9163e0045408-kube-api-access-vhglg\") pod \"certified-operators-2pj5g\" (UID: \"136e5670-c1db-4e24-a402-9163e0045408\") " pod="openshift-marketplace/certified-operators-2pj5g" Nov 25 12:21:19 crc kubenswrapper[4776]: I1125 12:21:19.906034 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/136e5670-c1db-4e24-a402-9163e0045408-catalog-content\") pod \"certified-operators-2pj5g\" (UID: \"136e5670-c1db-4e24-a402-9163e0045408\") " pod="openshift-marketplace/certified-operators-2pj5g" Nov 25 12:21:19 crc kubenswrapper[4776]: I1125 12:21:19.906108 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/136e5670-c1db-4e24-a402-9163e0045408-utilities\") pod \"certified-operators-2pj5g\" (UID: \"136e5670-c1db-4e24-a402-9163e0045408\") " pod="openshift-marketplace/certified-operators-2pj5g" Nov 25 12:21:19 crc kubenswrapper[4776]: I1125 12:21:19.906258 4776 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhglg\" (UniqueName: \"kubernetes.io/projected/136e5670-c1db-4e24-a402-9163e0045408-kube-api-access-vhglg\") pod \"certified-operators-2pj5g\" (UID: \"136e5670-c1db-4e24-a402-9163e0045408\") " pod="openshift-marketplace/certified-operators-2pj5g" Nov 25 12:21:19 crc kubenswrapper[4776]: I1125 12:21:19.907086 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/136e5670-c1db-4e24-a402-9163e0045408-catalog-content\") pod \"certified-operators-2pj5g\" (UID: \"136e5670-c1db-4e24-a402-9163e0045408\") " pod="openshift-marketplace/certified-operators-2pj5g" Nov 25 12:21:19 crc kubenswrapper[4776]: I1125 12:21:19.907332 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/136e5670-c1db-4e24-a402-9163e0045408-utilities\") pod \"certified-operators-2pj5g\" (UID: \"136e5670-c1db-4e24-a402-9163e0045408\") " pod="openshift-marketplace/certified-operators-2pj5g" Nov 25 12:21:19 crc kubenswrapper[4776]: I1125 12:21:19.936096 4776 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhglg\" (UniqueName: \"kubernetes.io/projected/136e5670-c1db-4e24-a402-9163e0045408-kube-api-access-vhglg\") pod \"certified-operators-2pj5g\" (UID: \"136e5670-c1db-4e24-a402-9163e0045408\") " pod="openshift-marketplace/certified-operators-2pj5g" Nov 25 12:21:19 crc kubenswrapper[4776]: I1125 12:21:19.984879 4776 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2pj5g" Nov 25 12:21:20 crc kubenswrapper[4776]: I1125 12:21:20.581939 4776 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2pj5g"] Nov 25 12:21:20 crc kubenswrapper[4776]: W1125 12:21:20.589660 4776 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod136e5670_c1db_4e24_a402_9163e0045408.slice/crio-d9eb7cfc5c4d66d505a16e287b816b9897943b0c6dc676d781cfaaef2d49182c WatchSource:0}: Error finding container d9eb7cfc5c4d66d505a16e287b816b9897943b0c6dc676d781cfaaef2d49182c: Status 404 returned error can't find the container with id d9eb7cfc5c4d66d505a16e287b816b9897943b0c6dc676d781cfaaef2d49182c Nov 25 12:21:21 crc kubenswrapper[4776]: I1125 12:21:21.427892 4776 generic.go:334] "Generic (PLEG): container finished" podID="136e5670-c1db-4e24-a402-9163e0045408" containerID="1ec34a41408d36dfd6e397c7560bc880790b2b7feb9e761704ad103b4cd8ddbd" exitCode=0 Nov 25 12:21:21 crc kubenswrapper[4776]: I1125 12:21:21.428655 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2pj5g" event={"ID":"136e5670-c1db-4e24-a402-9163e0045408","Type":"ContainerDied","Data":"1ec34a41408d36dfd6e397c7560bc880790b2b7feb9e761704ad103b4cd8ddbd"} Nov 25 12:21:21 crc kubenswrapper[4776]: I1125 12:21:21.428700 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2pj5g" event={"ID":"136e5670-c1db-4e24-a402-9163e0045408","Type":"ContainerStarted","Data":"d9eb7cfc5c4d66d505a16e287b816b9897943b0c6dc676d781cfaaef2d49182c"} Nov 25 12:21:21 crc kubenswrapper[4776]: I1125 12:21:21.436671 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gr69n" event={"ID":"14a73053-d4c1-4633-bbf0-bed6e1ef64da","Type":"ContainerStarted","Data":"2eb9060b8ab9c0cf5a5268d26c745cfc3dbbf88a3e919c5e3c182e77cac25cdd"} Nov 25 12:21:23 crc kubenswrapper[4776]: I1125 12:21:23.471153 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2pj5g" event={"ID":"136e5670-c1db-4e24-a402-9163e0045408","Type":"ContainerStarted","Data":"14415b152610d90a74f57b289bfebf4a708759c93e795861b4fd8d6499faf15f"} Nov 25 12:21:24 crc kubenswrapper[4776]: I1125 12:21:24.487500 4776 generic.go:334] "Generic (PLEG): container finished" podID="14a73053-d4c1-4633-bbf0-bed6e1ef64da" containerID="2eb9060b8ab9c0cf5a5268d26c745cfc3dbbf88a3e919c5e3c182e77cac25cdd" exitCode=0 Nov 25 12:21:24 crc kubenswrapper[4776]: I1125 12:21:24.487548 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gr69n" event={"ID":"14a73053-d4c1-4633-bbf0-bed6e1ef64da","Type":"ContainerDied","Data":"2eb9060b8ab9c0cf5a5268d26c745cfc3dbbf88a3e919c5e3c182e77cac25cdd"} Nov 25 12:21:25 crc kubenswrapper[4776]: I1125 12:21:25.505328 4776 generic.go:334] "Generic (PLEG): container finished" podID="136e5670-c1db-4e24-a402-9163e0045408" containerID="14415b152610d90a74f57b289bfebf4a708759c93e795861b4fd8d6499faf15f" exitCode=0 Nov 25 12:21:25 crc kubenswrapper[4776]: I1125 12:21:25.505512 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2pj5g" event={"ID":"136e5670-c1db-4e24-a402-9163e0045408","Type":"ContainerDied","Data":"14415b152610d90a74f57b289bfebf4a708759c93e795861b4fd8d6499faf15f"} Nov 25 12:21:26 crc kubenswrapper[4776]: I1125 12:21:26.518905 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gr69n" event={"ID":"14a73053-d4c1-4633-bbf0-bed6e1ef64da","Type":"ContainerStarted","Data":"ba7a7601816a233cdf00d3aaa9c2f92d02115792009e59c52ccd88962fc33928"} Nov 25 12:21:26 crc kubenswrapper[4776]: I1125 12:21:26.558861 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gr69n" podStartSLOduration=3.168172282 podStartE2EDuration="9.558829955s" podCreationTimestamp="2025-11-25 12:21:17 +0000 UTC" firstStartedPulling="2025-11-25 12:21:19.389204087 +0000 UTC m=+10624.430263630" lastFinishedPulling="2025-11-25 12:21:25.77986175 +0000 UTC m=+10630.820921303" observedRunningTime="2025-11-25 12:21:26.540428183 +0000 UTC m=+10631.581487726" watchObservedRunningTime="2025-11-25 12:21:26.558829955 +0000 UTC m=+10631.599889508" Nov 25 12:21:27 crc kubenswrapper[4776]: I1125 12:21:27.535733 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2pj5g" event={"ID":"136e5670-c1db-4e24-a402-9163e0045408","Type":"ContainerStarted","Data":"de6bb8a46e0d0912d4057990edf24947cac1988ceefc3b0fe7d3444fd867ee8b"} Nov 25 12:21:27 crc kubenswrapper[4776]: I1125 12:21:27.559309 4776 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2pj5g" podStartSLOduration=3.292043479 podStartE2EDuration="8.55928267s" podCreationTimestamp="2025-11-25 12:21:19 +0000 UTC" firstStartedPulling="2025-11-25 12:21:21.432059619 +0000 UTC m=+10626.473119172" lastFinishedPulling="2025-11-25 12:21:26.69929881 +0000 UTC m=+10631.740358363" observedRunningTime="2025-11-25 12:21:27.553514416 +0000 UTC m=+10632.594573979" watchObservedRunningTime="2025-11-25 12:21:27.55928267 +0000 UTC m=+10632.600342223" Nov 25 12:21:27 crc kubenswrapper[4776]: I1125 12:21:27.663927 4776 scope.go:117] "RemoveContainer" containerID="d4acaf62e3df4138de718bbb4ab1ab582c6c6cf75f1110ff504e82c463027563" Nov 25 12:21:27 crc kubenswrapper[4776]: I1125 12:21:27.750330 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gr69n" Nov 25 12:21:27 crc kubenswrapper[4776]: I1125 12:21:27.750656 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gr69n" Nov 25 12:21:28 crc kubenswrapper[4776]: I1125 12:21:28.553140 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-84dqb" event={"ID":"fa47ebcc-a95e-4693-876d-7284c28c3ade","Type":"ContainerStarted","Data":"0880704ded269f2841e3ab8f19a9c62b0057499e000802a1acdf6893046842cf"} Nov 25 12:21:28 crc kubenswrapper[4776]: I1125 12:21:28.814430 4776 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-gr69n" podUID="14a73053-d4c1-4633-bbf0-bed6e1ef64da" containerName="registry-server" probeResult="failure" output=< Nov 25 12:21:28 crc kubenswrapper[4776]: timeout: failed to connect service ":50051" within 1s Nov 25 12:21:28 crc kubenswrapper[4776]: > Nov 25 12:21:29 crc kubenswrapper[4776]: I1125 12:21:29.986023 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2pj5g" Nov 25 12:21:29 crc kubenswrapper[4776]: I1125 12:21:29.986670 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2pj5g" Nov 25 12:21:30 crc kubenswrapper[4776]: I1125 12:21:30.054335 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2pj5g" Nov 25 12:21:37 crc kubenswrapper[4776]: I1125 12:21:37.810329 4776 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gr69n" Nov 25 12:21:37 crc kubenswrapper[4776]: I1125 12:21:37.877662 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gr69n" Nov 25 12:21:38 crc kubenswrapper[4776]: I1125 12:21:38.049806 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gr69n"] Nov 25 12:21:39 crc kubenswrapper[4776]: I1125 12:21:39.685133 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gr69n" podUID="14a73053-d4c1-4633-bbf0-bed6e1ef64da" containerName="registry-server" containerID="cri-o://ba7a7601816a233cdf00d3aaa9c2f92d02115792009e59c52ccd88962fc33928" gracePeriod=2 Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:40.049695 4776 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2pj5g" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:40.344227 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gr69n" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:40.454127 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2pj5g"] Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:40.522495 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14a73053-d4c1-4633-bbf0-bed6e1ef64da-utilities\") pod \"14a73053-d4c1-4633-bbf0-bed6e1ef64da\" (UID: \"14a73053-d4c1-4633-bbf0-bed6e1ef64da\") " Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:40.522547 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hqr2\" (UniqueName: \"kubernetes.io/projected/14a73053-d4c1-4633-bbf0-bed6e1ef64da-kube-api-access-2hqr2\") pod \"14a73053-d4c1-4633-bbf0-bed6e1ef64da\" (UID: \"14a73053-d4c1-4633-bbf0-bed6e1ef64da\") " Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:40.522731 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14a73053-d4c1-4633-bbf0-bed6e1ef64da-catalog-content\") pod \"14a73053-d4c1-4633-bbf0-bed6e1ef64da\" (UID: \"14a73053-d4c1-4633-bbf0-bed6e1ef64da\") " Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:40.524084 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14a73053-d4c1-4633-bbf0-bed6e1ef64da-utilities" (OuterVolumeSpecName: "utilities") pod "14a73053-d4c1-4633-bbf0-bed6e1ef64da" (UID: "14a73053-d4c1-4633-bbf0-bed6e1ef64da"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:40.532123 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14a73053-d4c1-4633-bbf0-bed6e1ef64da-kube-api-access-2hqr2" (OuterVolumeSpecName: "kube-api-access-2hqr2") pod "14a73053-d4c1-4633-bbf0-bed6e1ef64da" (UID: "14a73053-d4c1-4633-bbf0-bed6e1ef64da"). InnerVolumeSpecName "kube-api-access-2hqr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:40.584576 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14a73053-d4c1-4633-bbf0-bed6e1ef64da-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "14a73053-d4c1-4633-bbf0-bed6e1ef64da" (UID: "14a73053-d4c1-4633-bbf0-bed6e1ef64da"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:40.626454 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14a73053-d4c1-4633-bbf0-bed6e1ef64da-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:40.626513 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14a73053-d4c1-4633-bbf0-bed6e1ef64da-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:40.626526 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hqr2\" (UniqueName: \"kubernetes.io/projected/14a73053-d4c1-4633-bbf0-bed6e1ef64da-kube-api-access-2hqr2\") on node \"crc\" DevicePath \"\"" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:40.703791 4776 generic.go:334] "Generic (PLEG): container finished" podID="14a73053-d4c1-4633-bbf0-bed6e1ef64da" containerID="ba7a7601816a233cdf00d3aaa9c2f92d02115792009e59c52ccd88962fc33928" exitCode=0 Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:40.703880 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gr69n" event={"ID":"14a73053-d4c1-4633-bbf0-bed6e1ef64da","Type":"ContainerDied","Data":"ba7a7601816a233cdf00d3aaa9c2f92d02115792009e59c52ccd88962fc33928"} Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:40.703991 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gr69n" event={"ID":"14a73053-d4c1-4633-bbf0-bed6e1ef64da","Type":"ContainerDied","Data":"314f2b648f07718b7b046759c107ed369b7ec2477eb77806a580ce5fa231cdaf"} Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:40.704034 4776 scope.go:117] "RemoveContainer" containerID="ba7a7601816a233cdf00d3aaa9c2f92d02115792009e59c52ccd88962fc33928" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:40.704105 4776 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2pj5g" podUID="136e5670-c1db-4e24-a402-9163e0045408" containerName="registry-server" containerID="cri-o://de6bb8a46e0d0912d4057990edf24947cac1988ceefc3b0fe7d3444fd867ee8b" gracePeriod=2 Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:40.704137 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gr69n" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:40.749144 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gr69n"] Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:40.754441 4776 scope.go:117] "RemoveContainer" containerID="2eb9060b8ab9c0cf5a5268d26c745cfc3dbbf88a3e919c5e3c182e77cac25cdd" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:40.759114 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gr69n"] Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:40.781783 4776 scope.go:117] "RemoveContainer" containerID="17e70e93c6d1dbb45c455d36fe165d434c26be1243daf949393d272fb2e8ce85" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:40.842056 4776 scope.go:117] "RemoveContainer" containerID="ba7a7601816a233cdf00d3aaa9c2f92d02115792009e59c52ccd88962fc33928" Nov 25 12:21:42 crc kubenswrapper[4776]: E1125 12:21:40.842781 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba7a7601816a233cdf00d3aaa9c2f92d02115792009e59c52ccd88962fc33928\": container with ID starting with ba7a7601816a233cdf00d3aaa9c2f92d02115792009e59c52ccd88962fc33928 not found: ID does not exist" containerID="ba7a7601816a233cdf00d3aaa9c2f92d02115792009e59c52ccd88962fc33928" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:40.842821 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba7a7601816a233cdf00d3aaa9c2f92d02115792009e59c52ccd88962fc33928"} err="failed to get container status \"ba7a7601816a233cdf00d3aaa9c2f92d02115792009e59c52ccd88962fc33928\": rpc error: code = NotFound desc = could not find container \"ba7a7601816a233cdf00d3aaa9c2f92d02115792009e59c52ccd88962fc33928\": container with ID starting with ba7a7601816a233cdf00d3aaa9c2f92d02115792009e59c52ccd88962fc33928 not found: ID does not exist" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:40.842864 4776 scope.go:117] "RemoveContainer" containerID="2eb9060b8ab9c0cf5a5268d26c745cfc3dbbf88a3e919c5e3c182e77cac25cdd" Nov 25 12:21:42 crc kubenswrapper[4776]: E1125 12:21:40.843545 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2eb9060b8ab9c0cf5a5268d26c745cfc3dbbf88a3e919c5e3c182e77cac25cdd\": container with ID starting with 2eb9060b8ab9c0cf5a5268d26c745cfc3dbbf88a3e919c5e3c182e77cac25cdd not found: ID does not exist" containerID="2eb9060b8ab9c0cf5a5268d26c745cfc3dbbf88a3e919c5e3c182e77cac25cdd" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:40.843850 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2eb9060b8ab9c0cf5a5268d26c745cfc3dbbf88a3e919c5e3c182e77cac25cdd"} err="failed to get container status \"2eb9060b8ab9c0cf5a5268d26c745cfc3dbbf88a3e919c5e3c182e77cac25cdd\": rpc error: code = NotFound desc = could not find container \"2eb9060b8ab9c0cf5a5268d26c745cfc3dbbf88a3e919c5e3c182e77cac25cdd\": container with ID starting with 2eb9060b8ab9c0cf5a5268d26c745cfc3dbbf88a3e919c5e3c182e77cac25cdd not found: ID does not exist" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:40.843906 4776 scope.go:117] "RemoveContainer" containerID="17e70e93c6d1dbb45c455d36fe165d434c26be1243daf949393d272fb2e8ce85" Nov 25 12:21:42 crc kubenswrapper[4776]: E1125 12:21:40.844807 4776 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17e70e93c6d1dbb45c455d36fe165d434c26be1243daf949393d272fb2e8ce85\": container with ID starting with 17e70e93c6d1dbb45c455d36fe165d434c26be1243daf949393d272fb2e8ce85 not found: ID does not exist" containerID="17e70e93c6d1dbb45c455d36fe165d434c26be1243daf949393d272fb2e8ce85" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:40.844901 4776 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17e70e93c6d1dbb45c455d36fe165d434c26be1243daf949393d272fb2e8ce85"} err="failed to get container status \"17e70e93c6d1dbb45c455d36fe165d434c26be1243daf949393d272fb2e8ce85\": rpc error: code = NotFound desc = could not find container \"17e70e93c6d1dbb45c455d36fe165d434c26be1243daf949393d272fb2e8ce85\": container with ID starting with 17e70e93c6d1dbb45c455d36fe165d434c26be1243daf949393d272fb2e8ce85 not found: ID does not exist" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:41.679584 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14a73053-d4c1-4633-bbf0-bed6e1ef64da" path="/var/lib/kubelet/pods/14a73053-d4c1-4633-bbf0-bed6e1ef64da/volumes" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:41.720576 4776 generic.go:334] "Generic (PLEG): container finished" podID="136e5670-c1db-4e24-a402-9163e0045408" containerID="de6bb8a46e0d0912d4057990edf24947cac1988ceefc3b0fe7d3444fd867ee8b" exitCode=0 Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:41.720610 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2pj5g" event={"ID":"136e5670-c1db-4e24-a402-9163e0045408","Type":"ContainerDied","Data":"de6bb8a46e0d0912d4057990edf24947cac1988ceefc3b0fe7d3444fd867ee8b"} Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:42.680493 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2pj5g" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:42.750586 4776 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2pj5g" event={"ID":"136e5670-c1db-4e24-a402-9163e0045408","Type":"ContainerDied","Data":"d9eb7cfc5c4d66d505a16e287b816b9897943b0c6dc676d781cfaaef2d49182c"} Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:42.750680 4776 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2pj5g" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:42.750699 4776 scope.go:117] "RemoveContainer" containerID="de6bb8a46e0d0912d4057990edf24947cac1988ceefc3b0fe7d3444fd867ee8b" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:42.776746 4776 scope.go:117] "RemoveContainer" containerID="14415b152610d90a74f57b289bfebf4a708759c93e795861b4fd8d6499faf15f" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:42.782706 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhglg\" (UniqueName: \"kubernetes.io/projected/136e5670-c1db-4e24-a402-9163e0045408-kube-api-access-vhglg\") pod \"136e5670-c1db-4e24-a402-9163e0045408\" (UID: \"136e5670-c1db-4e24-a402-9163e0045408\") " Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:42.783048 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/136e5670-c1db-4e24-a402-9163e0045408-utilities\") pod \"136e5670-c1db-4e24-a402-9163e0045408\" (UID: \"136e5670-c1db-4e24-a402-9163e0045408\") " Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:42.783170 4776 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/136e5670-c1db-4e24-a402-9163e0045408-catalog-content\") pod \"136e5670-c1db-4e24-a402-9163e0045408\" (UID: \"136e5670-c1db-4e24-a402-9163e0045408\") " Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:42.784196 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/136e5670-c1db-4e24-a402-9163e0045408-utilities" (OuterVolumeSpecName: "utilities") pod "136e5670-c1db-4e24-a402-9163e0045408" (UID: "136e5670-c1db-4e24-a402-9163e0045408"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:42.798431 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/136e5670-c1db-4e24-a402-9163e0045408-kube-api-access-vhglg" (OuterVolumeSpecName: "kube-api-access-vhglg") pod "136e5670-c1db-4e24-a402-9163e0045408" (UID: "136e5670-c1db-4e24-a402-9163e0045408"). InnerVolumeSpecName "kube-api-access-vhglg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:42.801318 4776 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhglg\" (UniqueName: \"kubernetes.io/projected/136e5670-c1db-4e24-a402-9163e0045408-kube-api-access-vhglg\") on node \"crc\" DevicePath \"\"" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:42.801348 4776 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/136e5670-c1db-4e24-a402-9163e0045408-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:42.821142 4776 scope.go:117] "RemoveContainer" containerID="1ec34a41408d36dfd6e397c7560bc880790b2b7feb9e761704ad103b4cd8ddbd" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:42.855101 4776 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/136e5670-c1db-4e24-a402-9163e0045408-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "136e5670-c1db-4e24-a402-9163e0045408" (UID: "136e5670-c1db-4e24-a402-9163e0045408"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 12:21:42 crc kubenswrapper[4776]: I1125 12:21:42.903798 4776 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/136e5670-c1db-4e24-a402-9163e0045408-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 12:21:43 crc kubenswrapper[4776]: I1125 12:21:43.085925 4776 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2pj5g"] Nov 25 12:21:43 crc kubenswrapper[4776]: I1125 12:21:43.106596 4776 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2pj5g"] Nov 25 12:21:43 crc kubenswrapper[4776]: I1125 12:21:43.682715 4776 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="136e5670-c1db-4e24-a402-9163e0045408" path="/var/lib/kubelet/pods/136e5670-c1db-4e24-a402-9163e0045408/volumes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515111317724024447 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015111317725017365 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015111272314016502 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015111272314015452 5ustar corecore